42 research outputs found
A New Computationally Simple Approach for Implementing Neural Networks with Output Hard Constraints
A new computationally simple method of imposing hard convex constraints on
the neural network output values is proposed. The key idea behind the method is
to map a vector of hidden parameters of the network to a point that is
guaranteed to be inside the feasible set defined by a set of constraints. The
mapping is implemented by the additional neural network layer with constraints
for output. The proposed method is simply extended to the case when constraints
are imposed not only on the output vectors, but also on joint constraints
depending on inputs. The projection approach to imposing constraints on outputs
can simply be implemented in the framework of the proposed method. It is shown
how to incorporate different types of constraints into the proposed method,
including linear and quadratic constraints, equality constraints, and dynamic
constraints, constraints in the form of boundaries. An important feature of the
method is its computational simplicity. Complexities of the forward pass of the
proposed neural network layer by linear and quadratic constraints are O(n*m)
and O(n^2*m), respectively, where n is the number of variables, m is the number
of constraints. Numerical experiments illustrate the method by solving
optimization and classification problems. The code implementing the method is
publicly available
Random Survival Forests Incorporated by the Nadaraya-Watson Regression
An attention-based random survival forest (Att-RSF) is presented in the paper. The first main idea behind this model is to adapt the Nadaraya-Watson kernel regression to the random survival forest so that the regression weights or kernels can be regarded as trainable attention weights under important condition that predictions of the random survival forest are represented in the form of functions, for example, the survival function and the cumulative hazard function. Each trainable weight assigned to a tree and a training or testing example is defined by two factors: by the ability of corresponding tree to predict and by the peculiarity of an example which falls into a leaf of the tree. The second main idea behind Att-RSF is to apply the Huber's contamination model to represent the attention weights as the linear function of the trainable attention parameters. The Harrell's C-index (concordance index) measuring the prediction quality of the random survival forest is used to form the loss function for training the attention weights. The C-index jointly with the contamination model lead to the standard quadratic optimization problem for computing the weights, which has many simple algorithms for its solution. Numerical experiments with real datasets containing survival data illustrate Att-RSF
Heterogeneous Treatment Effect with Trained Kernels of the Nadaraya-Watson Regression
A new method for estimating the conditional average treatment effect is
proposed in the paper. It is called TNW-CATE (the Trainable Nadaraya-Watson
regression for CATE) and based on the assumption that the number of controls is
rather large whereas the number of treatments is small. TNW-CATE uses the
Nadaraya-Watson regression for predicting outcomes of patients from the control
and treatment groups. The main idea behind TNW-CATE is to train kernels of the
Nadaraya-Watson regression by using a weight sharing neural network of a
specific form. The network is trained on controls, and it replaces standard
kernels with a set of neural subnetworks with shared parameters such that every
subnetwork implements the trainable kernel, but the whole network implements
the Nadaraya-Watson estimator. The network memorizes how the feature vectors
are located in the feature space. The proposed approach is similar to the
transfer learning when domains of source and target data are similar, but tasks
are different. Various numerical simulation experiments illustrate TNW-CATE and
compare it with the well-known T-learner, S-learner and X-learner for several
types of the control and treatment outcome functions. The code of proposed
algorithms implementing TNW-CATE is available in
https://github.com/Stasychbr/TNW-CATE
Droplet-shaped waves: Causal finite-support analogs of X-shaped waves
A model of steady-state X-shaped wave generation by a superluminal
(supersonic) pointlike source infinitely moving along a straight line is
extended to a more realistic causal scenario of a source pulse launched at time
zero and propagating rectilinearly at constant superluminal speed. In the case
of infinitely short (delta) pulse, the new model yields an analytical solution,
corresponding to the propagation-invariant X-shaped wave clipped by a
droplet-shaped support, which perpetually expands along the propagation and
transversal directions, thus tending the droplet-shaped wave to the X-shaped
one.Comment: 14 pages, 6 figure
BENK: The Beran Estimator with Neural Kernels for Estimating the Heterogeneous Treatment Effect
A method for estimating the conditional average treatment effect under
condition of censored time-to-event data called BENK (the Beran Estimator with
Neural Kernels) is proposed. The main idea behind the method is to apply the
Beran estimator for estimating the survival functions of controls and
treatments. Instead of typical kernel functions in the Beran estimator, it is
proposed to implement kernels in the form of neural networks of a specific form
called the neural kernels. The conditional average treatment effect is
estimated by using the survival functions as outcomes of the control and
treatment neural networks which consists of a set of neural kernels with shared
parameters. The neural kernels are more flexible and can accurately model a
complex location structure of feature vectors. Various numerical simulation
experiments illustrate BENK and compare it with the well-known T-learner,
S-learner and X-learner for several types of the control and treatment outcome
functions based on the Cox models, the random survival forest and the
Nadaraya-Watson regression with Gaussian kernels. The code of proposed
algorithms implementing BENK is available in https://github.com/Stasychbr/BENK