16,029 research outputs found
Learning classifier systems with memory condition to solve non-Markov problems
In the family of Learning Classifier Systems, the classifier system XCS has
been successfully used for many applications. However, the standard XCS has no
memory mechanism and can only learn optimal policy in Markov environments,
where the optimal action is determined solely by the state of current sensory
input. In practice, most environments are partially observable environments on
agent's sensation, which are also known as non-Markov environments. Within
these environments, XCS either fails, or only develops a suboptimal policy,
since it has no memory. In this work, we develop a new classifier system based
on XCS to tackle this problem. It adds an internal message list to XCS as the
memory list to record input sensation history, and extends a small number of
classifiers with memory conditions. The classifier's memory condition, as a
foothold to disambiguate non-Markov states, is used to sense a specified
element in the memory list. Besides, a detection method is employed to
recognize non-Markov states in environments, to avoid these states controlling
over classifiers' memory conditions. Furthermore, four sets of different
complex maze environments have been tested by the proposed method. Experimental
results show that our system is one of the best techniques to solve partially
observable environments, compared with some well-known classifier systems
proposed for these environments.Comment: 34 pages, 15 figures, 1 tabl
A Probabilistic Modeling Approach to One-Shot Gesture Recognition
Gesture recognition enables a natural extension of the way we currently
interact with devices. Commercially available gesture recognition systems are
usually pre-trained and offer no option for customization by the user. In order
to improve the user experience, it is desirable to allow end users to define
their own gestures. This scenario requires learning from just a few training
examples if we want to impose only a light training load on the user. To this
end, we propose a gesture classifier based on a hierarchical probabilistic
modeling approach. In this framework, high-level features that are shared among
different gestures can be extracted from a large labeled data set, yielding a
prior distribution for gestures. When learning new types of gestures, the
learned shared prior reduces the number of required training examples for
individual gestures. We implemented the proposed gesture classifier for a Myo
sensor bracelet and show favorable results for the tested system on a database
of 17 different gesture types. Furthermore, we propose and implement two
methods to incorporate the gesture classifier in a real-time gesture
recognition system
Hyperspectral Image Classification with Markov Random Fields and a Convolutional Neural Network
This paper presents a new supervised classification algorithm for remotely
sensed hyperspectral image (HSI) which integrates spectral and spatial
information in a unified Bayesian framework. First, we formulate the HSI
classification problem from a Bayesian perspective. Then, we adopt a
convolutional neural network (CNN) to learn the posterior class distributions
using a patch-wise training strategy to better use the spatial information.
Next, spatial information is further considered by placing a spatial smoothness
prior on the labels. Finally, we iteratively update the CNN parameters using
stochastic gradient decent (SGD) and update the class labels of all pixel
vectors using an alpha-expansion min-cut-based algorithm. Compared with other
state-of-the-art methods, the proposed classification method achieves better
performance on one synthetic dataset and two benchmark HSI datasets in a number
of experimental settings
A C++ library for Multimodal Deep Learning
MDL, Multimodal Deep Learning Library, is a deep learning framework that
supports multiple models, and this document explains its philosophy and
functionality. MDL runs on Linux, Mac, and Unix platforms. It depends on
OpenCV.Comment: 27 page
Supervised multiview learning based on simultaneous learning of multiview intact and single view classifier
Multiview learning problem refers to the problem of learning a classifier
from multiple view data. In this data set, each data points is presented by
multiple different views. In this paper, we propose a novel method for this
problem. This method is based on two assumptions. The first assumption is that
each data point has an intact feature vector, and each view is obtained by a
linear transformation from the intact vector. The second assumption is that the
intact vectors are discriminative, and in the intact space, we have a linear
classifier to separate the positive class from the negative class. We define an
intact vector for each data point, and a view-conditional transformation matrix
for each view, and propose to reconstruct the multiple view feature vectors by
the product of the corresponding intact vectors and transformation matrices.
Moreover, we also propose a linear classifier in the intact space, and learn it
jointly with the intact vectors. The learning problem is modeled by a
minimization problem, and the objective function is composed of a Cauchy error
estimator-based view-conditional reconstruction term over all data points and
views, and a classification error term measured by hinge loss over all the
intact vectors of all the data points. Some regularization terms are also
imposed to different variables in the objective function. The minimization
problem is solve by an iterative algorithm using alternate optimization
strategy and gradient descent algorithm. The proposed algorithm shows it
advantage in the compression to other multiview learning algorithms on
benchmark data sets
Distributed Machine Learning in Materials that Couple Sensing, Actuation, Computation and Communication
This paper reviews machine learning applications and approaches to detection,
classification and control of intelligent materials and structures with
embedded distributed computation elements. The purpose of this survey is to
identify desired tasks to be performed in each type of material or structure
(e.g., damage detection in composites), identify and compare common approaches
to learning such tasks, and investigate models and training paradigms used.
Machine learning approaches and common temporal features used in the domains of
structural health monitoring, morphable aircraft, wearable computing and
robotic skins are explored. As the ultimate goal of this research is to
incorporate the approaches described in this survey into a robotic material
paradigm, the potential for adapting the computational models used in these
applications, and corresponding training algorithms, to an amorphous network of
computing nodes is considered. Distributed versions of support vector machines,
graphical models and mixture models developed in the field of wireless sensor
networks are reviewed. Potential areas of investigation, including possible
architectures for incorporating machine learning into robotic nodes, training
approaches, and the possibility of using deep learning approaches for automatic
feature extraction, are discussed
Stochastic Security: Adversarial Defense Using Long-Run Dynamics of Energy-Based Models
The vulnerability of deep networks to adversarial attacks is a central
problem for deep learning from the perspective of both cognition and security.
The current most successful defense method is to train a classifier using
adversarial images created during learning. Another defense approach involves
transformation or purification of the original input to remove adversarial
signals before the image is classified. We focus on defending naturally-trained
classifiers using Markov Chain Monte Carlo (MCMC) sampling with an Energy-Based
Model (EBM) for adversarial purification. In contrast to adversarial training,
our approach is intended to secure pre-existing and highly vulnerable
classifiers.
The memoryless behavior of long-run MCMC sampling will eventually remove
adversarial signals, while metastable behavior preserves consistent appearance
of MCMC samples after many steps to allow accurate long-run prediction.
Balancing these factors can lead to effective purification and robust
classification. We evaluate adversarial defense with an EBM using the strongest
known attacks against purification. Our contributions are 1) an improved method
for training EBM's with realistic long-run MCMC samples, 2) an
Expectation-Over-Transformation (EOT) defense that resolves theoretical
ambiguities for stochastic defenses and from which the EOT attack naturally
follows, and 3) state-of-the-art adversarial defense for naturally-trained
classifiers and competitive defense compared to adversarially-trained
classifiers on Cifar-10, SVHN, and Cifar-100. Code and pre-trained models are
available at https://github.com/point0bar1/ebm-defense.Comment: ICLR 202
A new boosting algorithm based on dual averaging scheme
The fields of machine learning and mathematical optimization increasingly
intertwined. The special topic on supervised learning and convex optimization
examines this interplay. The training part of most supervised learning
algorithms can usually be reduced to an optimization problem that minimizes a
loss between model predictions and training data. While most optimization
techniques focus on accuracy and speed of convergence, the qualities of good
optimization algorithm from the machine learning perspective can be quite
different since machine learning is more than fitting the data. Better
optimization algorithms that minimize the training loss can possibly give very
poor generalization performance. In this paper, we examine a particular kind of
machine learning algorithm, boosting, whose training process can be viewed as
functional coordinate descent on the exponential loss. We study the relation
between optimization techniques and machine learning by implementing a new
boosting algorithm. DABoost, based on dual-averaging scheme and study its
generalization performance. We show that DABoost, although slower in reducing
the training error, in general enjoys a better generalization error than
AdaBoost.Comment: 8 pages, 3 figure
Patient Flow Prediction via Discriminative Learning of Mutually-Correcting Processes
Over the past decade the rate of care unit (CU) use in the United States has
been increasing. With an aging population and ever-growing demand for medical
care, effective management of patients' transitions among different care
facilities will prove indispensible for shortening the length of hospital
stays, improving patient outcomes, allocating critical care resources, and
reducing preventable re-admissions. In this paper, we focus on an important
problem of predicting the so-called "patient flow" from longitudinal electronic
health records (EHRs), which has not been explored via existing machine
learning techniques. By treating a sequence of transition events as a point
process, we develop a novel framework for modeling patient flow through various
CUs and jointly predicting patients' destination CUs and duration days. Instead
of learning a generative point process model via maximum likelihood estimation,
we propose a novel discriminative learning algorithm aiming at improving the
prediction of transition events in the case of sparse data. By parameterizing
the proposed model as a mutually-correcting process, we formulate the
estimation problem via generalized linear models, which lends itself to
efficient learning based on alternating direction method of multipliers (ADMM).
Furthermore, we achieve simultaneous feature selection and learning by adding a
group-lasso regularizer to the ADMM algorithm. Additionally, for suppressing
the negative influence of data imbalance on the learning of model, we
synthesize auxiliary training data for the classes with extremely few samples,
and improve the robustness of our learning method accordingly. Testing on
real-world data, we show that our method obtains superior performance in terms
of accuracy of predicting the destination CU transition and duration of each CU
occupancy.Comment: in IEEE Transactions on Knowledge and Data Engineering (TKDE), 201
A Credit Assignment Compiler for Joint Prediction
Many machine learning applications involve jointly predicting multiple
mutually dependent output variables. Learning to search is a family of methods
where the complex decision problem is cast into a sequence of decisions via a
search space. Although these methods have shown promise both in theory and in
practice, implementing them has been burdensomely awkward. In this paper, we
show the search space can be defined by an arbitrary imperative program,
turning learning to search into a credit assignment compiler. Altogether with
the algorithmic improvements for the compiler, we radically reduce the
complexity of programming and the running time. We demonstrate the feasibility
of our approach on multiple joint prediction tasks. In all cases, we obtain
accuracies as high as alternative approaches, at drastically reduced execution
and programming time
- …