1,794 research outputs found
ML4Chem: A Machine Learning Package for Chemistry and Materials Science
ML4Chem is an open-source machine learning library for chemistry and
materials science. It provides an extendable platform to develop and deploy
machine learning models and pipelines and is targeted to the non-expert and
expert users. ML4Chem follows user-experience design and offers the needed
tools to go from data preparation to inference. Here we introduce its atomistic
module for the implementation, deployment, and reproducibility of atom-centered
models. This module is composed of six core building blocks: data,
featurization, models, model optimization, inference, and visualization. We
present their functionality and easiness of use with demonstrations utilizing
neural networks and kernel ridge regression algorithms.Comment: 32 pages, 11 Figure
Domain Adaptation in Unmanned Aerial Vehicles Landing using Reinforcement Learning
Landing an unmanned aerial vehicle (UAV) on a moving platform is a challenging task that often requires exact models of the UAV dynamics, platform characteristics, and environmental conditions. In this thesis, we present and investigate three different machine learning approaches with varying levels of domain knowledge: dynamics randomization, universal policy with system identification, and reinforcement learning with no parameter variation. We first train the policies in simulation, then perform experiments both in simulation, making variations of the system dynamics with wind and friction coefficient, then perform experiments in a real robot system with wind variation. We initially expected that providing more information on environmental characteristics with system identification would improve the outcomes, however, we found that transferring a policy learned in simulation with domain randomization to the real robot system achieves the best result in the real robot and simulation. Although in simulation the universal policy with system identification is faster in some cases. In this thesis, we compare the results of multiple deep reinforcement learning approaches trained in simulation and transferred in robot experiments with the presence of external disturbances. We were able to create a policy to control a UAV completely trained in simulation and transfer to a real system with the presence of external disturbances. In doing so, we evaluate the performance of dynamics randomization and universal policy with system identification.
Adviser: Carrick Detweile
Classification of red blood cell shapes in flow using outlier tolerant machine learning
The manual evaluation, classification and counting of biological objects
demands for an enormous expenditure of time and subjective human input may be a
source of error. Investigating the shape of red blood cells (RBCs) in
microcapillary Poiseuille flow, we overcome this drawback by introducing a
convolutional neural regression network for an automatic, outlier tolerant
shape classification. From our experiments we expect two stable geometries: the
so-called `slipper' and `croissant' shapes depending on the prevailing flow
conditions and the cell-intrinsic parameters. Whereas croissants mostly occur
at low shear rates, slippers evolve at higher flow velocities. With our method,
we are able to find the transition point between both `phases' of stable shapes
which is of high interest to ensuing theoretical studies and numerical
simulations. Using statistically based thresholds, from our data, we obtain
so-called phase diagrams which are compared to manual evaluations.
Prospectively, our concept allows us to perform objective analyses of
measurements for a variety of flow conditions and to receive comparable
results. Moreover, the proposed procedure enables unbiased studies on the
influence of drugs on flow properties of single RBCs and the resulting
macroscopic change of the flow behavior of whole blood.Comment: 15 pages, published in PLoS Comput Biol, open acces
To Each Optimizer a Norm, To Each Norm its Generalization
We study the implicit regularization of optimization methods for linear
models interpolating the training data in the under-parametrized and
over-parametrized regimes. Since it is difficult to determine whether an
optimizer converges to solutions that minimize a known norm, we flip the
problem and investigate what is the corresponding norm minimized by an
interpolating solution. Using this reasoning, we prove that for
over-parameterized linear regression, projections onto linear spans can be used
to move between different interpolating solutions. For under-parameterized
linear classification, we prove that for any linear classifier separating the
data, there exists a family of quadratic norms ||.||_P such that the
classifier's direction is the same as that of the maximum P-margin solution.
For linear classification, we argue that analyzing convergence to the standard
maximum l2-margin is arbitrary and show that minimizing the norm induced by the
data results in better generalization. Furthermore, for over-parameterized
linear classification, projections onto the data-span enable us to use
techniques from the under-parameterized setting. On the empirical side, we
propose techniques to bias optimizers towards better generalizing solutions,
improving their test performance. We validate our theoretical results via
synthetic experiments, and use the neural tangent kernel to handle non-linear
models
DEVDAN: Deep Evolving Denoising Autoencoder
The Denoising Autoencoder (DAE) enhances the flexibility of the data stream
method in exploiting unlabeled samples. Nonetheless, the feasibility of DAE for
data stream analytic deserves an in-depth study because it characterizes a
fixed network capacity that cannot adapt to rapidly changing environments. Deep
evolving denoising autoencoder (DEVDAN), is proposed in this paper. It features
an open structure in the generative phase and the discriminative phase where
the hidden units can be automatically added and discarded on the fly. The
generative phase refines the predictive performance of the discriminative model
exploiting unlabeled data. Furthermore, DEVDAN is free of the problem-specific
threshold and works fully in the single-pass learning fashion. We show that
DEVDAN can find competitive network architecture compared with state-of-the-art
methods on the classification task using ten prominent datasets simulated under
the prequential test-then-train protocol.Comment: This paper has been accepted for publication in Neurocomputing 2019.
arXiv admin note: substantial text overlap with arXiv:1809.0908
Recommended from our members
Probabilistic Programming for Deep Learning
We propose the idea of deep probabilistic programming, a synthesis of advances for systems at the intersection of probabilistic modeling and deep learning. Such systems enable the development of new probabilistic models and inference algorithms that would otherwise be impossible: enabling unprecedented scales to billions of parameters, distributed and mixed precision environments, and AI accelerators; integration with neural architectures for modeling massive and high-dimensional datasets; and the use of computation graphs for automatic differentiation and arbitrary manipulation of probabilistic programs for flexible inference and model criticism.
After describing deep probabilistic programming, we discuss applications in novel variational inference algorithms and deep probabilistic models. First, we introduce the variational Gaussian process (VGP), a Bayesian nonparametric variational family, which adapts its shape to match complex posterior distributions. The VGP generates approximate posterior samples by generating latent inputs and warping them through random non-linear mappings; the distribution over random mappings is learned during inference, enabling the transformed outputs to adapt to varying complexity of the true posterior. Second, we introduce hierarchical implicit models (HIMs). HIMs combine the idea of implicit densities with hierarchical Bayesian modeling, thereby defining models via simulators of data with rich hidden structure
- …