294 research outputs found
Smoothed Analysis in Unsupervised Learning via Decoupling
Smoothed analysis is a powerful paradigm in overcoming worst-case
intractability in unsupervised learning and high-dimensional data analysis.
While polynomial time smoothed analysis guarantees have been obtained for
worst-case intractable problems like tensor decompositions and learning
mixtures of Gaussians, such guarantees have been hard to obtain for several
other important problems in unsupervised learning. A core technical challenge
in analyzing algorithms is obtaining lower bounds on the least singular value
for random matrix ensembles with dependent entries, that are given by
low-degree polynomials of a few base underlying random variables.
In this work, we address this challenge by obtaining high-confidence lower
bounds on the least singular value of new classes of structured random matrix
ensembles of the above kind. We then use these bounds to design algorithms with
polynomial time smoothed analysis guarantees for the following three important
problems in unsupervised learning:
1. Robust subspace recovery, when the fraction of inliers in the
d-dimensional subspace is at least for any constant integer . This contrasts with the known
worst-case intractability when , and the previous smoothed
analysis result which needed (Hardt and Moitra, 2013).
2. Learning overcomplete hidden markov models, where the size of the state
space is any polynomial in the dimension of the observations. This gives the
first polynomial time guarantees for learning overcomplete HMMs in a smoothed
analysis model.
3. Higher order tensor decompositions, where we generalize the so-called
FOOBI algorithm of Cardoso to find order- rank-one tensors in a subspace.
This allows us to obtain polynomially robust decomposition algorithms for
'th order tensors with rank .Comment: 44 page
Unsupervised discovery of temporal sequences in high-dimensional datasets, with applications to neuroscience.
Identifying low-dimensional features that describe large-scale neural recordings is a major challenge in neuroscience. Repeated temporal patterns (sequences) are thought to be a salient feature of neural dynamics, but are not succinctly captured by traditional dimensionality reduction techniques. Here, we describe a software toolbox-called seqNMF-with new methods for extracting informative, non-redundant, sequences from high-dimensional neural data, testing the significance of these extracted patterns, and assessing the prevalence of sequential structure in data. We test these methods on simulated data under multiple noise conditions, and on several real neural and behavioral datas. In hippocampal data, seqNMF identifies neural sequences that match those calculated manually by reference to behavioral events. In songbird data, seqNMF discovers neural sequences in untutored birds that lack stereotyped songs. Thus, by identifying temporal structure directly from neural data, seqNMF enables dissection of complex neural circuits without relying on temporal references from stimuli or behavioral outputs
Concentration inequalities for random tensors
We show how to extend several basic concentration inequalities for simple
random tensors where all are
independent random vectors in with independent coefficients. The
new results have optimal dependence on the dimension and the degree . As
an application, we show that random tensors are well conditioned: independent copies of the simple random tensor
are far from being linearly dependent with high probability. We prove this fact
for any degree and conjecture that it is true for any
.Comment: A few more typos were correcte
A hierarchical reduced-order model applied to nuclear reactors
Modelling the neutron transport of a nuclear reactor is a very computationally demanding task
that requires a large number of degrees of freedom to accurately capture all of the physics. For a
complete reactor picture, other physics must be incorporated, through coupling, further exacerbating the computational demand. Computational modelling has many benefits: optimisation,
real-time analysis, and safety analysis are some of the more important ones. However, nuclear modelling has yet to capitalise on these, and existing approaches are too computationally
demanding.
Machine Learning has seen incredible growth over the last decade, but it has yet to be utilised
within the nuclear modelling community to the same extent. The frameworks available represent incredibly efficient and optimised code, having been written to run on GPUs and AI
computers. Presented here is a physics-driven neural network that solves neutron transport,
first for the diffusion approximation and then extended to the whole transport problem.
One method that can potentially reduce the computational complexity is Reduced-Order Modelling (ROM), which is a way to define a low-dimensional space in which a high-dimensional
system can be approximated. These established methods can be used with machine learning
methods, potentially reducing computational costs further than either method individually. A
method to utilise autoencoders with a projection-based framework is also presented here.
The structure of a reactor can be broken down, forming a hierarchy which starts with the
reactor core, which is populated by fuel assemblies, which are then populated by fuel rods.
This hierarchy means that materials are repeated within a solution, and many existing methods
do not capitalise on this and instead resolve the entire global domain. This research presents
two ways to utilise this structure with ROM. The first involves combining it with domain
decomposition, producing ROMs for the sub-structures. The second presents a hierarchical
interpolating method, reducing the number of sub-domains within the solution that need to be
resolved.Open Acces
Recommended from our members
From multiscale modeling to metamodeling of geomechanics problems
In numerical simulations of geomechanics problems, a grand challenge consists of overcoming the difficulties in making accurate and robust predictions by revealing the true mechanisms in particle interactions, fluid flow inside pore spaces, and hydromechanical coupling effect between the solid and fluid constituents, from microscale to mesoscale, and to macroscale. While simulation tools incorporating subscale physics can provide detailed insights and accurate material properties to macroscale simulations via computational homogenizations, these numerical simulations are often too computational demanding to be directly used across multiple scales. Recent breakthroughs of Artificial Intelligence (AI) via machine learning have great potential to overcome these barriers, as evidenced by their great success in many applications such as image recognition, natural language processing, and strategy exploration in games. The AI can achieve super-human performance level in a large number of applications, and accomplish tasks that were thought to be not feasible due to the limitations of human and previous computer algorithms. Yet, machine learning approaches can also suffer from overfitting, lack of interpretability, and lack of reliability. Thus the application of machine learning into generation of accurate and reliable surrogate constitutive models for geomaterials with multiscale and multiphysics is not trivial. For this purpose, we propose to establish an integrated modeling process for automatic designing, training, validating, and falsifying of constitutive models, or "metamodeling". This dissertation focuses on our efforts in laying down step-by-step the necessary theoretical and technical foundations for the multiscale metamodeling framework.
The first step is to develop multiscale hydromechanical homogenization frameworks for both bulk granular materials and granular interfaces, with their behaviors homogenized from subscale microstructural simulations. For efficient simulations of field-scale geomechanics problems across more than two scales, we develop a hybrid data-driven method designed to capture the multiscale hydro-mechanical coupling effect of porous media with pores of various different sizes. By using sub-scale simulations to generate database to train material models, an offline homogenization procedure is used to replace the up-scaling procedure to generate path-dependent cohesive laws for localized physical discontinuities at both grain and specimen scales.
To enable AI in taking over the trial-and-error tasks in the constitutive modeling process, we introduce a novel âmetamodelingâ framework that employs both graph theory and deep reinforcement learning (DRL) to generate accurate, physics compatible and interpretable surrogate machine learning models. The process of writing constitutive models is simplified as a sequence of forming graph edges with the goal of maximizing the model score (a function of accuracy, robustness and forward prediction quality). By using neural networks to estimate policies and state values, the computer agent is able to efficiently self-improve the constitutive models generated through self-playing.
To overcome the obstacle of limited information in geomechanics, we improve the efficiency in utilization of experimental data by a multi-agent cooperative metamodeling framework to provide guidance on database generation and constitutive modeling at the same time. The modeler agent in the framework focuses on evaluating all modeling options (from domain expertsâ knowledge or machine learning) in a directed multigraph of elasto-plasticity theory, and finding the optimal path that links the source of the directed graph (e.g., strain history) to the target (e.g., stress). Meanwhile, the data agent focuses on collecting data from real or virtual experiments, interacts with the modeler agent sequentially and generates the database for model calibration to optimize the prediction accuracy. Finally, we design a non-cooperative meta-modeling framework that focuses on automatically developing strategies that simultaneously generate experimental data to calibrate model parameters and explore weakness of a known constitutive model until the strengths and weaknesses of the constitutive law on the application range can be identified through competition. These tasks are enabled by a zero-sum reward system of the metamodeling game and robust adversarial reinforcement learning techniques
Mecanismos biofĂsicos y fuentes de los potenciales extracelulares en el hipocampo
Tesis inĂ©dita de la Universidad Complutense de Madrid, Facultad de Ciencias FĂsicas, Departamento de FĂsica Aplicada III (Electricidad y ElectrĂłnica), leĂda el 20-11-2015Depto. de Estructura de la Materia, FĂsica TĂ©rmica y ElectrĂłnicaFac. de Ciencias FĂsicasTRUEunpu
- âŠ