294 research outputs found

    Smoothed Analysis in Unsupervised Learning via Decoupling

    Full text link
    Smoothed analysis is a powerful paradigm in overcoming worst-case intractability in unsupervised learning and high-dimensional data analysis. While polynomial time smoothed analysis guarantees have been obtained for worst-case intractable problems like tensor decompositions and learning mixtures of Gaussians, such guarantees have been hard to obtain for several other important problems in unsupervised learning. A core technical challenge in analyzing algorithms is obtaining lower bounds on the least singular value for random matrix ensembles with dependent entries, that are given by low-degree polynomials of a few base underlying random variables. In this work, we address this challenge by obtaining high-confidence lower bounds on the least singular value of new classes of structured random matrix ensembles of the above kind. We then use these bounds to design algorithms with polynomial time smoothed analysis guarantees for the following three important problems in unsupervised learning: 1. Robust subspace recovery, when the fraction α\alpha of inliers in the d-dimensional subspace T⊂RnT \subset \mathbb{R}^n is at least α>(d/n)ℓ\alpha > (d/n)^\ell for any constant integer ℓ>0\ell>0. This contrasts with the known worst-case intractability when α<d/n\alpha< d/n, and the previous smoothed analysis result which needed α>d/n\alpha > d/n (Hardt and Moitra, 2013). 2. Learning overcomplete hidden markov models, where the size of the state space is any polynomial in the dimension of the observations. This gives the first polynomial time guarantees for learning overcomplete HMMs in a smoothed analysis model. 3. Higher order tensor decompositions, where we generalize the so-called FOOBI algorithm of Cardoso to find order-ℓ\ell rank-one tensors in a subspace. This allows us to obtain polynomially robust decomposition algorithms for 2ℓ2\ell'th order tensors with rank O(nℓ)O(n^{\ell}).Comment: 44 page

    Unsupervised discovery of temporal sequences in high-dimensional datasets, with applications to neuroscience.

    Get PDF
    Identifying low-dimensional features that describe large-scale neural recordings is a major challenge in neuroscience. Repeated temporal patterns (sequences) are thought to be a salient feature of neural dynamics, but are not succinctly captured by traditional dimensionality reduction techniques. Here, we describe a software toolbox-called seqNMF-with new methods for extracting informative, non-redundant, sequences from high-dimensional neural data, testing the significance of these extracted patterns, and assessing the prevalence of sequential structure in data. We test these methods on simulated data under multiple noise conditions, and on several real neural and behavioral datas. In hippocampal data, seqNMF identifies neural sequences that match those calculated manually by reference to behavioral events. In songbird data, seqNMF discovers neural sequences in untutored birds that lack stereotyped songs. Thus, by identifying temporal structure directly from neural data, seqNMF enables dissection of complex neural circuits without relying on temporal references from stimuli or behavioral outputs

    Concentration inequalities for random tensors

    Get PDF
    We show how to extend several basic concentration inequalities for simple random tensors X=x1⊗⋯⊗xdX = x_1 \otimes \cdots \otimes x_d where all xkx_k are independent random vectors in Rn\mathbb{R}^n with independent coefficients. The new results have optimal dependence on the dimension nn and the degree dd. As an application, we show that random tensors are well conditioned: (1−o(1))nd(1-o(1)) n^d independent copies of the simple random tensor X∈RndX \in \mathbb{R}^{n^d} are far from being linearly dependent with high probability. We prove this fact for any degree d=o(n/log⁡n)d = o(\sqrt{n/\log n}) and conjecture that it is true for any d=O(n)d = O(n).Comment: A few more typos were correcte

    A hierarchical reduced-order model applied to nuclear reactors

    Get PDF
    Modelling the neutron transport of a nuclear reactor is a very computationally demanding task that requires a large number of degrees of freedom to accurately capture all of the physics. For a complete reactor picture, other physics must be incorporated, through coupling, further exacerbating the computational demand. Computational modelling has many benefits: optimisation, real-time analysis, and safety analysis are some of the more important ones. However, nuclear modelling has yet to capitalise on these, and existing approaches are too computationally demanding. Machine Learning has seen incredible growth over the last decade, but it has yet to be utilised within the nuclear modelling community to the same extent. The frameworks available represent incredibly efficient and optimised code, having been written to run on GPUs and AI computers. Presented here is a physics-driven neural network that solves neutron transport, first for the diffusion approximation and then extended to the whole transport problem. One method that can potentially reduce the computational complexity is Reduced-Order Modelling (ROM), which is a way to define a low-dimensional space in which a high-dimensional system can be approximated. These established methods can be used with machine learning methods, potentially reducing computational costs further than either method individually. A method to utilise autoencoders with a projection-based framework is also presented here. The structure of a reactor can be broken down, forming a hierarchy which starts with the reactor core, which is populated by fuel assemblies, which are then populated by fuel rods. This hierarchy means that materials are repeated within a solution, and many existing methods do not capitalise on this and instead resolve the entire global domain. This research presents two ways to utilise this structure with ROM. The first involves combining it with domain decomposition, producing ROMs for the sub-structures. The second presents a hierarchical interpolating method, reducing the number of sub-domains within the solution that need to be resolved.Open Acces

    Mecanismos biofĂ­sicos y fuentes de los potenciales extracelulares en el hipocampo

    Get PDF
    Tesis inédita de la Universidad Complutense de Madrid, Facultad de Ciencias Físicas, Departamento de Física Aplicada III (Electricidad y Electrónica), leída el 20-11-2015Depto. de Estructura de la Materia, Física Térmica y ElectrónicaFac. de Ciencias FísicasTRUEunpu
    • 

    corecore