4,181 research outputs found
Tensor decomposition with generalized lasso penalties
We present an approach for penalized tensor decomposition (PTD) that
estimates smoothly varying latent factors in multi-way data. This generalizes
existing work on sparse tensor decomposition and penalized matrix
decompositions, in a manner parallel to the generalized lasso for regression
and smoothing problems. Our approach presents many nontrivial challenges at the
intersection of modeling and computation, which are studied in detail. An
efficient coordinate-wise optimization algorithm for (PTD) is presented, and
its convergence properties are characterized. The method is applied both to
simulated data and real data on flu hospitalizations in Texas. These results
show that our penalized tensor decomposition can offer major improvements on
existing methods for analyzing multi-way data that exhibit smooth spatial or
temporal features
Tensor Graphical Lasso (TeraLasso)
This paper introduces a multi-way tensor generalization of the Bigraphical
Lasso (BiGLasso), which uses a two-way sparse Kronecker-sum multivariate-normal
model for the precision matrix to parsimoniously model conditional dependence
relationships of matrix-variate data based on the Cartesian product of graphs.
We call this generalization the {\bf Te}nsor g{\bf ra}phical Lasso (TeraLasso).
We demonstrate using theory and examples that the TeraLasso model can be
accurately and scalably estimated from very limited data samples of high
dimensional variables with multiway coordinates such as space, time and
replicates. Statistical consistency and statistical rates of convergence are
established for both the BiGLasso and TeraLasso estimators of the precision
matrix and estimators of its support (non-sparsity) set, respectively. We
propose a scalable composite gradient descent algorithm and analyze the
computational convergence rate, showing that the composite gradient descent
algorithm is guaranteed to converge at a geometric rate to the global minimizer
of the TeraLasso objective function. Finally, we illustrate the TeraLasso using
both simulation and experimental data from a meteorological dataset, showing
that we can accurately estimate precision matrices and recover meaningful
conditional dependency graphs from high dimensional complex datasets.Comment: accepted to JRSS-
Dynamic Tensor Clustering
Dynamic tensor data are becoming prevalent in numerous applications. Existing
tensor clustering methods either fail to account for the dynamic nature of the
data, or are inapplicable to a general-order tensor. Also there is often a gap
between statistical guarantee and computational efficiency for existing tensor
clustering solutions. In this article, we aim to bridge this gap by proposing a
new dynamic tensor clustering method, which takes into account both sparsity
and fusion structures, and enjoys strong statistical guarantees as well as high
computational efficiency. Our proposal is based upon a new structured tensor
factorization that encourages both sparsity and smoothness in parameters along
the specified tensor modes. Computationally, we develop a highly efficient
optimization algorithm that benefits from substantial dimension reduction. In
theory, we first establish a non-asymptotic error bound for the estimator from
the structured tensor factorization. Built upon this error bound, we then
derive the rate of convergence of the estimated cluster centers, and show that
the estimated clusters recover the true cluster structures with a high
probability. Moreover, our proposed method can be naturally extended to
co-clustering of multiple modes of the tensor data. The efficacy of our
approach is illustrated via simulations and a brain dynamic functional
connectivity analysis from an Autism spectrum disorder study.Comment: Accepted at Journal of the American Statistical Associatio
- …