6,451 research outputs found
A three domain covariance framework for EEG/MEG data
In this paper we introduce a covariance framework for the analysis of EEG and
MEG data that takes into account observed temporal stationarity on small time
scales and trial-to-trial variations. We formulate a model for the covariance
matrix, which is a Kronecker product of three components that correspond to
space, time and epochs/trials, and consider maximum likelihood estimation of
the unknown parameter values. An iterative algorithm that finds approximations
of the maximum likelihood estimates is proposed. We perform a simulation study
to assess the performance of the estimator and investigate the influence of
different assumptions about the covariance factors on the estimated covariance
matrix and on its components. Apart from that, we illustrate our method on real
EEG and MEG data sets.
The proposed covariance model is applicable in a variety of cases where
spontaneous EEG or MEG acts as source of noise and realistic noise covariance
estimates are needed for accurate dipole localization, such as in evoked
activity studies, or where the properties of spontaneous EEG or MEG are
themselves the topic of interest, such as in combined EEG/fMRI experiments in
which the correlation between EEG and fMRI signals is investigated.Comment: 25 pages, 8 figures, 1 tabl
Group Symmetry and non-Gaussian Covariance Estimation
We consider robust covariance estimation with group symmetry constraints.
Non-Gaussian covariance estimation, e.g., Tyler scatter estimator and
Multivariate Generalized Gaussian distribution methods, usually involve
non-convex minimization problems. Recently, it was shown that the underlying
principle behind their success is an extended form of convexity over the
geodesics in the manifold of positive definite matrices. A modern approach to
improve estimation accuracy is to exploit prior knowledge via additional
constraints, e.g., restricting the attention to specific classes of covariances
which adhere to prior symmetry structures. In this paper, we prove that such
group symmetry constraints are also geodesically convex and can therefore be
incorporated into various non-Gaussian covariance estimators. Practical
examples of such sets include: circulant, persymmetric and complex/quaternion
proper structures. We provide a simple numerical technique for finding maximum
likelihood estimates under such constraints, and demonstrate their performance
advantage using synthetic experiments
Penalized maximum likelihood for multivariate Gaussian mixture
In this paper, we first consider the parameter estimation of a multivariate
random process distribution using multivariate Gaussian mixture law. The labels
of the mixture are allowed to have a general probability law which gives the
possibility to modelize a temporal structure of the process under study. We
generalize the case of univariate Gaussian mixture in [Ridolfi99] to show that
the likelihood is unbounded and goes to infinity when one of the covariance
matrices approaches the boundary of singularity of the non negative definite
matrices set. We characterize the parameter set of these singularities. As a
solution to this degeneracy problem, we show that the penalization of the
likelihood by an Inverse Wishart prior on covariance matrices results to a
penalized or maximum a posteriori criterion which is bounded. Then, the
existence of positive definite matrices optimizing this criterion can be
guaranteed. We also show that with a modified EM procedure or with a Bayesian
sampling scheme, we can constrain covariance matrices to belong to a particular
subclass of covariance matrices. Finally, we study degeneracies in the source
separation problem where the characterization of parameter singularity set is
more complex. We show, however, that Inverse Wishart prior on covariance
matrices eliminates the degeneracies in this case too.Comment: Presented at MaxEnt01. To appear in Bayesian Inference and Maximum
Entropy Methods, B. Fry (Ed.), AIP Proceedings. 11pages, 3 Postscript figure
Foundational principles for large scale inference: Illustrations through correlation mining
When can reliable inference be drawn in the "Big Data" context? This paper
presents a framework for answering this fundamental question in the context of
correlation mining, with implications for general large scale inference. In
large scale data applications like genomics, connectomics, and eco-informatics
the dataset is often variable-rich but sample-starved: a regime where the
number of acquired samples (statistical replicates) is far fewer than the
number of observed variables (genes, neurons, voxels, or chemical
constituents). Much of recent work has focused on understanding the
computational complexity of proposed methods for "Big Data." Sample complexity
however has received relatively less attention, especially in the setting when
the sample size is fixed, and the dimension grows without bound. To
address this gap, we develop a unified statistical framework that explicitly
quantifies the sample complexity of various inferential tasks. Sampling regimes
can be divided into several categories: 1) the classical asymptotic regime
where the variable dimension is fixed and the sample size goes to infinity; 2)
the mixed asymptotic regime where both variable dimension and sample size go to
infinity at comparable rates; 3) the purely high dimensional asymptotic regime
where the variable dimension goes to infinity and the sample size is fixed.
Each regime has its niche but only the latter regime applies to exa-scale data
dimension. We illustrate this high dimensional framework for the problem of
correlation mining, where it is the matrix of pairwise and partial correlations
among the variables that are of interest. We demonstrate various regimes of
correlation mining based on the unifying perspective of high dimensional
learning rates and sample complexity for different structured covariance models
and different inference tasks
Estimation of a Covariance Matrix with Zeros
We consider estimation of the covariance matrix of a multivariate random
vector under the constraint that certain covariances are zero. We first present
an algorithm, which we call Iterative Conditional Fitting, for computing the
maximum likelihood estimator of the constrained covariance matrix, under the
assumption of multivariate normality. In contrast to previous approaches, this
algorithm has guaranteed convergence properties. Dropping the assumption of
multivariate normality, we show how to estimate the covariance matrix in an
empirical likelihood approach. These approaches are then compared via
simulation and on an example of gene expression.Comment: 25 page
Do We Really Need Both BEKK and DCC? A Tale of Two Multivariate GARCH Models
The management and monitoring of very large portfolios of financial assets are routine for many individuals and organizations. The two most widely used models of conditional covariances and correlations in the class of multivariate GARCH models are BEKK and DCC. It is well known that BEKK suffers from the archetypal “curse of dimensionalityâ€, whereas DCC does not. It is argued in this paper that this is a misleading interpretation of the suitability of the two models for use in practice. The primary purpose of this paper is to analyze the similarities and dissimilarities between BEKK and DCC, both with and without targeting, on the basis of the structural derivation of the models, the availability of analytical forms for the sufficient conditions for existence of moments, sufficient conditions for consistency and asymptotic normality of the appropriate estimators, and computational tractability for ultra large numbers of financial assets. Based on theoretical considerations, the paper sheds light on how to discriminate between BEKK and DCC in practical applications.forecasting;conditional correlations;Hadamard models;conditional covariances;diagonal models;generalized models;scalar models;targeting
- …