4,114,015 research outputs found
Longitudinal Functional Data Analysis
We consider analysis of dependent functional data that are correlated because
of a longitudinal-based design: each subject is observed at repeated time
visits and for each visit we record a functional variable. We propose a novel
parsimonious modeling framework for the repeatedly observed functional
variables that allows to extract low dimensional features. The proposed
methodology accounts for the longitudinal design, is designed for the study of
the dynamic behavior of the underlying process, and is computationally fast.
Theoretical properties of this framework are studied and numerical
investigation confirms excellent behavior in finite samples. The proposed
method is motivated by and applied to a diffusion tensor imaging study of
multiple sclerosis. Using Shiny (Chang et al., 2015) we implement interactive
plots to help visualize longitudinal functional data as well as the various
components and prediction obtained using the proposed method.Comment: 32 pages, 4 figure
Weakly dependent functional data
Functional data often arise from measurements on fine time grids and are
obtained by separating an almost continuous time record into natural
consecutive intervals, for example, days. The functions thus obtained form a
functional time series, and the central issue in the analysis of such data
consists in taking into account the temporal dependence of these functional
observations. Examples include daily curves of financial transaction data and
daily patterns of geophysical and environmental data. For scalar and vector
valued stochastic processes, a large number of dependence notions have been
proposed, mostly involving mixing type distances between -algebras. In
time series analysis, measures of dependence based on moments have proven most
useful (autocovariances and cumulants). We introduce a moment-based notion of
dependence for functional time series which involves -dependence. We show
that it is applicable to linear as well as nonlinear functional time series.
Then we investigate the impact of dependence thus quantified on several
important statistical procedures for functional data. We study the estimation
of the functional principal components, the long-run covariance matrix, change
point detection and the functional linear model. We explain when temporal
dependence affects the results obtained for i.i.d. functional observations and
when these results are robust to weak dependence.Comment: Published in at http://dx.doi.org/10.1214/09-AOS768 the Annals of
Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical
Statistics (http://www.imstat.org
Fast DD-classification of functional data
A fast nonparametric procedure for classifying functional data is introduced.
It consists of a two-step transformation of the original data plus a classifier
operating on a low-dimensional hypercube. The functional data are first mapped
into a finite-dimensional location-slope space and then transformed by a
multivariate depth function into the -plot, which is a subset of the unit
hypercube. This transformation yields a new notion of depth for functional
data. Three alternative depth functions are employed for this, as well as two
rules for the final classification on . The resulting classifier has
to be cross-validated over a small range of parameters only, which is
restricted by a Vapnik-Cervonenkis bound. The entire methodology does not
involve smoothing techniques, is completely nonparametric and allows to achieve
Bayes optimality under standard distributional settings. It is robust,
efficiently computable, and has been implemented in an R environment.
Applicability of the new approach is demonstrated by simulations as well as a
benchmark study
Nonparametric Hierarchical Clustering of Functional Data
In this paper, we deal with the problem of curves clustering. We propose a
nonparametric method which partitions the curves into clusters and discretizes
the dimensions of the curve points into intervals. The cross-product of these
partitions forms a data-grid which is obtained using a Bayesian model selection
approach while making no assumptions regarding the curves. Finally, a
post-processing technique, aiming at reducing the number of clusters in order
to improve the interpretability of the clustering, is proposed. It consists in
optimally merging the clusters step by step, which corresponds to an
agglomerative hierarchical classification whose dissimilarity measure is the
variation of the criterion. Interestingly this measure is none other than the
sum of the Kullback-Leibler divergences between clusters distributions before
and after the merges. The practical interest of the approach for functional
data exploratory analysis is presented and compared with an alternative
approach on an artificial and a real world data set
- …
