237 research outputs found
Nonparametric maximum likelihood approach to multiple change-point problems
In multiple change-point problems, different data segments often follow
different distributions, for which the changes may occur in the mean, scale or
the entire distribution from one segment to another. Without the need to know
the number of change-points in advance, we propose a nonparametric maximum
likelihood approach to detecting multiple change-points. Our method does not
impose any parametric assumption on the underlying distributions of the data
sequence, which is thus suitable for detection of any changes in the
distributions. The number of change-points is determined by the Bayesian
information criterion and the locations of the change-points can be estimated
via the dynamic programming algorithm and the use of the intrinsic order
structure of the likelihood function. Under some mild conditions, we show that
the new method provides consistent estimation with an optimal rate. We also
suggest a prescreening procedure to exclude most of the irrelevant points prior
to the implementation of the nonparametric likelihood method. Simulation
studies show that the proposed method has satisfactory performance of
identifying multiple change-points in terms of estimation accuracy and
computation time.Comment: Published in at http://dx.doi.org/10.1214/14-AOS1210 the Annals of
Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical
Statistics (http://www.imstat.org
Model-free controlled variable selection via data splitting
Addressing the simultaneous identification of contributory variables while
controlling the false discovery rate (FDR) in high-dimensional data is a
crucial statistical challenge. In this paper, we propose a novel model-free
variable selection procedure in sufficient dimension reduction framework via a
data splitting technique. The variable selection problem is first converted to
a least squares procedure with several response transformations. We construct a
series of statistics with global symmetry property and leverage the symmetry to
derive a data-driven threshold aimed at error rate control. Our approach
demonstrates the capability for achieving finite-sample and asymptotic FDR
control under mild theoretical conditions. Numerical experiments confirm that
our procedure has satisfactory FDR control and higher power compared with
existing methods.Comment: 55 pages, 5 figures, 6 table
Covariance Regression with High-Dimensional Predictors
In the high-dimensional landscape, addressing the challenges of covariance
regression with high-dimensional covariates has posed difficulties for
conventional methodologies. This paper addresses these hurdles by presenting a
novel approach for high-dimensional inference with covariance matrix outcomes.
The proposed methodology is illustrated through its application in elucidating
brain coactivation patterns observed in functional magnetic resonance imaging
(fMRI) experiments and unraveling complex associations within anatomical
connections between brain regions identified through diffusion tensor imaging
(DTI). In the pursuit of dependable statistical inference, we introduce an
integrative approach based on penalized estimation. This approach combines data
splitting, variable selection, aggregation of low-dimensional estimators, and
robust variance estimation. It enables the construction of reliable confidence
intervals for covariate coefficients, supported by theoretical confidence
levels under specified conditions, where asymptotic distributions are provided.
Through various types of simulation studies, the proposed approach performs
well for covariance regression in the presence of high-dimensional covariates.
This innovative approach is applied to the Lifespan Human Connectome Project
(HCP) Aging Study, which aims to uncover a typical aging trajectory and
variations in the brain connectome among mature and older adults. The proposed
approach effectively identifies brain networks and associated predictors of
white matter integrity, aligning with established knowledge of the human brain
Zipper: Addressing degeneracy in algorithm-agnostic inference
The widespread use of black box prediction methods has sparked an increasing
interest in algorithm/model-agnostic approaches for quantifying
goodness-of-fit, with direct ties to specification testing, model selection and
variable importance assessment. A commonly used framework involves defining a
predictiveness criterion, applying a cross-fitting procedure to estimate the
predictiveness, and utilizing the difference in estimated predictiveness
between two models as the test statistic. However, even after standardization,
the test statistic typically fails to converge to a non-degenerate distribution
under the null hypothesis of equal goodness, leading to what is known as the
degeneracy issue. To addresses this degeneracy issue, we present a simple yet
effective device, Zipper. It draws inspiration from the strategy of additional
splitting of testing data, but encourages an overlap between two testing data
splits in predictiveness evaluation. Zipper binds together the two overlapping
splits using a slider parameter that controls the proportion of overlap. Our
proposed test statistic follows an asymptotically normal distribution under the
null hypothesis for any fixed slider value, guaranteeing valid size control
while enhancing power by effective data reuse. Finite-sample experiments
demonstrate that our procedure, with a simple choice of the slider, works well
across a wide range of settings
CAP: A General Algorithm for Online Selective Conformal Prediction with FCR Control
We study the problem of post-selection predictive inference in an online
fashion. To avoid devoting resources to unimportant units, a preliminary
selection of the current individual before reporting its prediction interval is
common and meaningful in online predictive tasks. Since the online selection
causes a temporal multiplicity in the selected prediction intervals, it is
important to control the real-time false coverage-statement rate (FCR) which
measures the overall miscoverage level. We develop a general framework named
CAP (Calibration after Adaptive Pick) that performs an adaptive pick rule on
historical data to construct a calibration set if the current individual is
selected and then outputs a conformal prediction interval for the unobserved
label. We provide tractable procedures for constructing the calibration set for
popular online selection rules. We proved that CAP can achieve an exact
selection-conditional coverage guarantee in the finite-sample and
distribution-free regimes. To account for the distribution shift in online
data, we also embed CAP into some recent dynamic conformal prediction
algorithms and show that the proposed method can deliver long-run FCR control.
Numerical results on both synthetic and real data corroborate that CAP can
effectively control FCR around the target level and yield more narrowed
prediction intervals over existing baselines across various settings
Selective Conformal Inference with FCR Control
Conformal inference is a popular tool for constructing prediction intervals
(PI). We consider here the scenario of post-selection/selective conformal
inference, that is PIs are reported only for individuals selected from an
unlabeled test data. To account for multiplicity, we develop a general split
conformal framework to construct selective PIs with the false
coverage-statement rate (FCR) control. We first investigate the Benjamini and
Yekutieli (2005)'s FCR-adjusted method in the present setting, and show that it
is able to achieve FCR control but yields uniformly inflated PIs. We then
propose a novel solution to the problem, named as Selective COnditional
conformal Predictions (SCOP), which entails performing selection procedures on
both calibration set and test set and construct marginal conformal PIs on the
selected sets by the aid of conditional empirical distribution obtained by the
calibration set. Under a unified framework and exchangeable assumptions, we
show that the SCOP can exactly control the FCR. More importantly, we provide
non-asymptotic miscoverage bounds for a general class of selection procedures
beyond exchangeablity and discuss the conditions under which the SCOP is able
to control the FCR. As special cases, the SCOP with quantile-based selection or
conformal p-values-based multiple testing procedures enjoys valid coverage
guarantee under mild conditions. Numerical results confirm the effectiveness
and robustness of SCOP in FCR control and show that it achieves more narrowed
PIs over existing methods in many settings
Robust multicategory support matrix machines
We consider the classification problem when the input features are represented as matrices rather than vectors. To preserve the intrinsic structures for classification, a successful method is the support matrix machine (SMM) in Luo et al. (in: Proceedings of the 32nd international conference on machine learning, Lille, France, no 1, pp 938-947, 2015), which optimizes an objective function with a hinge loss plus a so-called spectral elastic net penalty. However, the issues of extending SMM to multicategory classification still remain. Moreover, in practice, it is common to see the training data contaminated by outlying observations, which can affect the robustness of existing matrix classification methods. In this paper, we address these issues by introducing a robust angle-based classifier, which boils down binary and multicategory problems to a unified framework. Benefitting from the use of truncated hinge loss functions, the proposed classifier achieves certain robustness to outliers. The underlying optimization model becomes nonconvex, but admits a natural DC (difference of two convex functions) representation. We develop a new and efficient algorithm by incorporating the DC algorithm and primal-dual first-order methods together. The proposed DC algorithm adaptively chooses the accuracy of the subproblem at each iteration while guaranteeing the overall convergence of the algorithm. The use of primal-dual methods removes a natural complexity of the linear operator in the subproblems and enables us to use the proximal operator of the objective functions, and matrix-vector operations. This advantage allows us to solve large-scale problems efficiently. Theoretical and numerical results indicate that for problems with potential outliers, our method can be highly competitive among existing methods
- …