10,273 research outputs found

    Algorithms for Learning Sparse Additive Models with Interactions in High Dimensions

    Get PDF
    A function f:Rdβ†’Rf: \mathbb{R}^d \rightarrow \mathbb{R} is a Sparse Additive Model (SPAM), if it is of the form f(x)=βˆ‘l∈SΟ•l(xl)f(\mathbf{x}) = \sum_{l \in \mathcal{S}}\phi_{l}(x_l) where SβŠ‚[d]\mathcal{S} \subset [d], ∣S∣β‰ͺd|\mathcal{S}| \ll d. Assuming Ο•\phi's, S\mathcal{S} to be unknown, there exists extensive work for estimating ff from its samples. In this work, we consider a generalized version of SPAMs, that also allows for the presence of a sparse number of second order interaction terms. For some S1βŠ‚[d],S2βŠ‚([d]2)\mathcal{S}_1 \subset [d], \mathcal{S}_2 \subset {[d] \choose 2}, with ∣S1∣β‰ͺd,∣S2∣β‰ͺd2|\mathcal{S}_1| \ll d, |\mathcal{S}_2| \ll d^2, the function ff is now assumed to be of the form: βˆ‘p∈S1Ο•p(xp)+βˆ‘(l,lβ€²)∈S2Ο•(l,lβ€²)(xl,xlβ€²)\sum_{p \in \mathcal{S}_1}\phi_{p} (x_p) + \sum_{(l,l^{\prime}) \in \mathcal{S}_2}\phi_{(l,l^{\prime})} (x_l,x_{l^{\prime}}). Assuming we have the freedom to query ff anywhere in its domain, we derive efficient algorithms that provably recover S1,S2\mathcal{S}_1,\mathcal{S}_2 with finite sample bounds. Our analysis covers the noiseless setting where exact samples of ff are obtained, and also extends to the noisy setting where the queries are corrupted with noise. For the noisy setting in particular, we consider two noise models namely: i.i.d Gaussian noise and arbitrary but bounded noise. Our main methods for identification of S2\mathcal{S}_2 essentially rely on estimation of sparse Hessian matrices, for which we provide two novel compressed sensing based schemes. Once S1,S2\mathcal{S}_1, \mathcal{S}_2 are known, we show how the individual components Ο•p\phi_p, Ο•(l,lβ€²)\phi_{(l,l^{\prime})} can be estimated via additional queries of ff, with uniform error bounds. Lastly, we provide simulation results on synthetic data that validate our theoretical findings.Comment: To appear in Information and Inference: A Journal of the IMA. Made following changes after review process: (a) Corrected typos throughout the text. (b) Corrected choice of sampling distribution in Section 5, see eqs. (5.2), (5.3). (c) More detailed comparison with existing work in Section 8. (d) Added Section B in appendix on roots of cubic equatio

    Sparse Additive Models

    Full text link
    We present a new class of methods for high-dimensional nonparametric regression and classification called sparse additive models (SpAM). Our methods combine ideas from sparse linear modeling and additive nonparametric regression. We derive an algorithm for fitting the models that is practical and effective even when the number of covariates is larger than the sample size. SpAM is closely related to the COSSO model of Lin and Zhang (2006), but decouples smoothing and sparsity, enabling the use of arbitrary nonparametric smoothers. An analysis of the theoretical properties of SpAM is given. We also study a greedy estimator that is a nonparametric version of forward stepwise regression. Empirical results on synthetic and real data are presented, showing that SpAM can be effective in fitting sparse nonparametric models in high dimensional data
    • …
    corecore