7,380 research outputs found
Four lectures on probabilistic methods for data science
Methods of high-dimensional probability play a central role in applications
for statistics, signal processing theoretical computer science and related
fields. These lectures present a sample of particularly useful tools of
high-dimensional probability, focusing on the classical and matrix Bernstein's
inequality and the uniform matrix deviation inequality. We illustrate these
tools with applications for dimension reduction, network analysis, covariance
estimation, matrix completion and sparse signal recovery. The lectures are
geared towards beginning graduate students who have taken a rigorous course in
probability but may not have any experience in data science applications.Comment: Lectures given at 2016 PCMI Graduate Summer School in Mathematics of
Data. Some typos, inaccuracies fixe
A Comparison between Fixed-Basis and Variable-Basis Schemes for Function Approximation and Functional Optimization
Fixed-basis and variable-basis approximation schemes are compared for the problems of function approximation and functional optimization (also known as infinite programming). Classes of problems are investigated for which variable-basis schemes with sigmoidal computational
units perform better than fixed-basis ones, in terms of the minimum number of computational units needed to achieve a desired error in function approximation or approximate optimization. Previously known bounds on the accuracy are extended, with better rates, to families o
Lepskii Principle in Supervised Learning
In the setting of supervised learning using reproducing kernel methods, we
propose a data-dependent regularization parameter selection rule that is
adaptive to the unknown regularity of the target function and is optimal both
for the least-square (prediction) error and for the reproducing kernel Hilbert
space (reconstruction) norm error. It is based on a modified Lepskii balancing
principle using a varying family of norms
- …