44,012 research outputs found
The Expected Norm of a Sum of Independent Random Matrices: An Elementary Approach
In contemporary applied and computational mathematics, a frequent challenge
is to bound the expectation of the spectral norm of a sum of independent random
matrices. This quantity is controlled by the norm of the expected square of the
random matrix and the expectation of the maximum squared norm achieved by one
of the summands; there is also a weak dependence on the dimension of the random
matrix. The purpose of this paper is to give a complete, elementary proof of
this important, but underappreciated, inequality.Comment: 20 page
A Stochastic View of Optimal Regret through Minimax Duality
We study the regret of optimal strategies for online convex optimization
games. Using von Neumann's minimax theorem, we show that the optimal regret in
this adversarial setting is closely related to the behavior of the empirical
minimization algorithm in a stochastic process setting: it is equal to the
maximum, over joint distributions of the adversary's action sequence, of the
difference between a sum of minimal expected losses and the minimal empirical
loss. We show that the optimal regret has a natural geometric interpretation,
since it can be viewed as the gap in Jensen's inequality for a concave
functional--the minimizer over the player's actions of expected loss--defined
on a set of probability distributions. We use this expression to obtain upper
and lower bounds on the regret of an optimal strategy for a variety of online
learning problems. Our method provides upper bounds without the need to
construct a learning algorithm; the lower bounds provide explicit optimal
strategies for the adversary
Low-Cost Learning via Active Data Procurement
We design mechanisms for online procurement of data held by strategic agents
for machine learning tasks. The challenge is to use past data to actively price
future data and give learning guarantees even when an agent's cost for
revealing her data may depend arbitrarily on the data itself. We achieve this
goal by showing how to convert a large class of no-regret algorithms into
online posted-price and learning mechanisms. Our results in a sense parallel
classic sample complexity guarantees, but with the key resource being money
rather than quantity of data: With a budget constraint , we give robust risk
(predictive error) bounds on the order of . Because we use an
active approach, we can often guarantee to do significantly better by
leveraging correlations between costs and data.
Our algorithms and analysis go through a model of no-regret learning with
arriving pairs (cost, data) and a budget constraint of . Our regret bounds
for this model are on the order of and we give lower bounds on the
same order.Comment: Full version of EC 2015 paper. Color recommended for figures but
nonessential. 36 pages, of which 12 appendi
Lower Bounds for Ground States of Condensed Matter Systems
Standard variational methods tend to obtain upper bounds on the ground state
energy of quantum many-body systems. Here we study a complementary method that
determines lower bounds on the ground state energy in a systematic fashion,
scales polynomially in the system size and gives direct access to correlation
functions. This is achieved by relaxing the positivity constraint on the
density matrix and replacing it by positivity constraints on moment matrices,
thus yielding a semi-definite programme. Further, the number of free parameters
in the optimization problem can be reduced dramatically under the assumption of
translational invariance. A novel numerical approach, principally a combination
of a projected gradient algorithm with Dykstra's algorithm, for solving the
optimization problem in a memory-efficient manner is presented and a proof of
convergence for this iterative method is given. Numerical experiments that
determine lower bounds on the ground state energies for the Ising and
Heisenberg Hamiltonians confirm that the approach can be applied to large
systems, especially under the assumption of translational invariance.Comment: 16 pages, 4 figures, replaced with published versio
Kullback-Leibler aggregation and misspecified generalized linear models
In a regression setup with deterministic design, we study the pure
aggregation problem and introduce a natural extension from the Gaussian
distribution to distributions in the exponential family. While this extension
bears strong connections with generalized linear models, it does not require
identifiability of the parameter or even that the model on the systematic
component is true. It is shown that this problem can be solved by constrained
and/or penalized likelihood maximization and we derive sharp oracle
inequalities that hold both in expectation and with high probability. Finally
all the bounds are proved to be optimal in a minimax sense.Comment: Published in at http://dx.doi.org/10.1214/11-AOS961 the Annals of
Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical
Statistics (http://www.imstat.org
Query Complexity of Derivative-Free Optimization
This paper provides lower bounds on the convergence rate of Derivative Free
Optimization (DFO) with noisy function evaluations, exposing a fundamental and
unavoidable gap between the performance of algorithms with access to gradients
and those with access to only function evaluations. However, there are
situations in which DFO is unavoidable, and for such situations we propose a
new DFO algorithm that is proved to be near optimal for the class of strongly
convex objective functions. A distinctive feature of the algorithm is that it
uses only Boolean-valued function comparisons, rather than function
evaluations. This makes the algorithm useful in an even wider range of
applications, such as optimization based on paired comparisons from human
subjects, for example. We also show that regardless of whether DFO is based on
noisy function evaluations or Boolean-valued function comparisons, the
convergence rate is the same
- …