4,425 research outputs found

    Sharper bounds for uniformly stable algorithms

    Get PDF
    Deriving generalization bounds for stable algorithms is a classical question in learning theory taking its roots in the early works by Vapnik and Chervonenkis (1974) and Rogers and Wagner (1978). In a series of recent breakthrough papers by Feldman and Vondrak (2018, 2019), it was shown that the best known high probability upper bounds for uniformly stable learning algorithms due to Bousquet and Elisseef (2002) are sub-optimal in some natural regimes. To do so, they proved two generalization bounds that significantly outperform the simple generalization bound of Bousquet and Elisseef (2002). Feldman and Vondrak also asked if it is possible to provide sharper bounds and prove corresponding high probability lower bounds. This paper is devoted to these questions: firstly, inspired by the original arguments of Feldman and Vondrak (2019), we provide a short proof of the moment bound that implies the generalization bound stronger than both recent results in Feldman and Vondrak (2018, 2019). Secondly, we prove general lower bounds, showing that our moment bound is sharp (up to a logarithmic factor) unless some additional properties of the corresponding random variables are used. Our main probabilistic result is a general concentration inequality for weakly correlated random variables, which may be of independent interest

    Improved Second-Order Bounds for Prediction with Expert Advice

    Full text link
    This work studies external regret in sequential prediction games with both positive and negative payoffs. External regret measures the difference between the payoff obtained by the forecasting strategy and the payoff of the best action. In this setting, we derive new and sharper regret bounds for the well-known exponentially weighted average forecaster and for a new forecaster with a different multiplicative update rule. Our analysis has two main advantages: first, no preliminary knowledge about the payoff sequence is needed, not even its range; second, our bounds are expressed in terms of sums of squared payoffs, replacing larger first-order quantities appearing in previous bounds. In addition, our most refined bounds have the natural and desirable property of being stable under rescalings and general translations of the payoff sequence

    Accuracy of numerical solutions using the eulers equation residuals

    Get PDF
    In this paper we derive sorne asymptotic properties on the accuracy of numerical solutions. We sIlow tIlat the approximation error of the policy function is of the same order of magnitude as the size of the Euler equation residuals. Moreover, for bounding this approximation error tIle most relevant parameters are the discount factor and the curvature of the return function. These findings provide theoretical foundations for the construction of tests that can assess the performance of alternative computational methods

    A Statistical Perspective on Randomized Sketching for Ordinary Least-Squares

    Full text link
    We consider statistical as well as algorithmic aspects of solving large-scale least-squares (LS) problems using randomized sketching algorithms. For a LS problem with input data (X,Y)Rn×p×Rn(X, Y) \in \mathbb{R}^{n \times p} \times \mathbb{R}^n, sketching algorithms use a sketching matrix, SRr×nS\in\mathbb{R}^{r \times n} with rnr \ll n. Then, rather than solving the LS problem using the full data (X,Y)(X,Y), sketching algorithms solve the LS problem using only the sketched data (SX,SY)(SX, SY). Prior work has typically adopted an algorithmic perspective, in that it has made no statistical assumptions on the input XX and YY, and instead it has been assumed that the data (X,Y)(X,Y) are fixed and worst-case (WC). Prior results show that, when using sketching matrices such as random projections and leverage-score sampling algorithms, with p<rnp < r \ll n, the WC error is the same as solving the original problem, up to a small constant. From a statistical perspective, we typically consider the mean-squared error performance of randomized sketching algorithms, when data (X,Y)(X, Y) are generated according to a statistical model Y=Xβ+ϵY = X \beta + \epsilon, where ϵ\epsilon is a noise process. We provide a rigorous comparison of both perspectives leading to insights on how they differ. To do this, we first develop a framework for assessing algorithmic and statistical aspects of randomized sketching methods. We then consider the statistical prediction efficiency (PE) and the statistical residual efficiency (RE) of the sketched LS estimator; and we use our framework to provide upper bounds for several types of random projection and random sampling sketching algorithms. Among other results, we show that the RE can be upper bounded when p<rnp < r \ll n while the PE typically requires the sample size rr to be substantially larger. Lower bounds developed in subsequent results show that our upper bounds on PE can not be improved.Comment: 27 pages, 5 figure

    Accuracy of numerical solutions using the eulers equation residuals.

    Get PDF
    In this paper we derive sorne asymptotic properties on the accuracy of numerical solutions. We sIlow tIlat the approximation error of the policy function is of the same order of magnitude as the size of the Euler equation residuals. Moreover, for bounding this approximation error tIle most relevant parameters are the discount factor and the curvature of the return function. These findings provide theoretical foundations for the construction of tests that can assess the performance of alternative computational methods.Accuracy; Euler equation residuals; value and policy functions;
    corecore