36 research outputs found

    The Loss Rank Principle for Model Selection

    Full text link
    We introduce a new principle for model selection in regression and classification. Many regression models are controlled by some smoothness or flexibility or complexity parameter c, e.g. the number of neighbors to be averaged over in k nearest neighbor (kNN) regression or the polynomial degree in regression with polynomials. Let f_D^c be the (best) regressor of complexity c on data D. A more flexible regressor can fit more data D' well than a more rigid one. If something (here small loss) is easy to achieve it's typically worth less. We define the loss rank of f_D^c as the number of other (fictitious) data D' that are fitted better by f_D'^c than D is fitted by f_D^c. We suggest selecting the model complexity c that has minimal loss rank (LoRP). Unlike most penalized maximum likelihood variants (AIC,BIC,MDL), LoRP only depends on the regression function and loss function. It works without a stochastic noise model, and is directly applicable to any non-parametric regressor, like kNN. In this paper we formalize, discuss, and motivate LoRP, study it for specific regression problems, in particular linear ones, and compare it to other model selection schemes.Comment: 16 page

    Considerations on Genre and Gender Conventions in Translating from Old English

    Get PDF
    The Old English poem The Wife's Lament is an extremely conventional and, at the same time, original text. It portrays a female character suffering for the absence of her loved one, through the framework of the so-called 'elegiac' style and a mainly heroic vocabulary. The traditional exile theme is, thus, interwoven with the uncommon motif of love sickness. While this appraisal of the poem is the most widely accepted one, disagreement still remains about the translation of some keywords, strictly related to the exile theme, such as sīþ or wræcsīþ. The aim of this paper is to examine diverging readings and glosses of the above mentioned 'exilic/elegiac' keywords, and to show that an accurate translation should not neglect a thorough appraisal of the text in its complexity and the association with related literary patterns and imagery in other poetic and prose texts

    Part Segmentation for Object Recognition

    No full text

    Convergence of Discrete MDL for Sequential Prediction

    No full text
    We study the properties of the Minimum Description Length principle for sequence prediction, considering a two-part MDL estimator which is chosen from a countable class of models. This applies in particular to the important case of universal sequence prediction, where the model class corresponds to all algorithms for some fixed universal Turing machine (this correspondence is by enumerable semimeasures, hence the resulting models are stochastic). We prove convergence theorems similar to Solomonoff’s theorem of universal induction, which also holds for general Bayes mixtures. The bound characterizing the convergence speed for MDL predictions is exponentially larger as compared to Bayes mixtures. We observe that there are at least three different ways of using MDL for prediction. One of these has worse prediction properties, for which predictions only converge if the MDL estimator stabilizes. We establish sufficient conditions for this to occur. Finally, some immediate consequences for complexity relations and randomness criteria are proven

    Probability Model Type Sufficiency

    No full text
    We investigate the role of sufficient statistics in generalized probabilistic data mining and machine learning software frameworks. Some issues involved in the specification of a statistical model type are discussed and we show that it is beneficial to explicitly include a sufficient statistic and functions for its manipulation in the model type's specification. Instances of such types can then be used by generalized learning algorithms while maintaining optimal learning time complexity. Examples are given for problems such as incremental learning and data partitioning problems (e.g. change-point problems, decision trees and mixture models)
    corecore