52,148 research outputs found

    Defensive forecasting for optimal prediction with expert advice

    Get PDF
    The method of defensive forecasting is applied to the problem of prediction with expert advice for binary outcomes. It turns out that defensive forecasting is not only competitive with the Aggregating Algorithm but also handles the case of "second-guessing" experts, whose advice depends on the learner's prediction; this paper assumes that the dependence on the learner's prediction is continuous.Comment: 14 page

    Universal Learning of Repeated Matrix Games

    Full text link
    We study and compare the learning dynamics of two universal learning algorithms, one based on Bayesian learning and the other on prediction with expert advice. Both approaches have strong asymptotic performance guarantees. When confronted with the task of finding good long-term strategies in repeated 2x2 matrix games, they behave quite differently.Comment: 16 LaTeX pages, 8 eps figure

    The truth, but not yet: Avoiding naïve skepticism via explicit communication of metadisciplinary aims

    Get PDF
    Introductory students regularly endorse naïve skepticism—unsupported or uncritical doubt about the existence and universality of truth—for a variety of reasons. Though some of the reasons for students’ skepticism can be traced back to the student—for example, a desire to avoid engaging with controversial material or a desire to avoid offense—naïve skepticism is also the result of how introductory courses are taught, deemphasizing truth to promote students’ abilities to develop basic disciplinary skills. While this strategy has a number of pedagogical benefits, it prevents students in early stages of intellectual development from understanding truth as a threshold concept. Using philosophy as a case study, I argue that we can make progress against naïve skepticism by clearly discussing how metadisciplinary aims differ at the disciplinary and course levels in a way that is meaningful, reinforced, and accessible

    On-line regression competitive with reproducing kernel Hilbert spaces

    Get PDF
    We consider the problem of on-line prediction of real-valued labels, assumed bounded in absolute value by a known constant, of new objects from known labeled objects. The prediction algorithm's performance is measured by the squared deviation of the predictions from the actual labels. No stochastic assumptions are made about the way the labels and objects are generated. Instead, we are given a benchmark class of prediction rules some of which are hoped to produce good predictions. We show that for a wide range of infinite-dimensional benchmark classes one can construct a prediction algorithm whose cumulative loss over the first N examples does not exceed the cumulative loss of any prediction rule in the class plus O(sqrt(N)); the main differences from the known results are that we do not impose any upper bound on the norm of the considered prediction rules and that we achieve an optimal leading term in the excess loss of our algorithm. If the benchmark class is "universal" (dense in the class of continuous functions on each compact set), this provides an on-line non-stochastic analogue of universally consistent prediction in non-parametric statistics. We use two proof techniques: one is based on the Aggregating Algorithm and the other on the recently developed method of defensive forecasting.Comment: 37 pages, 1 figur

    Prediction and Situational Option Generation in Soccer

    Get PDF
    Paul Ward, Michigan Technological University Naturalistic models of decision making, such as the Recognition- Primed Decision (RPD) model (e.g., Klein, Calderwood, & Clinton-Cirocco, 1986; Klein, 1997), suggest that as individuals become more experienced within a domain they automatically recognize situational patterns as familiar which, in turn, activates an associated situational response. Typically, this results in a workable course of action being generated first, and subsequent options generated only if the initial option proves ineffective

    Administrative Compensation for Medical Injuries: Lessons From Three Foreign Systems

    Get PDF
    Examines "no-fault" systems in New Zealand, Sweden, and Denmark, in which patients injured by medical negligence can file for compensation through governmental or private adjudicating organizations. Considers lessons for U.S. medical malpractice reform

    Competitive on-line learning with a convex loss function

    Get PDF
    We consider the problem of sequential decision making under uncertainty in which the loss caused by a decision depends on the following binary observation. In competitive on-line learning, the goal is to design decision algorithms that are almost as good as the best decision rules in a wide benchmark class, without making any assumptions about the way the observations are generated. However, standard algorithms in this area can only deal with finite-dimensional (often countable) benchmark classes. In this paper we give similar results for decision rules ranging over an arbitrary reproducing kernel Hilbert space. For example, it is shown that for a wide class of loss functions (including the standard square, absolute, and log loss functions) the average loss of the master algorithm, over the first NN observations, does not exceed the average loss of the best decision rule with a bounded norm plus O(N−1/2)O(N^{-1/2}). Our proof technique is very different from the standard ones and is based on recent results about defensive forecasting. Given the probabilities produced by a defensive forecasting algorithm, which are known to be well calibrated and to have good resolution in the long run, we use the expected loss minimization principle to find a suitable decision.Comment: 26 page

    Online Learning in Case of Unbounded Losses Using the Follow Perturbed Leader Algorithm

    Full text link
    In this paper the sequential prediction problem with expert advice is considered for the case where losses of experts suffered at each step cannot be bounded in advance. We present some modification of Kalai and Vempala algorithm of following the perturbed leader where weights depend on past losses of the experts. New notions of a volume and a scaled fluctuation of a game are introduced. We present a probabilistic algorithm protected from unrestrictedly large one-step losses. This algorithm has the optimal performance in the case when the scaled fluctuations of one-step losses of experts of the pool tend to zero.Comment: 31 pages, 3 figure

    Introduction: food relocalisation and knowledge dynamics for sustainability in rural areas

    Get PDF
    The chapter presents the literature on local food and local knowledge and introduces the case studies analysed in the volum
    • …
    corecore