88,745 research outputs found

    A Note on the Utility of Incremental Learning

    Get PDF
    ... This paper defines the notion of incrementality for learning tasks and algorithms. It then provides some motivation for incremental learning and argues in favour of the design of incremental learning algorithms for solving incremental learning tasks. A number of issues raised by such systems are outlined and the incremental learner ILA is used for illustration

    Utility Analysis for Multiple Selection Devices and Multiple Outcomes

    Get PDF
    Traditional utility analysis only calculates the value of a given selection procedure over random selection. This assumption is not only an inaccurate representation of staffing policy but leads to overestimates of a device\u27s value. This paper generates a new utility model that accounts for multiple selection devices and multiple criteria. The model is illustrated using previous utility analysis work and an actual case of secretarial employees with eight predictors and nine criteria. A final example also is provided which includes these advancements as well as other researchers\u27 advances in a combined utility model. Results reveal that accounting for multiple criteria and outcomes dramatically reduces the utility estimates of implementing new selection devices

    Private Incremental Regression

    Full text link
    Data is continuously generated by modern data sources, and a recent challenge in machine learning has been to develop techniques that perform well in an incremental (streaming) setting. In this paper, we investigate the problem of private machine learning, where as common in practice, the data is not given at once, but rather arrives incrementally over time. We introduce the problems of private incremental ERM and private incremental regression where the general goal is to always maintain a good empirical risk minimizer for the history observed under differential privacy. Our first contribution is a generic transformation of private batch ERM mechanisms into private incremental ERM mechanisms, based on a simple idea of invoking the private batch ERM procedure at some regular time intervals. We take this construction as a baseline for comparison. We then provide two mechanisms for the private incremental regression problem. Our first mechanism is based on privately constructing a noisy incremental gradient function, which is then used in a modified projected gradient procedure at every timestep. This mechanism has an excess empirical risk of ≈d\approx\sqrt{d}, where dd is the dimensionality of the data. While from the results of [Bassily et al. 2014] this bound is tight in the worst-case, we show that certain geometric properties of the input and constraint set can be used to derive significantly better results for certain interesting regression problems.Comment: To appear in PODS 201

    Truncating Temporal Differences: On the Efficient Implementation of TD(lambda) for Reinforcement Learning

    Full text link
    Temporal difference (TD) methods constitute a class of methods for learning predictions in multi-step prediction problems, parameterized by a recency factor lambda. Currently the most important application of these methods is to temporal credit assignment in reinforcement learning. Well known reinforcement learning algorithms, such as AHC or Q-learning, may be viewed as instances of TD learning. This paper examines the issues of the efficient and general implementation of TD(lambda) for arbitrary lambda, for use with reinforcement learning algorithms optimizing the discounted sum of rewards. The traditional approach, based on eligibility traces, is argued to suffer from both inefficiency and lack of generality. The TTD (Truncated Temporal Differences) procedure is proposed as an alternative, that indeed only approximates TD(lambda), but requires very little computation per action and can be used with arbitrary function representation methods. The idea from which it is derived is fairly simple and not new, but probably unexplored so far. Encouraging experimental results are presented, suggesting that using lambda &gt 0 with the TTD procedure allows one to obtain a significant learning speedup at essentially the same cost as usual TD(0) learning.Comment: See http://www.jair.org/ for any accompanying file

    Customer anger and incentives for quality provision

    Get PDF
    Emotions are a significant determinant of consumer behaviour. A customer may get angry if he feels that he is being treated unfairly by his supplier and that anger may make him more likely to switch to an alternative provider. We model the strategic interaction between firms that choose quality levels and anger-prone customers who pick their supplier based on their expectations of suppliers' quality. Strategic interaction can allow for multiple equilibria including some in which no firm invests in high quality. Allowing customers to voice their anger on peer-review fora can eliminate low-quality equilibria, and may even support a unique equilibrium in which all firms choose high quality
    • …
    corecore