23,586 research outputs found

    Session 3-4-C: Are Lottery Players Affected by Random Shocks? Evidence from China’s Individual Lottery Betting Panel Data

    Full text link
    I explore a unique, individual level, lottery betting panel data and show that lottery gambling is significantly affected by lottery winning history even though this winning history is shown to be merely an exogenous random shock. This panel data records lottery players’ collective lottery betting behaviors on a Chinese online lottery purchase website. This website lists each player’s lottery investment performance, the ratio between the lottery return and the lottery investment in the past three months, for lottery players’ reference and this ratio is shown to be an independent random shock across players. Based on the data with around 400,000 observations, I find that this random shock significantly affects lottery players’ purchasing behaviors. Specifically, collective lottery gamblers are significantly more likely to join a lottery package proposed by someone with a higher winning rate; lottery players are spending more money on the proposers with higher return rates

    The Merits of Sharing a Ride

    Full text link
    The culture of sharing instead of ownership is sharply increasing in individuals behaviors. Particularly in transportation, concepts of sharing a ride in either carpooling or ridesharing have been recently adopted. An efficient optimization approach to match passengers in real-time is the core of any ridesharing system. In this paper, we model ridesharing as an online matching problem on general graphs such that passengers do not drive private cars and use shared taxis. We propose an optimization algorithm to solve it. The outlined algorithm calculates the optimal waiting time when a passenger arrives. This leads to a matching with minimal overall overheads while maximizing the number of partnerships. To evaluate the behavior of our algorithm, we used NYC taxi real-life data set. Results represent a substantial reduction in overall overheads

    State-Augmentation Transformations for Risk-Sensitive Reinforcement Learning

    Full text link
    In the framework of MDP, although the general reward function takes three arguments-current state, action, and successor state; it is often simplified to a function of two arguments-current state and action. The former is called a transition-based reward function, whereas the latter is called a state-based reward function. When the objective involves the expected cumulative reward only, this simplification works perfectly. However, when the objective is risk-sensitive, this simplification leads to an incorrect value. We present state-augmentation transformations (SATs), which preserve the reward sequences as well as the reward distributions and the optimal policy in risk-sensitive reinforcement learning. In risk-sensitive scenarios, firstly we prove that, for every MDP with a stochastic transition-based reward function, there exists an MDP with a deterministic state-based reward function, such that for any given (randomized) policy for the first MDP, there exists a corresponding policy for the second MDP, such that both Markov reward processes share the same reward sequence. Secondly we illustrate that two situations require the proposed SATs in an inventory control problem. One could be using Q-learning (or other learning methods) on MDPs with transition-based reward functions, and the other could be using methods, which are for the Markov processes with a deterministic state-based reward functions, on the Markov processes with general reward functions. We show the advantage of the SATs by considering Value-at-Risk as an example, which is a risk measure on the reward distribution instead of the measures (such as mean and variance) of the distribution. We illustrate the error in the reward distribution estimation from the direct use of Q-learning, and show how the SATs enable a variance formula to work on Markov processes with general reward functions

    Functional Bandits

    Full text link
    We introduce the functional bandit problem, where the objective is to find an arm that optimises a known functional of the unknown arm-reward distributions. These problems arise in many settings such as maximum entropy methods in natural language processing, and risk-averse decision-making, but current best-arm identification techniques fail in these domains. We propose a new approach, that combines functional estimation and arm elimination, to tackle this problem. This method achieves provably efficient performance guarantees. In addition, we illustrate this method on a number of important functionals in risk management and information theory, and refine our generic theoretical results in those cases

    Economic Transition and the Motherhood Wage Penalty in Urban China: Investigation using Panel Data

    Get PDF
    China’s economic transition has fundamentally changed the mechanisms for allocating and compensating labor. This paper investigates how the economic transition has affected the wage gap between mothers and childless women in urban China using panel data for the period 1990-2005. The results show that overall, mothers earned considerably less than childless women; additionally, the wage penalties for motherhood went up substantially from the gradualist reform period (1990-1996) to the radical reform period (1999-2005). The results also show that that although motherhood does not appear to have a significant wage effect for the state sector, it imposes substantial wage losses for mothers in the non-state sector. These findings suggest that the economic transition has shifted part of the cost of child-bearing and -rearing from the state and employers back to women in the form of lower earnings for working mothers.
    • …
    corecore