72,941 research outputs found
Stability and convergence analysis of a class of continuous piecewise polynomial approximations for time fractional differential equations
We propose and study a class of numerical schemes to approximate time
fractional differential equations. The methods are based on the approximation
of the Caputo fractional derivative by continuous piecewise polynomials, which
is strongly related to the backward differentiation formulae for the
integer-order case. We investigate their theoretical properties, such as the
local truncation error and global error analyses with respect to a sufficiently
smooth solution, and the numerical stability in terms of the stability region
and -stability by refining the technique proposed in
\cite{LubichC:1986b}. Numerical experiments are given to verify the theoretical
investigations.Comment: 34 pages, 3 figure
Optimal No-regret Learning in Repeated First-price Auctions
We study online learning in repeated first-price auctions with censored
feedback, where a bidder, only observing the winning bid at the end of each
auction, learns to adaptively bid in order to maximize her cumulative payoff.
To achieve this goal, the bidder faces a challenging dilemma: if she wins the
bid--the only way to achieve positive payoffs--then she is not able to observe
the highest bid of the other bidders, which we assume is iid drawn from an
unknown distribution. This dilemma, despite being reminiscent of the
exploration-exploitation trade-off in contextual bandits, cannot directly be
addressed by the existing UCB or Thompson sampling algorithms in that
literature, mainly because contrary to the standard bandits setting, when a
positive reward is obtained here, nothing about the environment can be learned.
In this paper, by exploiting the structural properties of first-price
auctions, we develop the first learning algorithm that achieves
regret bound when the bidder's private values are
stochastically generated. We do so by providing an algorithm on a general class
of problems, which we call monotone group contextual bandits, where the same
regret bound is established under stochastically generated contexts. Further,
by a novel lower bound argument, we characterize an lower
bound for the case where the contexts are adversarially generated, thus
highlighting the impact of the contexts generation mechanism on the fundamental
learning limit. Despite this, we further exploit the structure of first-price
auctions and develop a learning algorithm that operates sample-efficiently (and
computationally efficiently) in the presence of adversarially generated private
values. We establish an regret bound for this algorithm,
hence providing a complete characterization of optimal learning guarantees for
this problem
- …