152,645 research outputs found
Feature Selection Using Regularization in Approximate Linear Programs for Markov Decision Processes
Approximate dynamic programming has been used successfully in a large variety
of domains, but it relies on a small set of provided approximation features to
calculate solutions reliably. Large and rich sets of features can cause
existing algorithms to overfit because of a limited number of samples. We
address this shortcoming using regularization in approximate linear
programming. Because the proposed method can automatically select the
appropriate richness of features, its performance does not degrade with an
increasing number of features. These results rely on new and stronger sampling
bounds for regularized approximate linear programs. We also propose a
computationally efficient homotopy method. The empirical evaluation of the
approach shows that the proposed method performs well on simple MDPs and
standard benchmark problems.Comment: Technical report corresponding to the ICML2010 submission of the same
nam
An approximate dynamic programming approach to solving dynamic oligopoly models
In this article, we introduce a new method to approximate Markov perfect equilibrium in large-scale Ericson and Pakes (1995)-style dynamic oligopoly models that are not amenable to exact solution due to the curse of dimensionality. The method is based on an algorithm that iterates an approximate best response operator using an approximate dynamic programming approach. The method, based on mathematical programming, approximates the value function with a linear combination of basis functions. We provide results that lend theoretical support to our approach. We introduce a rich yet tractable set of basis functions, and test our method on important classes of models. Our results suggest that the approach we propose significantly expands the set of dynamic oligopoly models that can be analyzed computationally
A linear programming methodology for approximate dynamic programming
[EN] The linear programming (LP) approach to solve the Bellman equation in dynamic programming is a well-known option for
finite state and input spaces to obtain an exact solution. However, with function approximation or continuous state spaces,
refinements are necessary. This paper presents a methodology to make approximate dynamic programming via LP work
in practical control applications with continuous state and input spaces. There are some guidelines on data and regressor
choices needed to obtain meaningful and well-conditioned value function estimates. The work discusses the introduction of
terminal ingredients and computation of lower and upper bounds of the value function. An experimental inverted-pendulum
application will be used to illustrate the proposal and carry out a suitable comparative analysis with alternative options in
the literature.The authors are grateful for the financial support of the Spanish Ministry of Economy and the European Union, grant DPI2016-81002-R (AEI/FEDER, UE), and the PhD grant from the Government of Ecuador (SENESCYT).Diaz, H.; Sala, A.; Armesto Ángel, L. (2020). A linear programming methodology for approximate dynamic programming. International Journal of Applied Mathematics and Computer Science (Online). 30(2):363-375. https://doi.org/10.34768/amcs-2020-0028S36337530
RAND Journal of Economics
approach to solving dynamic oligopoly models Vivek Farias ∗ Denis Saure∗ ∗ and Gabriel Y. Weintraub∗∗ ∗ In this article, we introduce a new method to approximate Markov perfect equilibrium in largescale Ericson and Pakes (1995)-style dynamic oligopoly models that are not amenable to exact solution due to the curse of dimensionality. The method is based on an algorithm that iterates an approximate best response operator using an approximate dynamic programming approach. The method, based on mathematical programming, approximates the value function with a linear combination of basis functions. We provide results that lend theoretical support to our approach. We introduce a rich yet tractable set of basis functions, and test our method on important classes of models. Our results suggest that the approach we propose significantly expands the set of dynamic oligopoly models that can be analyzed computationally. 1
Approximate Dynamic Programming via a Smoothed Linear Program
We present a novel linear program for the approximation of the dynamic programming cost-to-go function in high-dimensional stochastic control problems. LP approaches to approximate DP have typically relied on a natural “projection” of a well-studied linear program for exact dynamic programming. Such programs restrict attention to approximations that are lower bounds to the optimal cost-to-go function. Our program—the “smoothed approximate linear program”—is distinct from such approaches and relaxes the restriction to lower bounding approximations in an appropriate fashion while remaining computationally tractable. Doing so appears to have several advantages: First, we demonstrate bounds on the quality of approximation to the optimal cost-to-go function afforded by our approach. These bounds are, in general, no worse than those available for extant LP approaches and for specific problem instances can be shown to be arbitrarily stronger. Second, experiments with our approach on a pair of challenging problems (the game of Tetris and a queueing network control problem) show that the approach outperforms the existing LP approach (which has previously been shown to be competitive with several ADP algorithms) by a substantial margin
- …