Abstract
We present an accelerated algorithm for the solution of static Hamilton-Jacobi-Bellman equations related to optimal control problems and differential games. The new scheme combines the advantages of value iteration and policy iteration methods by means of an efficient coupling. The method starts with a value iteration phase on a coarse mesh and then switches to a policy iteration procedure over a finer mesh when a fixed error threshold is reached. We present numerical tests assessing the performance of the scheme.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
A. Alla, M. Falcone, D. Kalise, An efficient policy iteration algorithm for dynamic programming equations (submitted). Available at arXiv:1308.2087
M. Bardi, I. Capuzzo-Dolcetta, Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations (Birkhäuser, Boston, 1997)
R. Bellman, Dynamic Programming (Princeton University Press, Princeton, 1957)
O. Bokanowski, S. Maroso, H. Zidani, Some convergence results for Howard’s algorithm. SIAM J. Numer. Anal. 47, 3001–3026 (2009)
C. Chow, J. Tsitsiklis, An optimal one-way multigrid algorithm for discrete-time stochastic control. IEEE Trans. Autom. Control 36, 898–914 (1991)
R.A. Howard, Dynamic Programming and Markov Processes. Technology Press of the Massachusetts Institute of technology. New York; London: J. Wiley, cop. (1960)
R. Kalaba, On nonlinear differential equations, the maximum operation and monotone convergence. J. Math. Mech. 8, 519–574 (1959)
M. Pollatschek, B. Avi-Itzhak, Algorithms for stochastic games with geometrical interpretation. Manage. Sci. 15, 399–415 (1969)
M.L. Puterman, S.L. Brumelle, On the convergence of Policy iteration in stationary dynamic programming. Math. Oper. Res. 4, 60–69 (1979)
M.S. Santos, Numerical solution of dynamic economic models, in Handbook of Macroeconomics, ed. by J.B. Taylor, M. Woodford (Elsevier Science, Amsterdam/New York, 1999), pp. 311–386
M.S. Santos, J. Rust, Convergence properties of policy iteration. SIAM J. Control Optim. 42, 2094–2115 (2004)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Alla, A., Falcone, M., Kalise, D. (2015). An Accelerated Value/Policy Iteration Scheme for Optimal Control Problems and Games. In: Abdulle, A., Deparis, S., Kressner, D., Nobile, F., Picasso, M. (eds) Numerical Mathematics and Advanced Applications - ENUMATH 2013. Lecture Notes in Computational Science and Engineering, vol 103. Springer, Cham. https://doi.org/10.1007/978-3-319-10705-9_48
Download citation
DOI: https://doi.org/10.1007/978-3-319-10705-9_48
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-10704-2
Online ISBN: 978-3-319-10705-9
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)