When Inaccuracies in Value Functions Do Not Propagate on Optima and Equilibria

We study general classes of discrete time dynamic optimization problems and dynamic games with feedback controls. In such problems, the solution is usually found by using the Bellman or Hamilton–Jacobi–Bellman equation for the value function in the case of dynamic optimization and a set of such coup...

Full description

Bibliographic Details
Main Authors: Agnieszka Wiszniewska-Matyszkiel, Rajani Singh
Format: Article
Language:English
Published: MDPI AG 2020-07-01
Series:Mathematics
Subjects:
Online Access:https://www.mdpi.com/2227-7390/8/7/1109
Description
Summary:We study general classes of discrete time dynamic optimization problems and dynamic games with feedback controls. In such problems, the solution is usually found by using the Bellman or Hamilton–Jacobi–Bellman equation for the value function in the case of dynamic optimization and a set of such coupled equations for dynamic games, which is not always possible accurately. We derive general rules stating what kind of errors in the calculation or computation of the value function do not result in errors in calculation or computation of an optimal control or a Nash equilibrium along the corresponding trajectory. This general result concerns not only errors resulting from using numerical methods but also errors resulting from some preliminary assumptions related to replacing the actual value functions by some a priori assumed constraints for them on certain subsets. We illustrate the results by a motivating example of the Fish Wars, with singularities in payoffs.
ISSN:2227-7390