Jump to content

Value function

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Bender235 (talk | contribs) at 13:40, 29 July 2019 (top). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

The value function of an optimization problem gives the value attained by the objective function at a solution, while only depending on the parameters of the problem.[1] In an economic context, where the objective function usually represents utility, the value function is conceptually equivalent to the indirect utility function.[2]

In a problem of optimal control, the value function is defined as the supremum of the objective function taken over the set of admissible controls. Generally an optimal control problem may be written:

where the objective function is to be maximized over all admissible controls for which the corresponding trajectory of , .[3] In discrete time, i.e. , the integral would be replaced by a summation in an otherwise identical problem:

The parameters of the problem are the initial and terminal value of the state variable, and , as well as the initial time and the terminal time . Then the value function is defined as

where is the optimal control or policy function.[4]

By Bellman's principle of optimality, which roughly states that any optimal policy at time , taking the current state as "new" initial condition must be optimal for the remaining problem, gives rise to an important functional recurrence equation, known as the Bellman equation (in discrete time), or Hamilton–Jacobi–Bellman equation (in continuous time). The latter, for the above problem, can be written:

where means the partial derivative of wrt. the time variable . means the dot product of the vectors a and b and the gradient of wrt. the variables . The maximand on the right-hand side is equivalent to the Hamiltonian, with playing the role of the costate variables.[5]

Although unknown until a solution to the optimization problem is found, the value function itself can be used to find a solution.[6] Benveniste and Scheinkman established sufficient conditions for the differentiability of the value function,[7] which in turn allows the application of the envelope theorem to solve the Bellman equation.

References

  1. ^ Mas-Colell, Andreu; Whinston, Michael D.; Green, Jerry R. (1995). Microeconomic Theory. New York: Oxford University Press. p. 964. ISBN 0-19-507340-1.
  2. ^ Corbae, Dean; Stinchcombe, Maxwell B.; Zeman, Juraj (2009). An Introduction to Mathematical Analysis for Economic Theory and Econometrics. Princeton University Press. p. 145. ISBN 978-0-691-11867-3.
  3. ^ Kamien, Morton I.; Schwartz, Nancy L. (1991). Dynamic Optimization : The Calculus of Variations and Optimal Control in Economics and Management (2nd ed.). Amsterdam: North-Holland. p. 259. ISBN 0-444-01609-0.
  4. ^ Ljungqvist, Lars; Sargent, Thomas J. (2018). Recursive Macroeconomic Theory (Fourth ed.). Cambridge: MIT Press. p. 106. ISBN 978-0-262-03866-9.
  5. ^ Kirk, Donald E. (1970). Optimal Control Theory. Englewood Cliffs, NJ: Prentice-Hall. p. 88. ISBN 0-13-638098-0.
  6. ^ Stokey, Nancy L.; Lucas, Robert E. Jr. (1987). Recursive Methods in Economic Dynamics. Cambridge: Harvard University Press. pp. 13–14. ISBN 0-674-75096-9.
  7. ^ Benveniste, L. M.; Scheinkman, J. A. (1979). "On the Differentiability of the Value Function in Dynamic Models of Economics". Econometrica. 47 (3): 727–732. JSTOR 1910417.