Stochastic control
Stochastic control or stochastic
Certainty equivalence
An extremely well-studied formulation in stochastic control is that of
Any deviation from the above assumptions—a nonlinear state equation, a non-quadratic objective function, noise in the multiplicative parameters of the model, or decentralization of control—causes the certainty equivalence property not to hold. For example, its failure to hold for decentralized control was demonstrated in Witsenhausen's counterexample.
Discrete time
In a discrete-time context, the decision-maker observes the state variable, possibly with observational noise, in each time period. The objective may be to optimize the sum of expected values of a nonlinear (possibly quadratic) objective function over all the time periods from the present to the final period of concern, or to optimize the value of the objective function as of the final period only. At each time period new observations are made, and the control variables are to be adjusted optimally. Finding the optimal solution for the present time may involve iterating a
In the discrete-time case with uncertainty about the parameter values in the transition matrix (giving the effect of current values of the state variables on their own evolution) and/or the control response matrix of the state equation, but still with a linear state equation and quadratic objective function, a Riccati equation can still be obtained for iterating backward to each period's solution even though certainty equivalence does not apply.[2]ch.13[3] The discrete-time case of a non-quadratic loss function but only additive disturbances can also be handled, albeit with more complications.[4]
Example
A typical specification of the discrete-time stochastic linear quadratic control problem is to minimize[2]: ch. 13, [3][5]
where E1 is the
where y is an n × 1 vector of observable state variables, u is a k × 1 vector of control variables, At is the time t realization of the
Induction backwards in time can be used to obtain the optimal control solution at each time,[2]: ch. 13
with the symmetric positive definite cost-to-go matrix X evolving backwards in time from according to
which is known as the discrete-time dynamic Riccati equation of this problem. The only information needed regarding the unknown parameters in the A and B matrices is the expected value and variance of each element of each matrix and the covariances among elements of the same matrix and among elements across matrices.
The optimal control solution is unaffected if zero-mean, i.i.d. additive shocks also appear in the state equation, so long as they are uncorrelated with the parameters in the A and B matrices. But if they are so correlated, then the optimal control solution for each period contains an additional additive constant vector. If an additive constant vector appears in the state equation, then again the optimal control solution for each period contains an additional additive constant vector.
The steady-state characterization of X (if it exists), relevant for the infinite-horizon problem in which S goes to infinity, can be found by iterating the dynamic equation for X repeatedly until it converges; then X is characterized by removing the time subscripts from its dynamic equation.
Continuous time
If the model is in continuous time, the controller knows the state of the system at each instant of time. The objective is to maximize either an integral of, for example, a concave function of a state variable over a horizon from time zero (the present) to a terminal time T, or a concave function of a state variable at some future date T. As time evolves, new observations are continuously made and the control variables are continuously adjusted in optimal fashion.
Stochastic model predictive control
In the literature, there are two types of MPCs for stochastic systems; Robust model predictive control and Stochastic Model Predictive Control (SMPC). Robust model predictive control is a more conservative method which considers the worst scenario in the optimization procedure. However, this method, similar to other robust controls, deteriorates the overall controller's performance and also is applicable only for systems with bounded uncertainties. The alternative method, SMPC, considers soft constraints which limit the risk of violation by a probabilistic inequality.[6]
In finance
In a continuous time approach in a
The maximization, say of the expected logarithm of net worth at a terminal date T, is subject to stochastic processes on the components of wealth.[11] In this case, in continuous time Itô's equation is the main tool of analysis. In the case where the maximization is an integral of a concave function of utility over an horizon (0,T), dynamic programming is used. There is no certainty equivalence as in the older literature, because the coefficients of the control variables—that is, the returns received by the chosen shares of assets—are stochastic.
See also
- Backward stochastic differential equation
- Stochastic process
- Control theory
- Multiplier uncertainty
- Stochastic scheduling
References
- ^ Definition from Answers.com
- ^ ISBN 0-471-15616-7.
- ^ JSTOR 2296614.
- .
- JSTOR 1814888.
- Bibcode:2017arXiv170404710H.
- ^ Merton, Robert (1990). Continuous Time Finance. Blackwell.
- ISBN 0-387-90155-8.
- ^ Fleming, W.; Soner, M. (2006). Controlled Markov Processes and Viscosity Solutions. Springer.
- ^ Stein, J. L. (2012). Stochastic Optimal Control and the US Financial Crisis. Springer-Science.
- ISSN 2169-3536.
Further reading
- .
- Yong, Jiongmin; Zhou, Xun Yu (1999). Stochastic Controls : Hamiltonian Systems and HJB Equations. New York: Springer. ISBN 0-387-98723-1.