TY - JOUR
T1 - Dynamic consistency for stochastic optimal control problems
AU - Carpentier, Pierre
AU - Chancelier, Jean Philippe
AU - Cohen, Guy
AU - De Lara, Michel
AU - Girardeau, Pierre
PY - 2012/11/1
Y1 - 2012/11/1
N2 - For a sequence of dynamic optimization problems, we aim at discussing a notion of consistency over time. This notion can be informally introduced as follows. At the very first time step t0, the decision maker formulates an optimization problem that yields optimal decision rules for all the forthcoming time steps t0,t1,...,T; at the next time step t1, he is able to formulate a new optimization problem starting at time t1 that yields a new sequence of optimal decision rules. This process can be continued until the final time T is reached. A family of optimization problems formulated in this way is said to be dynamically consistent if the optimal strategies obtained when solving the original problem remain optimal for all subsequent problems. The notion of dynamic consistency, well-known in the field of economics, has been recently introduced in the context of risk measures, notably by Artzner et al. (Ann. Oper. Res. 152(1):5-22, 2007) and studied in the stochastic programming framework by Shapiro (Oper. Res. Lett. 37(3):143-147, 2009) and for Markov Decision Processes (MDP) by Ruszczynski (Math. Program. 125(2):235-261, 2010). We here link this notion with the concept of "state variable" in MDP, and show that a significant class of dynamic optimization problems are dynamically consistent, provided that an adequate state variable is chosen.
AB - For a sequence of dynamic optimization problems, we aim at discussing a notion of consistency over time. This notion can be informally introduced as follows. At the very first time step t0, the decision maker formulates an optimization problem that yields optimal decision rules for all the forthcoming time steps t0,t1,...,T; at the next time step t1, he is able to formulate a new optimization problem starting at time t1 that yields a new sequence of optimal decision rules. This process can be continued until the final time T is reached. A family of optimization problems formulated in this way is said to be dynamically consistent if the optimal strategies obtained when solving the original problem remain optimal for all subsequent problems. The notion of dynamic consistency, well-known in the field of economics, has been recently introduced in the context of risk measures, notably by Artzner et al. (Ann. Oper. Res. 152(1):5-22, 2007) and studied in the stochastic programming framework by Shapiro (Oper. Res. Lett. 37(3):143-147, 2009) and for Markov Decision Processes (MDP) by Ruszczynski (Math. Program. 125(2):235-261, 2010). We here link this notion with the concept of "state variable" in MDP, and show that a significant class of dynamic optimization problems are dynamically consistent, provided that an adequate state variable is chosen.
KW - Dynamic consistency
KW - Dynamic programming
KW - Risk measures
KW - Stochastic optimal control
UR - https://www.scopus.com/pages/publications/84867403941
U2 - 10.1007/s10479-011-1027-8
DO - 10.1007/s10479-011-1027-8
M3 - Article
AN - SCOPUS:84867403941
SN - 0254-5330
VL - 200
SP - 247
EP - 263
JO - Annals of Operations Research
JF - Annals of Operations Research
IS - 1
ER -