Spatial Decomposition/Coordination Methods for Stochastic Optimal Control Problems. Practical aspects and theoretical questions

Similar documents
The Heat Equation. Lectures INF2320 p. 1/88

Stochastic Gradient Method: Applications

Two-Stage Stochastic Linear Programs

B. Delyon. Stochastic approximation with decreasing gain: convergence and asymptotic theory. Unpublished lecture notes, Université de Rennes, 2000.

Duality of linear conic problems

Nonlinear Optimization: Algorithms 3: Interior-point methods

Further Study on Strong Lagrangian Duality Property for Invex Programs via Penalty Functions 1

Duality in General Programs. Ryan Tibshirani Convex Optimization /36-725

Big Data - Lecture 1 Optimization reminders

24. The Branch and Bound Method

Equilibrium computation: Part 1

Date: April 12, Contents

Recovery of primal solutions from dual subgradient methods for mixed binary linear programming; a branch-and-bound approach

Introduction to Support Vector Machines. Colin Campbell, Bristol University

CHAPTER 9. Integer Programming

GI01/M055 Supervised Learning Proximal Methods

Some representability and duality results for convex mixed-integer programs.

Separation Properties for Locally Convex Cones

ECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE

1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.

Moral Hazard. Itay Goldstein. Wharton School, University of Pennsylvania

Reinforcement Learning

Introduction to Stochastic Optimization in Supply Chain and Logistic Optimization

A NEW LOOK AT CONVEX ANALYSIS AND OPTIMIZATION

Integrating Benders decomposition within Constraint Programming

Practical Guide to the Simplex Method of Linear Programming

Cyber-Security Analysis of State Estimators in Power Systems

2.3 Convex Constrained Optimization Problems

The Advantages and Disadvantages of Online Linear Optimization

Lecture 10. Finite difference and finite element methods. Option pricing Sensitivity analysis Numerical examples

Statistical Machine Learning

Mathematical finance and linear programming (optimization)

4.6 Linear Programming duality

Reinforcement Learning

1 Portfolio mean and variance

Numerisches Rechnen. (für Informatiker) M. Grepl J. Berger & J.T. Frings. Institut für Geometrie und Praktische Mathematik RWTH Aachen

CSCI567 Machine Learning (Fall 2014)

Summer course on Convex Optimization. Fifth Lecture Interior-Point Methods (1) Michel Baes, K.U.Leuven Bharath Rangarajan, U.

Walrasian Demand. u(x) where B(p, w) = {x R n + : p x w}.

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

ALMOST COMMON PRIORS 1. INTRODUCTION

Revenue Management for Transportation Problems

Optimization Theory for Large Systems

Robust Optimization for Unit Commitment and Dispatch in Energy Markets

Linear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued.

Convex analysis and profit/cost/support functions

STRATEGIC CAPACITY PLANNING USING STOCK CONTROL MODEL

Scheduling Shop Scheduling. Tim Nieberg

Adaptive Online Gradient Descent

Functional Optimization Models for Active Queue Management

The Master s Degree with Thesis Course Descriptions in Industrial Engineering

GenOpt (R) Generic Optimization Program User Manual Version 3.0.0β1

Generating Random Numbers Variance Reduction Quasi-Monte Carlo. Simulation Methods. Leonid Kogan. MIT, Sloan , Fall 2010

Copositive Cones. Qingxia Kong. Department of Decision Sciences, NUS. Chung-Yee Lee

Linear Programming Notes V Problem Transformations

Decentralized control of stochastic multi-agent service system. J. George Shanthikumar Purdue University

Numerical methods for American options

10. Proximal point method

TD(0) Leads to Better Policies than Approximate Value Iteration

Statistical Machine Translation: IBM Models 1 and 2

6. Budget Deficits and Fiscal Policy

Contrôle dynamique de méthodes d approximation

Variational approach to restore point-like and curve-like singularities in imaging

Big Data Analytics. Lucas Rego Drumond

The sample space for a pair of die rolls is the set. The sample space for a random number between 0 and 1 is the interval [0, 1].

Abstract. 1. Introduction. Caparica, Portugal b CEG, IST-UTL, Av. Rovisco Pais, Lisboa, Portugal

Research Article Stability Analysis for Higher-Order Adjacent Derivative in Parametrized Vector Optimization

No: Bilkent University. Monotonic Extension. Farhad Husseinov. Discussion Papers. Department of Economics

Advanced Lecture on Mathematical Science and Information Science I. Optimization in Finance

Notes from Week 1: Algorithms for sequential prediction

A Programme Implementation of Several Inventory Control Algorithms

A Network Flow Approach in Cloud Computing

Supplement to Call Centers with Delay Information: Models and Insights

3. Linear Programming and Polyhedral Combinatorics

Lecture 3: Linear methods for classification

Offline sorting buffers on Line

Hedging bounded claims with bounded outcomes

Proximal mapping via network optimization

LECTURE 15: AMERICAN OPTIONS

Optimal shift scheduling with a global service level constraint

Distributed Machine Learning and Big Data

Resource Optimization of Spatial TDMA in Ad Hoc Radio Networks: A Column Generation Approach

Discussion on the paper Hypotheses testing by convex optimization by A. Goldenschluger, A. Juditsky and A. Nemirovski.

On Adaboost and Optimal Betting Strategies

Stochastic Inventory Control

5 INTEGER LINEAR PROGRAMMING (ILP) E. Amaldi Fondamenti di R.O. Politecnico di Milano 1

15 Kuhn -Tucker conditions

Introduction to General and Generalized Linear Models

Part II Redundant Dictionaries and Pursuit Algorithms

An interval linear programming contractor

Transcription:

Spatial Decomposition/Coordination Methods for Stochastic Optimal Control Problems Practical aspects and theoretical questions P. Carpentier, J-Ph. Chancelier, M. De Lara, V. Leclère École des Ponts ParisTech 16 December 2014 M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 1 / 60

Large scale storage systems stand as powerful motivation The Optimization and Systems team was created in 2000 at École des Ponts ParisTech with emphasis on stochastic optimization Since 2011, we witness a growing demand from (small and large) energy firms for stochastic optimization, fueled by a deep and fast transformation of power systems M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 2 / 60

Something is changing in power systems Renewable energies penetration, telecommunication technologies and markets remold power systems and challenge optimization More renewable energies more unpredictability + more variability more storage more dynamic optimization, optimal control more stochastic optimization hence, stochastic optimal control (SOC) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 3 / 60

Lecture outline 1 Decomposition and coordination A bird s eye view of decomposition methods A brief insight into Progressive Hedging Spatial decomposition methods in the deterministic case The stochastic case raises specific obstacles 2 Dual approximate dynamic programming (DADP) Problem statement DADP principle and implementation Numerical results on a small size problem 3 Theoretical questions Existence of a saddle point Convergence of the Uzawa algorithm Convergence w.r.t. information 4 Summary and research agenda M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 4 / 60

Decomposition and coordination A long-term effort in our group 1976 A. Benveniste, P. Bernhard, G. Cohen, On the decomposition of stochastic control problems, IRIA-Laboria research report, No. 187, 1976. 1996 P. Carpentier, G. Cohen, J.-C. Culioli, A. Renaud, Stochastic optimization of unit commitment: a new decomposition framework, IEEE Transactions on Power Systems, Vol. 11, No. 2, 1996. 2006 C. Strugarek, Approches variationnelles et autres contributions en optimisation stochastique, Thèse de l ENPC, mai 2006. 2010 K. Barty, P. Carpentier, P. Girardeau, Decomposition of large-scale stochastic optimal control problems, RAIRO Operations Research, Vol. 44, No. 3, 2010. 2014 V. Leclère, Contributions to decomposition methods in stochastic optimization, Thèse de l Université Paris-Est, juin 2014. M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 5 / 60

Decomposition and coordination Let us fix problem and notations A bird s eye view of decomposition methods min u,x risk-neutral {}}{ E ( N subject to dynamics constraints i=1 ( T 1 t=0 x i t+1 = ft i (x i }{{} t, u i t, w } {{ t+1 } ), x i 0 = f-1(w i 0 ) state uncertainty ) ) L i t(x i t, u i t, w t+1 ) + K i (x i T ) to measurability constraints on the control u i t ( u i t σ(w 0,..., w t ) u i t = E u i t ) w 0,..., w t and to instantaneous coupling constraints N θt(x i i t, u i t) = 0 i=1 (The letter u stands for the Russian word for control: upravlenie) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 6 / 60

Decomposition and coordination A bird s eye view of decomposition methods Couplings for stochastic problems unit min ω π ω L i t(x i t, u i t, w t+1 ) i t s.t. x i t+1 = f i t (x i t, u i t, w t+1 ) ( u i t = E u i t ) w 0,..., w t time Θ i t(x i t, u i t) = 0 i uncertainty M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 7 / 60

Decomposition and coordination A bird s eye view of decomposition methods Couplings for stochastic problems: in time unit min ω π ω L i t(x i t, u i t, w t+1 ) i t s.t. x i t+1 = f i t (x i t, u i t, w t+1 ) ( u i t = E u i t ) w 0,..., w t time Θ i t(x i t, u i t) = 0 i uncertainty M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 7 / 60

Decomposition and coordination A bird s eye view of decomposition methods Couplings for stochastic problems: in uncertainty unit min ω π ω L i t(x i t, u i t, w t+1 ) i t s.t. x i t+1 = f i t (x i t, u i t, w t+1 ) ( u i t = E u i t ) w 0,..., w t time Θ i t(x i t, u i t) = 0 i uncertainty M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 7 / 60

Decomposition and coordination A bird s eye view of decomposition methods Couplings for stochastic problems: in space unit min ω π ω L i t(x i t, u i t, w t+1 ) i t s.t. x i t+1 = f i t (x i t, u i t, w t+1 ) ( u i t = E u i t ) w 0,..., w t time Θ i t(x i t, u i t) = 0 i uncertainty M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 7 / 60

Decomposition and coordination A bird s eye view of decomposition methods Can we decouple stochastic problems? unit min ω π ω L i t(x i t, u i t, w t+1 ) i t s.t. x i t+1 = f i t (x i t, u i t, w t+1 ) ( u i t = E u i t ) w 0,..., w t time Θ i t(x i t, u i t) = 0 i uncertainty M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 7 / 60

Decomposition and coordination A bird s eye view of decomposition methods Decompositions for stochastic problems: in time unit min ω π ω L i t(x i t, u i t, w t+1 ) i t s.t. x i t+1 = f i t (x i t, u i t, w t+1 ) ( u i t = E u i t ) w 0,..., w t uncertainty time Θ i t(x i t, u i t) = 0 i Dynamic Programming Bellman (56) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 7 / 60

Decomposition and coordination A bird s eye view of decomposition methods Decompositions for stochastic problems: in uncertainty unit min ω π ω L i t(x i t, u i t, w t+1 ) i t s.t. x i t+1 = f i t (x i t, u i t, w t+1 ) ( u i t = E u i t ) w 0,..., w t uncertainty time Θ i t(x i t, u i t) = 0 i Progressive Hedging Rockafellar - Wets (91) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 7 / 60

Decomposition and coordination A bird s eye view of decomposition methods Decompositions for stochastic problems: in space unit min ω π ω L i t(x i t, u i t, w t+1 ) i t s.t. x i t+1 = f i t (x i t, u i t, w t+1 ) ( u i t = E u i t ) w 0,..., w t uncertainty time Θ i t(x i t, u i t) = 0 i Dual Approximate Dynamic Programming M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 7 / 60

Decomposition and coordination Outline of the presentation A brief insight into Progressive Hedging 1 Decomposition and coordination A bird s eye view of decomposition methods A brief insight into Progressive Hedging Spatial decomposition methods in the deterministic case The stochastic case raises specific obstacles 2 Dual approximate dynamic programming (DADP) Problem statement DADP principle and implementation Numerical results on a small size problem 3 Theoretical questions Existence of a saddle point Convergence of the Uzawa algorithm Convergence w.r.t. information 4 Summary and research agenda M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 8 / 60

Decomposition and coordination Non-anticipativity constraints are linear A brief insight into Progressive Hedging From tree to scenarios (comb) Equivalent formulations of the non-anticipativity constraints pairwise equalities all equal to their mathematical expectation Linear structure ) u t = E (u t w 0,..., w t t=0 t=1 t=2 t=3 t=t t=0 t=1 t=2 t=3 t=t N scenarios Scenarios tree M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 9 / 60

Decomposition and coordination A brief insight into Progressive Hedging Progressive Hedging stands as a scenario decomposition method We dualize the non-anticipativity constraints When the criterion is strongly convex, we use a Lagrangian relaxation (algorithm à la Uzawa ) to obtain a scenario decomposition When the criterion is linear, Rockafellar - Wets (91) propose to use an augmented Lagrangian, and obtain the Progressive Hedging algorithm M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 10 / 60

Decomposition and coordination Spatial decomposition methods in the deterministic case 1 Decomposition and coordination A bird s eye view of decomposition methods A brief insight into Progressive Hedging Spatial decomposition methods in the deterministic case The stochastic case raises specific obstacles 2 Dual approximate dynamic programming (DADP) Problem statement DADP principle and implementation Numerical results on a small size problem 3 Theoretical questions Existence of a saddle point Convergence of the Uzawa algorithm Convergence w.r.t. information 4 Summary and research agenda M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 11 / 60

Decomposition and coordination Spatial decomposition methods in the deterministic case Decomposition and coordination Unit 1 Unit N Unit 2 Unit 3 Interconnected units The system to be optimized consists of interconnected subsystems We want to use this structure to formulate optimization subproblems of reasonable complexity But the presence of interactions requires a level of coordination Coordination iteratively provides a local model of the interactions for each subproblem We expect to obtain the solution of the overall problem by concatenation of the solutions of the subproblems M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 12 / 60

Decomposition and coordination Example: the flower model Spatial decomposition methods in the deterministic case Unit 1 Unit 2 Coupling constraint Unit 3 Unit N min u s.t. N J i (u i ) i=1 N θ i (u i ) = 0 i=1 Unit Commitment Problem M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 13 / 60

Decomposition and coordination Intuition of spatial decomposition Spatial decomposition methods in the deterministic case Unit 2 Unit 1 Coordinator Unit 3 Purpose: satisfy a demand with N production units, at minimal cost Price decomposition the coordinator sets a price λ the units send their optimal decision u i the coordinator compares total production N i=1 θ i(u i ) and demand, and then updates the price accordingly and so on... M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 14 / 60

Decomposition and coordination Intuition of spatial decomposition Spatial decomposition methods in the deterministic case Unit 2 Unit 1 λ (0) Coordinator λ (0) λ (0) Unit 3 Purpose: satisfy a demand with N production units, at minimal cost Price decomposition the coordinator sets a price λ the units send their optimal decision u i the coordinator compares total production N i=1 θ i(u i ) and demand, and then updates the price accordingly and so on... M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 14 / 60

Decomposition and coordination Intuition of spatial decomposition Spatial decomposition methods in the deterministic case Unit 2 Unit 1 u (0) 1 Coordinator u (0) 2 u (0) 3 Unit 3 Purpose: satisfy a demand with N production units, at minimal cost Price decomposition the coordinator sets a price λ the units send their optimal decision u i the coordinator compares total production N i=1 θ i(u i ) and demand, and then updates the price accordingly and so on... M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 14 / 60

Decomposition and coordination Intuition of spatial decomposition Spatial decomposition methods in the deterministic case Unit 2 Unit 1 λ (1) Coordinator λ (1) λ (1) Unit 3 Purpose: satisfy a demand with N production units, at minimal cost Price decomposition the coordinator sets a price λ the units send their optimal decision u i the coordinator compares total production N i=1 θ i(u i ) and demand, and then updates the price accordingly and so on... M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 14 / 60

Decomposition and coordination Intuition of spatial decomposition Spatial decomposition methods in the deterministic case Unit 2 Unit 1 u (1) 1 Coordinator u (1) 2 u (1) 3 Unit 3 Purpose: satisfy a demand with N production units, at minimal cost Price decomposition the coordinator sets a price λ the units send their optimal decision u i the coordinator compares total production N i=1 θ i(u i ) and demand, and then updates the price accordingly and so on... M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 14 / 60

Decomposition and coordination Intuition of spatial decomposition Spatial decomposition methods in the deterministic case Unit 2 Unit 1 λ (2) Coordinator λ (2) λ (2) Unit 3 Purpose: satisfy a demand with N production units, at minimal cost Price decomposition the coordinator sets a price λ the units send their optimal decision u i the coordinator compares total production N i=1 θ i(u i ) and demand, and then updates the price accordingly and so on... M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 14 / 60

Decomposition and coordination Intuition of spatial decomposition Spatial decomposition methods in the deterministic case Unit 2 Unit 1 u (2) 1 Coordinator u (2) 2 u (2) 3 Unit 3 Purpose: satisfy a demand with N production units, at minimal cost Price decomposition the coordinator sets a price λ the units send their optimal decision u i the coordinator compares total production N i=1 θ i(u i ) and demand, and then updates the price accordingly and so on... M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 14 / 60

Decomposition and coordination Price decomposition relies on dualization Spatial decomposition methods in the deterministic case min u i U i,i=1...n N J i (u i ) i=1 subject to N θ i (u i ) = 0 i=1 1 Form the Lagrangian and assume that a saddle point exists max λ V min u i U i,i=1...n i=1 N (J i (u i ) + λ, θ i (u i ) ) 2 Solve this problem by the dual gradient algorithm à la Uzawa u (k+1) i arg min J i (u i ) + λ (k), θ i (u i ), i = 1..., N u i U i N θ i (u (k+1) i ) λ (k+1) = λ (k) + ρ i=1 M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 15 / 60

Decomposition and coordination Remarks on decomposition methods Spatial decomposition methods in the deterministic case The theory is available for infinite dimensional Hilbert spaces, and thus applies in the stochastic framework, that is, when the U i are spaces of random variables The minimization algorithm used for solving the subproblems is not specified in the decomposition process New variables λ (k) appear in the subproblems arising at iteration k of the optimization process min J i (u i ) + λ (k), θ i (u i ) u i U i These variables are fixed when solving the subproblems, and do not cause any difficulty, at least in the deterministic case M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 16 / 60

Decomposition and coordination Spatial decomposition methods in the deterministic case Price decomposition applies to various couplings DECOMPOSITION M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 17 / 60

Decomposition and coordination The stochastic case raises specific obstacles 1 Decomposition and coordination A bird s eye view of decomposition methods A brief insight into Progressive Hedging Spatial decomposition methods in the deterministic case The stochastic case raises specific obstacles 2 Dual approximate dynamic programming (DADP) Problem statement DADP principle and implementation Numerical results on a small size problem 3 Theoretical questions Existence of a saddle point Convergence of the Uzawa algorithm Convergence w.r.t. information 4 Summary and research agenda M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 18 / 60

Decomposition and coordination The stochastic case raises specific obstacles Stochastic optimal control (SOC) problem formulation Consider the following SOC problem ( N E ( T 1 ) ) L i t(x i t, u i t, w t+1 ) + K i (x i T ) min u,x i=1 subject to the constraints t=0 x i 0 = f-1(w i 0 ), x i t+1 = ft i (x i t, u i t, w t+1 ), i = 1... N t = 0... T 1, i = 1... N u i t F t = σ(w 0,..., w t ), t = 0... T 1, i = 1... N N θt(x i i t, u i t) = 0, i=1 t = 0... T 1 M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 19 / 60

Decomposition and coordination The stochastic case raises specific obstacles Stochastic optimal control (SOC) problem formulation Consider the following SOC problem N ( ( T 1 ) ) E L i t(x i t, u i t, w t+1 ) + K i (x i T ) min u,x i=1 subject to the constraints t=0 x i 0 = f-1(w i 0 ), x i t+1 = ft i (x i t, u i t, w t+1 ), i = 1... N t = 0... T 1, i = 1... N u i t F t = σ(w 0,..., w t ), t = 0... T 1, i = 1... N N θt(x i i t, u i t) = 0, i=1 t = 0... T 1 M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 19 / 60

Decomposition and coordination The stochastic case raises specific obstacles Dynamic programming yields centralized controls As we want to solve this SOC problem using dynamic programming (DP), we suppose to be in the Markovian setting, that is, w 0,..., w T are a white noise The system is made of N interconnected subsystems, with the control u i t and the state x i t of subsystem i at time t The optimal control u i t of subsystem i is a function of the whole system state ( x 1 t,..., x N ) t u i t = γt( i x 1 t,..., x N ) t Naive decomposition should lead to decentralized feedbacks u i t = γ i t(x i t) which are, in most cases, far from being optimal... M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 20 / 60

Decomposition and coordination The stochastic case raises specific obstacles Straightforward decomposition of dynamic programming? The crucial point is that the optimal feedback of a subsystem a priori depends on the state of all other subsystems, so that using a decomposition scheme by subsystems is not obvious... As far as we have to deal with dynamic programming, the central concern for decomposition/coordination purpose boils down to?????? how to decompose a feedback γ t w.r.t. its domain X t rather than its range U t? And the answer is impossible in the general case! M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 21 / 60

Decomposition and coordination The stochastic case raises specific obstacles Price decomposition and dynamic programming When applying price decomposition to the problem by dualizing the (almost sure) coupling constraint i θi t(x i t, u i t) = 0, multipliers Λ (k) t appear in the subproblems arising at iteration k min E u i,x i ( t L i t(x i t, u i t, w t+1 ) + Λ (k) t θ i t(x i t, u i t) The variables Λ (k) t are fixed random variables, so that the random process Λ (k) acts as an additional input noise in the subproblems But this process may be correlated in time, so that the white noise assumption has no reason to be fulfilled DP cannot be applied in a straightforward manner! Question: how to handle the coordination instruments Λ (k) t to obtain (an approximation of) the overall optimum? ) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 22 / 60

Dual approximate dynamic programming (DADP) 1 Decomposition and coordination 2 Dual approximate dynamic programming (DADP) 3 Theoretical questions 4 Summary and research agenda M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 23 / 60

Dual approximate dynamic programming (DADP) Problem statement 1 Decomposition and coordination A bird s eye view of decomposition methods A brief insight into Progressive Hedging Spatial decomposition methods in the deterministic case The stochastic case raises specific obstacles 2 Dual approximate dynamic programming (DADP) Problem statement DADP principle and implementation Numerical results on a small size problem 3 Theoretical questions Existence of a saddle point Convergence of the Uzawa algorithm Convergence w.r.t. information 4 Summary and research agenda M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 24 / 60

Dual approximate dynamic programming (DADP) Optimization problem Problem statement The SOC problem under consideration reads ( N ( T 1 E min u,x i=1 subject to dynamics constraints x i 0 = f-1(w i 0 ) x i t+1 = ft i (x i t, u i t, w t+1 ) to measurability constraints u i t σ(w 0,..., w t ) t=0 and to instantaneous coupling constraints ) ) L i t(x i t, u i t, w t+1 ) + K i (x i T ) N θt(x i i t, u i t) = 0 i=1 Constraints to be dualized M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 25 / 60

Dual approximate dynamic programming (DADP) Assumptions Problem statement Assumption 1 (White noise) Noises w 0,..., w T are independent over time Hence dynamic programming applies: there is no optimality loss to look after the controls u i t as functions of the state at time t Assumption 2 (Constraint qualification) A saddle point of the Lagrangian L exists (more on that later) L ( x, u, Λ ) = E ( N i=1 ( T 1 t=0 T 1 ) ) L i t(x i t, u i t, w t+1) + K i (x i T ) + Λ t θt(x i i t, u i t) where the Λ t are σ(w 0,..., w t )-measurable random variables Assumption 3 (Dual gradient algorithm) Uzawa algorithm applies... (more on that later) t=0 M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 26 / 60

Dual approximate dynamic programming (DADP) Uzawa algorithm Problem statement At iteration k of the algorithm, 1 Solve Subproblem i, i = 1,..., N, with Λ (k) fixed min E u i,x i subject to ( T 1 t=0 ( ) ) L i t(x i t, u i t, w t+1 ) + Λ (k) t θt(x i i t, u i t) + K i (x i T ) x i t+1 = f i t (x i t, u i t, w t+1 ) u i t σ(w 0,..., w t ) whose solution is denoted ( u i,(k+1), x i,(k+1)) 2 Update the multipliers Λ t Λ (k+1) t = Λ (k) t + ρ t ( N i=1 θ i t ( x i,(k+1) t, u i,(k+1) ) ) t M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 27 / 60

Dual approximate dynamic programming (DADP) Structure of a subproblem Problem statement Subproblem i reads subject to ( T 1 ( ) ) min E L i t(x i t, u i t, w t+1 u i,x i } {{ } ) + Λ (k) t θ }{{} t(x i i t, u i t) t=0 white??? x i t+1 = f i t (x i t, u i t, w t+1 ) u i t σ(w 0,..., w t ) Without some knowledge of the process Λ (k) (we just know that Λ (k) t (w 0,..., w t )), the informational state of this subproblem i at time t cannot be summarized by the physical state x i t M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 28 / 60

Dual approximate dynamic programming (DADP) DADP principle and implementation 1 Decomposition and coordination A bird s eye view of decomposition methods A brief insight into Progressive Hedging Spatial decomposition methods in the deterministic case The stochastic case raises specific obstacles 2 Dual approximate dynamic programming (DADP) Problem statement DADP principle and implementation Numerical results on a small size problem 3 Theoretical questions Existence of a saddle point Convergence of the Uzawa algorithm Convergence w.r.t. information 4 Summary and research agenda M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 29 / 60

Dual approximate dynamic programming (DADP) We outline the main idea in DADP DADP principle and implementation The core idea of DADP is to replace the multiplier Λ (k) t at iteration k by its conditional expectation E(Λ (k) t y t ) where we introduce a new adapted information process y = ( ) y 0,..., y T 1 (More on the interpretation later) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 30 / 60

Dual approximate dynamic programming (DADP) Let us go on with our trick DADP principle and implementation Using this idea, we replace Subproblem i by min E u i,x i subject to ( T 1 t=0 ( ) ) L i t(x i t, u i t, w t+1 ) + E(Λ (k) t y t ) θt(x i i t, u i t) + K i (x i T ) x i t+1 = f i t (x i t, u i t, w t+1 ) u i t σ(w 0,..., w t ) The conditional expectation E(Λ (k) t y t ) is an (updated) function of the variable y t so that Subproblem i involves the two noises processes w and y If y follows a dynamical equation, DP applies M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 31 / 60

Dual approximate dynamic programming (DADP) DADP principle and implementation We obtain a dynamic programming equation by subsystem Assuming a non-controlled dynamics yt+1 i = ( ) hi t yt, w t+1 for the information process y, the DP equation writes VT i (x, y) = K i (x) ( Vt i (x, y) = min E L i u t(x, u, w t+1 ) + E ( Λ (k) t y t = y ) θt(x, i u) + Vt+1 i ( ) ) x i t+1, y t+1 subject to the extended dynamics x i t+1 = f i t (x, u, w t+1 ) y i t+1 = h i t(y, w t+1 ) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 32 / 60

Dual approximate dynamic programming (DADP) DADP principle and implementation What have we done? Trick: DADP as an approximation of the optimal multiplier λ t E ( λ t yt ) Interpretation: DADP as a decision-rule approach in the dual max min L( λ, u ) max min L ( λ, u ) λ u λ t y t u Interpretation: DADP as a constraint relaxation in the primal n ( θt( i x i t, u i ) n t = 0 E θt( i x i t, u i ) ) t y t = 0 i=1 i=1 M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 33 / 60

Dual approximate dynamic programming (DADP) DADP principle and implementation A bunch of practical questions remains open How to choose the information variables y t? Perfect memory: y t = ( w 0,..., w t ) Minimal information: y t cste Static information: y t = h t ( wt ) Dynamic information: y t+1 = h t ( yt, w t+1 ) How to obtain a feasible solution from the relaxed problem? Use an appropriate heuristic! How to accelerate the gradient algorithm? Augmented Lagrangian More sophisticated gradient methods How to handle more complex structures than the flower model? M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 34 / 60

Dual approximate dynamic programming (DADP) Numerical results on a small size problem 1 Decomposition and coordination A bird s eye view of decomposition methods A brief insight into Progressive Hedging Spatial decomposition methods in the deterministic case The stochastic case raises specific obstacles 2 Dual approximate dynamic programming (DADP) Problem statement DADP principle and implementation Numerical results on a small size problem 3 Theoretical questions Existence of a saddle point Convergence of the Uzawa algorithm Convergence w.r.t. information 4 Summary and research agenda M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 35 / 60

Dual approximate dynamic programming (DADP) Numerical results on a small size problem We consider 3 dams in a row, amenable to DP DECOMPOSITION M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 36 / 60

Dual approximate dynamic programming (DADP) Problem specification Numerical results on a small size problem We consider a 3 dam problem, over 12 time steps We relax each constraint with a given information process y i that depends on the constraint All random variable are discrete (noise, control, state) We test the following information processes Constant information: equivalent to replace each a.s. constraint by the expected constraint Part of noise: the information process depends on the constraint and is the inflow of the above dam yt i = wt i 1 Phantom state: the information process mimicks the optimal trajectory of the state of the first dam (by statistical regression over the known optimal trajectory in this case) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 37 / 60

Dual approximate dynamic programming (DADP) Numerical results are encouraging Numerical results on a small size problem DADP - E DADP - w i 1 DADP - dyn. DP Nb of iterations 165 170 25 1 Time (min) 2 3 67 41 Lower Bound 1.386 10 6 1.379 10 6 1.373 10 6 Final Value 1.335 10 6 1.321 10 6 1.344 10 6 1.366 10 6 Loss 2.3% 3.3% 1.6% ref. PhD thesis of J.-C. Alais M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 38 / 60

Dual approximate dynamic programming (DADP) Numerical results on a small size problem Summing up DADP Choose an information process y following y t+1 = f t ( yt, w t+1 ) Relax the almost sure coupling constraint into a conditional expectation Then apply a price decomposition scheme to the relaxed problem The subproblems can be solved by dynamic programming with the modest state ( x i t, y t ) In the theoretical part, we give Conditions for the existence of an L 1 multiplier Convergence of the algorithm (fixed information process) Consistency result (family of information process) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 39 / 60

Theoretical questions 1 Decomposition and coordination 2 Dual approximate dynamic programming (DADP) 3 Theoretical questions 4 Summary and research agenda M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 40 / 60

Theoretical questions What are the issues to consider? We treat the coupling constraints in a stochastic optimization problem by duality methods Uzawa algorithm is a dual method which is naturally described in an Hilbert space, but we cannot guarantee the existence of an optimal multiplier in the space L 2( Ω, F, P; R n)! Consequently, we extend the algorithm to the non-reflexive Banach space L ( Ω, F, P; R n), by giving a set of conditions ensuring the existence of a L 1( Ω, F, P; R n) optimal multiplier, and by providing a convergence result of the algorithm We also have to deal with the approximation induced by the information variable: we give an epi-convergence result related to such an approximation PhD thesis of V. Leclère M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 41 / 60

Theoretical questions Abstract formulation of the problem We consider the following abstract optimization problem ( P ) min u U ad J(u) s.t. Θ(u) C where U and V are two Banach spaces, and J : U R is the objective function U ad is the admissible set Θ : U V is the constraint function to be dualized C V is the cone of constraint Here, U is a space of random variables, and J is defined by J(u) = E ( j(u, w) ) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 42 / 60

Theoretical questions Existence of a saddle point 1 Decomposition and coordination A bird s eye view of decomposition methods A brief insight into Progressive Hedging Spatial decomposition methods in the deterministic case The stochastic case raises specific obstacles 2 Dual approximate dynamic programming (DADP) Problem statement DADP principle and implementation Numerical results on a small size problem 3 Theoretical questions Existence of a saddle point Convergence of the Uzawa algorithm Convergence w.r.t. information 4 Summary and research agenda M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 43 / 60

Theoretical questions Standard duality in L 2 spaces Existence of a saddle point (I) Assume that U = L 2( Ω, F, P; R n) and V = L 2( Ω, F, P; R m) The standard sufficient constraint qualification condition ( 0 ri Θ ( U ad dom(j) ) ) + C is scarcely satisfied in such a stochastic setting Proposition 1 If the σ-algebra F is not finite modulo P, a then for any subset U ad R n that is not an affine subspace, the set { U ad = u L p( Ω, F, P; R n) } u U ad P a.s. has an empty relative interior in L p, for any p < + a If the σ-algebra is finite modulo P, U and V are finite dimensional spaces M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 44 / 60

Theoretical questions Existence of a saddle point Standard duality in L 2 spaces (II) Consider the following optimization problem (a variation on a linear example given by R. Wets) inf u u 0,u 0 2 + E ( (u 1 + α) 2) 1 s.t. u 0 a u 1 0 u 0 u 1 w to be dualized where w is a random variable uniform on [1, 2] For a < 2, we exhibit a maximizing sequence of multipliers for the dual problem that does not converge in L 2. (We are in the so-called non relatively complete recourse case, that is, the case where the constraints on u 1 induce a stronger constraint on u 0 ) The optimal multiplier is not in L 2, but in ( L ) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 45 / 60

Theoretical questions Constraint qualification in ( L, L 1) Existence of a saddle point From now on, we assume that U = L ( Ω, F, P; R n) V = L ( Ω, F, P; R m) C = {0} where the σ-algebra F is not finite modulo P We consider the pairing ( L, L 1) with the following topologies: ) σ (L, L 1 : weak topology on L (coarsest topology such that all the L 1 -linear forms are continuous), ) τ (L, L 1 : Mackey-topology on L (finest topology such that the continuous linear forms are only the L 1 -linear forms) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 46 / 60

Theoretical questions Existence of a saddle point Weak closedness of linear subspaces of L Proposition 2 Let Θ : L ( Ω, F, P; R n) L ( Ω, F, P; R m) be a linear operator, and assume that there exists a linear operator Θ : L 1( Ω, F, P; R m) L 1( Ω, F, P; R n) such that: v, Θ(u) = Θ (v), u, u, v Then the linear operator Θ is weak continuous Applications Θ(u) = u E ( u B ) : non-anticipativity constraints Θ(u) = Au with A M m,n (R): finite number of constraints M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 47 / 60

A duality theorem Theoretical questions Existence of a saddle point ( ) P min J(u) s.t. Θ(u) = 0 u U with J(u) = E ( j(u, w) ) Theorem 1 Assume that j is a convex normal integrand, that J is continuous in the Mackey topology at some point u 0 such that Θ(u 0 ) = 0, and that Θ is weak continuous on L ( Ω, F, P; R n) Then, u U is an optimal solution of Problem ( P ) if and only if there exists λ L 1( Ω, F, P; R m) such that u arg min E u U Θ(u ) = 0 ( ) j(u, w) + λ Θ(u) Extension of a result given by R. Wets for non-anticipativity constraints M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 48 / 60

Theoretical questions Convergence of the Uzawa algorithm 1 Decomposition and coordination A bird s eye view of decomposition methods A brief insight into Progressive Hedging Spatial decomposition methods in the deterministic case The stochastic case raises specific obstacles 2 Dual approximate dynamic programming (DADP) Problem statement DADP principle and implementation Numerical results on a small size problem 3 Theoretical questions Existence of a saddle point Convergence of the Uzawa algorithm Convergence w.r.t. information 4 Summary and research agenda M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 49 / 60

Uzawa algorithm Theoretical questions Convergence of the Uzawa algorithm ( ) P min J(u) s.t. Θ(u) = 0 u U with J(u) = E ( j(u, w) ) The standard Uzawa algorithm makes sense u (k+1) arg min u U ad J(u) + λ (k), Θ(u) λ (k+1) = λ (k) }{{} dual +ρ Θ ( u (k+1)) } {{ } primal Note that all the multipliers λ (k) belong to L ( Ω, F, P; R m), as soon as the initial multiplier λ (0) L ( Ω, F, P; R m) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 50 / 60

Convergence result Theoretical questions Convergence of the Uzawa algorithm Theorem 2 Assume that 1 J : U R is proper, weak l.s.c., differentiable and a-convex 2 Θ : U V is affine, weak continuous and κ-lipschitz 3 U ad is weak closed and convex, 4 an admissible u 0 dom J Θ 1 (0) U ad exists 5 an optimal L 1 -multiplier to the constraint Θ ( u ) = 0 exists 6 the step ρ is such that 0 < ρ < 2a κ Then, there exists a subsequence { u (n k) } of the sequence generated by k N the Uzawa algorithm converging in L toward the optimal solution u of the primal problem M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 51 / 60

Theoretical questions Convergence of the Uzawa algorithm Discussion The result is not as good as expected (convergence of a subsequence) Improvements and extensions (inequality constraint) needed The Mackey-continuity assumption forbids the use of extended functions In order to deal with almost sure bound constraints, we can turn towards the work of T. Rockafellar and R. Wets In a series of 4 papers (stochastic convex programming), they have detailed the duality theory on two-stage and multistage problems, with the focus on non-anticipativity constraints These papers require a strict feasability assumption a relatively complete recourse assumption M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 52 / 60

Theoretical questions Convergence w.r.t. information 1 Decomposition and coordination A bird s eye view of decomposition methods A brief insight into Progressive Hedging Spatial decomposition methods in the deterministic case The stochastic case raises specific obstacles 2 Dual approximate dynamic programming (DADP) Problem statement DADP principle and implementation Numerical results on a small size problem 3 Theoretical questions Existence of a saddle point Convergence of the Uzawa algorithm Convergence w.r.t. information 4 Summary and research agenda M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 53 / 60

Theoretical questions Convergence w.r.t. information Relaxed problems Following the interpretation of DADP in terms of a relaxation of the original problem, and given a sequence {F n } n N of subfields of the σ-field F, we replace the abstract problem ( P ) min J(u) s.t. Θ(u) = 0 u U by the sequence of approximated problems: ( Pn ) min u U J(u) s.t. E( Θ(u) F n ) = 0 We assume the Kudo convergence of {F n } n N toward F: F n F E ( z F n ) L 1 E ( z F ), z L 1 (Ω, F, P; R) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 54 / 60

Convergence result Theoretical questions Convergence w.r.t. information Theorem 3 Assume that U is a topological space V = L p (Ω, F, P; R m ) with p [1, + ) J and Θ are continuous operators {F n } n N Kudo converges toward F Then the sequence { J n } n N epi-converges toward J, with {J(u) if u satisfies the constraints of ( ) P n Jn = + otherwise M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 55 / 60

Theoretical questions Convergence w.r.t. information Summing up theoretical questions Conditions for the existence of an L 1 multiplier Convergence of the algorithm (fixed information process) Consistency result (family of information process) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 56 / 60

Summary and research agenda 1 Decomposition and coordination 2 Dual approximate dynamic programming (DADP) 3 Theoretical questions 4 Summary and research agenda M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 57 / 60

Summary and research agenda Discussing DADP DADP (Dual Approximate Dynamic Programming) is a method to design stochastic price signals allowing decentralized agents to act as a team Hence, DADP is especially adapted to tackle large-scale stochastic optimal control problems, such as those found in energy management A host of theoretical and practical questions remains open We would like to test DADP on network models (smart grids) extending the works already made on flower models (unit commitment problem) and on chained models (hydraulic valley management) M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 58 / 60

Summary and research agenda Let us move to broader stochastic optimization challenges Stochastic optimization requires to make risk attitudes explicit robust, worst case, risk measures, in probability, almost surely, etc. Stochastic dynamic optimization requires to make online information explicit State-based functional approach Scenario-based measurability approach Numerical walls in dynamic programming, the bottleneck is the dimension of the state in stochastic programming, the bottleneck is the number of stages M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 59 / 60

Summary and research agenda Here is our research agenda for stochastic decomposition Combining different decomposition methods time: dynamic programming scenario: progressive hedging space: dual approximate dynamic programming Designing risk criterion compatible with decomposition (time-consistent dynamic risk measures) Mixing decomposition with analytical properties (convexity, linearity) on costs, constraints and dynamics functions M. De Lara (École des Ponts ParisTech) CMM, Santiago de Chile, December 2014 16 December 2014 60 / 60