19 LINEAR QUADRATIC REGULATOR



Similar documents
Nonlinear Systems and Control Lecture # 15 Positive Real Transfer Functions & Connection with Lyapunov Stability. p. 1/?

Lecture 13 Linear quadratic Lyapunov theory

Methods for Finding Bases

MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.

Algebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions.

Solving Systems of Linear Equations

Linear-Quadratic Optimal Controller 10.3 Optimal Linear Control Systems

by the matrix A results in a vector which is a reflection of the given

Lecture 7: Finding Lyapunov Functions 1

x = + x 2 + x

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

Understanding Poles and Zeros

Nonlinear Programming Methods.S2 Quadratic Programming

4 Lyapunov Stability Theory

LS.6 Solution Matrices

Partial Fractions. p(x) q(x)

Introduction to Matrix Algebra

2x + y = 3. Since the second equation is precisely the same as the first equation, it is enough to find x and y satisfying the system

Practical Guide to the Simplex Method of Linear Programming

MATH APPLIED MATRIX THEORY

Zeros of a Polynomial Function

3. Reaction Diffusion Equations Consider the following ODE model for population growth

Homogeneous systems of algebraic equations. A homogeneous (ho-mo-geen -ius) system of linear algebraic equations is one in which

PYTHAGOREAN TRIPLES KEITH CONRAD

Systems of Linear Equations

T ( a i x i ) = a i T (x i ).

1 VECTOR SPACES AND SUBSPACES

Multi-variable Calculus and Optimization

Linear Algebra Notes

Lecture 2. Marginal Functions, Average Functions, Elasticity, the Marginal Principle, and Constrained Optimization

Chapter 12 Modal Decomposition of State-Space Models 12.1 Introduction The solutions obtained in previous chapters, whether in time domain or transfor

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

CONTROLLABILITY. Chapter Reachable Set and Controllability. Suppose we have a linear system described by the state equation

ECE 3510 Final given: Spring 11

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Solving Linear Systems, Continued and The Inverse of a Matrix

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8

Scalar Valued Functions of Several Variables; the Gradient Vector

Lecture L3 - Vectors, Matrices and Coordinate Transformations

MAT 200, Midterm Exam Solution. a. (5 points) Compute the determinant of the matrix A =

General Framework for an Iterative Solution of Ax b. Jacobi s Method

Inner Product Spaces

DERIVATIVES AS MATRICES; CHAIN RULE

5 Homogeneous systems

Chapter 19. General Matrices. An n m matrix is an array. a 11 a 12 a 1m a 21 a 22 a 2m A = a n1 a n2 a nm. The matrix A has n row vectors

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

Stability. Chapter 4. Topics : 1. Basic Concepts. 2. Algebraic Criteria for Linear Systems. 3. Lyapunov Theory with Applications to Linear Systems

HOMEWORK 5 SOLUTIONS. n!f n (1) lim. ln x n! + xn x. 1 = G n 1 (x). (2) k + 1 n. (n 1)!

MA107 Precalculus Algebra Exam 2 Review Solutions

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components

1 Sets and Set Notation.

Notes on Determinant

First, we show how to use known design specifications to determine filter order and 3dB cut-off

Root Locus. E(s) K. R(s) C(s) 1 s(s+a) Consider the closed loop transfer function:

Chapter 13 Internal (Lyapunov) Stability 13.1 Introduction We have already seen some examples of both stable and unstable systems. The objective of th

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

Unified Lecture # 4 Vectors

Row Echelon Form and Reduced Row Echelon Form

Vector and Matrix Norms

Partial Fractions. Combining fractions over a common denominator is a familiar operation from algebra:

Iterative Methods for Solving Linear Systems

Numerical Analysis Lecture Notes

Solution to Homework 2

Using the Theory of Reals in. Analyzing Continuous and Hybrid Systems

Lecture 1: Systems of Linear Equations

MATH2210 Notebook 1 Fall Semester 2016/ MATH2210 Notebook Solving Systems of Linear Equations... 3

Solving Systems of Linear Equations

1 Review of Least Squares Solutions to Overdetermined Systems

MATH 4330/5330, Fourier Analysis Section 11, The Discrete Fourier Transform

Math 4310 Handout - Quotient Vector Spaces

Recall that two vectors in are perpendicular or orthogonal provided that their dot

CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C.

DIFFERENTIABILITY OF COMPLEX FUNCTIONS. Contents

3.2 Sources, Sinks, Saddles, and Spirals

Modeling and Simulation of a Three Degree of Freedom Longitudinal Aero plane System. Figure 1: Boeing 777 and example of a two engine business jet

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).

3. INNER PRODUCT SPACES

Positive Feedback and Oscillators

Lecture 3: Finding integer solutions to systems of linear equations

7 Gaussian Elimination and LU Factorization

Partial Fractions. (x 1)(x 2 + 1)

Solutions to Math 51 First Exam January 29, 2015

Chapter 17. Orthogonal Matrices and Symmetries of Space

Orthogonal Projections

CURVE FITTING LEAST SQUARES APPROXIMATION

TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAA

5.3 The Cross Product in R 3

Σ _. Feedback Amplifiers: One and Two Pole cases. Negative Feedback:

The Method of Partial Fractions Math 121 Calculus II Spring 2015

Lecture 5 Principal Minors and the Hessian

THE DIMENSION OF A VECTOR SPACE

Linear Algebra: Vectors

Lecture Notes 2: Matrices as Systems of Linear Equations

3 Contour integrals and Cauchy s Theorem

MATHEMATICS FOR ENGINEERING BASIC ALGEBRA

Lecture 5: Singular Value Decomposition SVD (1)

10.2 ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS. The Jacobi Method

ROUTH S STABILITY CRITERION

Transcription:

19 LINEAR QUADRATIC REGULATOR 19.1 Introduction The simple form of loopshaping in scalar systems does not extend directly to multivariable (MIMO) plants, which are characterized by transfer matrices instead of transfer functions. The notion of optimality is closely tied to MIMO control system design. Optimal controllers, i.e., controllers that are the best possible, according to some figure of merit, turn out to generate only stabilizing controllers for MIMO plants. In this sense, optimal control solutions provide an automated design procedure we have only to decide what figure of merit to use. The linear quadratic regulator (LQR) is a well-known design technique that provides practical feedback gains. (Continued on next page)

19.2 Full-State Feedback 93 19.2 Full-State Feedback For the derivation of the linear quadratic regulator, we assume the plant to be written in state-space form ẋ = Ax + Bu, and that all of the n states x are available for the controller. The feedback gain is a matrix K, implemented as u = K(x x desired ). The system dynamics are then written as: ẋ = (A BK)x + BKx desired. (210) x desired represents the vector of desired states, and serves as the external input to the closedloop system. The A-matrix of the closed loop system is (A BK), and the B-matrix of the closed-loop system is BK. The closed-loop system has exactly as many outputs as inputs: n. The column dimension of B equals the number of channels available in u, and must match the row dimension of K. Pole-placement is the process of placing the poles of (A BK) in stable, suitably-damped locations in the complex plane. 19.3 The Maximum Principle First we develop a general procedure for solving optimal control problems, using the calculus of variations. We begin with the statemenf the problem for a fixed end time t f : choose u(t) to minimize J = ω(x(t f )) + L(x(t), u(t), t)dt (211) subject to ẋ = f(x(t), u(t), t) (212) x( ) = x o (213) where ω(x(t f ), t f ) is the terminal cost; the total cost J is a sum of the terminal cost and an integral along the way. We assume that L(x(t), u(t), t) is nonnegative. The first step is to augment the cost using the costate vector η(t). J = ω(x(t f )) + (L + η T (f ẋ))dt (214) Clearly, η(t) can be anything we choose, since it multiplies f ẋ = 0. Along the optimum trajectory, variations in J and hence J should vanish. This follows from the fact that J is chosen to be continuous in x, u, and t. We write the variation as ζj = ω x ζx(t f ) + Lx ζx + L u ζu + η T f x ζx + η T f u ζu η T ζẋ dt, (215) where subscripts denote partial derivatives. The last term above can be evaluated using integration by parts as and thus xdt = η T (t f )ζx(t f ) + η T T η T ζ ( )ζx( ) + η ζxdt, (216)

94 19 LINEAR QUADRATIC REGULATOR ζj = ω x (x(t f ))ζx(t f ) + (L u + η T f u )ζudt + (217) (Lx + η T f x + η T )ζxdt η T (t f )ζx(t f ) + η T ( )ζx( ). Now the last term is zero, since we cannot vary the initial condition of the state by changing something later in time - it is a fixed value. This way of writing J makes it clear that there are three components of the variation that must independently be zero (since we can vary any of x, u, or x(t f )): L u + η T f u = 0 (218) L x + η T T f x + η = 0 (219) ω x (x(t f )) η T (t f ) = 0. (220) The second and third requirements are met by explicitly setting T η = L x η T f x (221) η T (t f ) = ω x (x(t f )). (222) The evolution of η is given in reverse time, from a final state to the initial. Hence we see the primary difficulty of solving optimal control problems: the state propogates forward in time, while the costate propogates backward. The state and costate are coordinated through the above equations. 19.4 Gradient Method Solution for the General Case Numerical solutions to the general problem are iterative, and the simplest approach is the gradient method. It is outlined as follows: 1. For a given x o, pick a control history u(t). 2. Propogate ẋ = f(x, u, t) forward in time to create a state trajectory. 3. Evaluate ω x (x(t f )), and the propogate the costate backward in time from t f to, using Equation 221. 4. At each time step, choose ζu = K(L u + η T f u ), where K is a positive scalar or a positive definite matrix in the case of multiple input channels. 5. Let u = u + ζu. 6. Go back to step 2 and repeat loop until solution has converged.

19.5 LQR Solution 95 The first three steps are consistent in the sense that x is computed directly from x( and u, and η is computed directly from x and x(t f ). All of ζj in Equation 217 except the integral with ζu is therefore eliminated explicitly. The choice of ζu in step 4 then achieves ζj < 0, unless ζu = 0, in which case the problem is solved. The gradient method is quite popular in applications and for complex problems may be the only way to effectively develop a solution. The major difficulties are computational expense, and the requiremenf having a reasonable control trajectory to begin. 19.5 LQR Solution In the case of the Linear Quadratic Regulator (with zero terminal cost), we set ω = 0, and 1 L = 1 x T Qx + u T Ru, (223) 2 2 where the requirement that L 0 implies that both Q and R are positive definite. In the case of linear plant dynamics also, we have so that L x = x T Q (224) L u = u T R (225) f x = A (226) f u = B, (227) ẋ = Ax + Bu (228) x( ) = x o (229) η = Qx A T η (230) η(t f ) = 0 (231) Ru + B T η = 0. (232) Since the systems are clearly linear, we try a connection η = P x. Inserting this into the η equation, and then using the ẋ equation, and a substitution for u, we obtain P Ax + A T P x + Qx P BR 1 B T P x + P = 0. (233) This has to hold for all x, so in fact it is a matrix equation, the matrix Riccati equation. The steady-state solution is given satisfies P A + A T P + Q P BR 1 B T P = 0. (234)

96 19 LINEAR QUADRATIC REGULATOR 19.6 Optimal Full-State Feedback This equation is the matrix algebraic Riccati equation (MARE), whose solution P is needed to compute the optimal feedback gain K. The MARE is easily solved by standard numerical tools in linear algebra. The equation Ru + B T η = 0 gives the feedback law: u = R 1 B T P x. (235) 19.7 Properties and Use of the LQR Static Gain. The LQR generates a static gain matrix K, which is not a dynamical system. Hence, the order of the closed-loop system is the same as thaf the plant. Robustness. The LQR achieves infinite gain margin: k g =, implying that the loci of (P C) (scalar case) or (det(i +P C) 1) (MIMO case) approach the origin along the imaginary axis. The LQR also guarantees phase margin ρ 60 degrees. This is in good agreement with the practical guidelines for control system design. Output Variables. In many cases, it is not the states x which are to be minimized, but the output variables y. In this case, we set the weighting matrix Q = C T Q C, since y = Cx, and the auxiliary matrix Q weights the planutput. Behavior of Closed-Loop Poles: Expensive Control. When R >> C T Q C, the cost function is dominated by the control effort u, and so the controller minimizes the control action itself. In the case of a completely stable plant, the gain will indeed go to zero, so that the closed-loop poles approach the open-loop plant poles in a manner consistent with the scalar root locus. The optimal control must always stabilize the closed-loop system, however, so there should be some account made for unstable plant poles. The expensive control solution puts stable closed-loop poles at the mirror images of the unstable plant poles. Behavior of Closed-Loop Poles: Cheap Control. When R << C T Q C, the cost function is dominated by the output errors y, and there is no penalty for using large u. There are two groups of closed-loop poles. First, poles are placed at stable plant zeros, and at the mirror images of the unstable plant zeros. This part is akin to the high-gain limiting case of the root locus. The remaining poles assume a Butterworth pattern, whose radius increases to infinity as R becomes smaller and smaller. The Butterworth pattern refers to an arc in the stable left-half plane, as shown in the figure. The angular separation of n closed-loop poles on the arc is constant, and equal to 180 /n. An angle 90 /n separates the most lightly-damped poles from the imaginary axis.

19.8 Proof of the Gain and Phase Margins 97 φ /2 φ n = 4 φ = 45 degrees φ φ φ /2 19.8 Proof of the Gain and Phase Margins The above stated properties of the LQR control are not difficult to prove. We need an intermediate tool (which is very useful in its own right) to prove it: the Liapunov function. Consider a scalar, continuous, positive definite function of the state V (x). If this function is always decreasing, except at the origin, then the origin is asymptotically stable. There is an intuitive feel to the Liapunov function; it is like the energy of a physical system. If it can be shown that energy is leaving such a system, then it will settle. It suffices to find one Liapunov function to show that a system is stable. Now, in the case of the LQR control, pick 1 V (x) = x T P x. 2 It can be shown that P is positive definite; suppose that instead of the design gain B, we have actual gain B, giving (with constant P based on the design system) V = 2x T P ẋ (236) = 2x T P (Ax B R 1 B T Kx) (237) T = x ( Q + P BR 1 B T P )x 2x T KB R 1 B T P x. (238) where we used the Riccati equation in a substitution. Since we require V < 0 for x = 0, we must have Q + P BG 2P (B B )G > 0, (239) where G = R 1 B T P. Let B = BN, where N is a diagonal matrix of random, unknown gains on the control channels. The above equation becomes Q P B(I 2N)R 1 B T P > 0. (240) This is satisfied if, for all channels, N i,i > 1/2. Thus, we see that the LQR provides for one-half gain reduction, and infinite gain amplification, in all channels.

98 20 KALMAN FILTER The phase margin is seen (roughly) by allowing N to be a diagonal matrix of transfer functions N i,i = e jδ i. In this case, we require that the real parf N i,i is greater than onehalf. This is true for all δ i < 60, and hence sixty degrees of phase margin is provided in all channels.