fbpx
Wikipedia

Linear–quadratic–Gaussian control

In control theory, the linear–quadratic–Gaussian (LQG) control problem is one of the most fundamental optimal control problems, and it can also be operated repeatedly for model predictive control. It concerns linear systems driven by additive white Gaussian noise. The problem is to determine an output feedback law that is optimal in the sense of minimizing the expected value of a quadratic cost criterion. Output measurements are assumed to be corrupted by Gaussian noise and the initial state, likewise, is assumed to be a Gaussian random vector.

Under these assumptions an optimal control scheme within the class of linear control laws can be derived by a completion-of-squares argument.[1] This control law which is known as the LQG controller, is unique and it is simply a combination of a Kalman filter (a linear–quadratic state estimator (LQE)) together with a linear–quadratic regulator (LQR). The separation principle states that the state estimator and the state feedback can be designed independently. LQG control applies to both linear time-invariant systems as well as linear time-varying systems, and constitutes a linear dynamic feedback control law that is easily computed and implemented: the LQG controller itself is a dynamic system like the system it controls. Both systems have the same state dimension.

A deeper statement of the separation principle is that the LQG controller is still optimal in a wider class of possibly nonlinear controllers. That is, utilizing a nonlinear control scheme will not improve the expected value of the cost function. This version of the separation principle is a special case of the separation principle of stochastic control which states that even when the process and output noise sources are possibly non-Gaussian martingales, as long as the system dynamics are linear, the optimal control separates into an optimal state estimator (which may no longer be a Kalman filter) and an LQR regulator.[2][3]

In the classical LQG setting, implementation of the LQG controller may be problematic when the dimension of the system state is large. The reduced-order LQG problem (fixed-order LQG problem) overcomes this by fixing a priori the number of states of the LQG controller. This problem is more difficult to solve because it is no longer separable. Also, the solution is no longer unique. Despite these facts numerical algorithms are available[4][5][6][7] to solve the associated optimal projection equations[8][9] which constitute necessary and sufficient conditions for a locally optimal reduced-order LQG controller.[4]

LQG optimality does not automatically ensure good robustness properties.[10] The robust stability of the closed loop system must be checked separately after the LQG controller has been designed. To promote robustness some of the system parameters may be assumed stochastic instead of deterministic. The associated more difficult control problem leads to a similar optimal controller of which only the controller parameters are different.[5]

It is possible to compute the expected value of the cost function for the optimal gains, as well as any other set of stable gains.[11]

The LQG controller is also used to control perturbed non-linear systems.[12]

Mathematical description of the problem and solution edit

Continuous time edit

Consider the continuous-time linear dynamic system

 
 

where   represents the vector of state variables of the system,   the vector of control inputs and   the vector of measured outputs available for feedback. Both additive white Gaussian system noise   and additive white Gaussian measurement noise   affect the system. Given this system the objective is to find the control input history   which at every time   may depend linearly only on the past measurements   such that the following cost function is minimized:

 
 

where   denotes the expected value. The final time (horizon)   may be either finite or infinite. If the horizon tends to infinity the first term   of the cost function becomes negligible and irrelevant to the problem. Also to keep the costs finite the cost function has to be taken to be  .

The LQG controller that solves the LQG control problem is specified by the following equations:

 
 

The matrix   is called the Kalman gain of the associated Kalman filter represented by the first equation. At each time   this filter generates estimates   of the state   using the past measurements and inputs. The Kalman gain   is computed from the matrices  , the two intensity matrices   associated to the white Gaussian noises   and   and finally  . These five matrices determine the Kalman gain through the following associated matrix Riccati differential equation:

 
 

Given the solution   the Kalman gain equals

 

The matrix   is called the feedback gain matrix. This matrix is determined by the matrices   and   through the following associated matrix Riccati differential equation:

 
 

Given the solution   the feedback gain equals

 

Observe the similarity of the two matrix Riccati differential equations, the first one running forward in time, the second one running backward in time. This similarity is called duality. The first matrix Riccati differential equation solves the linear–quadratic estimation problem (LQE). The second matrix Riccati differential equation solves the linear–quadratic regulator problem (LQR). These problems are dual and together they solve the linear–quadratic–Gaussian control problem (LQG). So the LQG problem separates into the LQE and LQR problem that can be solved independently. Therefore, the LQG problem is called separable.

When   and the noise intensity matrices  ,   do not depend on   and when   tends to infinity the LQG controller becomes a time-invariant dynamic system. In that case the second matrix Riccati differential equation may be replaced by the associated algebraic Riccati equation.

Discrete time edit

Since the discrete-time LQG control problem is similar to the one in continuous-time, the description below focuses on the mathematical equations.

The discrete-time linear system equations are

 
 

Here   represents the discrete time index and   represent discrete-time Gaussian white noise processes with covariance matrices  , respectively, and are independent of each other.

The quadratic cost function to be minimized is

 
 

The discrete-time LQG controller is

 ,
 

and   corresponds to the predictive estimate  .

The Kalman gain equals

 

where   is determined by the following matrix Riccati difference equation that runs forward in time:

 

The feedback gain matrix equals

 

where   is determined by the following matrix Riccati difference equation that runs backward in time:

 

If all the matrices in the problem formulation are time-invariant and if the horizon   tends to infinity the discrete-time LQG controller becomes time-invariant. In that case the matrix Riccati difference equations may be replaced by their associated discrete-time algebraic Riccati equations. These determine the time-invariant linear–quadratic estimator and the time-invariant linear–quadratic regulator in discrete-time. To keep the costs finite instead of   one has to consider   in this case.

See also edit

References edit

  1. ^ Karl Johan Astrom (1970). Introduction to Stochastic Control Theory. Vol. 58. Academic Press. ISBN 0-486-44531-3.
  2. ^ Anders Lindquist (1973). "On Feedback Control of Linear Stochastic Systems". SIAM Journal on Control. 11 (2): 323–343. doi:10.1137/0311025..
  3. ^ Tryphon T. Georgiou and Anders Lindquist (2013). "The Separation Principle in Stochastic Control, Redux". IEEE Transactions on Automatic Control. 58 (10): 2481–2494. arXiv:1103.3005. doi:10.1109/TAC.2013.2259207. S2CID 12623187.
  4. ^ a b Van Willigenburg L.G.; De Koning W.L. (2000). "Numerical algorithms and issues concerning the discrete-time optimal projection equations". European Journal of Control. 6 (1): 93–100. doi:10.1016/s0947-3580(00)70917-4. Associated software download from Matlab Central.
  5. ^ a b Van Willigenburg L.G.; De Koning W.L. (1999). "Optimal reduced-order compensators for time-varying discrete-time systems with deterministic and white parameters". Automatica. 35: 129–138. doi:10.1016/S0005-1098(98)00138-1. Associated software download from Matlab Central.
  6. ^ Zigic D.; Watson L.T.; Collins E.G.; Haddad W.M.; Ying S. (1996). "Homotopy methods for solving the optimal projection equations for the H2 reduced order model problem". International Journal of Control. 56 (1): 173–191. doi:10.1080/00207179208934308.
  7. ^ Collins Jr. E.G; Haddad W.M.; Ying S. (1996). "A homotopy algorithm for reduced-order dynamic compensation using the Hyland-Bernstein optimal projection equations". Journal of Guidance, Control, and Dynamics. 19 (2): 407–417. doi:10.2514/3.21633.
  8. ^ Hyland D.C; Bernstein D.S. (1984). "The optimal projection equations for fixed order dynamic compensation" (PDF). IEEE Transactions on Automatic Control. AC-29 (11): 1034–1037. doi:10.1109/TAC.1984.1103418. hdl:2027.42/57875.
  9. ^ Bernstein D.S.; Davis L.D.; Hyland D.C. (1986). "The optimal projection equations for reduced-order discrete-time modeling estimation and control" (PDF). Journal of Guidance, Control, and Dynamics. 9 (3): 288–293. Bibcode:1986JGCD....9..288B. doi:10.2514/3.20105. hdl:2027.42/57880.
  10. ^ Green, Michael; Limebeer, David J. N. (1995). Linear Robust Control. Englewood Cliffs: Prentice Hall. p. 27. ISBN 0-13-102278-4.
  11. ^ Matsakis, Demetrios (March 8, 2019). "The effects of proportional steering strategies on the behavior of controlled clocks". Metrologia. 56 (2): 025007. Bibcode:2019Metro..56b5007M. doi:10.1088/1681-7575/ab0614.
  12. ^ Athans M. (1971). "The role and use of the stochastic Linear-Quadratic-Gaussian problem in control system design". IEEE Transactions on Automatic Control. AC-16 (6): 529–552. doi:10.1109/TAC.1971.1099818.

Further reading edit

  • Stengel, Robert F. (1994). Optimal Control and Estimation. New York: Dover. ISBN 0-486-68200-5.

linear, quadratic, gaussian, control, control, theory, linear, quadratic, gaussian, control, problem, most, fundamental, optimal, control, problems, also, operated, repeatedly, model, predictive, control, concerns, linear, systems, driven, additive, white, gau. In control theory the linear quadratic Gaussian LQG control problem is one of the most fundamental optimal control problems and it can also be operated repeatedly for model predictive control It concerns linear systems driven by additive white Gaussian noise The problem is to determine an output feedback law that is optimal in the sense of minimizing the expected value of a quadratic cost criterion Output measurements are assumed to be corrupted by Gaussian noise and the initial state likewise is assumed to be a Gaussian random vector Under these assumptions an optimal control scheme within the class of linear control laws can be derived by a completion of squares argument 1 This control law which is known as the LQG controller is unique and it is simply a combination of a Kalman filter a linear quadratic state estimator LQE together with a linear quadratic regulator LQR The separation principle states that the state estimator and the state feedback can be designed independently LQG control applies to both linear time invariant systems as well as linear time varying systems and constitutes a linear dynamic feedback control law that is easily computed and implemented the LQG controller itself is a dynamic system like the system it controls Both systems have the same state dimension A deeper statement of the separation principle is that the LQG controller is still optimal in a wider class of possibly nonlinear controllers That is utilizing a nonlinear control scheme will not improve the expected value of the cost function This version of the separation principle is a special case of the separation principle of stochastic control which states that even when the process and output noise sources are possibly non Gaussian martingales as long as the system dynamics are linear the optimal control separates into an optimal state estimator which may no longer be a Kalman filter and an LQR regulator 2 3 In the classical LQG setting implementation of the LQG controller may be problematic when the dimension of the system state is large The reduced order LQG problem fixed order LQG problem overcomes this by fixing a priori the number of states of the LQG controller This problem is more difficult to solve because it is no longer separable Also the solution is no longer unique Despite these facts numerical algorithms are available 4 5 6 7 to solve the associated optimal projection equations 8 9 which constitute necessary and sufficient conditions for a locally optimal reduced order LQG controller 4 LQG optimality does not automatically ensure good robustness properties 10 The robust stability of the closed loop system must be checked separately after the LQG controller has been designed To promote robustness some of the system parameters may be assumed stochastic instead of deterministic The associated more difficult control problem leads to a similar optimal controller of which only the controller parameters are different 5 It is possible to compute the expected value of the cost function for the optimal gains as well as any other set of stable gains 11 The LQG controller is also used to control perturbed non linear systems 12 Contents 1 Mathematical description of the problem and solution 1 1 Continuous time 1 2 Discrete time 2 See also 3 References 4 Further readingMathematical description of the problem and solution editContinuous time edit Consider the continuous time linear dynamic system x t A t x t B t u t v t displaystyle dot mathbf x t A t mathbf x t B t mathbf u t mathbf v t nbsp y t C t x t w t displaystyle mathbf y t C t mathbf x t mathbf w t nbsp where x displaystyle mathbf x nbsp represents the vector of state variables of the system u displaystyle mathbf u nbsp the vector of control inputs and y displaystyle mathbf y nbsp the vector of measured outputs available for feedback Both additive white Gaussian system noise v t displaystyle mathbf v t nbsp and additive white Gaussian measurement noise w t displaystyle mathbf w t nbsp affect the system Given this system the objective is to find the control input history u t displaystyle mathbf u t nbsp which at every time t displaystyle mathbf t nbsp may depend linearly only on the past measurements y t 0 t lt t displaystyle mathbf y t 0 leq t lt t nbsp such that the following cost function is minimized J E x T T F x T 0 T x T t Q t x t u T t R t u t d t displaystyle J mathbb E left mathbf x mathrm T T F mathbf x T int 0 T mathbf x mathrm T t Q t mathbf x t mathbf u mathrm T t R t mathbf u t dt right nbsp F 0 Q t 0 R t gt 0 displaystyle F geq 0 quad Q t geq 0 quad R t gt 0 nbsp where E displaystyle mathbb E nbsp denotes the expected value The final time horizon T displaystyle mathbf T nbsp may be either finite or infinite If the horizon tends to infinity the first term x T T F x T displaystyle mathbf x mathrm T T F mathbf x T nbsp of the cost function becomes negligible and irrelevant to the problem Also to keep the costs finite the cost function has to be taken to be J T displaystyle mathbf J T nbsp The LQG controller that solves the LQG control problem is specified by the following equations x t A t x t B t u t L t y t C t x t x 0 E x 0 displaystyle dot hat mathbf x t A t hat mathbf x t B t mathbf u t L t left mathbf y t C t hat mathbf x t right quad hat mathbf x 0 mathbb E left mathbf x 0 right nbsp u t K t x t displaystyle mathbf u t K t hat mathbf x t nbsp The matrix L t displaystyle mathbf L t nbsp is called the Kalman gain of the associated Kalman filter represented by the first equation At each time t displaystyle mathbf t nbsp this filter generates estimates x t displaystyle hat mathbf x t nbsp of the state x t displaystyle mathbf x t nbsp using the past measurements and inputs The Kalman gain L t displaystyle mathbf L t nbsp is computed from the matrices A t C t displaystyle mathbf A t C t nbsp the two intensity matrices V t W t displaystyle mathbf V t W t nbsp associated to the white Gaussian noises v t displaystyle mathbf v t nbsp and w t displaystyle mathbf w t nbsp and finally E x 0 x T 0 displaystyle mathbb E left mathbf x 0 mathbf x mathrm T 0 right nbsp These five matrices determine the Kalman gain through the following associated matrix Riccati differential equation P t A t P t P t A T t P t C T t W 1 t C t P t V t displaystyle dot P t A t P t P t A mathrm T t P t C mathrm T t mathbf W 1 t C t P t V t nbsp P 0 E x 0 x T 0 displaystyle P 0 mathbb E left mathbf x 0 mathbf x mathrm T 0 right nbsp Given the solution P t 0 t T displaystyle P t 0 leq t leq T nbsp the Kalman gain equals L t P t C T t W 1 t displaystyle mathbf L t P t C mathrm T t W 1 t nbsp The matrix K t displaystyle mathbf K t nbsp is called the feedback gain matrix This matrix is determined by the matrices A t B t Q t R t displaystyle mathbf A t B t Q t R t nbsp and F displaystyle mathbf F nbsp through the following associated matrix Riccati differential equation S t A T t S t S t A t S t B t R 1 t B T t S t Q t displaystyle dot S t A mathrm T t S t S t A t S t B t R 1 t B mathrm T t S t Q t nbsp S T F displaystyle mathbf S T F nbsp Given the solution S t 0 t T displaystyle mathbf S t 0 leq t leq T nbsp the feedback gain equals K t R 1 t B T t S t displaystyle mathbf K t R 1 t B mathrm T t S t nbsp Observe the similarity of the two matrix Riccati differential equations the first one running forward in time the second one running backward in time This similarity is called duality The first matrix Riccati differential equation solves the linear quadratic estimation problem LQE The second matrix Riccati differential equation solves the linear quadratic regulator problem LQR These problems are dual and together they solve the linear quadratic Gaussian control problem LQG So the LQG problem separates into the LQE and LQR problem that can be solved independently Therefore the LQG problem is called separable When A t B t C t Q t R t displaystyle mathbf A t B t C t Q t R t nbsp and the noise intensity matrices V t displaystyle mathbf V t nbsp W t displaystyle mathbf W t nbsp do not depend on t displaystyle mathbf t nbsp and when T displaystyle mathbf T nbsp tends to infinity the LQG controller becomes a time invariant dynamic system In that case the second matrix Riccati differential equation may be replaced by the associated algebraic Riccati equation Discrete time edit Since the discrete time LQG control problem is similar to the one in continuous time the description below focuses on the mathematical equations The discrete time linear system equations are x i 1 A i x i B i u i v i displaystyle mathbf x i 1 A i mathbf x i B i mathbf u i mathbf v i nbsp y i C i x i w i displaystyle mathbf y i C i mathbf x i mathbf w i nbsp Here i displaystyle mathbf i nbsp represents the discrete time index and v i w i displaystyle mathbf v i mathbf w i nbsp represent discrete time Gaussian white noise processes with covariance matrices V i W i displaystyle mathbf V i W i nbsp respectively and are independent of each other The quadratic cost function to be minimized is J E x N T F x N i 0 N 1 x i T Q i x i u i T R i u i displaystyle J mathbb E left mathbf x N mathrm T F mathbf x N sum i 0 N 1 mathbf x i mathrm T Q i mathbf x i mathbf u i mathrm T R i mathbf u i right nbsp F 0 Q i 0 R i gt 0 displaystyle F geq 0 Q i geq 0 R i gt 0 nbsp The discrete time LQG controller is x i 1 A i x i B i u i L i 1 y i 1 C i 1 A i x i B i u i x 0 E x 0 displaystyle hat mathbf x i 1 A i hat mathbf x i B i mathbf u i L i 1 left mathbf y i 1 C i 1 left A i hat mathbf x i B i mathbf u i right right qquad hat mathbf x 0 mathbb E mathbf x 0 nbsp u i K i x i displaystyle mathbf u i K i hat mathbf x i nbsp and x i displaystyle hat mathbf x i nbsp corresponds to the predictive estimate x i E x i y i u i 1 displaystyle hat mathbf x i mathbb E mathbf x i mathbf y i mathbf u i 1 nbsp The Kalman gain equals L i P i C i T C i P i C i T W i 1 displaystyle mathbf L i P i C i mathrm T C i P i C i mathrm T W i 1 nbsp where P i displaystyle mathbf P i nbsp is determined by the following matrix Riccati difference equation that runs forward in time P i 1 A i P i P i C i T C i P i C i T W i 1 C i P i A i T V i P 0 E x 0 x 0 x 0 x 0 T displaystyle P i 1 A i left P i P i C i mathrm T left C i P i C i mathrm T W i right 1 C i P i right A i mathrm T V i qquad P 0 mathbb E left mathbf x 0 hat mathbf x 0 right left mathbf x 0 hat mathbf x 0 right mathrm T nbsp The feedback gain matrix equals K i B i T S i 1 B i R i 1 B i T S i 1 A i displaystyle mathbf K i B i mathrm T S i 1 B i R i 1 B i mathrm T S i 1 A i nbsp where S i displaystyle mathbf S i nbsp is determined by the following matrix Riccati difference equation that runs backward in time S i A i T S i 1 S i 1 B i B i T S i 1 B i R i 1 B i T S i 1 A i Q i S N F displaystyle S i A i mathrm T left S i 1 S i 1 B i left B i mathrm T S i 1 B i R i right 1 B i mathrm T S i 1 right A i Q i quad S N F nbsp If all the matrices in the problem formulation are time invariant and if the horizon N displaystyle mathbf N nbsp tends to infinity the discrete time LQG controller becomes time invariant In that case the matrix Riccati difference equations may be replaced by their associated discrete time algebraic Riccati equations These determine the time invariant linear quadratic estimator and the time invariant linear quadratic regulator in discrete time To keep the costs finite instead of J displaystyle mathbf J nbsp one has to consider J N displaystyle mathbf J N nbsp in this case See also editStochastic control Witsenhausen s counterexampleReferences edit Karl Johan Astrom 1970 Introduction to Stochastic Control Theory Vol 58 Academic Press ISBN 0 486 44531 3 Anders Lindquist 1973 On Feedback Control of Linear Stochastic Systems SIAM Journal on Control 11 2 323 343 doi 10 1137 0311025 Tryphon T Georgiou and Anders Lindquist 2013 The Separation Principle in Stochastic Control Redux IEEE Transactions on Automatic Control 58 10 2481 2494 arXiv 1103 3005 doi 10 1109 TAC 2013 2259207 S2CID 12623187 a b Van Willigenburg L G De Koning W L 2000 Numerical algorithms and issues concerning the discrete time optimal projection equations European Journal of Control 6 1 93 100 doi 10 1016 s0947 3580 00 70917 4 Associated software download from Matlab Central a b Van Willigenburg L G De Koning W L 1999 Optimal reduced order compensators for time varying discrete time systems with deterministic and white parameters Automatica 35 129 138 doi 10 1016 S0005 1098 98 00138 1 Associated software download from Matlab Central Zigic D Watson L T Collins E G Haddad W M Ying S 1996 Homotopy methods for solving the optimal projection equations for the H2 reduced order model problem International Journal of Control 56 1 173 191 doi 10 1080 00207179208934308 Collins Jr E G Haddad W M Ying S 1996 A homotopy algorithm for reduced order dynamic compensation using the Hyland Bernstein optimal projection equations Journal of Guidance Control and Dynamics 19 2 407 417 doi 10 2514 3 21633 Hyland D C Bernstein D S 1984 The optimal projection equations for fixed order dynamic compensation PDF IEEE Transactions on Automatic Control AC 29 11 1034 1037 doi 10 1109 TAC 1984 1103418 hdl 2027 42 57875 Bernstein D S Davis L D Hyland D C 1986 The optimal projection equations for reduced order discrete time modeling estimation and control PDF Journal of Guidance Control and Dynamics 9 3 288 293 Bibcode 1986JGCD 9 288B doi 10 2514 3 20105 hdl 2027 42 57880 Green Michael Limebeer David J N 1995 Linear Robust Control Englewood Cliffs Prentice Hall p 27 ISBN 0 13 102278 4 Matsakis Demetrios March 8 2019 The effects of proportional steering strategies on the behavior of controlled clocks Metrologia 56 2 025007 Bibcode 2019Metro 56b5007M doi 10 1088 1681 7575 ab0614 Athans M 1971 The role and use of the stochastic Linear Quadratic Gaussian problem in control system design IEEE Transactions on Automatic Control AC 16 6 529 552 doi 10 1109 TAC 1971 1099818 Further reading editStengel Robert F 1994 Optimal Control and Estimation New York Dover ISBN 0 486 68200 5 Retrieved from https en wikipedia org w index php title Linear quadratic Gaussian control amp oldid 1161646560, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.