fbpx
Wikipedia

Linear–quadratic regulator

The theory of optimal control is concerned with operating a dynamic system at minimum cost. The case where the system dynamics are described by a set of linear differential equations and the cost is described by a quadratic function is called the LQ problem. One of the main results in the theory is that the solution is provided by the linear–quadratic regulator (LQR), a feedback controller whose equations are given below.

LQR controllers possess inherent robustness with guaranteed gain and phase margin,[1] and they also are part of the solution to the LQG (linear–quadratic–Gaussian) problem. Like the LQR problem itself, the LQG problem is one of the most fundamental problems in control theory.

General description Edit

The settings of a (regulating) controller governing either a machine or process (like an airplane or chemical reactor) are found by using a mathematical algorithm that minimizes a cost function with weighting factors supplied by a human (engineer). The cost function is often defined as a sum of the deviations of key measurements, like altitude or process temperature, from their desired values. The algorithm thus finds those controller settings that minimize undesired deviations. The magnitude of the control action itself may also be included in the cost function.

The LQR algorithm reduces the amount of work done by the control systems engineer to optimize the controller. However, the engineer still needs to specify the cost function parameters, and compare the results with the specified design goals. Often this means that controller construction will be an iterative process in which the engineer judges the "optimal" controllers produced through simulation and then adjusts the parameters to produce a controller more consistent with design goals.

The LQR algorithm is essentially an automated way of finding an appropriate state-feedback controller. As such, it is not uncommon for control engineers to prefer alternative methods, like full state feedback, also known as pole placement, in which there is a clearer relationship between controller parameters and controller behavior. Difficulty in finding the right weighting factors limits the application of the LQR based controller synthesis.

Versions Edit

Finite-horizon, continuous-time Edit

For a continuous-time linear system, defined on  , described by:

 

where   (that is,   is an  -dimensional real-valued vector) is the state of the system and   is the control input. Given a quadratic cost function for the system, defined as:

 

the feedback control law that minimizes the value of the cost is:

 

where   is given by:

 

and   is found by solving the continuous time Riccati differential equation:

 

with the boundary condition:

 

The first order conditions for Jmin are:

1) State equation

 

2) Co-state equation

 

3) Stationary equation

 

4) Boundary conditions

 

and  

Infinite-horizon, continuous-time Edit

For a continuous-time linear system described by:

 

with a cost function defined as:

 

the feedback control law that minimizes the value of the cost is:

 

where   is given by:

 

and   is found by solving the continuous time algebraic Riccati equation:

 

This can be also written as:

 

with

 

Finite-horizon, discrete-time Edit

For a discrete-time linear system described by: [2]

 

with a performance index defined as:

 , where   is the time horizon

the optimal control sequence minimizing the performance index is given by:

 

where:

 

and   is found iteratively backwards in time by the dynamic Riccati equation:

 

from terminal condition  .[3] Note that   is not defined, since   is driven to its final state   by  .

Infinite-horizon, discrete-time Edit

For a discrete-time linear system described by:

 

with a performance index defined as:

 

the optimal control sequence minimizing the performance index is given by:

 

where:

 

and   is the unique positive definite solution to the discrete time algebraic Riccati equation (DARE):

 .

This can be also written as:

 

with:

 .

Note that one way to solve the algebraic Riccati equation is by iterating the dynamic Riccati equation of the finite-horizon case until it converges.

Constraints Edit

In practice, not all values of   may be allowed. One common constraint is the linear one:

 

The finite horizon version of this is a convex optimization problem, and so the problem is often solved repeatedly with a receding horizon. This is a form of model predictive control.[4][5]

Related controllers Edit

Quadratic-quadratic regulator Edit

If the state equation is quadratic then the problem is known as the quadratic-quadratic regulator (QQR). The Al'Brekht algorithm can be applied to reduce this problem to one that can be solved efficiently using tensor based linear solvers.[6]

Polynomial-quadratic regulator Edit

If the state equation is polynomial then the problem is known as the polynomial-quadratic regulator (PQR). Again, the Al'Brekht algorithm can be applied to reduce this problem to a large linear one which can be solved with a generalization of the Bartels-Stewart algorithm; this is feasible provided that the degree of the polynomial is not too high.[7]

Model-predictive control Edit

Model predictive control and linear-quadratic regulators are two types of optimal control methods that have distinct approaches for setting the optimization costs. In particular, when the LQR is run repeatedly with a receding horizon, it becomes a form of model predictive control (MPC). In general, however, MPC does not rely on any assumptions regarding linearity of the system.

References Edit

  1. ^ Lehtomaki, N.; Sandell, N.; Athans, M. (1981). "Robustness results in linear-quadratic Gaussian based multivariable control designs". IEEE Transactions on Automatic Control. 26 (1): 75–93. doi:10.1109/TAC.1981.1102565. ISSN 0018-9286.
  2. ^ Chow, Gregory C. (1986). Analysis and Control of Dynamic Economic Systems. Krieger Publ. Co. ISBN 0-89874-969-7.
  3. ^ Shaiju, AJ, Petersen, Ian R. (2008). "Formulas for discrete time LQR, LQG, LEQG and minimax LQG optimal control problems". IFAC Proceedings Volumes. Elsevier. 41 (2): 8773–8778. doi:10.3182/20080706-5-KR-1001.01483.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  4. ^ "Ch. 8 - Linear Quadratic Regulators". underactuated.mit.edu. Retrieved 20 August 2022.
  5. ^ https://minds.wisconsin.edu/bitstream/handle/1793/10888/file_1.pdf;jsessionid=52A001EAADF4C22B901290B594BFDA8E?sequence=1. Retrieved 20 August 2022. {{cite web}}: Missing or empty |title= (help)
  6. ^ Borggaard, Jeff; Zietsman, Lizette (July 2020). The Quadratic-Quadratic Regulator Problem: Approximating feedback controls for quadratic-in-state nonlinear systems. pp. 818–823. arXiv:1910.03396. doi:10.23919/ACC45564.2020.9147286. ISBN 978-1-5386-8266-1. S2CID 203904925. Retrieved 20 August 2022. {{cite book}}: |website= ignored (help)
  7. ^ Borggaard, Jeff; Zietsman, Lizette (1 January 2021). "On Approximating Polynomial-Quadratic Regulator Problems". IFAC-PapersOnLine. 54 (9): 329–334. doi:10.1016/j.ifacol.2021.06.090. S2CID 221856517.
  • Kwakernaak, Huibert & Sivan, Raphael (1972). Linear Optimal Control Systems. First Edition. Wiley-Interscience. ISBN 0-471-51110-2.

External links Edit

  • MATLAB function for Linear Quadratic Regulator design
  • Mathematica function for Linear Quadratic Regulator design

linear, quadratic, regulator, theory, optimal, control, concerned, with, operating, dynamic, system, minimum, cost, case, where, system, dynamics, described, linear, differential, equations, cost, described, quadratic, function, called, problem, main, results,. The theory of optimal control is concerned with operating a dynamic system at minimum cost The case where the system dynamics are described by a set of linear differential equations and the cost is described by a quadratic function is called the LQ problem One of the main results in the theory is that the solution is provided by the linear quadratic regulator LQR a feedback controller whose equations are given below LQR controllers possess inherent robustness with guaranteed gain and phase margin 1 and they also are part of the solution to the LQG linear quadratic Gaussian problem Like the LQR problem itself the LQG problem is one of the most fundamental problems in control theory Contents 1 General description 2 Versions 2 1 Finite horizon continuous time 2 2 Infinite horizon continuous time 2 3 Finite horizon discrete time 2 4 Infinite horizon discrete time 3 Constraints 4 Related controllers 4 1 Quadratic quadratic regulator 4 2 Polynomial quadratic regulator 4 3 Model predictive control 5 References 6 External linksGeneral description EditThe settings of a regulating controller governing either a machine or process like an airplane or chemical reactor are found by using a mathematical algorithm that minimizes a cost function with weighting factors supplied by a human engineer The cost function is often defined as a sum of the deviations of key measurements like altitude or process temperature from their desired values The algorithm thus finds those controller settings that minimize undesired deviations The magnitude of the control action itself may also be included in the cost function The LQR algorithm reduces the amount of work done by the control systems engineer to optimize the controller However the engineer still needs to specify the cost function parameters and compare the results with the specified design goals Often this means that controller construction will be an iterative process in which the engineer judges the optimal controllers produced through simulation and then adjusts the parameters to produce a controller more consistent with design goals The LQR algorithm is essentially an automated way of finding an appropriate state feedback controller As such it is not uncommon for control engineers to prefer alternative methods like full state feedback also known as pole placement in which there is a clearer relationship between controller parameters and controller behavior Difficulty in finding the right weighting factors limits the application of the LQR based controller synthesis Versions EditFinite horizon continuous time Edit For a continuous time linear system defined on t t 0 t 1 displaystyle t in t 0 t 1 described by x A x B u displaystyle dot x Ax Bu where x R n displaystyle x in mathbb R n that is x displaystyle x is an n displaystyle n dimensional real valued vector is the state of the system and u R m displaystyle u in mathbb R m is the control input Given a quadratic cost function for the system defined as J x T t 1 F t 1 x t 1 t 0 t 1 x T Q x u T R u 2 x T N u d t displaystyle J x T t 1 F t 1 x t 1 int limits t 0 t 1 left x T Qx u T Ru 2x T Nu right dt the feedback control law that minimizes the value of the cost is u K x displaystyle u Kx where K displaystyle K is given by K R 1 B T P t N T displaystyle K R 1 B T P t N T and P displaystyle P is found by solving the continuous time Riccati differential equation A T P t P t A P t B N R 1 B T P t N T Q P t displaystyle A T P t P t A P t B N R 1 B T P t N T Q dot P t with the boundary condition P t 1 F t 1 displaystyle P t 1 F t 1 The first order conditions for Jmin are 1 State equation x A x B u displaystyle dot x Ax Bu 2 Co state equation l Q x N u A T l displaystyle dot lambda Qx Nu A T lambda 3 Stationary equation 0 R u N T x B T l displaystyle 0 Ru N T x B T lambda 4 Boundary conditions x t 0 x 0 displaystyle x t 0 x 0 and l t 1 F t 1 x t 1 displaystyle lambda t 1 F t 1 x t 1 Infinite horizon continuous time Edit For a continuous time linear system described by x A x B u displaystyle dot x Ax Bu with a cost function defined as J 0 x T Q x u T R u 2 x T N u d t displaystyle J int 0 infty left x T Qx u T Ru 2x T Nu right dt the feedback control law that minimizes the value of the cost is u K x displaystyle u Kx where K displaystyle K is given by K R 1 B T P N T displaystyle K R 1 B T P N T and P displaystyle P is found by solving the continuous time algebraic Riccati equation A T P P A P B N R 1 B T P N T Q 0 displaystyle A T P PA PB N R 1 B T P N T Q 0 This can be also written as A T P P A P B R 1 B T P Q 0 displaystyle mathcal A T P P mathcal A PBR 1 B T P mathcal Q 0 with A A B R 1 N T Q Q N R 1 N T displaystyle mathcal A A BR 1 N T qquad mathcal Q Q NR 1 N T Finite horizon discrete time Edit For a discrete time linear system described by 2 x k 1 A x k B u k displaystyle x k 1 Ax k Bu k with a performance index defined as J x H p T Q H p x H p k 0 H p 1 x k T Q x k u k T R u k 2 x k T N u k displaystyle J x H p T Q H p x H p sum limits k 0 H p 1 left x k T Qx k u k T Ru k 2x k T Nu k right where H p displaystyle H p is the time horizonthe optimal control sequence minimizing the performance index is given by u k F k x k displaystyle u k F k x k where F k R B T P k 1 B 1 B T P k 1 A N T displaystyle F k R B T P k 1 B 1 B T P k 1 A N T and P k displaystyle P k is found iteratively backwards in time by the dynamic Riccati equation P k 1 A T P k A A T P k B N R B T P k B 1 B T P k A N T Q displaystyle P k 1 A T P k A A T P k B N left R B T P k B right 1 B T P k A N T Q from terminal condition P H p Q H p displaystyle P H p Q H p 3 Note that u H p displaystyle u H p is not defined since x displaystyle x is driven to its final state x H p displaystyle x H p by A x H p 1 B u H p 1 displaystyle Ax H p 1 Bu H p 1 Infinite horizon discrete time Edit For a discrete time linear system described by x k 1 A x k B u k displaystyle x k 1 Ax k Bu k with a performance index defined as J k 0 x k T Q x k u k T R u k 2 x k T N u k displaystyle J sum limits k 0 infty left x k T Qx k u k T Ru k 2x k T Nu k right the optimal control sequence minimizing the performance index is given by u k F x k displaystyle u k Fx k where F R B T P B 1 B T P A N T displaystyle F R B T PB 1 B T PA N T and P displaystyle P is the unique positive definite solution to the discrete time algebraic Riccati equation DARE P A T P A A T P B N R B T P B 1 B T P A N T Q displaystyle P A T PA A T PB N left R B T PB right 1 B T PA N T Q This can be also written as P A T P A A T P B R B T P B 1 B T P A Q displaystyle P mathcal A T P mathcal A mathcal A T PB left R B T PB right 1 B T P mathcal A mathcal Q with A A B R 1 N T Q Q N R 1 N T displaystyle mathcal A A BR 1 N T qquad mathcal Q Q NR 1 N T Note that one way to solve the algebraic Riccati equation is by iterating the dynamic Riccati equation of the finite horizon case until it converges Constraints EditIn practice not all values of x k u k displaystyle x k u k may be allowed One common constraint is the linear one C x D u e displaystyle C mathbf x D mathbf u leq mathbf e The finite horizon version of this is a convex optimization problem and so the problem is often solved repeatedly with a receding horizon This is a form of model predictive control 4 5 Related controllers EditQuadratic quadratic regulator Edit If the state equation is quadratic then the problem is known as the quadratic quadratic regulator QQR The Al Brekht algorithm can be applied to reduce this problem to one that can be solved efficiently using tensor based linear solvers 6 Polynomial quadratic regulator Edit If the state equation is polynomial then the problem is known as the polynomial quadratic regulator PQR Again the Al Brekht algorithm can be applied to reduce this problem to a large linear one which can be solved with a generalization of the Bartels Stewart algorithm this is feasible provided that the degree of the polynomial is not too high 7 Model predictive control Edit Main article Model predictive control MPC vs LQR Model predictive control and linear quadratic regulators are two types of optimal control methods that have distinct approaches for setting the optimization costs In particular when the LQR is run repeatedly with a receding horizon it becomes a form of model predictive control MPC In general however MPC does not rely on any assumptions regarding linearity of the system References Edit Lehtomaki N Sandell N Athans M 1981 Robustness results in linear quadratic Gaussian based multivariable control designs IEEE Transactions on Automatic Control 26 1 75 93 doi 10 1109 TAC 1981 1102565 ISSN 0018 9286 Chow Gregory C 1986 Analysis and Control of Dynamic Economic Systems Krieger Publ Co ISBN 0 89874 969 7 Shaiju AJ Petersen Ian R 2008 Formulas for discrete time LQR LQG LEQG and minimax LQG optimal control problems IFAC Proceedings Volumes Elsevier 41 2 8773 8778 doi 10 3182 20080706 5 KR 1001 01483 a href Template Cite journal html title Template Cite journal cite journal a CS1 maint multiple names authors list link Ch 8 Linear Quadratic Regulators underactuated mit edu Retrieved 20 August 2022 https minds wisconsin edu bitstream handle 1793 10888 file 1 pdf jsessionid 52A001EAADF4C22B901290B594BFDA8E sequence 1 Retrieved 20 August 2022 a href Template Cite web html title Template Cite web cite web a Missing or empty title help Borggaard Jeff Zietsman Lizette July 2020 The Quadratic Quadratic Regulator Problem Approximating feedback controls for quadratic in state nonlinear systems pp 818 823 arXiv 1910 03396 doi 10 23919 ACC45564 2020 9147286 ISBN 978 1 5386 8266 1 S2CID 203904925 Retrieved 20 August 2022 a href Template Cite book html title Template Cite book cite book a website ignored help Borggaard Jeff Zietsman Lizette 1 January 2021 On Approximating Polynomial Quadratic Regulator Problems IFAC PapersOnLine 54 9 329 334 doi 10 1016 j ifacol 2021 06 090 S2CID 221856517 Kwakernaak Huibert amp Sivan Raphael 1972 Linear Optimal Control Systems First Edition Wiley Interscience ISBN 0 471 51110 2 Sontag Eduardo 1998 Mathematical Control Theory Deterministic Finite Dimensional Systems Second Edition Springer ISBN 0 387 98489 5 External links EditMATLAB function for Linear Quadratic Regulator design Mathematica function for Linear Quadratic Regulator design Retrieved from https en wikipedia org w index php title Linear quadratic regulator amp oldid 1171783797, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.