Linear–quadratic regulator

The theory of optimal control is concerned with operating a dynamic system at minimum cost.

The case where the system dynamics are described by a set of linear differential equations and the cost is described by a quadratic function is called the LQ problem.

One of the main results in the theory is that the solution is provided by the linear–quadratic regulator (LQR), a feedback controller whose equations are given below.

LQR controllers possess inherent robustness with guaranteed gain and phase margin,[1] and they also are part of the solution to the LQG (linear–quadratic–Gaussian) problem.

[2] The settings of a (regulating) controller governing either a machine or process (like an airplane or chemical reactor) are found by using a mathematical algorithm that minimizes a cost function with weighting factors supplied by the operator.

The cost function is often defined as a sum of the deviations of key measurements, like altitude or process temperature, from their desired values.

The algorithm thus finds those controller settings that minimize undesired deviations.

The magnitude of the control action itself may also be included in the cost function.

However, the engineer still needs to specify the cost function parameters, and compare the results with the specified design goals.

The LQR algorithm is essentially an automated way of finding an appropriate state-feedback controller.

Difficulty in finding the right weighting factors limits the application of the LQR based controller synthesis.

is found by solving the continuous time algebraic Riccati equation: This can be also written as: with For a discrete-time linear system described by: [3] with a performance index defined as: the optimal control sequence minimizing the performance index is given by: where: and

is found iteratively backwards in time by the dynamic Riccati equation: from terminal condition

For a discrete-time linear system described by: with a performance index defined as: the optimal control sequence minimizing the performance index is given by: where: and

is the unique positive definite solution to the discrete time algebraic Riccati equation (DARE): This can be also written as: with: Note that one way to solve the algebraic Riccati equation is by iterating the dynamic Riccati equation of the finite-horizon case until it converges.

One common constraint is the linear one: The finite horizon version of this is a convex optimization problem, and so the problem is often solved repeatedly with a receding horizon.

[5][6] If the state equation is quadratic then the problem is known as the quadratic-quadratic regulator (QQR).

The Al'Brekht algorithm can be applied to reduce this problem to one that can be solved efficiently using tensor based linear solvers.

[7] If the state equation is polynomial then the problem is known as the polynomial-quadratic regulator (PQR).

Again, the Al'Brekht algorithm can be applied to reduce this problem to a large linear one which can be solved with a generalization of the Bartels-Stewart algorithm; this is feasible provided that the degree of the polynomial is not too high.

[8] Model predictive control and linear-quadratic regulators are two types of optimal control methods that have distinct approaches for setting the optimization costs.

In particular, when the LQR is run repeatedly with a receding horizon, it becomes a form of model predictive control (MPC).

In general, however, MPC does not rely on any assumptions regarding linearity of the system.