7
Synthesis of Robust Model Predictive Control

The methods for synthesizing stable predictive controllers introduced in the last chapter are with respect to systems with an accurate model and without disturbance. However, in real applications, model uncertainty or unknown disturbance always exists, which may affect the control performance of the predictive control system, and may even make the closed‐loop system unstable. Over the last two decades, how to design a predictive controller with guaranteed stability for uncertain systems has become an important topic of the qualitative synthesis theory of predictive control. Many novel methods were proposed and important results have been obtained. In this chapter, we will introduce the basic philosophy of synthesizing robust model predictive controllers (RMPCs) for systems with some typical categories of uncertainties. According to most of the literature, the uncertain systems discussed here are classified into two main categories: systems subjected to model uncertainty, particularly with polytopic type uncertainties, and systems subjected to external disturbance.

7.1 Robust Predictive Control for Systems with Polytopic Uncertainties

7.1.1 Synthesis of RMPC Based on Ellipsoidal Invariant Sets

Many uncertain systems can be described by the following linear time varying (LTV) system with polytopic uncertainty:

where A(k) ∈ Rn × n, B(k) ∈ Rn × m, and

where Co means a convex hull, i.e. for any time‐varying A(k) and B(k), [A(k) B(k)] ∈ Π means that there exists a group of non‐negative parameters λl(k), l = 1, …, L satisfying

The uncertainty described by (7.2) and (7.3) is called polytopic uncertainty, which can be formulated by a convex hull with L given vertices [AlBl]. The real system parameters [A(k) B(k)] will always be in the convex hull and λl are the convex combination parameters, either time‐varying or time‐invariant.

The system model with polytopic uncertainty can describe both nonlinear systems which are linearized at several equilibrium points and some linear systems with time‐varying parameters. For example, the linear parameter varying (LPV) system can be formulated by (7.2) and (7.3). Therefore, much RMPC literature focuses on this kind of uncertain system.

In addition, consider the constraints on system inputs and outputs:

Note that in order to highleight the basic idea of RMPC synthesis, (7.4) and (7.5) are taken as examples of system constraints. For other kinds of constraint, such as the peak bounds on individual components of system input or output, as discussed in Section 6.1.4, the results derived here only need to be adjusted accordingly.

For this kind of uncertain system described by (7.1) to (7.3), and with input and output constraints (7.4) and (7.5), referring to the linear robust control theory, the following infinite horizon “min‐max” problem is proposed at time k:

equation

Different from the optimization problem in previous chapters, the min‐max problem (7.6) results from the uncertainty of the system model. This means that the RMPC will optimize the future performance index with an infinite horizon so as to achieve an optimal solution for the worst case in all possible model variations in Π.

The min‐max problem (7.6) is a reasonable formulation of RMPC for an uncertain system (7.1) to (7.5), but is difficult to solve. Firstly, the objective of the minimization in (7.6) results from the maximization procedure and is heavily affected by system uncertainty, while the objective of the usual optimization method is explicitly given by analytical functions or variables. Secondly, the infinite terms in the objective function make the optimization computationally intractable. A specific strategy should be adopted to make the optimization problem solvable. In the following, referring to [1], we introduce the synthesis method of RMPC for this kind of polytopic uncertain system, explaining how to use an effective strategy to overcome these difficulties and how to implement it using LMIs and ellipsoidal invariant sets.

(1) Strategy and key ideas

1) Introducing an upper bound for the maximization problem Consider the quadratic function V(i, k) = x(k + i ∣ k)TPx(k + i ∣ k), where P is a symmetric positive definite matrix. At time k, for all x(k + i ∣ k) and u(k + i ∣ k) satisfying (7.1) to (7.3), impose the following condition on the corresponding function V:

This condition implies that the closed‐loop system is stable. Therefore x(∞ ∣ k) = 0 and thus V(∞, k) = 0. Summing (7.7) from i = 0 to i = ∞, it follows that

equation

Then define an upper bound γ > 0 on the robust performance index

In this way the min‐max problem (7.6) can be converted into a minimization problem with a simple linear objective γ.

2) Adopting a single feedback control law for the infinite horizon

In order to handle the infinite number of control actions in problem (7.6), introduce the ellipsoid controlled invariant set

with a corresponding state feedback control law

Control invariance means that if the system state x(k ∣ k) = x(k) ∈ Ω, then under the control law (7.10), x(k + 1 ∣ k) ∈ Ω. This can be ensured by

equation

and furthermore

Use of the above state feedback control law and the corresponding ellipsoid controlled invariant set make it possible to parametrize the control actions to a tractable form as well as to keep the system state within the invariant set.

(2) Constraints formulated in LMIs

Using the above key ideas, the min‐max problem (7.6) can be converted into an optimization problem of minimizing the upper bound γ by using the state feedback control law (7.10) with the following constraints: the upper bound constraint (7.8), controlled invariant set conditions (7.9) and (7.11), and input/output constraints (7.4) and (7.5), where γ, P, Q, F should be determined. Note that the quadratic form of xTPx in the upper bound constraint (7.8) is in accordance with the ellipsoid set representation in (7.9). Use P = γQ−1 to reduce the complexity. Then, with the help of the Schur Complement of LMIs (6.11) and the properties of the convex hull (7.3), the details of synthesizing the RMPC can be given as follows.

  1. The upper bound of the maximum of the objective function (7.8). From (7.8) it follows that
    equation

    With P = γQ−1 and x(k ∣ k) = x(k) the above inequality is exactly the invariant set condition (7.9) and can be rewritten in LMI form as

  1. Imposing V(i, k) descending and satisfying (7.7). With V(i, k) = x(k + i ∣ k)TPx(k + i ∣ k) and (7.1), (7.10), the condition (7.7) can be expressed as

    ∀[A(k + i) B(k + i)] ∈ Π, i ≥ 0

    equation

    This inequality can be ensured by the following matrix inequality:

    With P = γQ−1, left‐ and right‐multiplying Q on both sides will yield

    equation

    Let Y = FQ and rewrite the above matrix inequality as

    equation

    According to (6.11d) and (6.11e), this condition can be expressed in the following LMI form:

    equation

    where the symbol * is used to induce a symmetric structure of the matrix for simplicity.

    Since the uncertain parameters A(k + i) and B(k + i) satisfy the polytopic condition (7.3) and can be represented by a linear combination of parameters of the vertices [AlBl] in the polytopic model, the above condition is satisfied if the following group of LMIs holds:

  1. x (k) in the ellipsoid set Ω = {x ∈ Rn ∣ xTQ−1x ≤ 1}. This is also described by (7.12) because the condition xT(k)Q−1x(k) ≤ 1 is identical to xT(k)Px(k) ≤ γ under the selection P = γQ−1 and γ > 0.
  2. Ω is an invariant set of system (7.1) under the feedback control law (7.10), i.e. (7.11):
    equation

    It is obviously covered by the V(i, k) descending condition (7.13) with P = γQ−1 and γ > 0. Thus the corresponding LMI has already been included in (7.14).

  1. Constraints on control inputs (7.4). Referring to (6.13) in Section 6.1.4, it can be given by
  1. Constraints on system outputs (7.5). According to (7.5) and referring to (6.15) in Section 6.1.4, it follows that

(3) RMPC algorithm with robust stability

Summing the above derivations, at time k, solving the min‐max problem (7.6) can be converted into solving the following optimization problem:

After Q and Y are solved, we can obtain F = YQ−1 and u(k) = Fx(k), i.e. the current control action at time k. Note that {γ, Q, Y} are solved at time k, and strictly they should be denoted by {γ(k), Q(k), Y(k)}.

Now the RMPC algorithm based on ellipsoidal invariant set can be given as folllows.

For Algorithm 7.1, the following theorem on the stability of the closed‐loop system can be given.

The above proposed RMPC should solve the optimization problem (7.17) online by using LMIs toolboxes. The online computational burden is determined by the number of optimization variables and the number of rows of LMIs. It is easy to see that both of these two numbers will increase very quickly with the increase of the system dimensions and the number of vertices.

7.1.2 Improved RMPC with Parameter‐Dependent Lyapunov Functions

During the RMPC design in the last section, each time when solving the optimization problem (7.6) a Lyapunov function V(x) = xTPx with fixed P is adopted and the strictly descending condition of V(x) is imposed for all admissible model parameters. The use of this unique Lyapunov function for all possible models with parameters in the convex hull makes the problem solving easy but is conservative. To improve the control performance, Cuzzola et al. [2] and Mao [3] proposed that multiple Lyapunov functions could be used to design RMPC. When solving the optimization problem (7.6) at a fixed time, Lyapunov functions are defined for all the models corresponding to the vertices of the polytope and then a Lyapunov function is constructed by weighting these Lyapunov functions with the model combination coefficients λl. This kind of Lyapunov function is called a parameter‐dependent Lyapunov function. In the following, the main idea of synthesizing RMPC based on parameter‐dependent Lyapunov functions proposed in [2] and [3] is briefly introduced.

Consider the LTV system with polytopic uncertainty and constraints, (7.1) to (7.5). To solve the optimization problem (7.6) at time k, instead of the single Lyapunov function V(x) = xTPx with fixed P, as presented in Section 7.1.1, firstly define the Lyapunov functions

equation

for each linear time‐invariant system

equation

where [AlBl] represents a vertex of the polytope Π. Then construct the following Lyapunov function:

(7.18)equation

with

where λl(k + i) for k ≥ 0, i ≥ 0 represents the combination coefficient of [AlBl] in [A(k + i) B(k + i)] (see (7.3)). It is obvious that P(i, k) is time‐varying along with the combination coefficients and can handle uncertainty more finely to reduce conservatism.

Along the same way as presented in Section 7.1.1, for all x(k + i ∣ k) and u(k + i ∣ k) satisfying (7.1) to (7.3), impose the following condition on the corresponding function V(i, k):

By summing (7.20) from i = 0 to ∞, we can similarly obtain

equation

According to (7.19), this inequality holds if

equation

Let images and noting that x(k ∣ k) = x(k), the above inequalities can be represented by

(7.21)equation

Still assume the state feedback control law u(k + i ∣ k) = Fx(k + i ∣ k), i ≥ 0, with a fixed feedback gain F; the condition (7.20) holds if

(7.22)equation

With the help of the Schur complement, it can be written as the following matrix inequality:

where the symbol * in the matrix expression is used to induce a symmetric structure.

In order to distinguish the combination coefficients λ at time k + i and k + i + 1, use different subscript l and j, respectively, to mark them, i.e.

equation
equation

Then, according to (7.3), it is easy to show that the matrix inequality (7.23) holds if

equation

Let images and images, this can be rewritten as

By multiplying diag (γ−1/2Ql, γ−1/2Qj, γ1/2I, γ1/2I) on both sides of (7.24), the following matrix inequalities can be equivalently obtained:

equation

If all Qi are the same and denoted by Q, this is the case of using a single Lyapunov function discussed in Section 7.1.1. By letting Y = FQ, these matrix inequalities can be easily written in the form of LMIs with γ, Q, Y to be determined (see (7.14)). The feedback control gain F can then be obtained by F = YQ−1 with solved Y, Q. However, in the case of adopting multiple Lyapunov functions, in the above matrix inequalities the feedback control gain F appears in the product terms FQl for all Ql, l = 1, … , L. By similarly defining Yl = FQl these matrix inequalities might be rewritten into LMIs with γ, Ql, Yl, l = 1, … , L to be determined. However, with the solved pairs {Yl, Ql}, l = 1, … , L, it is impossible to find a feedback control gain F satisfying all the L matrix equations Yl = FQl.

In order to get an available feedback gain F, when defining a new matrix variable while converting the matrix inequality into LMI, it should be ensured that F could be uniquely solved by it. In spite of different Ql, refer to the single Q case and define F = YG−1 with G an invertible matrix to be determined. Multiplying diag (γ−1/2GT, γ−1/2Qj, γ1/2I, γ1/2I) and diag (γ−1/2G, γ−1/2Qj, γ1/2I, γ1/2I) on the left and right side of (7.24), respectively, we get

equation

Using Y instead of FG, then the only nonlinear term in the matrix is images. Note that with matrix inequality images, it follows that images. Therefore the above matrix inequalities are satisfied if the following LMIs hold:

Thus, the descending condition (7.20) of the parameter‐dependent Lyapunov function can be guaranteed by LMIs (7.25) if there exist L symmetric matrices Ql, l = 1, … , L, a pair of matrices {Y, G}, and a positive γ satisfying (7.25).

Then consider the constraints on control inputs (7.4). Note that the feedback control gain F is now F = YG−1, and at time k the input constraint (7.4) can be specified as

equation

Due to xT(k + i ∣ k)Q−1(i, k)x(k + i ∣ k) ≤ 1, this condition can be ensured by

equation

Since images, and considering images, the above condition holds if the following condition holds:

equation

which can be written in LMI form as

(7.26)equation

Finally, consider the constraints on system outputs (7.5), i.e. at time k:

equation

Since at time k, xT(k + i ∣ k)Q−1(i, k)x(k + i ∣ k) ≤ 1, this constraint can be satisfied if the following condition holds:

equation

which can be written in the form of the following matrix inequality:

equation

Due to (7.3) and

equation

also consider images. The above matrix inequalities can be ensured by the following LMIs:

(7.27)equation

According to the above discussion, the RMPC synthesis with parameter‐dependent Lyapunov functions can be formulated as solving the following optimization problem at each time k:

With Y, G solved, the feedback gain at time k can be obtained by F(k) = YG−1 and the control action at time k can then be calculated by u(k) = F(k)x(k) and implemented.

Compared with the design in Section 7.1.1, it is easy to show that Problem (7.17) is a special case of Problem (7.28) if we select a single Lyapunov function with L = 1 and set G = Q = Ql, with which the constraints in (7.28) are exactly reduced into those given in (7.17). The use of parameter‐dependent Lyapunov functions brings a greater degree of freedom to RMPC design and a better control performance could be achieved. Of course, the less conservative result is obtained at the cost of increasing the number of optimization variables and more rows of LMIs, which may cause the online computational burden to be heavier than that in Section 7.1.1.

7.1.3 Synthesis of RMPC with Dual‐Mode Control

It is easy to find that the RMPC designs in Section 7.1.1 and 7.1.2 adopt a single feedback control law to approximate the future control strategy for the infinite horizon optimization problem at time k. In order to improve the control performance, in addition to introducing parameter‐dependent Lyapunov functions, as presented in the last section, considering more reasonable future control strategies is also an attractive way. Some researchers explored bringing dual‐mode control [4] (see Section 6.2.3) into the RMPC design, such as Wan and Kothare [5] and Ding et al. [6]. In this section, we will briefly introduce this RMPC design method.

Consider the same RMPC problem in Section 7.1.1. At each time, an optimization problem (7.6) should be solved, with respect to the model (7.1) with polytopic uncertainties (7.2) and (7.3) and input and output constraints (7.4) and (7.5). According to the dual‐mode control, at time k, the future control inputs are assumed to be

where N is the length of the control horizon, which is the same as the optimization horizon. With this strategy, N free control inputs are firstly adopted to steer the system state into the terminal set and F is the gain of the state feedback control law in the terminal set.

Consider the min‐max performance index in the optimization problem (7.6) at time k:

equation

Since two different control modes are adopted, J(k) can accordingly be divided into two parts: the input and state items before and after the state entering the terminal set, respectively, i.e.

equation

Firstly we derive the upper bounds of the maximum of the above two objective functions under polytopic uncertainty, respectively.

For the first part, using the system model (7.1) starting from x(k ∣ k) = x(k), it is easy to get the system states controlled by the N free control inputs as shown in (7.29):

equation
equation

or in the matrix form

where

Note that all the coefficients of matrices images, images in (7.30) are composed of A(k + i), B(k + i), which belong to the convex hull Π described by (7.2) and (7.3). If the combination parameters λl(k + i), l = 1, … , L, are time‐varying, the system uncertainty will be expanded with time. We use time‐related subscripts to distinguish time‐varying combination parameters and denote the combination parameters at time k + i as images, l = 1, … , L, i = 0, 1, …, li = 1, … , L. Then images.

To illustrate how the system uncertainty expands with time, start from the fixed x(k) at time k. For x(k + 1 ∣ k), uncertainty is only caused by [A(k) B(k)] and can be described by the polytope with L vertices images, l0 = 1, … , L. For x(k + 2 ∣ k), with the same procedure for each possible x(k + 1 ∣ k), uncertainty will be expanded due to [A(k + 1) B(k + 1)], which can be described by the same type of polytope with L vertices images, l1 = 1, … , L, but with independent combination parameters. Therefore the uncertainty of x(k + 2 ∣ k) should be described by a composite polytope with L2 vertices images, l0 = 1, … , L, l1 = 1, … , L. For example, the term A(k + 1)A(k)x(k) in x(k + 2 ∣ k) can be expressed by

equation

With the above understanding, a convex hull images with images vertices can now be used to describe the polytopic uncertainties of images corresponding to x(k + 1 ∣ k) to x(k + N − 1 ∣ k), i.e.

where images is the images vertex of the polytope images. Similarly, a convex hull images with images vertices can be used to describe the polytopic uncertainties of images corresponding to the terminal state x(k + N ∣ k):

(7.34)equation
(7.35)equation

Now we can define an upper bound for J1(k). According to (7.30), it follows that

equation

where images, images are the diagonal matrices, respectively, composed of images, images with proper dimensions. Remove the fixed term images and define an upper bound γ1 for the remaining part. It follows that

equation

Due to the polytopic characters of the system parameters (7.32) and (7.33), it can be ensured by

The input and output constriants (7.4) and (7.5) can be expressed by

equation
equation

where Lu, i + 1(Lx, i) is a proper matrix that selects u(k + i ∣ k)(x(k + i ∣ k)) from U(k)(X(k)) in (7.31). These two conditions can be ensured if the following LMIs are satisfied:

(7.37)equation
(7.38)equation

For the second part, the optimization problem with performance index J2(k) is almost the same as that discussed in Section 7.1.1, with the only difference being that the initial state x(k ∣ k) is replaced by x(k + N ∣ k). We can similarly define a quadratic function V(x) = xTPx to obtain an upper bound γ for the maximization problem, and define a controlled invariant set Ω = {x ∈ Rn ∣ xTQ−1x ≤ 1} with a corresponding feedback control law u = Fx for all k + i, i ≥ N. Ω is just the terminal set that the system state should be steered to by the first N free control actions. Let P = γQ−1 and F = YQ−1; the following conditions similar to (7.12) and (7.14) to (7.16) can be obtained:

(7.41)equation

Then the min‐max problem (7.6) to be solved at time k can be converted into the following optimization problem:

Pluymer et al. thus proposed a control policy based on a tree structure spanned by free control inputs according to the character of polytopic uncertainty [7]. With this structure and the system model (7.1) to (7.3), the system state at any future time is uncertain but can be directly described by a polytope with corresponding state vertices (see Figure 7.1), where at each time a node means that a possible state vertex is driven by a corresponding input. In this way, the uncertainty expansion of system inputs and states could be explicitly described in accordance with the model uncertainty.

Image described by caption and surrounding text.

Figure 7.1 State trajectory in a tree structure with polytopic uncertainty.

In Figure 7.1, at time k, the system state x(k) is known. According to the polytopic uncertainty of system model (7.1) to (7.3), the state under the control of u(k ∣ k) can be given by

equation

which indicates that the system state x(k + 1 ∣ k) must be the convex combination of L state vertices images, l0 = 1, … , L (the nodes at time k + 1 in Figure 7.1), each corresponding to the fixed model with the vertex parameter images, respectively.

For the sake of simplification, in the following we use the symbol l0, … , lm = * to represent l0 =  * , ⋯, lm = *, and images to represent images. For the system state x(k + 2 ∣ k), in terms of L state vertices images, define different inputs images for each state vertex and let

equation

Similarly, under the control of images, each state vertex images will evolve to L state vertices images, l0, l1 = 1, … , L. We have

equation

where the system state x(k + 2 ∣ k) must be the convex combination of L2 state vertices images, l1, l0 = 1, … , L (i.e. the nodes at time k + 2 in Figure 7.1), where each corresponds to the fixed model with parameters of the vertex images, respectively.

This procedure is continued until time k + N. As the result, the uncertainty of the system state x(k + i ∣ k) can still be described by polytope but the number of state vertices expands with time, as shown by the tree in Figure 7.1. Note that here a control policy of using different control inputs corresponding to each expanded state vertex is adopted, so the control inputs and the states in (7.31) are given by

where

Each time the control input is not a single one but is represented in a more detailed form corresponding to the tree structure

Then the state evolution (7.30) under the new control input U(k) with 1 + L + ⋯ + LN − 1 components and new matrices images with proper dimensions can be similarly derived. According to the new form of U(k) and new matrices, the optimization problem with constraint conditions similar to (7.43) and (7.36) to (7.42) can be established.

For the new formulated optimization problem, assume that it has an optimal solution U*(k) at time k. Let li(k) be the subscript li defined at time k. Due to vertices expansion in Figure 7.1, images has Li − 1 components images while images has Li components images. Obviously, for U(k) with this new structure, during proof of recursive feasibility, the candidate solution U(k + 1) at time k + 1 cannot be constructed by using a shifting operation to set images as usual. However, as shown in Figure 7.1, u(k + 1 ∣ k + 1) corresponds to L images with different l0(k), and each images corresponds to L images resulting from images with different l0(k). Therefore each component in images, i.e. images, can be set by the weighted sum of images, i.e.

(7.47a)equation

Similarly, set

(7.47b)equation

Since all images are fixed at time k, the components of images given by (7.47) are also fixed and the constructed U(k + 1) is available. Furthermore, with this setting and (7.44), it is easy to prove that

equation
equation

This indicates that the use of U(k) with new components in (7.46), on the one hand, makes the constructed U(k + 1) available at time k + 1 and, on the other hand, implies that the control inputs u(k + i ∣ k + 1) have the above relationship with u*(k + i ∣ k) and x*(k + N ∣ k), which is just what we originally expect to guarantee on the feasibility of the optimization problem (7.43) at time k + 1. For more details, refer to [8].

The design in this section uses the dual‐mode control concept. With additional free control inputs this kind of RMPC can improve the control performance and enlarge the initial feasible region, but at the cost of bringing a heavy online computational burden and difficuty on recursive feasibility, particularly when the combination parameters describing uncertainty are time varying.

7.1.4 Synthesis of RMPC with Multistep Control Sets

The RMPC described in the last section improves the control performance through adding free control inputs before employing a fixed‐state feedback control law, which brings more degrees of freedom for control but also increases the complexity to guarantee the recursive feasibility because of uncertainty expansion. This contradiction was resolved by introducing an open‐loop tree structure for state and control expansion. From another perspective, note that the state feedback law is often helpful to ensure recursive feasibility, such as its use in the terminal set. Li and Xi [9] proposed the concept of the multistep control set and used it to design a sequence of feedback control laws instead of free control inputs used in the dual‐mode framework. It seems an efficient way to guarantee recursive feasibility and thus is helpful to RMPC synthesis. In this section, we will briefly introduce the synthesis method of RMPC using multistep control sets.

(1) Multistep control set

From the definition, it is easy to see that any controlled invariant set is a one‐step control set. Furthermore, it is also easy to prove that a set is an s‐step control set if all the system states in it can be steered into an (s − 1)‐step control set by an admissible feedback control law. With this property, in the following we discuss how to design an s‐step ellipsoidal control set for the constrained polytopic uncertain system (7.1) to (7.5).

An s‐step control set can be designed by a series of ellipsoidal sets images, i = 0, ⋯, s − 1. When the system state x is in the ellipsoidal set Si − 1, one can use an admissible feedback control law Fi − 1x to steer the system state x into the ellipsoidal set Si, while the last ellipsoidal set images is a controlled invariant set with the feedback control law Fs − 1x in usual sense. Let images. Conditions for S0 as an s‐step control set can be given by

with

where conditions (7.48) ensure that the states in Si − 1 enter into set Si at the next time instant under the feedback control law u = Fi − 1x and meanwhile the set Ss − 1 is a controlled invariant set with the control law u = Fs − 1x. In addition, conditions (7.49) are used to ensure the input and state constraints, which are similar to (7.15), (7.16) in Section 7.1.1.

According to Definition 7.1 and the conditions (7.48) and (7.49), it is easy to show the following properties of a multistep control set.

(2) RMPC synthesis with a multistep control set

Now for the constrained uncertain system (7.1) to (7.5), consider the online optimization problem (7.6) at time k. Since the current system state x(k) is determined, its quadratic term can be firstly removed from J(k). Then rewrite the performance index in the optimization problem (7.6) as

equation

Using the concept of a multistep control set, the following control strategy is adopted:

which implies that a series of ellipsoidal sets images, i = 1, …, N is available to construct a multistep control set. At first, use a free control action u(k) to steer the system state x(k) into the ellipsoidal set S1. Then for i = 1, …, N − 1, step by step use the feedback control law Fix to steer the system state from the ellipsoidal set Si to Si + 1. The last ellipsoidal set SN is a controlled invariant set where the feedback control law FNx is adopted. The control strategy π is to mix the quasi‐min‐max RMPC in [10] and the multistep control set, where the multistep control set is used as the terminal set of the quasi‐min‐max RMPC.

With control strategy π, choose quadratic functions V(i, k) = xT(k + i ∣ k)Pix(k + i ∣ k), i = 1, …, N − 1 for feedback control laws with gain Fi. Let Pi = PN when i ≥ N. Similar to that in Section 7.1.1, the optimization problem to be solved at time k is derived as follows.

  1. Imposing V(i, k) descending and satisfying the following conditions:
  2. Obtaining the upper bound of the objective function. Summing Eq. (7.51) from i = 1 to i = ∞, it follows that
    (7.52)equation

    Define the upper bounds for both images and V(1, k):

    Then it follows that

    equation
  3. Using u(k ∣ k) to steer the state x(k) into the ellipsoidal set images:
  4. Steering the system state into the invariant set SN by the Nstep control set:

    Now all of the above conditions will be converted into LMI forms. Let images, images, i = 1, …, N. Then the bound condition (7.54) is equivalent to the ellipsoidal set condition (7.55), and the multistep control set conditions (7.58) and (7.59) could be covered by the condition (7.51) together with (7.55). Therefore we only need to derive LMIs for other conditions. Similar to (7.14) but with a different Pi, the V(i, k) descending condition (7.51) can be written as

    (7.63)equation

    The bound condition (7.53) will be

    The condition (7.55) of x(k) steered into the ellipsoidal set S1 can be written as

    The input/output constraints at the first step (7.56) and (7.57) can be guaranteed by

    Referring to (6.13) and (6.15), the input/output constraints when x enters the multistep control set, i.e. (7.60) and (7.61), can be guaranteed by

    Now the RMPC algorithm based on the multistep control set can be given.

(3) Algorithm and robust stability

The following theorem is given for the stability of the above algorithm.

7.2 Robust Predictive Control for Systems with Disturbances

In the last section, we focused on synthesizing robust predictive controllers for systems with parameter uncertainty. In real applications, in addition to such unmodeled dynamics, external disturbance always exists and affects the controlled system away from satisfactory performance. How to deal with external disturbance is thus also a hot topic in RMPC study and applications. In this section, two catalogs of synthesis approaches for such types of RMPC are introduced.

7.2.1 Synthesis with Disturbance Invariant Sets

Consider the following linear time‐invariant system with external disturbance

where A ∈ Rn × n, B ∈ Rn × m, and ω(k) ∈ Rn represent the external disturbance. For the controlled system, the state and input constraints are given in general forms as

where Ωx and Ωu are the admissible sets of states and inputs, respectively. The notations of Minkowski set operations are also given as follows, which will be used later. For two sets A, B, set addition A ⊕ B ≜ {a + b ∣ a ∈ A, b ∈ B}, and set subtraction A ⊖ B ≜ {a ∣ a ⊕ B ⊆ A}.

If the disturbance is unbounded and varies arbitrarily, it is impossible to guarantee the constraints satisfied. In practical applications, the disturbance with a too large amplitude is often handled by some other measures. It is therefore generally assumed that the disturbance in (7.71) is bounded as

For system (7.71) to (7.74), the appearance of the disturbance makes the constraint handling much more complicated. If the approaches for RMPC synthesis in Section 7.1 are adopted, the influence caused by disturbance will be time varying, so the conditions that ensure the satisfaction of the input and state constraints are also time varying. In order to simplify the disturbance handling, Mayne et al. [11] proposed an approach to synthesize RMPC by using disturbance invariant sets. In the following, the main idea of this approach is introduced.

(1) Disturbance invariant set

Firstly, similar to the definition of the invariant set given in Chapter 6, the following definition on the disturbance invariant set is given.

It is obvious that if the system x(k + 1) = Ax(k) is unstable, no disturbance invariant set exists. For a control system given by x(k + 1) = Ax(k) + Bu(k) + ω(k), a feedback control law u(k) = Kx(k), often taken as the optimal controller of the unconstrained LQ problem with images, can be used to make x(k + 1) = AKx(k) with AK = A + BK stable. The disturbance invariant set Z designed for the resultant system x(k + 1) = AKx(k) + ω(k) makes x(k + 1) ∈ Z for all ω(k) ∈ Ω when x(k) ∈ Z and u(k) = Kx(k).

(2) RMPC synthesis by dividing the uncertain system into nominal and disturbed subsystems

Like the invariant set, the disturbance invariant set can keep the system state in it always within it by using the corresponding feedback control law even when the system is disturbed. This is convenient when handling constraints. Since system (7.71) is linear and of an additive character, [11] proposed an RMPC synthesis approach by dividing the system (7.71) into two subsystems:

where subsystem (7.75a) is the nominal system of the original system (7.71). The basic idea to synthesize RMPC is as follows. Firstly, as mentioned above, use the feedback control law images for the subsystem (7.75b) to get images and find a disturbance invariant set Z for it. That is, under the control of images, the state component images will be kept in Z when the subsystem is disturbed by ω(k) ∈ Ω, i.e. AKZ ⊕ Ω ⊆ Z. Then, for the nominal system (7.75a), use the conventional predictive control approach and optimize the nominal control component images in order to steer the nominal state images to the origin under tighter constraints, which take the influence of the disturbance into account. The real control input will be composed of images. With this basic idea, when designing an RMPC using the disturbance invariant set, the admissible input set can be simply divided into two parts: one for designing the disturbance invariant set while the other for stabilizing the nominal system. After the disturbance has been separated, the nominal system will be stabilized by usual predictive control synthesis methods, which simplifies the design.

Synthesis of the disturbed subsystem

For subsystem (7.75b), a disturbance invariant set Z would be found such that with the feedback control law images, images if images. The disturbance invariant set Z should be as small as possible to reduce conservativeness. Its form should also be suitable for optimization in the RMPC design. It has been suggested [11] that an algorithm be used to compute a polytopic, disturbance invariant, outer approximation of the minimal disturbance set, which can be taken as the disturbance invariant set Z here. If the nominal state reaches zero, the system state will always be in set Z. Therefore the disturbance invariant set Z serves as the “origin” for the original system with a bounded but unknown disturbance.

Synthesis of the nominal subsystem

Now turn to control of the nominal subsystem (7.75a). It is a linear time‐invariant system without uncertainty and disturbance, and thus can be easily handled by conventional predictive control approaches as presented in Chapter 6. The dual‐mode control strategy with N free control actions images followed by the feedback control law images is adopted. At time k, the cost function of the online optimization problem is given by

where the last term in (7.76) is the terminal cost function. This optimization problem is very similar to Problem PC4, (6.31) in Section 6.2.3, and could be solved in the same way. However, some important issues different from conventional predictive control should be pointed out.

  1. Tighter state and input constraints. Since the optimization is only for the nominal subsystem (7.75a) with state images and control input images, the constraints (7.72) and (7.73) for the real system state x(k) and input u(k) cannot be directly used here. According to (7.75c), the admissible set Ωx is shared by images and images. Therefore the admissible set of images should be constructed by subtracting the admissible set of images from Ωx. A similar consideration is also for constructing the admissible set of images as well as for setting the constraint on the terminal set Xf used in dual‐mode control. Therefore, we have tighter state, input, and terminal constraints as follows:
    (7.79)equation
  2. Initial state images as an optimization variable. When solving the optimization problem at time k, the real system state x(k) is assumed available and is used as the initial state of the optimization in conventional predictive control. However, the optimization here is with respect to images and images. Nothing is known about images and thus images cannot be set as the initial state in the optimization, but would be taken as an additional optimization variable and handled as a parameter of the control law. It can vary but would be constrained by the following initial condition with the known system state x(k):

    Thus, the optimization problem at time k can be formulated as

    where images. the terminal set Xf and the weighting matrix P in the terminal cost function should satisfy the following conditions:

    (7.82b)equation

    where the first condition is to ensure that the terminal set Xf is a controlled invariant set for the nominal system (7.75a) with state component images and feedback control law images when images enters Xf, while the second condition is to ensure that the value of the objective function is nonincreasing within the terminal set; see condition A4 in (6.38).

    Note that the disturbance invariant set Z, the terminal set Xf, and set operations can be calculated off‐line. Thus the online optimization problem (7.81) is simple and straightforward.

    After the optimal images and U*(k) are solved, the following control law will be applied to the real system:

    (7.83)equation

    The associated optimal state sequence images as well as the optimal value of the cost function images can also be obtained.

(3) Theoretical properties of the RMPC controller

For the optimization problem (7.81), the optimal solution and the control law depend on the system state x(k). The following important results were given in [11] after investigating the theoretical properties of the solution and the controlled system:

  1. For all x(k) ∈ Z, J*(x) = 0, images, U*(x) = {0, … , 0}, X*(x) = {0, … , 0}, and the control law images.
  2. Let XN be the set of x(k) for which there always exists images satisfying (7.77) to (7.80). Suppose that x(k) ∈ XN so that (images, U*(k)) exists and is feasible for Problem (7.81) at time k. Then, for all images, (images, images) with images is feasible at time k + 1 and images.
  3. The set Z is robustly exponentially stable for the uncertain controlled system images, ω(k) ∈ Ω with a region of attraction XN, which means there exists a c > 0 and a γ ∈ (0, 1) such that for the initial state x(0) ∈ XN and admissible disturbance sequence ω(k) ∈ Ω, k ≥ 0, any solution x(k) of above system satisfies d(x(k), Z) ≤ kd(x(0), Z) for all k ≥ 0, i.e. the real system state x(k) will exponentially approach the disturbance invariant set Z, where d(x, Z) is the minimal distance of a point x from the set Z.

The RMPC approach presented above makes use of the linear and additive character of the original system to divide it into two subsystems. A modified optimization problem for the nominal subsystem is solved online where the initial nominal state is taken as an additional optimization variable. A suitable Lyapunov function is accordingly selected that has value zero in the set Z, with which the robust exponential stability of Z can be proved. It is shown that the real state of the controlled system is always along a tunnel with the nominal system state images as the center and the disturbance invariant set Z as the cross‐section. This idea led to the tube technique, which was later widely used in RMPC synthesis for systems with uncertainty and disturbance.

7.2.2 Synthesis with Mixed H2/Himages Performances

In this section, we introduce another kind of RMPC for systems with disturbance, which adopts H2 and H performance in optimization. Consider the following disturbed system:

where x(k) ∈ Rn, u(k) ∈ Rm, and z(k) ∈ Rr are system state, control input, and controlled output, respectively. A(k) ∈ Rn × n, B(k) ∈ Rn × m, and [A(k) B(k)] ∈ Π with

(7.85)equation

The disturbance ω(k) ∈ Rs is bounded as

where images and q > 0 are known.

The constraints on the system inputs and maesurable states are given by

where Ψj is the jth row of the state constraint matrix Ψ ∈ Rs × n. Note that (7.87) is just the peak bound constraint on the input components, as discussed in Section 6.1.4, while (7.88) is the bound constraint on the linear combination of the states.

Unlike system (7.71) discussed in the last section, system (7.84) includes not only the external disturbance but also the model uncertainty. Although the disturbance invariant set can also be designed for the system with model uncertainty, the resultant design may be conservative.

It is well known that the LQ performance index corresponds to an H2 optimal control problem, which is generally useful for optimal controller design and has traditionally been adopted as the objective function in RMPC to achieve a better control performance, while the H performance index is known as a useful criterion on a control system for its ability to reject disturbance. In order to achieve both the desired disturbance rejection and a good dynamic performance, it is a natural idea to combine these two performances, i.e. using a mixed H2/H performance, to synthesize RMPC. This idea actually came from research of robust control and was later adopted in RMPC study. In this section we briefly introduce this kind of RMPC synthesis approach presented in [12].

For system (7.84) to (7.88), consider the following two performance requirements:

  1. H performance requirement: under the zero‐initial condition, the controlled output z(k) satisfies
  2. H2 performance requirement: the controlled output z(k) satisfies

To optimize both performance indices, it is usual to firstly assign an upper bound on one performance index and then design RMPC by online optimizing the other performance index with the above upper bound condition as an additional constraint. Here the H2 performance index α is taken as the online optimization objective of RMPC and the H performance index γ > 0 is set as a fixed scalar to deal with (7.89) as a constraint.

Following the above idea, choose the RMPC control law as u(k) = Fx(k). At time k, based on the initial state x(k), an optimization problem with the performance indices (7.89) and (7.90), rewritten with the initial time k, can be solved online. Define the quadratic function images. Denote ϕ(k) = A(k) + B(k)F and ϕl = Al + BlF, l = 1, … , L. We can get

where

equation

Note that from (7.86) the disturbance is energy‐bounded. This means that, if the closed‐loop system is stable, then images. Thus, summing (7.91) from k to ∞ can yield

equation

It follows that

From (7.92), it can be seen that:

  1. When x(k) = 0, the H performance requirement (7.89), i.e.
    equation

    can be ensured if M(k + i) ≤ 0.

  1. When M(k + i) ≤ 0, the H2 performance requirement (7.90) can be guaranteed if the following condition holds:

    where M(k + i) ≤ 0 can be rewritten as

    (7.94)equation

    which can be reformulated as the LMIs

    equation

    Let Q = αP−1, F = YQ−1. Note that ϕ(k + i) is a convex combination of ϕl. Left‐ and right‐multiply the above inequality by diag (α1/2P−1, α1/2I, α1/2I, α1/2I, α1/2I). Then the following condition can be obtained which can ensure the above inequality:

    Meanwhile, (7.93) can be rewritten into LMI as

    (7.96)equation

    which implies that images, and thus the current system state x(k) belongs to an ellipsoidal set Ω = {x ∣ xTQ−1x ≤ 1}. Therefore, as (6.14) in Section 6.1.4, the input constraints (7.87) can be written as

    (7.97)equation

    The state constraints (7.88) can be written as

    (7.98)equation

    However, due to the existence of disturbance, the condition M(k + i) ≤ 0, i.e. (7.95), is not enough to ensure V(x(k + 1)) − V(x(k)) ≤ 0 in (7.91), and Ω = {x ∣ xTQ−1x ≤ 1} is not a controlled invariant set, i.e. the system state x(k + 1) cannot be ensured to be in Ω even when the current state x(k) ∈ Ω. This also leads to a problem with recursive feasibility and will be solved by adding some new conditions as follows.

    For the current state x(k) ∈ Ω, it follows that x(k + 1) = ϕ(k)x(k) + Bωω(k). To ensure the state x(k + 1) ∈ Ω, divide it into two parts: the nominal substate images and the substate v(k) = Bωω(k) affected by disturbance. The condition of x(k + 1) in the set Ω = {x ∣ xTQ−1x ≤ 1} can then be represented by

    equation

    Let 0 < b < 1 be a parameter chosen in advance. The above matrix inequality can be ensured if both of the following two conditions hold:

    where (7.100) indicates that ω(k) should be in an ellipsoidal set images. According to (7.86), all the disturbances ω(k) ∈ Πq ≔ {ω ∣ ωTq−1ω ≤ 1}. Therefore (7.100) can be ensured if images, which can be rewritten as

    Meanwhile, (7.99) means images, where Qb = (1 − b)(Q − Qω). Since x(k) ∈ Ω = {x ∣ xTQ−1x ≤ 1}, this can be ensured if

    Let images. Referring to (7.91), we get

    where Mb(k + i) ≤ 0 can be ensured by

    Note that images is the nominal substate with ω(k) = 0; from (7.103) it is known that (7.104) can guarantee (7.102) and then (7.99). It is also tighter than condition (7.95) due to Qb < Q and thus can cover (7.95). With conditions (7.101) and (7.104), the recursive feasibility, i.e. state x(k + 1) ∈ Ω if x(k) ∈ Ω, can be ensured even if a disturbance exists, and the decreasing of V(x(k)) can be guaranteed.

    As mentioned above, here the H performance γ is used as a constraint and the H2 performance α would be optimized. By fixing γ and choosing the parameter b in advance, the online optimization problem of RMPC can be formulated as follows:

    (7.105)equation

    In the above, the basic idea of the RMPC synthesis with mixed H2/H performance is introduced. It adopts the control startegy with the unique feedback control law, which may be conservative. New design approaches were also developed to improve the performance, which can be found in the literature.

7.3 Strategies for Improving Robust Predictive Controller Design

7.3.1 Difficulties for Robust Predictive Controller Synthesis

The RMPC synthesis approach presented in Section 7.1.1 can also be applied to systems with structural uncertainties. In [1], a kind of system with structural feedback uncertainty is converted into an uncertain LTV system and RMPC for it is synthesized in the same way but with different LMI forms. The basic idea of synthesizing RMPC and the specific technique of handling constraints in [1] have an important impact on subsequent researches of robust predictive control, which have been shown, for example, in the other parts of Section 7.1. However, the approach presented in [1] still has some limitations:

  1. To make the design simple, a unique feedback control law is adopted in [1], which strictly restricts the structure of the future control sequence. As the result, the control performance could not be further improved because of the lack of more degrees of freedom for control.
  2. In practical applications, the feasible region of the predictive controller, i.e. the region of system initial states with guaranteed stability, is preferred to be as large as possible such that the closed‐loop stability can be guaranteed for a larger region of system states. However, with the unique feedback control law as in Section 7.1.1, the closed‐loop system is robustly stable only when x(0) ∈ Ω, i.e. the feasible region is obviously restricted by Ω.
  3. The optimization problem in [1] should be solved online according to the rolling style of predictive control. The problem scale increases rapidly with the system dimension and the number of vertices of the uncertain model, which will lead to a heavy computational burden and make it difficult to be implemented in real time. Furthermore, the optimization problem with the LMI formulation cannot be solved as efficiently as that with the QP formulation.

In fact, the above limitations reflect the main difficulty in RMPC synthesis, i.e. the contradiction among the requirements on feasible region (reflecting the region of the system states with guaranteed stability), online computational burden (reflecting whether the algorithm can be applied in real time), and control performance (reflecting the optimization effect of the designed predictive controller). In order to illustrate this, consider the general method adopted in the stable and robust predictive controller synthesis shown in Figure 7.2.

Graphical representation displaying a bigger vertical ellipse representing the initial feasible region (left) having a curve connected to a small vertical ellipse representing the terminal constraint set (right).

Figure 7.2 Free control actions, initial feasible region, and terminal constraint set in predictive control synthesis.

At each step of rolling optimization, take the following predictive control strategy: start from the initial system state, adopt N free control inputs in the optimization horizon to steer the system state into the terminal constraint set, and then use the fixed feedback control law in the terminal constraint set to drive the system state to zero.

According to Section 6.2.3, the terminal set constraint is a relaxation of the zero terminal constraint and is easier to implement than it. However, the zero terminal constraint means that the system state will remain at the origin when it arrives at the terminal, while the terminal set constraint means that the system state is not zero when it enters the terminal set and a fixed state feedback control law is still needed to steer the state asymptotically to the origin. In this sense, the performance of the control system corresponding to the terminal set constraint is in general inferior to that of the zero terminal constraint. The larger the terminal set, the worse the control performance will be. From the point of view of improving the control performance, a smaller terminal set is often preferable.

The feasible region of the predictive controller characterizes the applicable region of the system states starting from which the stability or robust stability of the closed‐loop system can be theoretically guaranteed. In practice, the initial state outside the feasible region may also lead to stable results, but without a theoretical guarantee. From the point of view of the applicable range of the designed predictive controller, a larger feasible region is often preferable.

With the general control strategy mentioned above, N free control inputs before entering the terminal set and the feedback control law after entering the terminal set need to be solved online. The larger the optimization horizon N, the heavier the online computational burden will be. To reduce the online computational burden, a smaller N is more acceptable.

However, the above three requirements are often contradictory in a specific synthesis algorithm of predictive control. Referring to Figure 7.2, for example,

  • A larger feasible region can be achieved by enlarging the terminal set or increasing the optimization horizon, but the control performance will be degenerated or the computational burden will be increased, respectively.
  • Reducing the terminal set may improve the control performance, but will reduce the feasible region. If at the same time keeping the feasible region unchanged, the optimization horizon must be increased and thus the computational burden will be increased.
  • Decreasing the optimization horizon can reduce the online computational burden, but the feasible region will be reduced if the terminal set is unchanged to maintain the control performance. If the feasible region must be maintained, the terminal set has to be enlarged and the control performance will be sacrificed.

The above‐mentioned contradiction between control performance, feasible region, and online computational burden is the critical issue faced by predictive control synthesis, especially in the synthesis of robust predictive controllers. Some typical approaches to improve the design in [1] have been introduced in Section 7.1. In the following, we further introduce some other approaches that also show the variety of the RMPC research.

7.3.2 Efficient Robust Predictive Controller

In order to enlarge the feasible region and reduce the online computational burden, Kouvaritakis et al. proposed an RMPC algorithm named efficient robust predictive control (ERPC) [13]. The basic idea of ERPC is as follows. Firstly, an off‐line design using a fixed feedback control law is proposed that satisfies the input constraints and can robustly stabilize the controlled system. Then, additional degrees of freedom for control are introduced and an invariant set for the augmented system is solved off‐line, which has the maximum projection set in the original state space. When online, the additional variables are optimized in order to achieve a better control performance.

For the system (7.1) to (7.3) with polytopic uncertainties, consider the input constraints

where uj(k) is the jth component of control input u(k) ∈ Rm. The RMPC for this system is formulated as solving an infinite horizon optimization problem at each time k with the cost function

Due to the polytopic uncertainty of the model, the optimization for (7.107) is indeed a “min‐max” problem as in (7.6). In the following, the main idea of ERPC is briefly introduced.

(1) Design the feedback control law and handle the input constraints

Firstly, a fixed feedback control law u(k) = Kx(k) is designed that stabilizes all the models of system (7.1) to (7.3) without considering the input constraints (7.106). Define the controlled invariant set of system (7.1) to (7.3) with this feedback control law as

Let Φ(k) = A(k) + B(k)K; then we have images. Due to the polytopic property of the model uncertainty, it follows that

where Φl = Al + BlK, l = 1, …, L. With the control law u(k) = Kx(k), the input constraints (7.106) can be written as

equation

where images is the jth row of K. If the system state x(k) belongs to the invariant set Ωx, it follows that

equation

Thus the input constraints can be satisfied if

Then, based on (7.109) and (7.110), adopt the method in Section 7.1.1 to design RMPC and obtain K, Qx, such that system (7.1) to (7.3) can be robustly stabilized and the input constraints (7.106) are satisfied. This is simply the RMPC strategy presented in [1].

(2) Introduce perturbations to form the augmented system

In order to enlarge the feasible region and improve the control performance, it was suggested in [13] that extra degrees of freedom should be introduced by using perturbations c(k + i) on the fixed state feedback law. The control law was then modified as

where c(k + i) = 0, i ≥ nc. This means that c(k + i), i = 0, …, nc − 1 as additional input variables can be used to increase the degree of freedom for RMPC synthesis. Then the closed‐loop system can be expressed by

equation

By introducing the augmented state,

(7.112)equation

it can be formulated as an autonomous system

where

equation

For the autonomous system (7.113), get the controlled invariant set like (7.108):

where Qz satisfies

Meanwhile, note that the control law is now (7.111), i.e.

equation

Referring to (7.110), the input constraints (7.106) can be ensured if

where images is the jth row of the identity matrix with dimension m.

(3) Design the invariant set to enlarge the feasible region

Now consider the influence of introducing additional input variables to the invariant set of augmented system, which is also the feasible region of the RMPC. The matrix images in (7.114) can be divided into four blocks corresponding to x and f: images, images, images, and images, where images. Then, from (7.114) it follows that

Therefore, if we let images, then for all images, the above inequality is degenerated to the condition in (7.108). However, if f ≠ 0, from (7.117), it is possible to make the projection set of Ωz on x space larger than Ωx by properly choosing f. For example, if we let images, the projection of Ωz on x space can be given by

equation

Since images, if images, it follows that Ωx ⊆ Ωxz. That is, the invariant set on x space can be enlarged by the additional input variables. Figure 7.3 shows the enlarged invariant set in the x space after the state space has been augmented.

In order to achieve a general formulation, denote the projection set of the invariant set Ωz, i.e. (7.114) on x space as

Let x = Tz with T = [I  0] of dimension n × (n + mnc) and make a transformation for the LMIs implied in (7.114):

equation

Thus it is known that Qxz = TQzTT in (7.118). If images is expressed by four blocks defined as above, we can calculate images exactly. Therefore, a larger feasible region than Ωx can be achieved by making use of the additional input variables to design Ωz such that Qxz ≥ Qx.

(4) Implementation and algorithm

Then, according to the above analysis, the ERPC synthesis can be summarized as:

  1. Design the optimal feedback control law u(k) = Kx(k) for the unconstrained system (7.1) to (7.3) by using the usual robust control method with a specific optimization requirement, such as minimizing a worst‐case cost or optimizing a nominal performance.
  2. Introduce f and construct the augmented autonomous system (7.113) with K and design the invariant set Ωz of the autonomous system such that its projection set Ωxz on x space is as large as possible and Qz satisfies the invariant set condition (7.115) as well as the input constraint condition (7.116).
  3. The current augmented state z(k) belongs to set Ωz, i.e. condition (7.114) holds.
  4. The degree of freedom of f is used to optimize the system performance.
Graph of c vs. x displaying an ellipse titled to the right and centering at the origin having two dimensional arrows labeled Ωx and Ωxz. Ωz is also indicated.

Figure 7.3 Invariant set Ωx, Ωxz, and Ωz [13].

Source: Reproduced with permission from Kouvaritakis, Rossiter and Schuurmans of IEEE.

Note that (1) and (2) are independent from the system state x(k) and thus can be solved offline to reduce the online computational burden. Therefore the control algorithm can be divided into two parts: offline and online.

It was proved in [13] that if K, Qz, and nc exist such that x(0) ∈ Ωxz, ERPC with Algorithm 7.3 will robustly stabilize system (7.1) to (7.3) with satisfaction of the input constraints. Kouvaritakis et al. also pointed out in [14] that the optimization problem (7.120) can be efficiently solved by well‐known techniques, such as the Newton–Raphson (NR) method. Compared with the conventional RMPC method in [1], the control performance can be improved and the feasible region is enlarged by using the extra degrees of freedom introduced by f, while the online computational burden is greatly reduced because the main part of the computation has been moved to off‐line. Thus, ERPC makes a good trade‐off among the control performance, feasible region, and online computational burden.

7.3.3 Off‐Line Design and Online Synthesis

The EPRC presented in the last section greatly reduces the online computational burden by an off‐line design, which shows the possibility to make a trade‐off among the control performance, feasible region, and online computational burden by combining off‐line and online computations. In view of the heavy online computational burden of the RMPC synthesis approach in [1], Wan and Kothare [15] proposed an offline design/online synthesis method for an RMPC design based on the ellipsoidal invariant set.

Consider the RMPC synthesis problem in Section 7.1.1. For the uncertain system (7.1) to (7.3) with input and state constraints (7.4) and (7.5), given an initial state x, if it is feasible for Problem (7.17), then we can solve Problem (7.17) to get the optimal solution γ, Q, Y, with which an ellipsoidal invariant set Ω = {x ∈ Rn ∣ xTQ−1x ≤ 1} and its corresponding feedback control law u(k) = Fx(k) with F = YQ−1 can be obtained. Since all the system states in the set Ω can be steered to the origin by the feedback control law, the design of an ellipsoidal invariant set is independent from the current system state x(k). Thus, the core idea of [15] is to use off‐line to construct a series of ellipsoidal invariant sets and calculate their corresponding feedback control laws and online to determine the most suitable ellipsoidal invariant set and the corresponding feedback control law in terms of an actual system state and then to calculate the required control action. This can be described by the following algorithm in detail.

Using Algorithm 7.4, the online computational burden can be greatly reduced. However, from the online Algorithm 7.1 it is known that the optimal control law and the corresponding invariant set depend on the system state. While in Algorithm 7.4, with a limited number of preserved invariant sets, the same state feedback control law may be used for different states that belong to the same region between two adjacent invariant sets. This means that such a control law is not necessarily the optimal one and the control performance might be degenerated. It is obvious that increasing the number of presented invariant sets may achieve a better control performance. However, due to the limitation of storage as well as the off‐line computational burden, the number of chosen initial feasible states is restricted. In the following, two ways to further improve the control performance for Algorithm 7.4 are briefly introduced.

In [15], an additional constraint is imposed on the invariant set obtained by the off‐line part of Algorithm 7.4 as follows:

If the sets in the look‐up table can satisfy conditions (7.121), then the online part of Algorithm 7.4 can be modified as follows.

Robust stability of the closed‐loop system controlled by Algorithm 7.4 or its modified version was proved in [15]. In both cases, the obtained feedback control law between sets Ωi and Ωi+1 is guaranteed to keep x(k) within Ωi and converge it into Ωi+1 (see [15] for details). Since the feedback control law is online adjusted according to the current system state, a better control performance can be achieved.

Another way to improve Algorithm 7.1 that is different from the method in [15] is based on the fact that the convex combination of multiple ellipsoidal invariant sets is also an invariant set. For n ellipsoidal invariant sets images, h = 1, …, n with Qh, Yh, γh satisfying (7.14) to (7.16), their convex combination images is also an ellipsoidal invariant set with images, images, images satisfying (7.14) to (7.16), where λh ≥ 0, h = 1, …, n, images. This can be easily shown by multiplying a parameter λh to the inequalities in (7.14) to (7.16):

equation
equation

and then sum them for h = 1, …, n. Note that the condition (7.14) also implies V(i, k) descending and satisfying (7.7) and is tighter than needed by the invariant set.

Since conditions (7.14) to (7.16) are independent from the real‐time system state x(k), it is possible to off‐line design a series of invariant sets with different performance indices like that in Algorithm 7.4 and then to online optimize the convex combination coefficients to achieve a better control performance. The off‐line design is often formulated as several optimization problems under constraints (7.14) to (7.16), each with a specific performance index such as maximizing the covered region of the ellipsoidal invariant set Sh with a fixed γh or pursuing a better control performance by setting a much smaller γh [9]. With respect to Algorithm 7.1, the online synthesis should take the constraint (7.12) into account with off‐line solved Qh, Yh, γh, h = 1, …, n, and the real‐time state x(k), formulated as follows.

With the above modified version 2, although an optimization problem with LMI formulation still needs to be solved online, the number of online optimization variables and the rows of LMIs decrease greatly because the optimization variables are just the combination coefficients. Therefore, with this algorithm the online computational burden can be greatly reduced. The basic idea of using the convex combination of multiple invariant sets as an invariant set and the synthesis approach of off‐line designing multiple invariant sets and online optimizing the combination coefficients can also be applied to other similar cases such as the multistep control sets presented in Section 7.1.4.

7.3.4 Synthesis of the Robust Predictive Controller by QP

In the above sections, LMI is clearly a main tool used to formulate the online optimization problem for RMPC synthesis. This means that the semi‐definite programming (SDP) solver should be adopted online, which may lead to a heavy online computational burden. Furthermore, when dealing with asymmetrical constraints, there is often a need to reformulate them into symmetrical ones such that LMIs can be derived through ellipsoidal invariant sets, which obviously makes the result conservative. To overcome these shortcomings, it was also suggested to formulate the RMPC as a QP problem with which a good trade‐off among feasible region, control performance, and online computational burden could be achieved. In the following, we briefly introduce the synthesis method used by Li and Xi [16].

Consider the uncertain system (7.1) to (7.3). In order to simplify the presentation, the considered constraints are given in a very general form:

where images and images are the column vectors with appropriate dimensions and the number of constraints is p.

Consider the objective function of RMPC with an infinite horizon as

(7.124)equation

where images. To guarantee the closed‐loop stability, the control strategy is chosen as

where K is a feedback gain and c(k + i) is the additional control input. The control strategy in (7.125) is the same as (7.111) in ERPC (see Section 7.3.2). However, in order to enlarge the feasible region, here a dual‐mode control strategy is adopted, i.e. to use free control inputs to steer the system state to a terminal invariant set Ωf in N steps and then in Ωf to use the feedback control law u = Kx to stabilize the system. Using a similar method to that in Section 7.1.1, a terminal cost function can be used as an upper bound to replace the sum of infinite terms after N steps in the cost function. Thus, the online optimization problem can be formulated as

where the feedback control gain K and the weighting matrix of the terminal cost function Pf should satisfy the following conditions:

(7.127)equation

which implies that V(i, k) = x(k + i ∣ k)TPf  x(k + i ∣ k) is imposed to descend and to satisfy conditions (7.7) for all possible [A(k + i) B(k + i)] ∈ Π and J(k) after i = N has an upper bound similar to (7.8), images.

The terminal invariant set Ωf and its corresponding feedback control gain K are necessary for RMPC synthesis. They can be calculated off‐line by the following algorithm.

After Pf, Ωf, K are solved off‐line by using Algorithm 7.5, the online optimization problem (7.126) can be formulated as an QP problem, where the constraints of future control inputs should be linear and the objective function should be quadratic.

(1) Linear constraints

The constraints in the optimization problem (7.126) are all linear if the terminal set Ωf is selected as a polyhedral one. All these constraints are with respect to the future system states and control inputs. Due to the polytopic uncertainty of the system parameters, if the control strategy (7.125) is adopted, similar to the discussion in Section 7.1.3, the future system state trajectory will belong to a convex hull, whose vertices are the state trajectories along with the vertex trajectory of polytopic models (see (7.45)). For example, at time k + 1,

equation

where Φl = Al + BlK and ζi(k) = [xT(k), cT(k), … , cT(k + i)]T. Similarly,

(7.129)equation
equation

Due to the polytopic property of the system uncertainty, it is obvious that the future states x(k + i + 1 ∣ k) in the state/input constraints and the terminal set constraint in (7.126) can be replaced by all corresponding state vertices images. These constraints can be rewritten as

(7.130)equation
(7.131)equation

where

equation

(2) Quadratic cost function

Since the optimization problem (7.126) is a min‐max problem, the upper bound of J(k) is used as the objective function when converting it to a QP problem. In [16] it has been shown that if a vector w ∈ Co{wi, i = 1, … , L}, then images, where Q > 0. From time k to k + N, the future state trajectory belongs to a convex hull and thus the objective function J(k) of the predicted state trajectory must satisfy J(k) ≤ max {Ji(k)}, where images and Pi can be calculated according to the vertex sequences of different future state trajectories. Therefore, to get the upper bound of J(k), the following optimization problem should be solved to obtain the matrix P:

(7.132)equation

where Pi is determined by the quadratic terms in J(k) and corresponding images, l0, … , li = 1, …, L, i = 1, …, N − 1. If P has been solved, images can be used as the quadratic cost function for the minimization problem in (7.126).

(3) QP formulation

With the off‐line solved feedback gain K, terminal set Ωf, terminal weighting matrix Pf, and matrix P in the upper bound of the cost function, the optimization problem (7.126) can be formulated as the following QP problem:

From the above analysis it can be seen that since the constraints are commonly linear, the critical issue to transfer the RMPC optimization problem to a QP problem is to adopt the polyhedral invariant set as the terminal set and to design a quadratic function as the upper bound of the objective function.

The condition that ensures the closed‐loop system robustly stable is also given in [16], where a Lyapunov function images is adopted and the recursive feasibility is easily proved using the feasible solution at time k + 1 constructed by shifting forward the last N − 1 terms in images and setting its last term as zero.

For the QP problem (7.133), in addition to the usual QP solving methods such as the active set method, the multiparameter programming method is also helpful in reducing the online computational burden. Note that the optimization problem (7.133) is given with respect to the optimization variables c(k + i) and the current system state x(k), i.e. it has the form

equation

Therein, the state x(k) is time‐varying and determines the constraints. According to Bemporad et al. [18], the current system state x(k) can be handled as parameters and the explicit solution c(k) = Fix(k) + gi can be obtained for the parameter x(k) within the ith subset. Thus, by the multiparameter programing method, the x space is divided into several subsets and then the corresponding explicit solutions can be obtained and stored off‐line. When online, instead of solving the QP problem, it is only required to search the subset to which the state x(k) belongs and then to look up the corresponding feedback gain (Fi, gi) to construct the current control input.

7.4 Summary

The research of RMPC is an important part of the qualitative synthesis theory of predictive control. In particular, it has become the mainstream of the theoretical research of predictive control since the beginning of this century. The RMPC makes full use of the rich theoretical results on stable predictive controller synthesis and robust controller design, and has been investigated for a wide range of different types of systems. In this chapter, we introduced the fundamental concepts and typical methods for synthesizing robust predictive controllers, focusing on the RMPC for polytopic uncertain systems and the systems with external disturbance.

The RMPC synthesis using LMI in [1] is a classic work in the RMPC field, which triggered much subsequent research and has achieved fruitful theoretical results. In this chapter, the basic version of this method was firstly introduced and how to formulate the problem and how to handle the main difficulties were explained in detail. Based on this, some of its extensions were further introduced in subsequent subsections, i.e. using parameter‐dependent Lyapunov functions, adopting dual‐mode control schemes with either additional free control sequence or multiple feedback control laws. These improved approaches can reduce conservativeness of the basic RMPC version and improve the control performance, but often at the cost of increasing the computational complexity.

For disturbed systems, how to efficiently handle the external disturbance is the critical issue in RMPC deign. Two catalogs of synthesis approaches for such kinds of RMPC are introduced. The first one adopts the disturbance invariant set and divides the disturbed system into two subsystems. With a proper design for the nominal subsystem and the disturbed subsystem, the real state of the controlled system can be guaranteed to move along a tunnel with the nominal system state as the center and the disturbance invariant set as the cross‐section. The second one borrows the idea in robust control and adopts a mixed H2/H performance to synthesize RMPC. In this design, an upper bound on the H performance index is firstly assigned and RMPC is then designed by online optimizing the H2 performance index with the upper bound as an additional constraint.

Based on analyzing the limitations of the basic RMPC method in [1], we point out the main difficulty in RMPC synthesis, i.e. the contradiction among the requirements of a feasible region, the online computational burden, and control performance. In the remainder of this chapter, some representative approaches were introduced to illustrate how to handle this contradiction, showing the variety of the RMPC research. It can be seen that when using some strategies with high computational complexity to improve the control performance and enlarge the feasible region, the off‐line design and online synthesis is possibly a realistic way to reduce the online computational burden.

References

  1. 1 Kothare, M.V., Balakrishnan, V., and Morari, M. (1996). Robust constrained model predictive control using linear matrix inequalities. Automatica 32 (10): 1361–1379.
  2. 2 Cuzzola, F.A., Geromel, J.C., and Morari, M. (2002). An improved approach for constrained robust model predictive control. Automatica 38 (7): 1183–1189.
  3. 3 Mao, W.J. (2003). Robust stabilization of uncertain time‐varying discrete systems and comments on “an improved approach for constrained robust model predictive control”. Automatica 39 (6): 1109–1112.
  4. 4 Michalska, H. and Mayne, D.Q. (1993). Robust receding horizon control of constrained nonlinear systems. IEEE Transactions on Automatic Control 38 (11): 1623–1633.
  5. 5 Wan, Z. and Kothare, M.V. (2003). Efficient robust constrained model predictive control with a time varying terminal constraint set. Systems and Control Letters 48: 375–383.
  6. 6 Ding, B., Xi, Y., and Li, S. (2004). A synthesis approach of on‐line constrained robust model predictive control. Automatica 40 (1): 163–167.
  7. 7 Pluymers, B., Suykens, J.A.K., and De Moor, B. (2005). Min–max feedback MPC using a time‐varying terminal constraint set and comments on “Efficient robust constrained model predictive control with a time‐varying terminal constraint set”. Systems and Control Letters 54: 1143–1148.
  8. 8 Ding, B.C. (2010). Modern Predictive Control. Boca Raton: CRC Press.
  9. 9 Li, D. and Xi, Y. (2009). Design of robust model predictive control based on multi‐step control set. Acta Automatica Sinica 35 (4): 433–437.
  10. 10 Lu, Y. and Arkun, Y. (2000). Quasi‐min‐max MPC algorithms for LPV systems. Automatica 36 (4): 527–540.
  11. 11 Mayne, D.Q., Seron, M.M., and Raković, S.V. (2005). Robust model predictive control of constrained linear systems with bounded disturbances. Automatica 41 (2): 219–224.
  12. 12 Huang, H., Li, D., and Xi, Y. (2012). Design and ISpS analysis of the mixed H2/H feedback robust model predictive control. IET Control Theory and Applications 6 (4): 498–505.
  13. 13 Kouvaritakis, B., Rossiter, J.A., and Schuurmans, J. (2000). Efficient robust predictive control. IEEE Transactions on Automatic Control 45 (8): 1545–1549.
  14. 14 Kouvaritakis, B., Cannon, M., and Rossiter, J.A. (2002). Who needs QP for linear MPC anyway? Automatica 38 (5): 879–884.
  15. 15 Wan, Z. and Kothare, M.V. (2003). An efficient off‐line formulation of robust model predictive control using linear matrix inequalities. Automatica 39 (5): 837–846.
  16. 16 Li, D. and Xi, Y. (2011). The synthesis of robust model predictive control with QP formulation. International Journal of Modeling, Identification and Control 13 (1–2): 1–8.
  17. 17 Pluymers, B., Rossiter, J.A., Suykens, J.A.K., and Moor, B.D. The efficient computation of polyhedral invariant sets for linear systems with polytopic uncertainty, Proceedings of 2005 American Control Conference, Portland, OR (8–10 June 2005), 804–809.
  18. 18 Bemporad, A., Morari, M., Dua, V., and Pistikopoulos, E.N. (2002). The explicit linear quadratic regulator for constrained systems. Automatica 38 (1): 3–20.
..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset