Maximal terminal region approach for MPC using subsets sequence

Yafeng WANG , Fuchun SUN , Huaping LIU , Dongfang YANG

Front. Electr. Electron. Eng. ›› 2012, Vol. 7 ›› Issue (2) : 270 -278.

PDF (327KB)
Front. Electr. Electron. Eng. ›› 2012, Vol. 7 ›› Issue (2) :270 -278. DOI: 10.1007/s11460-012-0178-y
RESEARCH ARTICLE
RESEARCH ARTICLE

Maximal terminal region approach for MPC using subsets sequence

Author information +
History +
PDF (327KB)

Abstract

For the sake of enlarging the domain of attraction of model predictive control (MPC), a novel approach of gradually approximating the maximal terminal state region is proposed in this paper. Given the terminal cost, both surrounding set sequence and subsets sequence of the maximal terminal region were constructed by employing one-step set expansion iteratively. It was theoretically proved that when the iteration time goes to infinity, both the surrounding set sequence and the subsets sequence will converge to the maximal terminal region. All surrounding and subsets sets in these sequences are extracted from the state space by exploiting support vector machine (SVM) classifier. Finally, simulations are implemented to validate the effectiveness of this approach.

Keywords

model predictive control (MPC) / terminal region / domain of attraction / support vector machine (SVM)

Cite this article

Download citation ▾
Yafeng WANG, Fuchun SUN, Huaping LIU, Dongfang YANG. Maximal terminal region approach for MPC using subsets sequence. Front. Electr. Electron. Eng., 2012, 7(2): 270-278 DOI:10.1007/s11460-012-0178-y

登录浏览全文

4963

注册一个新账户 忘记密码

Introduction

Model predictive control (MPC), which approximates the infinite time-domain optimal control by finite horizon optimization, is well known as a suboptimal control algorithm. Due to the ability to handle control with state constraints, MPC has become quite popular recently. During the approximating process in MPC, the stability problem arises. To guarantee the stability of MPC, a terminal constraint and a terminal cost are added to the online optimization problem such that the terminal region is a positively invariant set for the system and the terminal cost is an associated Lyapunov function [1,2]. It is well accepted that the attraction domain of MPC can be enlarged either by increasing the prediction horizon or by enlarging the terminal region. Whereas, the former will subsequently increase the computational complexity, see Ref. [3] for example. Thus, the latter has attracted much more attentions in previous literatures.

In Refs. [4] and [5], an ellipsoidal set and a polytopic set, which were included in the stabilizable region of using linear feedback controller, were designed as the terminal region, respectively. While in Ref. [6], a saturated local control law was utilized to enlarge the terminal region. In Ref. [7], support vector machine (SVM) was employed to estimate the stabilizable region of linear feedback controller, and the subsequently estimated stabilizable region was utilized as the terminal region; thus, enlarged the terminal region dramatically. Moreover, it was proved in Ref. [8] that without terminal constraints, the terminal region can be enlarged by weighting the terminal cost. In Ref. [9], the enlargement of the domain of attraction was obtained by employing a contractive terminal constraint. Most recently, the domain of attraction was enlarged by the inclusion of an appropriate set of slacked terminal constraints into the control problem, as described in Ref. [2]. These previous literatures are of the same framework that the control law is supposed as known, and the corresponding terminal state region is calculated. Thus, there is no guarantee that the consequent terminal region is the maximal one in different control laws.

According to the given terminal cost-function, this paper proposes a novel approach that gradually approximates the maximal terminal state region based on the stability constraints. The fundamental principle of this method is the iterative procedure: At first, the sufficient conditions to guarantee the stability of MPC are presented and the maximal terminal region satisfying these conditions is defined. Then, given the terminal cost and an initial surrounding set or subset of the maximal terminal region according to the Lyapunov condition, a subsets sequence is obtained by using one-step set expansion iteratively. Moreover, it is proved that when the iteration time goes to infinity, this subsets sequence will converge to the maximal terminal region. Finally, the surrounding sets or subsets sequence are sequentially separated from the state space by exploiting SVM classifier (see Refs. [10,11] for details of SVM), and the maximal terminal state region is approximated gradually.

Model predictive control

Consider the discrete-time system as follows:
xk+1=f(xk,uk),
where xkRn, ukRm are the state and the input of the system at the sampling time k, respectively. xk+1Rn is the successor state and the mapping f:Rn+mRn satisfying f(0,0)=0 is known. The system is subject to constraints on both state and control action. They are given by xkX, ukU, where X is a closed and bounded set, U is a compact set. Both of them contain the origin.

The online optimization problem of MPC at the sample time k, denoted by PN(xk), is stated as
minu(i,xk)UJN(u,xk)=i=0N-1q(x(i,xk),u(i,xk))+F(x(N,xk)),  s.t. x(i+1,xk)=f(x(i,xk),u(i,xk)),    x(i+1,xk)X, u(i,xk)U, x(N,xk)Xf,
where x(0,xk)=xk is the state at the sample time k, q(x,u) denotes the stage cost and it is positive definite, N is the prediction horizon, Xf denotes the terminal region and it is closed and satisfies 0XfX, F() satisfying F(0)=0 is the terminal cost and it is continuous and positive definite.

Consider the following assumption at first.

Assumption 1 For the terminal region and the terminal cost, the following two conditions are satisfied [1]:

(C1) For any xXf, there exists a Lyapunov function F() that satisfying the inequality
F(x)minuU{q(x,u)+F(f(x,u))}.

(C2) Xf is a positively invariant set. For any xXf, by using the optimal control resulting from the minimization problem showed in condition (C1), denoted by uopt, we have f(x,uopt)Xf.

Let JN*(xk) be the minimum of PN(xk) and uN*(xk)={uN*(0,xk),uN*(1,xk),,uN*(N-1,xk)} be the optimal control trajectory. The control strategy of MPC is that, at the sample time k, uN*(0,xk) is inputted into the real system, and at the sample time k+1, the control inputted into the system is not uN*(1,xk) but the first element of the optimal control trajectory resulting from the similar online optimization problem. At the sample time k+1, the state is xk+1=f(xk,uN*(0,xk)) and the online optimization problem, denoted by PN(xk+1), is the same as Eq. (2) except that xk is replaced by xk+1. Similarly, let JN*(xk+1) be the minimum of PN(xk+1) and uN*(xk+1)={uN*(0,xk+1),uN*(1,xk+1),,uN*(N-1,xk+1)} be the optimal control trajectory. The control inputted into the system at the sample time k+1 is uN*(0,xk+1). Thus, the control law of MPC can be stated as
uRH(xk)=uN*(0,xk), k=0,1,2,,.

The closed-loop stability of the controlled system is shown in Lemma 1.

Lemma 1 For any x0X, if x0 satisfies xN*(N,x0)Xf and Assumption 1 is satisfied, it is guaranteed that x0 will be steered to0by using the control law of MPC.

The proof can be found in Ref. [1].

Gradually approximation of the maximal terminal region

Using SVM classifier to estimate the terminal region, as described in Ref. [7], is considered as an effective approach for solving the approximating problem in MPC. However, the method in Ref. [7] is somewhat conservative. The reason is that, the obtained terminal region actually is the stabilizable region of using a predetermined linear feedback controller.

In this section, a novel method of computing a terminal region is proposed. Given the terminal cost function and an initial set of the maximal terminal region, both surrounding sequence and subsets sequence can be constructed by using one-step set expansion iteratively. When some conditions are satisfied, the iteration ends and the consequent set can be seen as the suboptimal estimation of the maximal terminal region.

Construction of surrounding set sequence

The proposed approach constructs the surrounding set sequence directly from conditions (C1) and (C2). The terminal region Xf can be defined as
Xf:={xX|F(x)FXf*(x)},
where FXf*(x) is the solution of the following optimization problem:
minuUFXf(x)=q(x,u)+F(f(x,u)),  s.t.  f(x,u)Xf.

Remark 1 The definition of Xf has two meanings: (I) the optimization problem (4) has feasible solution, that is to say, uU, s.t. f(x,u)Xf; (II) the minimum of the optimization problem satisfies that FXf*(x)F(x).

Remark 2 From the definition of Xf, it is obvious that, the terminal region is essentially a positively invariant set of using the optimal control resulting from the optimization problem (4) when F() is given.

Remark 3 In Refs. [4,5,7], the linear feedback control is attached to the construction of Xf, and Xf is the stabilizable region of using the linear feedback controller. In Ref. [6], a saturated local control law was used. However, in this paper, there is no explicit control attached to the definition of Xf. Thus, the requirement on Xf is lower than that in Refs. [4-7] while it guarantees the stability of the controlled system.

From the definition of Xf, it cannot be determined whether a state point belongs to Xf. The difficulty arises that the Xf acts as the constraint in the optimization problem (4). To solve this problem, the method of using one-step set expansion iteratively is adopted.

The iterative approximating process begins with an initial estimation satisfying Xf0Xf, and from which, a more accurate estimation Xf1 is achieved. Repeatedly, the surrounding set sequence Xf2,Xf3,…,Xf+, which satisfies Xf0Xf1XfjXf, can be estimated sequentially. And the convergence of this sequence is proved in previous section. The gradual approximation process can be depicted in Fig. 1, the specific of this approach can be found in the authors’ previous literature [12].

The initial surrounding set is determined by
Xf0:={xX|F(x)FXf0*(x)}.

The iterative construction of surrounding set sequence is enforced based on the following assumptions.

Assumption 2 A surrounding set of Xf,max, denoted by Xf0 and containing the origin, is initialized, where Xf,max is defined as the largest terminal region.

Assumption 3 Xf0 is a positively invariant set, that is to say, for any xXf0, uU, s.t.F(x)q(x,u)+F(f(x,u)) and f(x,u)Xf0.

According to the known Xf0, another surrounding set of Xf,max, denoted by Xf1, can be constructed as
Xf1:={xXf0|F(x)FXf1*(x)},
where FXf1*(x) is the minimum of
minuUFXf1(x)=q(x,u)+F(f(x,u)),  s.t.  f(x,u)Xf0.

As mentioned in Remark 1, the construction of Xf1 contains two meanings: (I) for any xXf1, uU, s.t. f(x,u)Xf0; (II) the minimum of Eq. (7) satisfies FXf0*(x)F(x). The constructions of Xfj in sequel have the similar meanings. And the convergence of this sequence can be proved from the following theorem.

Theorem 1 If Assumption 2 and Assumption 3 are satisfied, the sequence {Xfj}, similarly defined as Xf0 and Xf1, will converge to Xf,max when j goes to infinity.

Proof This theorem is proved by contradiction.

Supposes there exists XspoXf,max to guarantee that XfjXspo. Then, for xXspo, satisfies
F(x)minuU{q(x,u)+F(f(x,u))},
where f(x,u)Xspo. Moreover, it is contractive to the fact that Xf,max is the maximal set, which satisfies conditions (C1) and (C2) as listed in Assumption 1.

And vice versa, supposes there exists XspoXf,max to guarantee that XfjXspo. Then, there exists a certain integer N that XfNXf,max and XfN+1Xf,max, for xXf,max\XfN+1, we have
F(x)minuU{q(x,u)+F(f(x,u))},
where f(x,u)Xf,maxXfN. Because xXfN, then, xXfN+1, which leads to paradox.

End proof.

From the previous description, we can find that the surrounding approximating approach will probably contain some unstable point by mistake. Then, we naturally proposed another gradually extending-approximation approach, which will be detailedly illustrated in the following subsection.

Construction of subsets sequence

Given F() and from conditions (C1) and (C2), the terminal region Xf can be defined as
Xf:={xX|F(x)FXf*(x)},
where FXf*(x) is the minimum of the following optimization problem:
minuUFXf(x)=q(x,u)+F(f(x,u)),  s.t.  f(x,u)Xf.

Remind here that Remark 1 and Remark 2 are also consistent in the problem (9). From the definition of Xf, it cannot be determined whether a state point belongs to Xf. The difficulty lies in that, the Xf itself acts as the constraint in the optimization problem (9).

To avoid this problem, the method of using one-step set expansion iteratively is adopted. The gradual approximation process can be depicted in Fig. 2.

Assumption 4 A subset of Xf,max, denoted by Xf0 and containing the origin, is known, and Xf,max is denoted as the largest terminal region.

Assumption 5 Xf0 is a positively invariant set, that is to say, for any xXf0, uU, s.t. F(x)q(x,u)+F(f(x,u))and f(x,u)Xf0.

Given Xf0, another subset of Xf,max, denoted by Xf1, can be constructed as
Xf1:={xX|F(x)FXf0*(x)},
where FXf0*(x) is the minimum of
minuUFXf0(x)=q(x,u)+F(f(x,u)),  s.t.  f(x,u)Xf0.

As mentioned in Remark 1, the construction of Xf1 contains two meanings: (I) for any xXf1, uU, s.t. f(x,u)Xf0; (II) the minimum of Eq. (11) satisfies FXf0*(x)F(x). The constructions of Xfj in sequel have the similar meanings.

Lemma 2 If Assumption 3 is satisfied, there is Xf0Xf1.

Proof If Assumption 5 is satisfied, it is obvious that, for any xXf0, uU, s.t. F(x)q(x,u)+F(f(x,u)) and f(x,u)Xf0. It follows that, F(x)FXf0*(x). From the construction of Xf1, we can know xXf1, namely, Xf0Xf1.

End proof.

Remark 4 From the construction of Xf1, it is obvious that, if Assumption 5 is satisfied, Xf1 is a positively invariant set. We know that, for any xXf1, uU, s.t. F(x)q(x,u)+F(f(x,u)) and f(x,u)Xf0. Because of Xf0Xf1 as shown in Lemma 2, we have f(x,u)Xf1.

Similarly, by replacing Xf0 with Xf1 in the constraint of Eq. (11), another subset, denoted by Xf2, can be obtained as follows:
Xf2:={xX|F(x)FXf1*(x)},
where FXf1*(x) is the minimum of
minuUFXf1(x)=q(x,u)+F(f(x,u)),  s.t.  f(x,u)Xf1.

Repeatedly, Xfj, j=3,4,,, can be constructed as
Xfj:={xX|F(x)FXfj-1*(x)},
where FXfj-1*(x) is the minimum of
minuUFXfj-1(x)=q(x,u)+F(f(x,u)),  s.t.  f(x,u)Xfj-1.

This method of constructing Xfj given Xfj-1 is defined as one-step set expansion in this paper. By employing it iteratively, a subsets sequence of largest terminal region, denoted by {Xfj}, j=1,2,,, can be achieved.

Remark 5 Similar with Lemma 2 and Remark 3, any subset in this sequence is positively invariant and any two neighboring subsets satisfy Xfj-1Xfj.

As j increases, {Xfj} will converge to a set, denoted by Xf+. Theorem 2 will show that, Xf+ is equal to the largest terminal region.

Theorem 2 If Assumption 4 and Assumption 5 are satisfied, forXfj constructed in Eqs. (14) and (15), whenj goes to infinity, {Xfj}will converge toXf,max.

Proof This theorem is proved by contradiction.

(A) Assume that, there exists a set which is denoted by Xspo satisfying XspoXf,max and XfjXspo when j+. From Remark 5, we can know Xf0Xspo. It is obvious that 0Xspo because of 0Xf0 as shown in Assumption 4. It follows that, 0Xf,max\Xspo and for any xXf,max\Xspo, we have F(x)>0 since F() is positive definite. Define ξ as the minimum of {F(x)|xXf,max\Xspo}, it is satisfied that, ξ>0.

From the construction of Xfj, we know that, for any x0Xf,max\Xspo, there exists no such a uU satisfying F(x0)q(x0,u)+F(f(x0,u)) and f(x0,u)Xspo because of XspoXf,max. However, from conditions (C1) and (C2), we know that, u(x0)U, s.t. F(x0)q(x0,u(x0))+F(x1) and x1Xf,max, where x1=f(x0,u(x0)). It is obvious that, x1Xspo. Thus, we have x1Xf,max\Xspo. Similarly, we can know, u(x1)U, s.t. F(x1)q(x1,u(x1))+F(x2) and x2Xf,max\Xspo, where x2=f(x1,u(x1)), since x1Xf,max\Xspo.

Repeatly, for xiXf,max\Xspo,u(xi)U, s.t. F(xi)q(xi,u(xi))+F(xi+1) and xi+1Xf,max\Xspo, where xi+1=f(xi,u(xi)), i=2,3,,. It is clear that, F(xi)0 when i. We know that, for the minimum of {F(x)|xXf,max\Xspo}, defined as ξ, there is a positive real number δ satisfying 0<δ<ξ. Since F(xi)0 when i, Nδ>0, s.t. for any iNδ, we have F(xi)<δ. Obviously, this is contradicted with that ξ is the minimum of {F(x)|xXf,max\Xspo}.

(B) Similarly, assume that, there exists an Xspo satisfying XspoXf,max and XfjXspo when j+. For any xXspo, we have that F(x)minuU{q(x,u)+F(f(x,u))} and f(x,u)Xspo. Obviously, this is contradicted with that Xf,max is the largest one satisfying conditions (C1) and (C2).

End proof.

Remark 6 In this paper, the largest terminal region means the positively invariant set satisfying conditions (C1) and (C2). However, conditions (C1) and (C2) are sufficient conditions to guarantee the stability of the controlled system, not the necessary conditions. There may be a set larger than Xf,max and the stability of the controlled system can be guaranteed by using this set as the terminal region.

Remark 7 In the calculation of Xf,max, it is impossible to keep iteration computation until j+. When the iteration time goes to j=E (E is a positive integer), if XfE is equal to XfE-1 in principle, it can be deemed that {Xfj} converges to XfE in rough. Hence, XfE can be taken as the terminal region and it is a good approximation to Xf,max.

Remark 8 If the iteration time does not go to infinity, the obtained set may be just a large positively invariant subset of Xf,max. This has no effect on the stability of the controlled system. The only negative influence is that its corresponding domain of attraction is smaller than that corresponding to Xf,max.

Until now, it seems that we can choose any Xfj in the subsets sequence as the terminal region. This is infeasible. Since Xfj is not described in explicit expression, it cannot serve as the terminal constraint in the optimization problem (2) directly. Then, an estimated one described in explicit expression is needed. Due to the strong optimizing ability of SVM, SVM is exploited to separate each Xfj from the state space.

Support vector machine

SVM is the youngest part in the statistical learning theory. It is an effective approach for pattern recognition. In SVM approach, the main aim is to obtain a function, which determines the decision boundary or hyperplane. This hyperplane optimally separates two classes of input data points.

Take the example of separating X into A and X\A. For each xiA, an additional variable yi=+1 is introduced. Similarly, for each xiX\A, yi=-1 is introduced. Define I+:={i:yi=+1} and I-:={i:yi=-1}, SVM will find a separating hyperplane, denoted by O(x):=wϕ(xi)+b=0, between A and X\A. Therefore, A can be estimated as A^={xX|O(x)0}, where O(x) is determined by solving the following problem:
minα 12i j αiαjyiyjker(xi,xj)-i αi, s.t.   i αiyi=0,     0αiC, iI+; αi0, iI-,
where ker(,) denotes the kernel function and the Gaussian kernel as follows is adopted in this paper:
ker(x,xi)=exp(-x-xi22σ2),
with σ being the positive Gaussian kernel width.

When {αi} are computed out, some support vectors are chosen from {xi} and the optimal hyperplane can be determined with these support vectors and their relevant weights. Denote Ps as the number of support vectors and Xs as the support vectors set, the optimal hyperplane is described as
O(x)=i=1Pswiker(xi,x)+b
where xiXs is the support vector and wi=αiyi satisfying i=1Pswi=0 is the relevant weight.

There are many software packages of SVM available on internet. They can be downloaded and used directly. To save space, it is not introduced in detail in this paper. For more details, please refer to Refs. [10] and [11].

Estimating the approximating sequence by employing SVM

From Subsection 3.3, we know that, SVM find a separating hyperplane between {xi|iI+} and {xi|iI-}. This hyperplane is used to separate X into A and X\A. All of {xi} and their relevant {yi} compose a set, named the training points set. This subsection will show how to achieve the training points set when estimating Xfj and how to determine Xfj when the separating hyperplane is known.

First, choose arbitrary points xiX, i=1,2,...,P (P is the number of training points); then, assign yi to each xi by implementing the following procedure:

IF (I) The following optimization problem has feasible solution
minuUFXfj(xi)=q(xi,u)+F(f(xi,u)),  s.t.  f(xi,u)X^fj-1.
(When j=1, X^f0=Xf0.)

(II) Its minimum satisfies
F(xi)FXfj*(xi).

THEN  yi=+1

ELSE  yi=-1

END IF

By implementing this procedure for every xi, each yi is known. Input {xi} and {yi} into SVM classifier, an optimal hyperplane Oj(x)=0 will be obtained. Therefore, the estimated set of Xfj can be achieved as X^fj={xX|Oj(x)0}.

When X^fj is known, the training points for separating Xfj+1 from X can be computed by the similar procedure. By inputting them into SVM classifier, a hyperplane Oj+1(x)=0 and an estimated set of Xfj+1, denoted by X^fj+1={xX|Oj+1(x)0}, will be obtained. Repeatedly, {Oj(x)}, j=1,2,,, and {X^fj} can be achieved by the similar technology.

Estimating the terminal region

Section 3 showed how to achieve the subsets sequence by employing SVM. Theoretically, the larger the iteration time j is, the higher the precision of X^fj approaching to Xf,max is. However, it is impossible to keep computation until j+. To avoid this problem, the iteration should be ended when some conditions are satisfied.

When j=E, if it is satisfied that, for xiXs,E-1, i=1,2,,Ps,E-1, there exists
i=1Ps,E-1OE(xi)-OE-1(xi)ϵPs,E-1,
and we can find that X^fE is equal to X^fE-1 in principle and X^fj converges to X^fE. In Eq. (19), Xs,E-1 is the support vectors set at j=E-1, Ps,E-1 is the number of support vectors and ϵ is a tunable threshold. The smaller ϵ is, the higher the precision of X^fE approximating to Xf,max is. Finally, X^fE is used to serve as the terminal region.

Remark 9 Here, we used the information that, in SVM classifier, the hyperplanes are only determined on the support vectors.

Now, the concrete algorithm of estimating the largest terminal region is displayed as follows.

Step 1 Set the number of training points P used in SVM and the tunable threshold ϵ.

Step 2 For j=1,2,,, use SVM to achieve the optimal hyperplane Oj(x)=0 and the estimated set of Xfj, denoted by X^fj.

Substep 1 Choose arbitrary points xiX, i=1,2,...,P.

Substep 2 Assign yi to each xi by implementing the procedure in Subsection 3.4.

Substep 3 Input {xi,yi} into the SVM. An optimal hyperplane Oj(x)=0 will be obtained and Xfj can be approximated by X^fj={xX|Oj(x)0}, where
Oj(x)=i=1Ps,jwiker(xi,x)+bj,

with Ps,j denoting the number of support vectors, xi being the support vector, wi denoting its relevant weight, and bj denoting the classifier threshold.

Step 3 Check the iteration status. When j=E, if inequality (14) is satisfied, end iteration and take X^fE as the largest terminal region.

Remark 10 It is obvious that, X^fj is achieved one by one. Namely, X^fj can only be achieved when X^fj-1 is known.

Simulation experiment

In this section, we implemented both surrounding and subset approximating approach in simulated model, and the comparison to representative alternatives are shown in results. The model is a discrete-time realization of the continuous-time system used in Refs. [4,7]:
[x1(k+1)x2(k+1)]=[1TT1][x1(k)x2(k)]+[TμTμ]u(k)      +[T(1-μ)00-4T(1-μ)][x1(k)x2(k)]u(k),
where μ=0.5, T=0.1 s, and the state constraint and control constraint are X={x|x14}, U={u||u|2}, respectively.

The stage cost is q(x,u)=xTQx+uTRu, where Q=0.5I and R=1. The terminal cost is chosen as F(x)=xTGx, where G=[1107.356 857.231; 857.231 1107.356] and Xf0 is given as the terminal region in Ref. [4], which is
Xf0={xX|xT[16.592611.592611.592616.5926]x0.7}
.

To estimate each Xfj, 4000 training points are generated, and set ϵ=2.5. From the simulating process we find that in the surrounding approximating approach, when j is iterated to 15, Xf15 and Xf14 are near enough. There only exists minor difference between Xf15 and Xfj, then,
i=1Ps,14O15(xi)-O14(xi)ϵPs,14,
where xiXs,14, Xs,14 is the support vectors set, and Ps,14 is the number of support vectors at j=14. For j>15, so we can choose Xf15 as the terminal estimation of Xf,max. The approximating process can be well identified in Fig. 3. On the other hand, if the subsets approximating approach is adopted, if j is as large as 22, the following iterations various slowly enough. It can be rearranged as
i=1Ps,21O22(xi)-O21(xi)ϵPs,21,
where xiXs,21, Xs,21 is the support vectors set, and Ps,21 is the number of support vectors at j=21. Thus, X^f22 can be seen as the final estimation of Xf,max, and the approximating process is well depicted in Fig. 4.

In Figs. 3 and 4, the blue circles, which stem from the method that presented in Ref. [4], are selected as the initial set of surrounding and subset approaches, respectively. We can find that the converged sets in both approaches are nearly the same, as depicted by black solid lines, so in the following illustration, the estimation results of surrounding method is neglected for clarity.

Figure 5 depicts the convergent trajectory of some state points, which are selected to implement the comparison between representative alternatives and the proposed approach. These state points are (-42), (-44), (-31.5), (-23.5), (0.15-4), (0.30) and (3.7-0.5), as denoted by “*” in Fig. 5.

In Fig. 5, the blue ellipsoid is the terminal region in Ref. [4] and the region encompassed by black dash lines is the result in Ref. [7]. The region encompassed by black solid lines is the terminal region in this paper. From this figure, we can find that all the selected states, which obviously lie in the domain of attraction, are out of the estimating regions of both Refs. [4] and [7], but they are all included in the corresponding results of this paper. The terminal region in this paper contains the solution that obtained in Ref. [4], but not entirely contains the result in Ref. [7] although it is much larger than that in Ref. [7]. The reason is that, the terminal region in this paper is the largest one satisfying conditions (C1) and (C2). However, conditions (C1) and (C2) are just the sufficient conditions to guarantee the stability of the controlled system, not the necessary conditions as shown in Remark 9. The red solid lines denote the closed-loop trajectories of the selected points. Note that, with the same sampling interval and prediction horizon as those in this paper, these points are not in the regions of attraction of MPC in Refs. [4] and [7]. However, they can be leaded to the origin by using the control law of MPC in this paper.

Conclusion

Given the terminal cost, a sequence of surrounding sets and subsets of the maximal terminal region is extracted from state space iteratively by employing SVM classifier. The convergence of these approximation processes are proved in analytical way. Simulations also show that the convergent region is better approximation to the maximal terminal region, when compared to some mature representative alternatives.

References

[1]

Mayne D Q, Rawlings J B, Rao C V, Scokaert P O M. Constrained model predictive control: Stability and optimality. Automatica, 2000, 36(6): 789-814

[2]

González A H, Odloak D. Enlarging the domain of attraction of stable MPC controllers, maintaining the output performance. Automatica, 2009, 45(4): 1080-1085

[3]

Magni L, De Nicolao G, Magnani L, Scattolini R. A stabilizing model based predictive control algorithm for nonlinear systems. Automatica, 2001, 37(9): 1351-1362

[4]

Chen H, Allgower F. A quasi-infinite horizon nonlinear model predictive control scheme with guaranteed stability. Automatica, 1998, 34(10): 1205-1217

[5]

Cannon M, Deshmukh V, Kouvaritakis B. Nonlinear model predictive control with polytopic invariant sets. Automatica, 2003, 39(8): 1487-1494

[6]

De Doná J A, Seron M M, Mayne D Q, Goodwin G C. Enlarged terminal sets guaranteeing stability of receding horizon control. Systems & Control Letters, 2002, 47(1): 57-63

[7]

Ong C J, Sui D, Gilbert E G. Enlarging the terminal region of nonlinear model predictive control using the support vector machine method. Automatica, 2006, 42(6): 1011-1016

[8]

Limon D, Alamo T, Salas F, Camacho E F. On the stability of constrained MPC without terminal constraint. IEEE Transactions on Automatic Control, 2006, 51(5): 832-836

[9]

Limon D, Alamo T, Camacho E F. Enlarging the domain of attraction of MPC controllers. Automatica, 2005, 41(4): 629-635

[10]

Burges C C. A tutorial on support vector machines for pattern recognition. Data Mining and Knowledge Discovery, 1998, 2(2): 121-167

[11]

Vapnik V. The Nature of Statistical Learning Theory. New York: Springer, 1995

[12]

Wang Y F, Sun F C, Zhang Y A, Liu H P, Min H B. Getting a suitable terminal cost and maximizing the terminal region for MPC. Mathematical Problems in Engineering, 2010, 2010: 1-16

RIGHTS & PERMISSIONS

Higher Education Press and Springer-Verlag Berlin Heidelberg

AI Summary AI Mindmap
PDF (327KB)

756

Accesses

0

Citation

Detail

Sections
Recommended

AI思维导图

/