1. Tsinghua National Laboratory for Information Science and Technology, Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China
2. The Aviation General Office in Guangzhou Bureau of Naval General Armaments Department, Anshun 561018, China
fcsun@mail.tsinghua.edu.cn
Show less
History+
Received
Accepted
Published Online
2011-04-28
2011-10-12
2012-06-05
PDF
(327KB)
Abstract
For the sake of enlarging the domain of attraction of model predictive control (MPC), a novel approach of gradually approximating the maximal terminal state region is proposed in this paper. Given the terminal cost, both surrounding set sequence and subsets sequence of the maximal terminal region were constructed by employing one-step set expansion iteratively. It was theoretically proved that when the iteration time goes to infinity, both the surrounding set sequence and the subsets sequence will converge to the maximal terminal region. All surrounding and subsets sets in these sequences are extracted from the state space by exploiting support vector machine (SVM) classifier. Finally, simulations are implemented to validate the effectiveness of this approach.
Model predictive control (MPC), which approximates the infinite time-domain optimal control by finite horizon optimization, is well known as a suboptimal control algorithm. Due to the ability to handle control with state constraints, MPC has become quite popular recently. During the approximating process in MPC, the stability problem arises. To guarantee the stability of MPC, a terminal constraint and a terminal cost are added to the online optimization problem such that the terminal region is a positively invariant set for the system and the terminal cost is an associated Lyapunov function [1,2]. It is well accepted that the attraction domain of MPC can be enlarged either by increasing the prediction horizon or by enlarging the terminal region. Whereas, the former will subsequently increase the computational complexity, see Ref. [3] for example. Thus, the latter has attracted much more attentions in previous literatures.
In Refs. [4] and [5], an ellipsoidal set and a polytopic set, which were included in the stabilizable region of using linear feedback controller, were designed as the terminal region, respectively. While in Ref. [6], a saturated local control law was utilized to enlarge the terminal region. In Ref. [7], support vector machine (SVM) was employed to estimate the stabilizable region of linear feedback controller, and the subsequently estimated stabilizable region was utilized as the terminal region; thus, enlarged the terminal region dramatically. Moreover, it was proved in Ref. [8] that without terminal constraints, the terminal region can be enlarged by weighting the terminal cost. In Ref. [9], the enlargement of the domain of attraction was obtained by employing a contractive terminal constraint. Most recently, the domain of attraction was enlarged by the inclusion of an appropriate set of slacked terminal constraints into the control problem, as described in Ref. [2]. These previous literatures are of the same framework that the control law is supposed as known, and the corresponding terminal state region is calculated. Thus, there is no guarantee that the consequent terminal region is the maximal one in different control laws.
According to the given terminal cost-function, this paper proposes a novel approach that gradually approximates the maximal terminal state region based on the stability constraints. The fundamental principle of this method is the iterative procedure: At first, the sufficient conditions to guarantee the stability of MPC are presented and the maximal terminal region satisfying these conditions is defined. Then, given the terminal cost and an initial surrounding set or subset of the maximal terminal region according to the Lyapunov condition, a subsets sequence is obtained by using one-step set expansion iteratively. Moreover, it is proved that when the iteration time goes to infinity, this subsets sequence will converge to the maximal terminal region. Finally, the surrounding sets or subsets sequence are sequentially separated from the state space by exploiting SVM classifier (see Refs. [10,11] for details of SVM), and the maximal terminal state region is approximated gradually.
Model predictive control
Consider the discrete-time system as follows:where , are the state and the input of the system at the sampling time , respectively. is the successor state and the mapping satisfying is known. The system is subject to constraints on both state and control action. They are given by , , where is a closed and bounded set, is a compact set. Both of them contain the origin.
The online optimization problem of MPC at the sample time , denoted by , is stated aswhere is the state at the sample time , denotes the stage cost and it is positive definite, is the prediction horizon, denotes the terminal region and it is closed and satisfies , satisfying is the terminal cost and it is continuous and positive definite.
Consider the following assumption at first.
Assumption 1 For the terminal region and the terminal cost, the following two conditions are satisfied [1]:
(C1) For any , there exists a Lyapunov function that satisfying the inequality
(C2) is a positively invariant set. For any , by using the optimal control resulting from the minimization problem showed in condition (C1), denoted by , we have .
Let be the minimum of and be the optimal control trajectory. The control strategy of MPC is that, at the sample time , is inputted into the real system, and at the sample time , the control inputted into the system is not but the first element of the optimal control trajectory resulting from the similar online optimization problem. At the sample time , the state is and the online optimization problem, denoted by , is the same as Eq. (2) except that is replaced by . Similarly, let be the minimum of and be the optimal control trajectory. The control inputted into the system at the sample time is . Thus, the control law of MPC can be stated as
The closed-loop stability of the controlled system is shown in Lemma 1.
Lemma 1 For any , if satisfies and Assumption 1 is satisfied, it is guaranteed that will be steered to0by using the control law of MPC.
Gradually approximation of the maximal terminal region
Using SVM classifier to estimate the terminal region, as described in Ref. [7], is considered as an effective approach for solving the approximating problem in MPC. However, the method in Ref. [7] is somewhat conservative. The reason is that, the obtained terminal region actually is the stabilizable region of using a predetermined linear feedback controller.
In this section, a novel method of computing a terminal region is proposed. Given the terminal cost function and an initial set of the maximal terminal region, both surrounding sequence and subsets sequence can be constructed by using one-step set expansion iteratively. When some conditions are satisfied, the iteration ends and the consequent set can be seen as the suboptimal estimation of the maximal terminal region.
Construction of surrounding set sequence
The proposed approach constructs the surrounding set sequence directly from conditions (C1) and (C2). The terminal region can be defined aswhere is the solution of the following optimization problem:
Remark 1 The definition of has two meanings: (I) the optimization problem (4) has feasible solution, that is to say, , s.t. ; (II) the minimum of the optimization problem satisfies that .
Remark 2 From the definition of , it is obvious that, the terminal region is essentially a positively invariant set of using the optimal control resulting from the optimization problem (4) when is given.
Remark 3 In Refs. [4,5,7], the linear feedback control is attached to the construction of , and is the stabilizable region of using the linear feedback controller. In Ref. [6], a saturated local control law was used. However, in this paper, there is no explicit control attached to the definition of . Thus, the requirement on is lower than that in Refs. [4-7] while it guarantees the stability of the controlled system.
From the definition of , it cannot be determined whether a state point belongs to . The difficulty arises that the acts as the constraint in the optimization problem (4). To solve this problem, the method of using one-step set expansion iteratively is adopted.
The iterative approximating process begins with an initial estimation satisfying , and from which, a more accurate estimation is achieved. Repeatedly, the surrounding set sequence ,,…,, which satisfies , can be estimated sequentially. And the convergence of this sequence is proved in previous section. The gradual approximation process can be depicted in Fig. 1, the specific of this approach can be found in the authors’ previous literature [12].
The initial surrounding set is determined by
The iterative construction of surrounding set sequence is enforced based on the following assumptions.
Assumption 2 A surrounding set of , denoted by and containing the origin, is initialized, where is defined as the largest terminal region.
Assumption 3 is a positively invariant set, that is to say, for any , , s.t. and .
According to the known , another surrounding set of , denoted by , can be constructed aswhere is the minimum of
As mentioned in Remark 1, the construction of contains two meanings: (I) for any , , s.t. ; (II) the minimum of Eq. (7) satisfies . The constructions of in sequel have the similar meanings. And the convergence of this sequence can be proved from the following theorem.
Theorem 1 If Assumption 2 and Assumption 3 are satisfied, the sequence , similarly defined as and , will converge to when j goes to infinity.
Proof This theorem is proved by contradiction.
Supposes there exists to guarantee that . Then, for , satisfieswhere . Moreover, it is contractive to the fact that is the maximal set, which satisfies conditions (C1) and (C2) as listed in Assumption 1.
And vice versa, supposes there exists to guarantee that . Then, there exists a certain integer that and , for , we havewhere . Because , then, , which leads to paradox.
End proof.
From the previous description, we can find that the surrounding approximating approach will probably contain some unstable point by mistake. Then, we naturally proposed another gradually extending-approximation approach, which will be detailedly illustrated in the following subsection.
Construction of subsets sequence
Given and from conditions (C1) and (C2), the terminal region can be defined aswhere is the minimum of the following optimization problem:
Remind here that Remark 1 and Remark 2 are also consistent in the problem (9). From the definition of , it cannot be determined whether a state point belongs to . The difficulty lies in that, the itself acts as the constraint in the optimization problem (9).
To avoid this problem, the method of using one-step set expansion iteratively is adopted. The gradual approximation process can be depicted in Fig. 2.
Assumption 4 A subset of , denoted by and containing the origin, is known, and is denoted as the largest terminal region.
Assumption 5 is a positively invariant set, that is to say, for any , , s.t. and .
Given , another subset of , denoted by , can be constructed aswhere is the minimum of
As mentioned in Remark 1, the construction of contains two meanings: (I) for any , , s.t. ; (II) the minimum of Eq. (11) satisfies . The constructions of in sequel have the similar meanings.
Lemma 2 If Assumption 3 is satisfied, there is .
Proof If Assumption 5 is satisfied, it is obvious that, for any , , s.t. and . It follows that, . From the construction of , we can know , namely, .
End proof.
Remark 4 From the construction of , it is obvious that, if Assumption 5 is satisfied, is a positively invariant set. We know that, for any , , s.t. and . Because of as shown in Lemma 2, we have .
Similarly, by replacing with in the constraint of Eq. (11), another subset, denoted by , can be obtained as follows:where is the minimum of
Repeatedly, , , can be constructed aswhere is the minimum of
This method of constructing given is defined as one-step set expansion in this paper. By employing it iteratively, a subsets sequence of largest terminal region, denoted by , , can be achieved.
Remark 5 Similar with Lemma 2 and Remark 3, any subset in this sequence is positively invariant and any two neighboring subsets satisfy .
As j increases, will converge to a set, denoted by . Theorem 2 will show that, is equal to the largest terminal region.
Theorem 2 If Assumption 4 and Assumption 5 are satisfied, for constructed in Eqs. (14) and (15), when goes to infinity, will converge to.
Proof This theorem is proved by contradiction.
(A) Assume that, there exists a set which is denoted by satisfying and when . From Remark 5, we can know . It is obvious that because of as shown in Assumption 4. It follows that, and for any , we have since is positive definite. Define as the minimum of , it is satisfied that, .
From the construction of , we know that, for any , there exists no such a satisfying and because of . However, from conditions (C1) and (C2), we know that, , s.t. and , where . It is obvious that, . Thus, we have . Similarly, we can know, , s.t. and , where , since .
Repeatly, for ,, s.t. and , where , . It is clear that, when . We know that, for the minimum of , defined as , there is a positive real number satisfying . Since when , , s.t. for any , we have . Obviously, this is contradicted with that is the minimum of .
(B) Similarly, assume that, there exists an satisfying and when . For any , we have that and . Obviously, this is contradicted with that is the largest one satisfying conditions (C1) and (C2).
End proof.
Remark 6 In this paper, the largest terminal region means the positively invariant set satisfying conditions (C1) and (C2). However, conditions (C1) and (C2) are sufficient conditions to guarantee the stability of the controlled system, not the necessary conditions. There may be a set larger than and the stability of the controlled system can be guaranteed by using this set as the terminal region.
Remark 7 In the calculation of , it is impossible to keep iteration computation until . When the iteration time goes to ( is a positive integer), if is equal to in principle, it can be deemed that converges to in rough. Hence, can be taken as the terminal region and it is a good approximation to .
Remark 8 If the iteration time does not go to infinity, the obtained set may be just a large positively invariant subset of . This has no effect on the stability of the controlled system. The only negative influence is that its corresponding domain of attraction is smaller than that corresponding to .
Until now, it seems that we can choose any in the subsets sequence as the terminal region. This is infeasible. Since is not described in explicit expression, it cannot serve as the terminal constraint in the optimization problem (2) directly. Then, an estimated one described in explicit expression is needed. Due to the strong optimizing ability of SVM, SVM is exploited to separate each from the state space.
Support vector machine
SVM is the youngest part in the statistical learning theory. It is an effective approach for pattern recognition. In SVM approach, the main aim is to obtain a function, which determines the decision boundary or hyperplane. This hyperplane optimally separates two classes of input data points.
Take the example of separating into and . For each , an additional variable is introduced. Similarly, for each , is introduced. Define and , SVM will find a separating hyperplane, denoted by , between and . Therefore, can be estimated as , where is determined by solving the following problem:where denotes the kernel function and the Gaussian kernel as follows is adopted in this paper:with being the positive Gaussian kernel width.
When are computed out, some support vectors are chosen from and the optimal hyperplane can be determined with these support vectors and their relevant weights. Denote as the number of support vectors and as the support vectors set, the optimal hyperplane is described aswhere is the support vector and satisfying is the relevant weight.
There are many software packages of SVM available on internet. They can be downloaded and used directly. To save space, it is not introduced in detail in this paper. For more details, please refer to Refs. [10] and [11].
Estimating the approximating sequence by employing SVM
From Subsection 3.3, we know that, SVM find a separating hyperplane between and . This hyperplane is used to separate into and . All of and their relevant compose a set, named the training points set. This subsection will show how to achieve the training points set when estimating and how to determine when the separating hyperplane is known.
First, choose arbitrary points , ( is the number of training points); then, assign to each by implementing the following procedure:
IF (I) The following optimization problem has feasible solution
(II) Its minimum satisfies
THEN
ELSE
END IF
By implementing this procedure for every , each is known. Input and into SVM classifier, an optimal hyperplane will be obtained. Therefore, the estimated set of can be achieved as .
When is known, the training points for separating from can be computed by the similar procedure. By inputting them into SVM classifier, a hyperplane and an estimated set of , denoted by will be obtained. Repeatedly, , and can be achieved by the similar technology.
Estimating the terminal region
Section 3 showed how to achieve the subsets sequence by employing SVM. Theoretically, the larger the iteration time is, the higher the precision of approaching to is. However, it is impossible to keep computation until . To avoid this problem, the iteration should be ended when some conditions are satisfied.
When , if it is satisfied that, for , , there existsand we can find that is equal to in principle and converges to . In Eq. (19), is the support vectors set at , is the number of support vectors and is a tunable threshold. The smaller is, the higher the precision of approximating to is. Finally, is used to serve as the terminal region.
Remark 9 Here, we used the information that, in SVM classifier, the hyperplanes are only determined on the support vectors.
Now, the concrete algorithm of estimating the largest terminal region is displayed as follows.
Step 1 Set the number of training points used in SVM and the tunable threshold .
Step 2 For , use SVM to achieve the optimal hyperplane and the estimated set of , denoted by .
Substep 1 Choose arbitrary points , .
Substep 2 Assign to each by implementing the procedure in Subsection 3.4.
Substep 3 Input into the SVM. An optimal hyperplane will be obtained and can be approximated by , where
with denoting the number of support vectors, being the support vector, denoting its relevant weight, and denoting the classifier threshold.
Step 3 Check the iteration status. When , if inequality (14) is satisfied, end iteration and take as the largest terminal region.
Remark 10 It is obvious that, is achieved one by one. Namely, can only be achieved when is known.
Simulation experiment
In this section, we implemented both surrounding and subset approximating approach in simulated model, and the comparison to representative alternatives are shown in results. The model is a discrete-time realization of the continuous-time system used in Refs. [4,7]:where , , and the state constraint and control constraint are , , respectively.
The stage cost is where and . The terminal cost is chosen as , where and is given as the terminal region in Ref. [4], which is.
To estimate each , 4000 training points are generated, and set . From the simulating process we find that in the surrounding approximating approach, when is iterated to 15, and are near enough. There only exists minor difference between and , then,where , is the support vectors set, and is the number of support vectors at . For , so we can choose as the terminal estimation of . The approximating process can be well identified in Fig. 3. On the other hand, if the subsets approximating approach is adopted, if is as large as 22, the following iterations various slowly enough. It can be rearranged aswhere , is the support vectors set, and is the number of support vectors at . Thus, can be seen as the final estimation of , and the approximating process is well depicted in Fig. 4.
In Figs. 3 and 4, the blue circles, which stem from the method that presented in Ref. [4], are selected as the initial set of surrounding and subset approaches, respectively. We can find that the converged sets in both approaches are nearly the same, as depicted by black solid lines, so in the following illustration, the estimation results of surrounding method is neglected for clarity.
Figure 5 depicts the convergent trajectory of some state points, which are selected to implement the comparison between representative alternatives and the proposed approach. These state points are , , , , , and , as denoted by “*” in Fig. 5.
In Fig. 5, the blue ellipsoid is the terminal region in Ref. [4] and the region encompassed by black dash lines is the result in Ref. [7]. The region encompassed by black solid lines is the terminal region in this paper. From this figure, we can find that all the selected states, which obviously lie in the domain of attraction, are out of the estimating regions of both Refs. [4] and [7], but they are all included in the corresponding results of this paper. The terminal region in this paper contains the solution that obtained in Ref. [4], but not entirely contains the result in Ref. [7] although it is much larger than that in Ref. [7]. The reason is that, the terminal region in this paper is the largest one satisfying conditions (C1) and (C2). However, conditions (C1) and (C2) are just the sufficient conditions to guarantee the stability of the controlled system, not the necessary conditions as shown in Remark 9. The red solid lines denote the closed-loop trajectories of the selected points. Note that, with the same sampling interval and prediction horizon as those in this paper, these points are not in the regions of attraction of MPC in Refs. [4] and [7]. However, they can be leaded to the origin by using the control law of MPC in this paper.
Conclusion
Given the terminal cost, a sequence of surrounding sets and subsets of the maximal terminal region is extracted from state space iteratively by employing SVM classifier. The convergence of these approximation processes are proved in analytical way. Simulations also show that the convergent region is better approximation to the maximal terminal region, when compared to some mature representative alternatives.
Mayne D Q, Rawlings J B, Rao C V, Scokaert P O M. Constrained model predictive control: Stability and optimality. Automatica, 2000, 36(6): 789-814
[2]
González A H, Odloak D. Enlarging the domain of attraction of stable MPC controllers, maintaining the output performance. Automatica, 2009, 45(4): 1080-1085
[3]
Magni L, De Nicolao G, Magnani L, Scattolini R. A stabilizing model based predictive control algorithm for nonlinear systems. Automatica, 2001, 37(9): 1351-1362
[4]
Chen H, Allgower F. A quasi-infinite horizon nonlinear model predictive control scheme with guaranteed stability. Automatica, 1998, 34(10): 1205-1217
[5]
Cannon M, Deshmukh V, Kouvaritakis B. Nonlinear model predictive control with polytopic invariant sets. Automatica, 2003, 39(8): 1487-1494
[6]
De Doná J A, Seron M M, Mayne D Q, Goodwin G C. Enlarged terminal sets guaranteeing stability of receding horizon control. Systems & Control Letters, 2002, 47(1): 57-63
[7]
Ong C J, Sui D, Gilbert E G. Enlarging the terminal region of nonlinear model predictive control using the support vector machine method. Automatica, 2006, 42(6): 1011-1016
[8]
Limon D, Alamo T, Salas F, Camacho E F. On the stability of constrained MPC without terminal constraint. IEEE Transactions on Automatic Control, 2006, 51(5): 832-836
[9]
Limon D, Alamo T, Camacho E F. Enlarging the domain of attraction of MPC controllers. Automatica, 2005, 41(4): 629-635
[10]
Burges C C. A tutorial on support vector machines for pattern recognition. Data Mining and Knowledge Discovery, 1998, 2(2): 121-167
[11]
Vapnik V. The Nature of Statistical Learning Theory. New York: Springer, 1995
[12]
Wang Y F, Sun F C, Zhang Y A, Liu H P, Min H B. Getting a suitable terminal cost and maximizing the terminal region for MPC. Mathematical Problems in Engineering, 2010, 2010: 1-16
RIGHTS & PERMISSIONS
Higher Education Press and Springer-Verlag Berlin Heidelberg
AI Summary 中Eng×
Note: Please be aware that the following content is generated by artificial intelligence. This website is not responsible for any consequences arising from the use of this content.