INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL ISSN 1841-9836, 13(6), 927-937, December 2018. Model Predictive Control of Stochastic Linear Systems with Probability Constraints C.F. Caruntu, C.C. Velandia-Cardenas, X. Liu, A.N. Vargas Constantin F. Caruntu Department of Automatic Control and Applied Informatics, Gheorghe Asachi Technical University of Iasi, Str. Prof. D. Mangeron no. 27, Iasi, Romania. Cristian C. Velandia-Cardenas Research Group MEM, Faculty of Electronic Engineering, Universidad Santo Tomás, Cra. 9 no. 51-11, Bogotá, Colombia. Xinghua Liu Department of Electrical Engineering, School of Automation and Information Engineering, Xi’an University of Technology, Xian 710048, China. Alessandro N. Vargas* Universidade Tecnológica Federal do Paraná, UTFPR, Av. Alberto Carazzai 1640, 86300-000 Cornelio Procópio-PR, Brazil. *Corresponding author: avargas@utfpr.edu.br Abstract: This paper presents a strategy for computing model predictive control of linear Gaussian noise systems with probability constraints. As usual, constraints are taken on the system state and control input. The novelty relies on setting bounds on the underlying cumulative probability distribution, and showing that the model pre- dictive control can be computed in an efficient manner through these novel bounds— an application confirms this assertion. Indeed real-time experiments were carried out to control a direct current (DC) motor. The corresponding data show the effectiveness and usefulness of the approach. Keywords: probability constraints, stochastic systems, linear systems, control. 1 Introduction Engineering applications show that constraints are not always met with probability one [4]. It is then reasonable to account constraints, together with performance indicators, under certain probability levels, e.g. [9]—this setup is referred to as model predictive control (MPC) with probability constraints, as detailed next. Probability constraints on linear systems constitute a research topic that has attracted at- tention in recent years, in particular when it is combined with the MPC method, see for in- stance [2, 5, 10, 15, 20, 25], just to name a few. In this scenario, the model accounts additive disturbances [5, 15, 20, 25], multiplicative disturbances [2], and systems with mixed features; namely, systems with known: 1. probability distribution of the system output [15,20]; 2. Gaussian distribution and general probability distribution of the disturbance [5]; 3. observable probability distribution of the disturbance at each time step [2]; 4. covariance and mean of the inputs and outputs [10]. Copyright ©2018 CC BY-NC 928 C.F. Caruntu, C.C. Velandia-Cardenas, X. Liu, A.N. Vargas In particular, probabilistic constraints on inputs [5, 10, 11], states [2, 5, 10, 11, 25], and outputs [15,20] were transformed into deterministic constraints on the process input. To handle probability constraints is a task that the literature suggests the evaluation of element-wise probability constraints [2, 10, 11, 15, 16, 20, 25]; here, we follow the literature with a minor adaptation, that is, the element-wise evaluation yields nonlinear equations, difficult to compute. To overcome this difficulty, we propose an upper bound for the underlying nonlinear equa- tions. As a byproduct, we obtain linear matrix inequalities—computing linear matrix inequalities increases efficiency on solving optimization problems [4, Ch. 1]. The main contribution of this paper is to convert nonlinear equations, drawn from probability constraints, into linear matrix inequalities useful for real-time implementation. This assertion is illustrated in practice through the MPC applied in a direct current (DC) motor—experimental data are shown in Section 4. These findings illustrate the practical benefits of our approach. The paper is organized as follows. Section 2 states the problem to be solved and the derived solution. Section 4 provides a performance evaluation of the proposed MPC with probabil- ity constraints based on the experimental results obtained on a direct current motor. Finally, concluding remarks are given in Section 5. Notation and basic definitions. R, R+, Z and Z+ are the real, non-negative real, integer and non-negative integer numbers. Z≥c1 and Z(c1,c2] denote the sets {k ∈ Z+ | k ≥ c1} and {k ∈ Z+ | c1 < k ≤ c2}, respectively, for some c1,c2 ∈ Z+. || · ||∞ denotes the infinity norm; given x ∈ Rn and c ∈ R, by definition of the infinity norm, for ||x||∞ ≤ c to be satisfied, it is necessary and sufficient that ±[x]j ≤ c for all j ∈ Z[1,n]. In is the identity matrix of order n and M ′ denotes the transpose of M. P [A] denotes the probability of event A and p(x) denotes the probability distribution function of random variable x. x̄ denotes the mean of the random variable x and Sx denotes its covariance. 2 Problem statement For some fixed and filtered probability space (Ω,F,P), consider the following discrete-time linear time invariant system xk+1 = Axk + Buk + Hwk, k ∈ Z+, (1) where xk denotes an n-dimensional system state, uk represents an m-dimensional control input, and wk represents a q-dimensional noise or disturbance driven by an independent and identically distributed (i.i.d.) Gaussian process. It is assumed that the initial state x0 is a Gaussian random variable uncorrelated with wk. Although it has been intensively studied over the last decades, (1) presents so far some interesting challenges. For instance, controlling (1) under probabilistic constraints remains an open problem [9]. The main contribution of this paper lies in showing a condition for controlling the system (1) under probabilistic constraints. We revisit some results from the literature, and show how to convert them into linear matrix inequalities, a useful scheme for numerical evaluations. To turn our contribution clear, we present some notation borrowed from [3]. Stacking the Model Predictive Control of Stochastic Linear Systems with Probability Constraints 929 elements of (1) for N steps in the form X =   x1 x2 ... xN   , U =   u0 u1 ... uN−1   , W =   w1 w2 ... wN   , (2) we obtain the Nn-dimensional vector representation X = Gxxx0 + GxuU + GxwW. (3) where the matrices Gxx, Gxu and Gxw are fixed, and their entries depend only on A, B, and H. Here, N is called prediction horizon. Note that X in (3) is a multivariate normal random vector, a direct consequence of the i.i.d. Gaussian assumption on W. Set X̄ as the mean value of X and let SX be its covariance, both given by [3] X̄ = Gxxx̄0 + GxuU + GxwW̄ SX = GxxSx0G ′ xx + GxwSWG ′ xw. (4) The probabilistic constraints to be imposed on (1) are defined as follows. Let FX be the set made up by the intersection of M linear inequality constraints, i.e., FX := M⋂ i=1 {X : a′iX ≤ bi}, (5) where ai and bi, i = 1, . . . ,M, represent vectors on RnN and R, respectively. Moreover, consider FU as a set defined in similar fashion. Assume for the moment that a given functional J : FX ×FU 7→ R+, convex with respect to its arguments, is associated with (1). Hereafter, this functional is referred to as cost function. Problem 1. Given some δ ∈ R(0,1), find a solution for the next optimization problem. min U∈FU J(X̄,U) s.t. Pr[X /∈ FX] ≤ δ, X = Gxxx0 + GxuU + GxwW (6) As a matter of fact, Problem 1 means that the control input minimizes the cost function while ensuring that the system state may leave the feasible region with a probability less or equal than δ. To the best of the authors’ knowledge, Problem 1 remains unresolved. The difficulty to solve such a problem stems from the fact that the probability in (6) represents a value taken over by multivariate integrals, requiring their evaluation, a difficult task to accomplish [12,17]. To avoid the evaluation of multivariate integrals, the authors of [3] suggest the element-wise evaluation of the set FX in probability; namely, find constants εi’s such that ∑M i=1 εi ≤ δ, and Pr [ a′iX > bi ] ≤ εi, ∀i ∈ Z[1,M]. (7) Note that εi’s are slack variables, now in association with a univariate probabilistic inequality, which can be computed via the standard cumulative distribution function. Even so, despite the progress researchers have made so far, the problem still remains difficult to solve because checking (7) may be cumbersome from the numerical point of view, as detailed next. 930 C.F. Caruntu, C.C. Velandia-Cardenas, X. Liu, A.N. Vargas 0 0.2 0.4 0.6 0.8 1 −3 −2 −1 0 1 2 3 x y Figure 1: The inverse cumulative distribution function (thick line) and upper linear approxima- tions (straight lines tangent to the icdf function). Let cdf(·) be the standard Gaussian cumulative distribution function cdf(γ) = 1 √ 2π ∫ γ −∞ e− z2 2 dz, (8) and let icdf(·) be the corresponding inverse. One can show that (7) holds if and only if both inequalities ∑M i=1 εi ≤ δ and bi −a′iX̄ si ≥ icdf (1 −εi) , ∀i ∈ Z[1,M], (9) hold true, where s1, . . . ,sM, are fixed, positive constants. Even though the inequalities in (9) are convex with respect to εi’s, as shown in [3], finding a solution for (9) may be cumbersome—there is no closed, algebraic expression for computing a feasible solution for (9). Consequently, to evaluate (9), it is necessary to use iterative algorithms, usually slow for computing the point of convergence [1, Ch. 7]. But slow algorithms cannot be applied in some online control schemes, such as model predictive control [6,7,14,19,24,26]. In order to speed up the computation of a feasible solution for (9), we convert (9) into an affine function so as to obtain linear matrix inequalities (LMI)—it is known that LMI solvers outperform classical convex optimization algorithms [4, Ch. 1]. The idea behind such conversion is now detailed. As described in Fig. 1, the icdf function can be bounded from above by an affine function on some interval. This fact allows us to define an affine function h(·) such that h(1 −εi) ≥ icdf (1 −εi) , ∀i ∈ Z[1,M]. (10) The novel condition hinges on checking whether the inequalities bi −a′iX̄ si ≥ h (1 −εi) , ∀i ∈ Z[1,M], (11) hold true. In the positive case (9) is assured; as a byproduct, we obtain U that suffices to (6) in Problem 1, the main condition pursued in this paper. The main contribution of this paper is capable of showing (11). Yet derived from a simple idea of imposing upper bounds on the icdf function, the condition in (11) represents linear matrix inequalities, useful for real-time experiments—the paper shows a real-time application for a direct current motor (see Section 4). This application illustrates the practical benefits of our approach. Model Predictive Control of Stochastic Linear Systems with Probability Constraints 931 3 Model predictive control with probability constraints The problem we solve in this paper is presented in what follows. Define yi = a′ix, and note that yi is a univariate Gaussian random variable with mean and variance given by ȳi = a ′ iGxxx̄0 + a ′ iGxuU + a ′ iGxwW̄ Syi = a ′ iGxxSx0G ′ xxai + a ′ iGxwSWG ′ xwai. (12) Using the novel condition in (11), we recast the Problem 1 as the next problem. Problem 2. Given some δ ∈ R(0,1), find the solution to the optimization problem min U∈FU J(X̄,U) s.t. M∑ i=1 εi ≤ δ, bi − ȳi√ Syi ≥ h (1 −εi) , ȳi = a ′ iGxxx̄0 + a ′ iGxuU + a ′ iGxwW̄, Syi = a ′ iGxxSx0G ′ xxai + a ′ iGxwSWG ′ xwai, ∀i ∈ Z[1,M]. (13) The next result is a direct consequence of the aforementioned facts. Theorem 3. A feasible solution to Problem 2 is a feasible solution to Problem 1. Proof: Recall that Pr[X 6∈ FX] = Pr [ X 6∈ M⋂ i=1 {X : a′iX ≤ bi} ] = Pr [ X ∈ M⋃ i=1 {X : a′iX > bi} ] . (14) The rightmost term in (14) is bounded from above by ∑M i=1 Pr[a ′ iX > bi] due to the Boole’s inequality. When Problem 2 has a solution, we have from (10) and (13) that bi − ȳi√ Syi ≥ h (1 −εi) ≥ icdf (1 −εi) , ∀i ∈ Z[1,M], which shows (9). Both (9) and ∑M i=1 εi ≤ δ assure Pr[a ′ iX > bi] ≤ εi for each i = 1, . . . ,M from (7), which in turn assures that Pr[X 6∈ FX] ≤ δ. This argument completes the proof. 2 Remark 4. Theorem 3 assures that any optimal solution to Problem 2 is a sub-optimal solution to Problem 1—it means that the optimality of Problem 1 is lost; however, the sub-optimal strategy turns the problem tractable from the numerical viewpoint. Theorem 3 then contributes towards finding a solution to the so-far unsolved Problem 1. Recall that the model predictive control (MPC) works under the receding horizon control principle [14, Ch. 1]; namely, the cost function J(·) in Problem 2 should be minimized at each time instant k = k0, . . . ,k1. The current input uk in (1) is obtained by determining the input sequence {ûk, . . . , ûk+N−1} that solves Problem 2, and next by setting uk = ûk. The remaining elements of the sequence are discarded, and this iterative procedure is repeated [14, Ch. 1]. The aforementioned iterative procedure, key for the MPC scheme, was implemented in a laboratory testbed as detailed in the next section. 932 C.F. Caruntu, C.C. Velandia-Cardenas, X. Liu, A.N. Vargas Figure 2: Equipment overview. 4 Application of model predictive control to a direct current mo- tor subject to probability constraints The goal of this section is to illustrate the proposed model predictive control scheme with probability constraints. To do so, we present data collected from real-time experiments—the experiments were carried out in a laboratory tested designed to control a direct current (DC) motor. 4.1 Identification The laboratory testbed was assembled with “Modulo 2208 servo-mecanismo”, a DC motor device assembled by Datapool-Brazil [18,21], which worked in association with a data-acquisition board (model USB-6008 DAQ) from National Instruments (see Fig. 2). The board was configured to work with fixed sampling time of 20 ms. The DC motor was equipped with a tachometer—it generated a voltage in the range [0, 5] V for an angular velocity in the range of [0, 26.17] rad/s, in a linear relationship. Likewise, a current sensor generated the current consumed by the motor in the range [0, 5] A. Pseudo random binary signals (PRBS) with amplitudes between 0 and 5 V were generated and applied in a sixth-order Butterworth low-pass filter with a cutoff frequency randomly chosen between 0.01Hz and 70Hz. The resulting signal was then applied in the input of the DC motor, and the corresponding data was collected and stored. Each experiment for sake of identification was performed in chunks of 60s, that is, each chunk contained three thousand points. The data were then split into smaller chunks, each with 1300 points. These chunks were used to obtain a model in the observability canonical form through the subspace identification method [8,13]. The canonical model was then modified to cope with the prediction-error minimization algorithm. Finally, the obtained model was validated through unused, uncorrelated data (see Fig. 3 for illustration of a sample). This procedure produced the model (≈ 85% fitting): [ ωk+1 ik+1 ] = [ 0.9814 0.4582 −0.0087 0.7667 ][ ωk ik ] + [ 0.0064 0.0207 ] (uk + dk), (15) Model Predictive Control of Stochastic Linear Systems with Probability Constraints 933 0 1 2 3 4 5 6 7 8 9 0 5 q 0 1 2 3 4 5 6 7 8 9 0 5 10 15 20 25 z Real Simulation 0 1 2 3 4 5 6 7 8 9 0 1 2 y x Figure 3: Real and simulated data: experiment suggests that the model attained for the DC motor seem appropriate for representing the real-time equipment. −2.5 −2 −1.5 −1 −0.5 0 0.5 1 1.5 2 2.5 0 1 2 3 4 5 x 10 4 x y Figure 4: Histogram of disturbances. Dashed curve shows the Gaussian distribution. where ω, i, u, and d represent the angular velocity of the shaft, the electrical current consumed by the motor, the armature voltage applied in the terminals of the DC motor, and disturbance, respectively. The disturbance d stands for an i.i.d. stochastic process under Gaussian probability distribution— the corresponding experimental data is summarized in Fig. 4. 4.2 Probability constraints Armature voltage (control signal) that drove the motor was restricted by lower and upper bounds as umin ≤ uk ≤ umax, (16) where umin and umax are the minimum and maximum voltages. The rotor angular velocity and the armature current were bounded as well, i.e., ωmin ≤ ωk ≤ ωmax, imin ≤ ik, and Pr[ik ≥ imax] ≤ δ, (17) where ωmin, ωmax, imin, and imax denote the corresponding limits. The control objective was to reach a desired value of the rotor angular velocity, say ωss, while fulfilling the constraints given in (16) and (17). 934 C.F. Caruntu, C.C. Velandia-Cardenas, X. Liu, A.N. Vargas 0 0.2 0.4 0.6 0.8 1 −3 −2 −1 0 1 2 3 x y Figure 5: Inverse cumulative distribution function and its upper linear approximation (dashed line). 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0 5 10 15 x y a Reference 0.95 0.955 0.96 0.965 0.97 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 −0.1 0 0.1 0.2 0.3 x z b 0 2 4 6 8 10 12 14 −0.1 0 0.1 0.2 0.3 y z c 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0 1 2 3 4 5 x q d Figure 6: Data from real-time experiments. Problem 2 was evaluated for each δ = 0.95, 0.955, 0.96, 0.965, 0.97, and every curve shows the mean of one-hundred sample paths. The curves show (a) angular velocity of the DC motor; (b) armature current of the DC motor; (c) phase-portrait of the system; and (d) armature voltage of the DC motor. 4.3 Experimental results Recall that Problem 2 requires a linear approximation for the inverse cumulative distribu- tion function. To fulfill this requirement, the linear approximation h(1 − εi) = 2.38 − 4εi was considered, valid for εi ∈ [0.01, 1) (see Fig. 5). Associated with Problem 2, we consider the cost J(X̄,U) = 1000||X̄||∞ + 2||U||∞. (18) Problem 2 was evaluated with the following additional parameters: N = 2, Sx0 = I2 and SW = I2. The limits for the constraints in (16) and (17) were set as umin = 0 V, umax = 5 V, ωmin = 0 rad/s, ωmax = 26.17 rad/s, imin = 0 A and imax = 0.38 A. The reference for the angular velocity was fixed at ωss = 13 rad/s. Problem 2 was evaluated with δ taking values 0.95, 0.955, 0.96, 0.965, and 0.97, and every corresponding solution was checked in practice. Indeed, one-hundred experiments were conducted for each value of δ—the experimental outcome is subsumed under the curves shown Fig. 6. Model Predictive Control of Stochastic Linear Systems with Probability Constraints 935 The curves in Fig. 6 show the statistical mean of the stored one-hundred samples. Fig. 6(a) shows the angular velocity of the DC motor; it can be observed that as δ decreases, settling time increases. Armature current is represented in Fig. 6(b), which shows that the upper bound on the current is satisfied and the maximum value of the current decreases when the value of δ decreases. Phase portrait of the process is illustrated in Fig. 6(c)—data suggest that the DC model be second moment stable [22]. The curves of the armature voltage (control input) in Fig. 6(d) suggest that as long as δ decreases, the upper bound on the armature voltage also decreases; as a result, δ successfully actuated as a constraint on the armature voltage. In summary, the experimental data indicate that our approach seems to be useful to handle real-time applications subject to physical constraints. 5 Conclusion The paper presents an easy-to-check method to handle linear stochastic systems subject to probability constraints. Here, constraints hinge upon Gaussian distributions. However, nonlinear equations arise naturally from that Gaussian-based evaluations, turning the problem difficult to solve. To overcome such a drawback, we have converted those nonlinear equations into simple-to- check linear matrix inequalities (conversion guarantees only sub-optimal solutions, see Remark 4). The derived conversion shows appropriate for real-time applications—specially when fast algorithms are necessary—as illustrated in the paper. Acknowledgments This work was partially supported by the Thousand Talents Plan of Shaanxi Province for Young Professionals, and the Brazilian agencies FAPESP Grants 03/06736-7; CNPq Grant 304856/2007-0; and CAPES Grant Program PVE 88881.030423/2013-01. Bibliography [1] Bazaraa, M.S.; Sherali, H.D.; Shetty, C.M. (2006). Nonlinear programming: theory and algorithms, 3rd edn., Wiley-Interscience, New Jersey, 2006. [2] Bernardini, D.; Bemporad, A. (2012). Stabilizing model predictive control of stochastic constrained linear systems, IEEE Trans. Autom. Control, 57, 1468–1480, 2012. [3] Blackmore, L.; Ono, M. (2009). Convex chance constrained predictive control without sam- pling, In: AIAA Guidance, Navigation and Control Conference, Chicago, Illinois, USA, 1-17, 2009. [4] Boyd, S.; El Ghaoui, L.; Feron, E.; Balakrishnan, V. (1994). Linear matrix inequalities in system and control theory, SIAM, Philadelphia, 1994. [5] Cannon, M.; Kouvaritakis, B.; Rakovic, S.; Cheng Q. (2011). Stochastic tubes in model predictive control with probabilistic constraints, IEEE Trans. Autom. Control, 56, 194–200, 2011. [6] Cao, G.; Lai, E.M.-K.; Alam, F. (2017). Gaussian process model predictive control of un- known non-linear systems, IET Control Theory Appl., 11, 703–713, 2017. 936 C.F. Caruntu, C.C. Velandia-Cardenas, X. Liu, A.N. Vargas [7] Caruntu, C.F.; Balau, A.E.; Lazar, M.; van den Bosch, P.P.J.; Di Cairano, S. (2016). Driveline oscillations damping: A tractable predictive control solution based on a piecewise affine model, Nonlinear Analysis: Hybrid Systems, 19, 168–185, 2016. [8] Costa Junior, A.G.; Riul J.A.; Montenegro, P.H.M. (2016). Application of the subspace identification method using the N4SID technique for a robotic manipulator, IEEE Latin America Transactions, 14, 1588–1993, 2016. [9] Farina, M.; Giulioni, L.; Scattolini, R. (2016). Stochastic linear model predictive control with chance constraints-A review, Journal of Process Control, 44, 53–67, 2016. [10] Farina, M.; Giulioni, L.; Magni, L.; Scattolini, R. (2015). An approach to output-feedback MPC of stochastic linear discrete-time systems, Automatica, 55, 140–149, 2015. [11] Farina, M.; Scattolini, R. (2016). Model predictive control of linear systems with multiplica- tive unbounded uncertainty and chance constraints, Automatica, 70, 258–265, 2016. [12] Hashorva, E.; Hüsler, J. (2003). On multivariate Gaussian tails, Annals of the Institute of Statistical Mathematics, 55, 507–522, 2003. [13] Katayama, T. (2005). Subspace methods for system identification, communications and con- trol engineering, Springer-Verlag, London, 2005. [14] Kwon, W.H., Han, S.H. (2005). Receding horizon control: model predictive control for state models, Springer-Verlag, New York, 2005. [15] Li, P.; Wendt, M.; Wozny, G. (2002). A probabilistically constrained model predictive con- troller, Automatica, 38, 1171–1176, 2002. [16] Li, J.W.; Li, D.W.; Xi, Y.G. (2017). H∞ predictive control with probability constraints for linear stochastic systems, IET Control Theory Appl., 11, 557–566, 2017. [17] Lu, D.; Li, W.V. (2009). A note on multivariate Gaussian estimates, Journal of Mathematical Analysis and Applications,354, 704–707, 2009. [18] Oliveira, R.C.L.; Vargas, A.N.; do Val, J.B.R.; Peres, P.L.D. (2014). Mode-independent H2-control of a DC motor modeled as a Markov jump linear system, IEEE Transactions on Control Systems Technology, 22, 1915–1919, 2014. [19] Rubagotti, M.; Patrinos, P.; Guiggiani, A.; Bemporad, A. (2016). Real-time model predictive control based on dual gradient projection: theory and fixed-point FPGA implementation, Int. J. Robust Nonlinear Control, 26, 3292–3310, 2016. [20] Schwarm, A.T.; Nikolaou, M. (1999). Chance-constrained model predictive control, AIChE Journal, 45, 1743–1752, 1999. [21] Vargas, A.N.; Costa, E.F.; do Val, J.B.R. (2013). On the control of Markov jump linear sys- tems with no mode observation: application to a DC motor device, Int. J. Robust Nonlinear Control, 23, 1136–1950, 2013. [22] Vargas, A. N.; do Val, J.B.R. (2010). Average cost and stability of time-varying linear systems, IEEE Trans. Autom. Control, 55, 714–720, 2010. [23] Wang, S.; Yu, M.; Sun, X. (2015). Robust H∞ control for time-delay networked control systems with probability constraints, IET Control Theory Appl., 9, 482–2489, 2015. Model Predictive Control of Stochastic Linear Systems with Probability Constraints 937 [24] Yang, H.; Guo, M.C.; Xia, Y.; Cheng, L. (2018). Trajectory tracking for wheeled mobile robots via model predictive control with softening constraints, IET Control Theory Appl., 12, 206–214, 2018. [25] Yan, J.; Bitmead, R.R. (2005). Incorporating state estimation into model predictive control and its application to network traffic control, Automatica, 41, 595–604, 2005. [26] Zeilinger, M.N.; Raimondo, D.M.; Domahidi, A.; Morari, M.; Jones, C.N. (2014). Flocking of multi-agents with a virtual leader, Automatica, 50, 683–694, 2014.