Next Article in Journal
Fronthaul Design for Wireless Networks
Next Article in Special Issue
Multi-Channel and Multi-Agent Signal Processing
Previous Article in Journal
Measurement Uncertainty Analysis of a Stitching Linear-Scan Method for the Evaluation of Roundness of Small Cylinders
Previous Article in Special Issue
An Enhanced Precoder for Multi User Multiple-Input Multiple-Output Downlink Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Dropout Compensation ILC Method for Formation Tracking of Heterogeneous Multi-Agent Systems with Loss of Multiple Communication Packets

Shanghai Institute for Advanced Communication and Data Science, Key Laboratory of Specialty Fiber Optics and Optical Access Networks, Joint International Research Laboratory of Specialty Fiber Optics and Advanced Communication, Shanghai University, Shanghai 200444, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(14), 4752; https://doi.org/10.3390/app10144752
Submission received: 22 April 2020 / Revised: 26 June 2020 / Accepted: 7 July 2020 / Published: 10 July 2020
(This article belongs to the Special Issue Multi-Channel and Multi-Agent Signal Processing)

Abstract

:
In this paper, the formation tracking problem for heterogeneous multi-agent systems with loss of multiple communication packets is considered using the iterative learning control (ILC) method. A dropout compensation ILC method is presented to construct effective distributed iterative learning protocols. The convergence conditions are given based on the frequency-domain analysis by using the general Nyquist stability criterion and Greshgorin’s disk theorem. The results show that the multi-agent system with different packet loss rate can achieve formation tracking without reducing the convergence speed. Numerical simulation results show the effectiveness of the proposed dropout compensation ILC method.

1. Introduction

In recent years, with the rapid development of communication and computing technology, people pay more and more attention to the control of multi-agent systems [1,2,3]. “Agent” can be broadly defined as a computing entity that is adaptive, autonomous, and has a certain life cycle in a particular environment. Drones, driverless cars, robots, etc., can all be called “Agent”. Apart from these, single-agent has played an important role in aerospace, military, and real life. Furthermore, the development of intelligent control, modern communication technologies, artificial intelligence concepts, etc., as well as people inspired by the phenomenon of animal cluster activity in nature, have pushed agent control research into a new stage of development.
By working together, multi-agent can accomplish some parallel, complex tasks, more efficiently than single-agent. For instance, multiple satellites orbiting the Earth to complete designated tasks and multiple drones can work together to transport goods in disaster relief.
In recent years, many researchers have considered applying iterative learning control (ILC) to multi-agent systems with repetitive work. ILC is a tracking control method for systems that require high precision and repeat the same actions. In a study by Ahn and Chen [4], ILC was applied to the formation control of a multi-agent system for the first time. Since then, more and more researchers have applied ILC to the practical application of multi-agent system. Ahn, Moore, and Chen [5] proposed an ILC law to maintain the flight path of satellite formation. Since the satellite rotates the earth periodically, the position related interference can be considered as the time periodic interference. According to Chen and Jia [6], multiple robots can guarantee formation using iterative learning. Furthermore, in order to ensure the safe running interval of trains, Sun, Hou, and Li [7] studied the corresponding ILC scheme of multi train cooperation. In studies by Hock and Schoellig [8] and Schoellig and colleagues [9], a group of four rotor aircrafts kept a given formation and accurately tracked the desired trajectory with ILC.
However, considering the limited capacity of the communication channel, the actual control design based on wireless communication is difficult. In addition, various industrial production processes and equipment are very complex. For these reasons, communication delay, packet loss, quantization and so on will cause the ILC of multi-agent to not converge [10,11]. There are specific researches (see [12,13,14,15,16,17]) on communication delay and quantization, but few on communication packet loss. Zhang and Li [18] designed an asynchronous event triggering protocol to solve the consistency problem of multi-agent system in the case of external interference, parameter uncertainty, time-delay, and packet loss by ΣΔ quantizer. By using the extended Lyapunov–Krasovskii functional, a consistency criterion in the form of matrix inequality was derived to ensure the consistency of multi-agent system with delay and packet loss under H1 controller. Firstly, Zhang and Li [19] considered the packet loss problem in the quantitative ILC of a multi-agent system. Considering the nonlinearity of quantization and the randomness of packet loss, a method of reducing learning gain to obtain accurate tracking was proposed. However, the results showed that the convergence rate slows down with the increase of the packet loss rate. All the results from their study [19] are for homogeneous multi-agent systems and assume that multiple agents have the same packet loss rate. The packet loss compensation for heterogeneous multi-agent systems with different packet loss rates is a more challenging problem.
This paper explores the ILC method with packet loss compensation for a heterogeneous multi-agent system. The question of under what conditions, with the increase of packet loss rate, the convergence speed of iterative learning algorithm will not slow down is solved. Firstly, the ILC model of a heterogeneous multi-agent system is established, and then the frequency domain model is obtained from the time domain model. When different communication packets are lost between agents, the previous iterative data is used to form the current iterative data. ILC in frequency domain is a two-dimensional system in which information propagates on two independent axes: frequency axis and iterative axis. The convergence of ILC is reduced to the convergence of any frequency in the iterative dimension, which is realized by the transformation of frequency and parameter iteration. After the exchange of variables and parameters, Z-transform is performed in the iterative domain. The convergence of the system error is analyzed by using the eigenvalue of the characteristic equation. Since the characteristic equation is in matrix form, it is difficult to get the eigenvalue of the characteristic equation. We use the generalized Nyquist criterion and Gershgorin’s disk theorem to analyze the eigenvalue range of the system and obtain the condition of the system convergence. The simulation results show that the compensation method proposed in this paper can make the system error convergence, and the convergence speed will not slow down.
This paper consists of the following parts. In Section 2, an ILC model in frequency domain is established for heterogeneous multi-agent systems and the control objective for formation tracking of a multi-agent system is given. A dropout compensation ILC method is proposed in Section 3. In Section 4, the convergence conditions are analyzed. The results are verified by numerical simulation in Section 5.

2. Problem Formulation and Preliminaries

2.1. Preliminaries

The interaction topologies in a multi-agent system are modeled by an undirected graph G = ( V , E , A ), where V = { v 1 ,   ,   v n } is the set of vertices, E V × V is the set of the edge, and A = [ a i j ] ,   a i j 0 is the weighted adjacency matrix of the graph G . If ( v i ,   v j ) E , a i j = a j i > 0 , otherwise, a i j = a j i = 0 . Each edge ( v i ,   v j ) represents information that can be transmitted between the i th agent and the j th agent. In addition, we assume that a i i = 0 . The neighboring set of agent, v i , is N i = { j :   ( v i ,   v j ) E } . The Laplacian matrix L = [ l i j ] n × n of graph G is denoted as L = D A , where D diag { d 1 ,   ,   d n } with d i = j = 1 n a i j . A path in the undirected graph G is denoted as a finite sequence { v i 1 , v i 2 , , v i j } and ( v i l , v i l + 1 ) E , l = 1 , 2 , , j 1 .

2.2. Problem Formulation

2.2.1. Control Objective

Consider the heterogeneous linear SISO (Single Input and Single Output) discrete-time multi-agent system of n agents with topology graph G. The dynamic of the i th agent is modeled by the following dynamic system.
{ x i , k ( t + 1 ) = A i x i , k ( t ) + B i u i , k ( t ) y i , k ( t ) = C i x i , k ( t )  
where i = 1 ,   2 ,   ,   n is the index of the i th agent, t = 0 ,   1 ,   2 ,   ,   T is the discrete-time index, and k = 1 ,   2 , is the index of iteration. x i , k ( t ) is the state of the i th agent in time t and k th iteration. Similarly,   u i , k ( t ) is the control input signal and y i , k ( t ) is the output signal. Multiple agents have different dynamics which are denoted as A i , B i , and C i . Taking the Z-transformation of Equation (1), we get
Y i , k ( z ) = C i [ B i U i , k ( z ) + x i , k ( 0 ) ] z A i
where z = e j ω ,   ω [ π , π ) with j = 1 and ω is frequency, and x i , k ( 0 ) is the initial state of i th agent in the k th iteration; let G i ( z ) = C i B i z A i , γ i , k ( z ) = C i z A i x i , k ( 0 ) , then (2) can be rewritten as
Y i , k ( z ) = G i ( z ) U i , k ( z ) + γ i , k ( z )
The desired formation is defined as the relative state values between agents. The relative state values could be position, velocity, and so on. The control objective for the multi-agent systems in this paper is to find an appropriate control input sequence u i , k ( t ) ,   i = 1 ,   2 ,   ,   n ,   t = 0 ,   1 ,   2 ,   ,   T such that every agent can move in the desired formation tracking trajectory accurately in the sense of expectation as the iterative number k tends to infinity. That is, lim k y i , k ( t ) = y i d ( t ) . y i d ( t ) is the desired output of the i th agent and denoted as
y i d ( t ) = r ( t ) + d i ( t )
where r ( t ) , t = 0 ,   1 ,   2 , , T is the desired reference trajectory of multi-agent systems. r ( t ) can also be called the desired trajectory of a virtual leader. In practice, not all agents can obtain r ( t ) , only some of them can. Let Φ = diag { φ 1 ,   φ 2 ,   , φ n } denote the reference-accessibility matrix, which is a diagonal, non-negative, real matrix. If the i th agent can obtain direct information about the r ( t ) , then φ i > 0 ,   i = 1 ,   2 ,   ,   n ; otherwise, φ i = 0 ,   i = 1 ,   2 ,   ,   n . In (4), d i ( t ) is the desired output deviation from desired reference trajectory of the i th agent. d i j ( t ) = d i ( t ) d j ( t ) represents the desired relative formation between the i th agent and the j th agent. The formation tracking error δ i , k ( t ) is denoted as δ i , k ( t ) = d i ( t ) y i , k ( t ) and the output error of the i th agent in k th iteration can be denoted as
e i , k ( t ) = r ( t ) + δ i , k ( t )
By applying the Z-transform to Equation (5), we get
E i , k ( z ) = R ( z ) + Δ i , k ( z )
If the following condition (7) is met
lim k E i , k ( z ) = 0
that is, l i m k Y i , k ( z ) = Y i d ( z ) , the multi-agent systems realize formation tracking control.

2.2.2. Structure of ILC for Networked Multi-Agent Systems

The following ILC law is used to achieve control objectives (7)
u i , k + 1 ( t ) = u i , k ( t ) + Γ i { j ϵ N i a i , j [ δ i , k ( t + 1 ) δ j , k ( t + 1 ) ] + φ i e i , k ( t + 1 ) }
where Γ i is the learning gain for ith agent. The formulation of control input of the ith agent requires the control input of the last iteration, error, and information of adjacent agents. Due to the limited bandwidth of communication channel, noise interference, and signal fading, when a multi-agent transmits information through the wireless network, random packet loss is introduced into the multi-agent. For the communication graph of multi-agent systems shown in Figure 1, a block diagram of ILC for multi-agent systems with loss of multiple communication packets is illustrated in Figure 2. When the i th agent and the j th agent exchange information with each other through the wireless network, the i th agent receives a signal from the j th agent, which is denoted as
δ ˙ j , k i ( t ) = ξ k , i j δ j , k ( t )
where ξ k , i j obeys Bernoulli distribution and δ j , k ( t ) is the formation tracking error of the i th agent. When the j th agent sends a signal to the i th agent and there is no packet loss in the k th iteration, let ξ k , i j = 1 . Otherwise, ξ k , i j = 0 . Since the i th agent can obtain its own output y i , k ( t ) and the desired output deviation d i ( t ) without transmission network, there is no packet loss for the i th agent itself. That is, δ i , k ( t + 1 ) does not have packet loss for the i th agent. It is worth noting that, unlike a previous study [18], the scenario discussed in this paper is closer to the engineering practice, assuming that the packet loss rates of communication between agents are different.

3. Dropout Compensation ILC Method for Multi-Agent Systems with Loss of Multiple Communication Packets

In this section, firstly, because the iterative data are changing in the iterative domain, the previous iterative data are used to compensate for the missing data in the iterative process. Secondly, considering that ILC is a two-dimensional system, the convergence of any frequency in the iterative dimension is analyzed by variable iteration and parameter iteration. We make the following assumptions:
Assumption 1.
The dynamics G i for each agent is stable, that is, | A i | < 1 , i = 1 ,   2 , ,   n . Without loss of generality, the system relative degree is one, that is, G i 0 .
Remark 1.
If an agent’s dynamics G i is not stable, we can use the local feedback of its output y i to make it stable.
Assumption 2.
The initial reset condition satisfies all agents and the desired input for each iteration, that is, x i , k ( 0 ) can be abbreviated to x i ( 0 ) and C i x i , k ( 0 ) = y i d ( 0 ) ,   k 1 ,   2 ,   3 ,   i = 1 ,   2 , ,   n .
Remark 2.
Assumption 2 is common in ILC of multi-agent systems (see [13,16,19]). If the initial reset condition is not met, we can regard it as the problem of robustness to the initial shifts. In our future work, the ILC with initial-state learning for formation tracking control of multi-agent systems will be explored.
Assumption 3.
The graph G of multi-agent system (1) is an undirected graph. In addition, its graph is connected, that is, there is a path between any two agents so that they can exchange information.
Lemma 1.
[20]: If an irreducible matrix M = ( a i , j ) n × n , n 1 is weakly generalized diagonally dominant and at least one of the rows is strictly diagonally dominant, M is nonsingular.
Remark 3.
Assumption 3 implies that the ( L + Φ ) n × n . In L + Φ , the magnitude of the diagonal entry is φ i + j ϵ N i a i , j ,   i = 1 , 2 , , n and the sum of the magnitudes of all non-diagonal entries is j ϵ N i a i , j ,   i = 1 , 2 , , n . Since φ i + j ϵ N i a i , j j ϵ N i a i , j ,   i = 1 , 2 , , n and at least one of the rows is satisfied φ i + j ϵ N i a i , j > j ϵ N i a i , j , matrix L + Φ is nonsingular by Lemma.

3.1. Dropout Compensate ILC Method

To achieve the control objective (7), we propose the ILC law to compensate the loss of packets.
u i , k + 1 ( t ) = u i , k ( t ) + Γ i { j ϵ N i a i , j [ δ i , k ( t + 1 ) δ ˜ j , k i ( t + 1 ) ] + φ i e i , k ( t + 1 ) }
When there are packet losses during multiple agents’ communication, we utilize the previous iterative data to compensate this iterative data. In (10), δ ˜ j , k ( t + 1 ) is data which the ith agent received from the jth agent
δ ˜ j , k i ( t + 1 ) = ξ k , i j δ j , k ( t + 1 ) + [ 1 ξ k , i j ] δ j , k 1 ( t + 1 )
where ξ k , i j is mentioned above. In practice, we can utilize a register to store each iterative data. If there is no packet loss, the next iterative data replace the old data. Otherwise, the old data are retained.
Remark 4.
In order to make the multi-agent system closer to the practical engineering applications, ξ k , i j could be different for different agents. That means the multi-agent system is heterogeneous in this paper.
Since Δ i , k ( z ) = D i ( z ) Y i , k ( z ) , (6) can be rewritten as
E i , k ( z ) = R ( z ) + Δ i , k ( z )
By combining the Z-transform of (10) with (12), we get
U i , k + 1 ( z ) = U i , k ( z ) + z Γ i { j ϵ N i a i , j [ E i , k ( z ) E ˜ j , k ( z ) ] + φ i E i , k ( z ) }
where E ˜ j , k ( z ) = ξ k , i j E j , k ( z ) ( 1 ξ k , i j ) E j , k 1 ( z ) . According to Equations (3), (11)–(13) and Assumption 2, we have
E i , k + 1 ( z ) E i , k ( z ) = Y i , k + 1 ( z ) + Y i , k ( z ) = G i ( z ) [ U i , k ( z ) U i , k + 1 ( z ) ] = z G i ( z ) Γ i { j ϵ N i a i , j [ E i , k ( z ) E ˜ j , k ( z ) ] + φ i E i , k ( z ) } = z G i ( z ) Γ i { j ϵ N i a i , j [ E i , k ( z ) ξ k , i j E j , k ( z ) ( 1 ξ k , i j ) E j , k 1 ( z ) ] + φ i E i , k ( z ) }
Let H i ( z ) = G i ( z ) Γ i , H ( z ) = d i a g { H 1 ( z ) , H 2 ( z ) H n ( z ) } , H i ( z ) , H ( z ) n × n . Let E k ( z ) = [ E 1 , k ( z ) , E 2 , k ( z ) , , E n , k ( z ) ] T , E k ( z ) n × 1 , then (14) can be rewritten as
E k + 1 ( z ) E k ( z ) = z H ( z ) [ ( L ξ + Φ ) E k ( z ) A ξ E k 1 ( z ) ]
where L ξ = [ l i j , ξ ] n × n and A ξ = [ a i j , ξ ] n × n ,
l i j , ξ = { a i , j ξ k , i j ,   j ϵ N i j ϵ N i a i , j ,   i = j       0 ,   o t h e r s   , a i j , ξ = { a i , j [ 1 ξ k , i j ] ,   j ϵ N i 0 ,   o t h e r s .
The relationship among the (k − 1)th, the kth, and the (k + 1)th system error is shown in (15). Our goal is to converge the system error when the number of iterations tends to infinity. Thus, (15) will be further analyzed in the iterative domain.

3.2. Analysis in Iterative Domain

ILC is a method to control systems operating in a repetitive mode over a finite interval with the requirement that the system accurately follows the desired output trajectory. The same actions are executed again and again. By using the previously executed information, the tracking accuracy is improved from repetition to repetition. It has long been recognized that ILC is a two-dimensional (2D) system in which information travels on two independent axes: the time axis and the iteration axis. Chow and Fang [21,22] proposed ILC 2D analysis methods for discrete-time and continue-time systems. After Z-transformation of system dynamics, ILC is considered as a 2D system in frequency domain and iteration domain in this paper. The following analysis is based on the decomposition of 2D dynamics into two separate 1D problems. E k ( z ) is considered as a 2D function E ( k , z ) . Exchanging variable frequency and parameter iteration, the convergence of the system is considered as a 1D problem in the iteration domain. Thus, we exchange variable z and parameter k . E k ( z ) can be rewritten as E z ( k ) , that is, E z ( k ) = [ E 1 , z ( k ) , E 2 , z ( k ) , , E n , z ( k ) ] T , ξ k , i j can be rewritten as ξ i j ( k ) . Then, (15) can be rewritten as
E z ( k + 1 ) E z ( k ) = z H ( z ) [ ( L ( k ) + Φ ) E z ( k ) A ( k ) E z ( k 1 ) ]
where L ( k ) = [ l i j ( k ) ] n × n and A ( k ) = [ a i j ( k ) ] n × n ,
l i j ( k ) = { a i , j ξ i j ( k ) ,   j ϵ N i j ϵ N i a i , j ,   i = j       0 ,   o t h e r s   , a i j ( k ) = { a i , j [ 1 ξ i j ( k ) ] ,   j ϵ N i           0 ,   o t h e r s .
Take Z-transformation for (17)
𝕫 E z ( 𝕫 ) E z ( 𝕫 ) = z H ( z ) [ ( L ( 𝕫 ) + Φ ) E z ( 𝕫 ) 𝕫 1 A ( 𝕫 ) E z ( 𝕫 ) ]
where L ( 𝕫 ) = [ l i j ( 𝕫 ) ] n × n and A ( 𝕫 ) = [ a i j ( 𝕫 ) ] n × n ,
l i j ( 𝕫 ) = { a i , j ξ i j ( 𝕫 ) ,   j ϵ N i j ϵ N i a i , j ,   i = j       0 ,   o t h e r s   , a i j ( 𝕫 ) = { a i , j [ 1 ξ i j ( 𝕫 ) ] ,   j ϵ N i           0 ,   o t h e r s .
(19) leads to
{ ( 𝕫 1 ) I + z H ( z ) [ L ( 𝕫 ) + Φ ] } E z ( 𝕫 ) = 0
where L ( 𝕫 ) = [ l i j ( 𝕫 ) ] n × n ,
l i j ( 𝕫 ) = { a i , j 𝕫 1 [ 𝕫 ξ i j ( 𝕫 ) + 1 ξ i j ( 𝕫 ) ] ,   j ϵ N i   j ϵ N i a i , j ,   i = j         0 ,   o t h e r s
From Equation (21), we can obtain characteristic equation of the system in iterative domain. In the next section, the convergence of the learning system is discussed based on the characteristic equation.

4. Convergence Analysis

As we know, when we say that an iterative learning algorithm converges, it means that the system error trends to zero as the number of iterations approaches infinity. For the proposed dropout compensation ILC method, we will analyze the convergence of formation tracking for the multi-agent systems in this section.
Theorem 1.
Consider the multi-agent system (1), let Assumptions 1–3 hold and the learning law (10) be applied. Given that any one of the following conditions is satisfied for all i = 1 , 2 , , n ,
1 .   ( 1 > A i > 0 ) ( P i , 1 > 0 ) ( P i , 2 > 0 ) ( P i , 3 > 0 )   ( P i , 4 > 0 ) ( P i , 5 > 0 ) , 2 .   ( 1 < A i < 0 ) ( P i , 1 > 0 ) ( P i , 2 < 0 ) ( P i , 3 > 0 ) ( P i , 4 < 0 ) ( P i , 5 > 0 ) ,
where
P i , 1 = 2 Q i , P i , 2 = 16 ( 1 A i ) 2 + 4 ( 1 A i ) ( 3 A i ) Q i + ( A i 2 ) Q i 2 , P i , 3 = 2 ( 1 A i ) 2 ( 1 A i ) Q i , P i , 4 = 4 ( 1 A i ) 2 + 2 ( 1 A i ) ( 3 A i ) Q i + ( A i 2 ) Q i 2 , P i , 5 = 4 ( 1 A i ) 2 4 ( 1 A i ) Q i + S i , Q i = Γ i B i C i ( K i + φ i ) , S i = ( Γ i B i C i ) 2 ( 2 K i φ i + φ i 2 ) , K i = j N i a i , j .
then the formation tracking objective (7) is achieved.
Proof. 
Let S ( 𝕫 , z ) = det { ( 𝕫 1 ) I + z H ( z ) [ L ( 𝕫 ) + Φ ] } . The characteristic equation is S ( 𝕫 , z ) = 0 . When 𝕫 = 1 , L ( 1 ) = L . Thus, S ( 1 , z ) = det [ z H ( z ) ( L + Φ ) ] . In Remark 3, L + Φ proves to be nonsingular. Therefore, matrix z H ( z ) ( L + Φ ) is nonsingular also and S ( 1 , z ) 0 . □
When 𝕫 1 , we need to prove that the modules of all roots of (25) are smaller than units.
det { I + z H ( z ) [ L ( 𝕫 ) + Φ ] 𝕫 1 } = 0
Let F ( 𝕫 , z ) = z H ( z ) [ L ( 𝕫 ) + Φ ] 𝕫 1 , based on the general Nyquist stability criterion [23], the modules of all roots of (25) are smaller than units, if the eigenloci λ [ F ( ω , β ) ] of
F ( ω , β ) = e j β H ( e j β ) [ L ( e j ω ) + Φ ] e j ω 1
does not enclose the point ( 1 , j 0 ) for   β [ π , π ) , ω [ π , π ) , and ω 0 . Based on Greshgorin’s disk theorem, we have λ [ F ( ω , β ) ] U i = 1 n F i for all β [ π , π ) , ω [ π , π ) , and ω 0 , where
F i = { ς : ς C , | ς e j β H i ( e j β ) ( j N i a i , j + φ i ) e j ω 1 | j N i | a i , j e j β H i ( e j β ) [ e j ω ξ i j ( e j ω ) + 1 ξ i j ( e j ω ) ] e j ω ( e j ω 1 ) | }
Since ξ i j ( e j ω ) can only be 0 or 1 , e j ω ξ i j ( e j ω ) + 1 ξ i j ( e j ω ) is 1 or e j ω , respectively. Taking H i ( e j β ) as Γ i B i C i ( e j β A i ) , we can rewrite (27) as
F i = { ς : ς C , | ς e j β Γ i B i C i ( K i + φ i ) ( e j ω 1 ) ( e j β A i ) | | K i Γ i B i C i ( e j ω 1 ) ( e j β A i ) | }
So the eigenloci λ [ F ( ω , β ) ] does not enclose the point ( 1 , j 0 ) for   β [ π , π ) , ω [ π , π ) , and ω 0 when the point ( a , j 0 ) with a 1 is not in the disc F i for all i = 1 , 2 , , n , β [ π , π ) , ω [ π , π ) , and ω 0 . That is, | a + e j β Γ i B i C i ( K i + φ i ) [ ( e j ω 1 ) ( e j β A i ) ] 1 | > | K i Γ i B i C i [ ( e j ω 1 ) ( e j β A i ) ] 1 | for all i = 1 , 2 , , n , β [ π , π ) , ω [ π , π ) , and ω 0 , when a 1 . We denote f i ( a ) as
f i ( a ) = | a ( e j ω 1 ) ( e j β A i ) + e j β Γ i B i C i ( K i + φ i ) | 2
Note that | a + e j β Γ i B i C i ( K i + φ i ) ( e j ω 1 ) ( e j β A i ) | 2 | K i Γ i B i C i ( e j ω 1 ) ( e j β A i ) | 2 = | 1 ( e j ω 1 ) ( e j β A i ) | 2 f i ( a ) for all a 1 , β [ π , π ) , ω [ π , π ) , and ω 0 . So, | a + e j β Γ i B i C i ( K i + φ i ) ( e j ω 1 ) ( e j β A i ) | 2 | K i Γ i B i C i ( e j ω 1 ) ( e j β A i ) | 2 > 0 as long as f i ( a ) > 0 for all a 1 , β [ π , π ) , ω [ π , π ) , and ω 0 . Further, we can show that
f i ( a ) = 2 ( 1 cos ω ) ( 1 + A i 2 2 A i cos β ) a 2 + 2 a Γ i B i C i ( K i + φ i ) [ cos ω 1 A i cos ( ω β ) + A i cos ( β ) ] + ( Γ i B i C i ) 2 ( 2 K i φ i + φ i 2 ) .
Using the conditions of Theorem 1, we can prove f i ( a ) > 0 for all a 1 , β [ π , π ) , ω [ π , π ) , and ω 0 (see Appendix of [17]). Then the eigenloci of λ [ F ( ω , β ) ] for all β [ π , π ) , ω [ π , π ) , and ω 0 does not enclose the point ( 1 , j 0 ) . The modules of all roots of (25) are smaller than units. That is, the system achieves a consensus asymptotically.
Multiply 𝕫 1 on both sides of Equation (21) and take the limit
lim 𝕫 1 { 𝕫 ( 𝕫 1 ) I + z H ( z ) [ L ( 𝕫 ) + Φ ] } ( 𝕫 1 ) E z ( 𝕫 ) = 0
According to the final value theorem, we get
{ z H ( z ) [ L + Φ ] } E z ( ) = 0
As the previous certificate, d e t { z H ( z ) [ L + Φ ] } 0 . Thus, r a n k { z H ( z ) [ L + Φ ] } = n . Based on Sylvester inequality, we get r a n k { z H ( z ) [ L + Φ ] } + r a n k ( E z ( ) ) n . Thus r a n k [ E z ( ) ] = 0 and the final value of E z ( k ) is 0 n . Theorem 1 is thus proved.
When considering random communication packet dropout, the ILC algorithm can converge using the compensation method we proposed. When considering the random packet loss, the ILC algorithm can use this proposed compensation method to converge. Theorem 1 gives the convergence condition of learning gain design and proves that the convergence condition is independent of communication packet loss. According to the final value theorem, the error of formation tracking converges to zero as the number of iterations goes to infinity.

5. Simulation

Firstly, consider the multi-agent system in Section 2.2.1. and let Assumptions 1–3 hold. Next, the learning law (10) is applied and the learning gains satisfy the condition (23). The results of numerical simulations are discussed in this section.
In order to illustrate the effectiveness of the proposed dropout compensation ILC method, we considered a heterogeneous multi-agent system consisting of four agents in the undirected graph as shown in Figure 3. It can be seen that the graph is a connected graph and satisfies Assumption 3. Only the second agent and the third agent can obtain the desired reference trajectory.
The Laplacian matrix L of graph G and the reference-accessibility matrix Φ were
L = [ 4 1.7 2.3 0 1.7 1.7 0   0 2.3 0 0 0   3.8 1.5 1.5 1.5 ] , Φ = [ 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 ] .
Then it was obtained that the information interaction weight between agents can be obtained, as shown in Table 1.
For every agent in Figure 3, the system parameters were defined in Table 2.
Let Γ 1 = 0.813 , Γ 2 = 1.31 , Γ 3 = 0.57 , and Γ 4 = 0.92 . Then, we can obtain the parameters in Table 3 for Theorem 1.
Therefore, the learning gains satisfied the second condition in Theorem 1. To perform our simulation tests, we considered the desired reference trajectory as
r ( t ) = 1.5 + sin π t 25
and the desired output deviation of each agent as
d 1 ( t ) = 0 , d 2 ( t ) = 1.5 sin π t 25 , d 3 ( t ) = 1.5 + sin π t 25 , d 4 ( t ) = 0.03 t .
Set the initial state of four agents as:   x 1 ( 0 ) = 4.277 , x 2 ( 0 ) = 0 , x 3 ( 0 ) = 6.1333 , and x 4 ( 0 ) = 2.806 , which satisfies Assumption 2. When the iteration step is 500 and t = 1 , 2 , , 150 , the trajectory of the multi-agent system without packet loss is shown as Figure 4. It can be seen that the desired formation is well achieved to track the desired reference trajectory by the desired output deviation.
In order to measure the formation accuracy quantitatively, the disagreement among all agents on their output errors was defined as: e r r o r ( k ) = 1 n T i = 1 n t = 1 T | e i , k ( t ) | . When lim k e r r o r ( k ) = 0 , multi-agent system achieved formation tacking. Figure 5 shows that formation performance of the system at the first 500 iterations. Obviously, when the iteration is around 350 the system error can converge to zero and achieve formation tracking.
It is assumed that there is no compensation when considering communication packet loss, that is
u i , k + 1 ( t ) = u i , k ( t ) + Γ i { j ϵ N i a i , j [ δ i , k ( t + 1 ) ξ k , i j δ j , k ( t + 1 ) ] + φ i e i , k ( t + 1 ) }
w here   ξ k , i j = 1 when there is no packet loss, otherwise, ξ k , i j = 0 . In Figure 6, it can be seen that the system error cannot converge to zero in different packet loss rates as iteration increases. When iteration is 500 and packet dropout rate is 10%, the trajectory of the multi-agent system without packet loss compensation is shown as Figure 7.
When there are different packet loss rates in a multi-agent system (Figure 8), it can be seen that system errors can converge to zero using (10) to compensate packet loss. The convergence speed hardly gets slower as the packet loss rates increase, as shown in Figure 8b. In addition, the multi-agent system error still can converge to zero around the 350th iteration. When iteration is 350 and packet loss rate is 10%, the trajectory of the multi-agent system with packet loss compensation is shown as in Figure 9.
In order to demonstrate the effectiveness of proposed method, the comparison results with the method of Zhang and Li [19] are shown in Figure 10. If there is data loss, the information for the current iteration will be discarded without any compensation. Therefore, although the error curve can converge to 0, the convergence speed gets slower as the packet loss rate increases [19], while the convergence speed hardly gets slower, as shown in Figure 8b, by the proposed method.

6. Conclusions

ILC-based formation tracking for heterogeneous multi-agent systems is a challenging problem since different communication packets are lost between agents. When packet loss occurs at different rates, the tracking error of ILC does not converge. This paper explores the ILC method of packet loss compensation for a heterogeneous multi-agent system. Since the iterative data are constantly changing in the iterative domain, it is effective to use previous data to compensate for the lost data in this iterative process. Based on the ILC model for multi-agent systems in frequency domain, the convergence of the system is considered as a 1D problem in the iterative domain, and the convergence of the system is analyzed. By using the generalized Nyquist criterion and Gershgorin’s disk theorem, the eigenvalue range of the system characteristic equation is analyzed, and the convergence condition of the system error is obtained. The dropout compensation ILC method proposed in this paper can not only make the error converge to zero with the increase of iteration times, but also the maintain the convergence speed under the packet loss scenario. The numerical simulation results demonstrate the effectiveness of the proposed method.

Author Contributions

Conceptualization, Y.W. and Y.F.; data curation, Y.W. and Y.F.; formal analysis, Z.S. and Y.F.; methodology, Y.W.; supervision, Y.F.; writing—original draft, Y.W.; writing—review and editing, Y.W., J.Z., Y.G., and Y.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by research grants from the National Natural Science Foundation of China (61673253, 61901254).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Olfati-Saber, R.; Fax, J.A.; Murray, R.M. Consensus and cooperation in networked multi-agent systems. Proc. IEEE 2007, 95, 215–233. [Google Scholar] [CrossRef] [Green Version]
  2. Zheng, Y.; Zhu, Y.; Wang, L. Consensus of heterogeneous multi-agent systems. IET Control. Theory Appl. 2011, 5, 1881–1888. [Google Scholar] [CrossRef]
  3. Nedic, A.; Ozdaglar, A. Distributed subgradient methods for multi-agent optimization. IEEE Trans. Autom. Control 2009, 54, 48–61. [Google Scholar] [CrossRef]
  4. Ahn, H.S.; Chen, Y.Q. Iterative learning control for multi-agent formation. In Proceedings of the ICROS-SICE International Joint Conference, Fukuoka, Japan, 18–21 August 2009. [Google Scholar]
  5. Ahn, H.S.; Moore, K.L.; Chen, Y.Q. Trajectory-keeping in satellite formation flying via robust periodic learning control. Int. J. Robust Nonlinear Control 2010, 20, 1655–1666. [Google Scholar] [CrossRef]
  6. Chen, X.; Jia, Y. Stereo vision-based formation control of mobile robots using iterative learning. In Proceedings of the International Conference on Humanized Systems, Kyoto, Japan, 17–19 September 2010. [Google Scholar]
  7. Sun, H.; Hou, Z.; Li, D. Coordinated iterative learning control schemes for train trajectory tracking with overspeed protection. IEEE Trans. Autom. Sci. Eng. 2013, 10, 323–333. [Google Scholar] [CrossRef]
  8. Hock, A.; Schoellig, A.P. Distributed iterative learning control for a team of quadrotors. In Proceedings of the IEEE Conference on Decision and Control (CDC), Las Vegas, NV, USA, 12–14 December 2016. [Google Scholar]
  9. Schoellig, A.P.; Mueller, F.L.; Andrea, R.D. Optimization-based iterative learning for precise quadrocopter trajectory tracking. Auton. Rob. 2012, 33, 103–127. [Google Scholar] [CrossRef] [Green Version]
  10. Yang, S.; Xu, J.; Li, X.; Shen, D. Iterative Learning Control for Multi-agent Systems Coordination; John Wiley & Sons: Hoboken, NJ, USA, 2017. [Google Scholar]
  11. Bu, X.; Hou, Z.; Jin, S.; Chi, R. An iterative learning control design approach for networked control systems with data dropouts. Int. J. Robust Nonlinear Control 2016, 26, 91–109. [Google Scholar]
  12. Dong, L.; Chai, S.; Zhang, B.; Nguang, S.K. Finite interval tracking algorithm for nonlinear multi-agent systems with communication delays. Int. J. Syst. Sci. 2016, 47, 3509–3517. [Google Scholar] [CrossRef]
  13. Meng, D.; Jia, Y.; Du, J. Consensus seeking via iterative learning for multi-agent systems with switching topologies and communication time-delays. Int. J. Robust Nonlinear Control 2016, 26, 3772–3790. [Google Scholar] [CrossRef]
  14. Zhang, T.; Li, J. Event-triggered iterative learning control for multi-agent systems with quantization. Asian J. Control 2018, 20, 1088–1101. [Google Scholar] [CrossRef]
  15. Zhang, R.; Chi, R.; Hou, Z. Consensus tracking of multi-agent systems with time-delays using adaptive iterative learning control. In Proceedings of the IEEE Data Driven Control and Learning Systems Conference (DDCLS), Chongqing, China, 26–27 May 2017. [Google Scholar]
  16. Zhang, J.; Fang, Y.; Wu, Y. An ILC method of formation control for multi-agent system with one-step random time-delay. In Proceedings of the IEEE Data Driven Control and Learning Systems Conference (DDCLS), Chongqing, China, 26–27 May 2017. [Google Scholar]
  17. Zhang, J.; Fang, Y.; Li, C.; Zhu, W. Formation tracking via iterative learning control for multiagent systems with diverse communication time-delays. Math. Probl. Eng. 2019. [Google Scholar] [CrossRef] [Green Version]
  18. Zhang, T.; Li, J. Asynchronous event-triggered control of multi-agent systems with Sigma-Delta quantizer and packet losses. J. Frankl. I 2016, 353, 1781–1808. [Google Scholar] [CrossRef]
  19. Zhang, T.; Li, J. Iterative learning control for multi-agent systems with finite-leveled sigma-delta quantization and random packet losses. IEEE Trans. Circuits Syst. I Regul. Pap. 2017, 64, 2171–2181. [Google Scholar] [CrossRef]
  20. Kolotilina, L.Y. The singularity/nonsingularity problem for matrices satisfying diagonal dominance conditions in terms of directed graphs. J. Math. Sci. 2006, 132, 166–189. [Google Scholar] [CrossRef]
  21. Chow, T.W.S.; Fang, Y. An iterative learning control method for continuous-time systems based on 2-D system theory. IEEE Trans. Circuits Syst. I Fundam. Theory Appl. 1998, 45, 683–689. [Google Scholar] [CrossRef] [Green Version]
  22. Fang, Y.; Chow, T.W.S. 2-D analysis for iterative learning controller for discrete-time systems with variable initial conditions. IEEE Trans. Circuits Syst. I Fundam. Theory Appl. 2003, 50, 722–727. [Google Scholar] [CrossRef]
  23. Desoer, C.A.; Wang, Y.T. On the generalized Nyquist stability criterion. IEEE Trans. Autom. Control 1980, 25, 187–196. [Google Scholar] [CrossRef]
Figure 1. Communication graph of multi-agent system.
Figure 1. Communication graph of multi-agent system.
Applsci 10 04752 g001
Figure 2. The structure of iterative learning control (ILC) for multi-agent systems with loss of multiple communication packets.
Figure 2. The structure of iterative learning control (ILC) for multi-agent systems with loss of multiple communication packets.
Applsci 10 04752 g002
Figure 3. Undirected graph of multi-agent system.
Figure 3. Undirected graph of multi-agent system.
Applsci 10 04752 g003
Figure 4. Multi-agent system trajectory at iteration 500.
Figure 4. Multi-agent system trajectory at iteration 500.
Applsci 10 04752 g004
Figure 5. Convergence of formation tracking errors for multi-agent system.
Figure 5. Convergence of formation tracking errors for multi-agent system.
Applsci 10 04752 g005
Figure 6. System error of multi-agent system with different packet loss rates: (a) 1% packet dropout, (b) 2% packet dropout, (c) 5% packet dropout, (d) 10% packet dropout.
Figure 6. System error of multi-agent system with different packet loss rates: (a) 1% packet dropout, (b) 2% packet dropout, (c) 5% packet dropout, (d) 10% packet dropout.
Applsci 10 04752 g006aApplsci 10 04752 g006b
Figure 7. Trajectory of the multi-agent system without packet loss compensation at iteration 500.
Figure 7. Trajectory of the multi-agent system without packet loss compensation at iteration 500.
Applsci 10 04752 g007
Figure 8. Convergence of system errors for multi-agent system with different packet loss rates with compensation: (a) k = 1, 2, …, 500; (b) k = 100, 101, …, 150.
Figure 8. Convergence of system errors for multi-agent system with different packet loss rates with compensation: (a) k = 1, 2, …, 500; (b) k = 100, 101, …, 150.
Applsci 10 04752 g008
Figure 9. Trajectory of the multi-agent system with dropout compensation at iteration 350.
Figure 9. Trajectory of the multi-agent system with dropout compensation at iteration 350.
Applsci 10 04752 g009
Figure 10. Comparison results with other method: (a) 20% packet loss; (b) 50% packet loss.
Figure 10. Comparison results with other method: (a) 20% packet loss; (b) 50% packet loss.
Applsci 10 04752 g010
Table 1. Interaction weight of agent.
Table 1. Interaction weight of agent.
K 1 = 4 K 2 = 1.7 K 3 = 3.8 K 4 = 1.5
φ 1 = 0 φ 2 = 1 φ 3 = 1 φ 4 = 0
Table 2. System parameters.
Table 2. System parameters.
A 1 = 0.16 A 2 = 0.18 A 3 = 0.12 A 4 = 0.32
B 1 = 0.416 B 2 = 0.15 B 3 = 0.6 B 4 = 0.05
C 1 = 0.38 C 2 = 0.7 C 3 = 0.53 C 4 = 0.59
Table 3. Convergence condition parameter.
Table 3. Convergence condition parameter.
Q 1 = 0.514 Q 2 = 0.371 Q 3 = 0.870 Q 4 = 0.041
S 1 = 0 S 2 = 0.083 S 3 = 0.283 S 4 = 0
P 1 , 1 = 1.486 P 2 , 1 = 1.629 P 3 , 1 = 1.130 P 4 , 1 = 1.959
P 1 , 2 = 14.563 P 2 , 2 = 17.005 P 3 , 2 = 9.514 P 4 , 2 = 27.168
P 1 , 3 = 2.095 P 2 , 3 = 2.347 P 3 , 3 = 1.534 P 4 , 3 = 3.431
P 1 , 4 = 2.184 P 2 , 4 = 3.083 P 3 , 4 = 0.542 P 4 , 4 = 6.617
P 1 , 5 = 2.997 P 2 , 5 = 3.900 P 3 , 5 = 1.402 P 4 , 5 = 6.755

Share and Cite

MDPI and ACS Style

Wu, Y.; Zhang, J.; Ge, Y.; Sheng, Z.; Fang, Y. A Dropout Compensation ILC Method for Formation Tracking of Heterogeneous Multi-Agent Systems with Loss of Multiple Communication Packets. Appl. Sci. 2020, 10, 4752. https://doi.org/10.3390/app10144752

AMA Style

Wu Y, Zhang J, Ge Y, Sheng Z, Fang Y. A Dropout Compensation ILC Method for Formation Tracking of Heterogeneous Multi-Agent Systems with Loss of Multiple Communication Packets. Applied Sciences. 2020; 10(14):4752. https://doi.org/10.3390/app10144752

Chicago/Turabian Style

Wu, Yuzhou, Jialu Zhang, Yu Ge, Zhichao Sheng, and Yong Fang. 2020. "A Dropout Compensation ILC Method for Formation Tracking of Heterogeneous Multi-Agent Systems with Loss of Multiple Communication Packets" Applied Sciences 10, no. 14: 4752. https://doi.org/10.3390/app10144752

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop