Cooperative Control of Multi-Agent Systems: Theory and Applications
By Yue Wang
()
About this ebook
A comprehensive review of the state of the art in the control of multi-agent systems theory and applications
The superiority of multi-agent systems over single agents for the control of unmanned air, water and ground vehicles has been clearly demonstrated in a wide range of application areas. Their large-scale spatial distribution, robustness, high scalability and low cost enable multi-agent systems to achieve tasks that could not successfully be performed by even the most sophisticated single agent systems.
Cooperative Control of Multi-Agent Systems: Theory and Applications provides a wide-ranging review of the latest developments in the cooperative control of multi-agent systems theory and applications. The applications described are mainly in the areas of unmanned aerial vehicles (UAVs) and unmanned ground vehicles (UGVs). Throughout, the authors link basic theory to multi-agent cooperative control practice — illustrated within the context of highly-realistic scenarios of high-level missions — without losing site of the mathematical background needed to provide performance guarantees under general working conditions. Many of the problems and solutions considered involve combinations of both types of vehicles. Topics explored include target assignment, target tracking, consensus, stochastic game theory-based framework, event-triggered control, topology design and identification, coordination under uncertainty and coverage control.
- Establishes a bridge between fundamental cooperative control theory and specific problems of interest in a wide range of applications areas
- Includes example applications from the fields of space exploration, radiation shielding, site clearance, tracking/classification, surveillance, search-and-rescue and more
- Features detailed presentations of specific algorithms and application frameworks with relevant commercial and military applications
- Provides a comprehensive look at the latest developments in this rapidly evolving field, while offering informed speculation on future directions for collective control systems
The use of multi-agent system technologies in both everyday commercial use and national defense is certain to increase tremendously in the years ahead, making this book a valuable resource for researchers, engineers, and applied mathematicians working in systems and controls, as well as advanced undergraduates and graduate students interested in those areas.
Related to Cooperative Control of Multi-Agent Systems
Related ebooks
Aircraft Control Allocation Rating: 0 out of 5 stars0 ratingsOperations Research for Unmanned Systems Rating: 0 out of 5 stars0 ratingsMathematical Modelling: A Graduate Textbook Rating: 0 out of 5 stars0 ratingsIntroduction to UAV Systems Rating: 5 out of 5 stars5/5An Introduction to Discrete-Valued Time Series Rating: 0 out of 5 stars0 ratingsMetaheuristics for Air Traffic Management Rating: 5 out of 5 stars5/5Modeling and Estimation of Structural Damage Rating: 0 out of 5 stars0 ratingsSmall Unmanned Fixed-wing Aircraft Design: A Practical Approach Rating: 0 out of 5 stars0 ratingsAnalytical Modeling of Wireless Communication Systems Rating: 0 out of 5 stars0 ratingsRobot Learning by Visual Observation Rating: 0 out of 5 stars0 ratingsDesign and Implementation of Large-Range Compliant Micropositioning Systems Rating: 0 out of 5 stars0 ratingsHybridization, Diagnostic and Prognostic of PEM Fuel Cells: Durability and Reliability Rating: 0 out of 5 stars0 ratingsFormation Control of Multiple Autonomous Vehicle Systems Rating: 0 out of 5 stars0 ratingsPractical Applications of Bayesian Reliability Rating: 0 out of 5 stars0 ratingsStructural Reliability Analysis and Prediction Rating: 0 out of 5 stars0 ratingsVulkan Expert: Mastering High-Performance Graphics: Vulcan Fundamentals Rating: 0 out of 5 stars0 ratingsImage, Video and 3D Data Registration: Medical, Satellite and Video Processing Applications with Quality Metrics Rating: 0 out of 5 stars0 ratingsDynamics and Control of Robotic Manipulators with Contact and Friction Rating: 0 out of 5 stars0 ratingsComputer Vision in Vehicle Technology: Land, Sea, and Air Rating: 0 out of 5 stars0 ratingsFundamental Statistical Inference: A Computational Approach Rating: 0 out of 5 stars0 ratingsEvolutionary Algorithms for Mobile Ad Hoc Networks Rating: 0 out of 5 stars0 ratingsMathematical and Computational Modeling: With Applications in Natural and Social Sciences, Engineering, and the Arts Rating: 0 out of 5 stars0 ratingsComputational Acoustics: Theory and Implementation Rating: 0 out of 5 stars0 ratingsIntroduction to Operational Modal Analysis Rating: 0 out of 5 stars0 ratingsDynamic System Reliability: Modeling and Analysis of Dynamic and Dependent Behaviors Rating: 0 out of 5 stars0 ratingsAdvanced UAV Aerodynamics, Flight Stability and Control: Novel Concepts, Theory and Applications Rating: 5 out of 5 stars5/5Multidisciplinary Design Optimization Supported by Knowledge Based Engineering Rating: 0 out of 5 stars0 ratingsReal-Time Embedded Systems Rating: 0 out of 5 stars0 ratingsMachine Learning in the AWS Cloud: Add Intelligence to Applications with Amazon SageMaker and Amazon Rekognition Rating: 0 out of 5 stars0 ratingsOpenGL Deep Dive: Expert Techniques and Performance Optimization: OpenGL Rating: 0 out of 5 stars0 ratings
Mechanical Engineering For You
Small Gas Engine Repair, Fourth Edition Rating: 0 out of 5 stars0 ratingsZinn & the Art of Mountain Bike Maintenance: The World's Best-Selling Guide to Mountain Bike Repair Rating: 0 out of 5 stars0 ratingsHow to Repair Briggs and Stratton Engines, 4th Ed. Rating: 0 out of 5 stars0 ratingsBasic Engineering Mechanics Explained, Volume 1: Principles and Static Forces Rating: 5 out of 5 stars5/5The CIA Lockpicking Manual Rating: 5 out of 5 stars5/5Mechanical Engineering Rating: 5 out of 5 stars5/5Basic Machines and How They Work Rating: 4 out of 5 stars4/5Plane Sense: A Beginner's Guide to Owning and Operating Private Aircraft FAA-H-8083-19A Rating: 0 out of 5 stars0 ratingsA Dynamical Theory of the Electromagnetic Field Rating: 0 out of 5 stars0 ratingsAirplane Flying Handbook: FAA-H-8083-3C (2024) Rating: 4 out of 5 stars4/5EPA 608 Study Guide: HVAC, #1 Rating: 4 out of 5 stars4/5Einstein's Fridge: How the Difference Between Hot and Cold Explains the Universe Rating: 4 out of 5 stars4/5How to Walk on Water and Climb up Walls: Animal Movement and the Robots of the Future Rating: 3 out of 5 stars3/5Machining for Hobbyists: Getting Started Rating: 5 out of 5 stars5/5Introduction to Thermodynamics Rating: 0 out of 5 stars0 ratings301 Top Tips for Design Engineers: To Help You 'Measure Up' in the World of Engineering Rating: 5 out of 5 stars5/5FreeCAD Basics Tutorial Rating: 3 out of 5 stars3/5The Art of Welding: Featuring Ryan Friedlinghaus of West Coast Customs Rating: 0 out of 5 stars0 ratingsTides: The Science and Spirit of the Ocean Rating: 4 out of 5 stars4/5Aircraft Weight and Balance Handbook: FAA-H-8083-1A Rating: 5 out of 5 stars5/51,001 Questions & Answers for the CWI Exam: Welding Metallurgy and Visual Inspection Study Guide Rating: 4 out of 5 stars4/5Dialogues Concerning Two New Sciences Rating: 0 out of 5 stars0 ratingsQuantum Gravity in a Nutshell1 Second Edition: Beyond Einstein, #9 Rating: 0 out of 5 stars0 ratingsMachinery's Handbook Guide: A Guide to Tables, Formulas, & More in the 31st Edition Rating: 5 out of 5 stars5/5Watch Repair for Beginners: An Illustrated How-To Guide for the Beginner Watch Repairer Rating: 4 out of 5 stars4/5Quantum Mechanics 1: Particles & Waves Rating: 4 out of 5 stars4/5Elements of Mechanical Design Rating: 5 out of 5 stars5/5
Reviews for Cooperative Control of Multi-Agent Systems
0 ratings0 reviews
Book preview
Cooperative Control of Multi-Agent Systems - Yue Wang
To our advisors, students, and family
List of Contributors
Yongcan Cao
Department of Electrical and Computer Engineering
University of Texas at San Antonio
San Antonio, TX
USA
David Casbeer
The Control Science Center of Excellence
Air Force Research Laboratory
Wright-Patterson AFB, OH
USA
Doo-Hyun Cho
Department of Aerospace Engineering
Korea Advanced Institute of Science and Technology
Daejeon
South Korea
Han-Lim Choi
Department of Aerospace Engineering
Korea Advanced Institute of Science and Technology
Daejeon
South Korea
David A. Copp
Center for Control, Dynamical Systems, and Computation
University of California
Santa Barbara, CA
USA
Jorge Cortes
Department of Mechanical and Aerospace Engineering
University of California
San Diego, CA
USA
Ran Dai
The Aerospace Engineering Department
Iowa State University
Ames, IA
USA
Eloy Garcia
InfoSciTex Corp, USA and
Air Force Research Laboratory
Wright-Patterson AFB, OH
USA
João P. Hespanha
Center for Control, Dynamical Systems, and Computation
University of California
Santa Barbara, CA
USA
Derek Kingston
The Control Science Center of Excellence
Air Force Research Laboratory
Wright-Patterson AFB, OH
USA
Cameron Nowzari
Department of Electrical and Systems Engineering
University of Pennsylvania
Pennsylvania, PA
USA
George J. Pappas
Department of Electrical and Systems Engineering
University of Pennsylvania
Pennsylvania, PA
USA
Steven A. P. Quintero
Center for Control, Dynamical Systems, and Computation
University of California
Santa Barbara, CA
USA
Sivakumar Rathinam
Department of Mechanical Engineering
Texas A&M University
College Station, TX
USA
Corey Schumacher
Air Force Research Laboratory
Wright-Patterson AFB, OH
USA
Rajnikant Sharma
Department of Electrical and Computer Engineering
Utah State University
Logan, UT
USA
Chuangchuang Sun
The Aerospace Engineering Department
Iowa State University
Ames, IA
USA
Li Wang
Department of Mechanical Engineering
Clemson University
Clemson, SC
USA
Xiaofeng Wang
Department of Electrical Engineering
University of South Carolina
Columbia, SC
USA
Yue Wang
Department of Mechanical Engineering
Clemson University
Clemson, SC
USA
Fumin Zhang
School of Electrical and Computer Engineering
Georgia Institute of Technology
Atlanta, GA
USA
Zheqing Zhou
Department of Electrical Engineering
University of South Carolina
Columbia, SC
USA
Preface
This book presents new developments in both the fundamental research and applications in the field of multi-agent systems where a team of agents cooperatively achieve a common goal. Multi-agent systems play an important role in defense and civilian sectors and have the potential to impact on areas such as search and rescue, surveillance, and transportation. Cooperative control algorithms are essential to the coordination among multiple agents and hence realization of an effective multi-agent system. The contents of this book aim at linking basic research and cooperative control methodologies with more advanced applications and real-world problems.
The chapters in this book seek to provide recent developments in the cooperative control of multi-agent systems from a practical perspective. Chapter 1 provides an overview of the state of the art in multi-agent systems and summarizes existing works in consensus control, formation control, synchronization and output regulation, leader and/or target tracking, optimal control, coverage control, passivity-based control, and event-triggered control. Chapter 2 develops sensor placement algorithms for a team of autonomous unmanned vehicles (AUVs) for a path covering problem with monitoring applications in GPS-denied environments. Chapter 3 proposes vision-based output-feedback MPC algorithms with moving horizon estimation for target tracking using fixed-wing unmanned aerial vehicles (UAVs) in measurements gathering and real-time decision-making tasks. Chapter 4 presents the continuous-time projection-based consensus algorithms for multi-UAV simultaneous arrival problem under velocity constraints and finds the convergence rate of the proposed consensus algorithms. Chapter 5 discusses the asset-based weapon-target assignment (WTA) problem to find the optimal launching time of a weapon to maximize the sum of asset values with time-dependent rewards. Chapter 6 presents a coordinated decision algorithm where a group of UAVs is assigned to a set of targets to minimize some cost terms associated with the mission. Chapter 7 provides a formal analysis of event-triggered control and communication techniques for multi-agent average consensus problems. Chapter 8 solves network topology design and identification problems for dynamic networks. Chapter 9 discusses stochastic interaction for distributed multi-agent systems and presents results about the probabilities to achieve multi-agent coordination. Finally, Chapter 10 addresses a cooperative coverage control problem employing wheeled mobile robots (WMRs) and UAVs.
August 2016
Yue Wang
Clemson University
Eloy Garcia
Air Force Research Laboratory,
Wright-Patterson AFB
David Casbeer
Air Force Research Laboratory,
Wright-Patterson AFB
Fumin Zhang
Georgia Institute of Technology
Acknowledgment
The editors would like to thank the authors of all the chapters and reviewers who worked together on this book. A special acknowledgment goes to all the graduate students in the Interdisciplinary & Intelligent Research (I²R) Laboratory in the Mechanical Engineering Department at Clemson University, who assisted the editors to review chapters and provided useful feedbacks to improve the quality of the book. The first editor would like to thank the support from the National Science Foundation under Grant No. CMMI-1454139. The editors would also like to thank Wiley and its staff for the professional support.
Chapter 1
Introduction
Yue Wang¹, Eloy Garcia², David Casbeer² and Fumin Zhang³
¹Department of Mechanical Engineering, Clemson University, Clemson, SC, USA
²The Control Science Center of Excellence, Air Force Research Laboratory, Wright-Patterson AFB, OH, USA
³School of Electrical and Computer Engineering, Georgia Institute of Technology, Atlanta, GA, USA
1.1 Introduction
Many military and civilian applications require a team of agents to coordinate with each other to perform specific tasks without human intervention. In those systems, individual agents (e.g., unmanned underwater/ground/aerial vehicles) have limited capabilities due to short sensing and communication ranges, and small computational power. However, their collective behavior exhibits significant advantages compared to a single sophisticated agent, including large-scale spatial distribution, robustness, high scalability, and low cost [1]. The deployment of large-scale multi-agent systems with constrained costs and smaller sizes can thus achieve tasks that are otherwise unable to be finished by a single agent. Teams of engineered multi-agent systems can collect and process data and perform tasks cooperatively [2–8]. Multi-agent systems play an important role in a wide range of applications such as search and rescue [9], tracking/classification [10–14], surveillance [15, 16], space exploration [17], and radiation shielding and site clearing [18]. Multi-agent systems have also been considered and utilized in fields such as cooperative mobile robotics [19], distributed artificial intelligence and computing [20–22], wireless sensor networks [23], biology [24], social study [25], smart grids [26], traffic management [27, 28], and supply-chain management [29]. Therefore, the use of multi-agent system technologies in both everyday modern society and national defense and homeland security is bound to tremendously increase. In this book, we aim to provide an overview of recent progresses made in the cooperative control of multi-agent systems on both fundamental theory development as well as applications.
In the control community, multi-agent system theory has focused on developing vehicle motion control laws for various tasks including consensus and formation control [2, 30–43], coverage control [44–48], target search and tracking [3–5, 49, 50], task allocation problems [25, 51–53], sensor management problems [14], output regulation [54, 55], optimization [56], and estimation. Three types of control schemes for multi-agent systems have been proposed in the open literature, that is, centralized [57], decentralized [58], and distributed multi-agent control [1]. The centralized control scheme assumes global knowledge of the multi-agent system and seeks to achieve some control objective considering all agents' states, which inevitably suffers from the scalability issue. The decentralized control scheme computes control actions based only on an agent's local information while the more popular distributed control scheme takes both the agent's own information and neighboring agents' information into account to calculate the control action. Both the decentralized and distributed control algorithms provide scalable solutions and can be implemented under minimal connectivity properties. On the other hand, connectivity preserving protocols are developed for multi-agent systems to keep connected and hence guarantee motion stability [59, 60]. The problem has been considered in scenarios such as flocking [61, 62], rendezvous [59, 63], and formation control [64, 65]. The control hierarchy for multi-agent systems can be categorized into two classes, that is, top-down and bottom-up methodologies [66]. The top-down scheme assigns an overarching objective for the multi-agent system and designs control action for each individual agent to achieve this objective. The top-down multi-agent task decomposition is often difficult. While the bottom-up scheme directly defines each individual agent's local control action and their cooperation protocol, which however cannot guarantee any global objective. The paper [67] provides an overview of progresses made in the distributed multi-agent coordination. The books [64, 68] provide an introduction to the distributed control of multi-agent systems. The book [1] discusses the distributed control of multi-agent systems from four main themes, or dimensions: distributed control and computation, adversarial interactions, uncertain evolution, and complexity management. A special category of multi-agent systems, multi-robot systems, has become one of the most important areas of research in robotics [19]. Significant advance has been made in distributed control and collaboration of multi-robot systems in control theory and artificial intelligence [68–70]. There are a considerable amount of works on multi-agent consensus and formation control, and synchronization. We briefly summarize the main results as follows.
The multi-agent consensus control problem ensures that a group of mobile agents stays connected and reaches agreement while achieving some performance objective [64]. The papers [71, 72] provide a good survey of consensus problems in multi-agent cooperative control. In [64], the consensus problem is considered over dynamic interaction graphs by adding appropriate weights to the edges in the graphs. Theoretical results regarding consensus seeking under both time-invariant and dynamically changing information exchange topologies are summarized. Applications of consensus protocols to multi-agent coordination are investigated. In [73, 74], consensus algorithms are extended for second-order nonlinear dynamics in a dynamic proximity network. Necessary and sufficient conditions are given to ensure second-order consensus. In [75], leader-following consensus algorithms are developed for a linear multi-agent system on a switching network, where the input of each agent is subject to saturation. In [76], multi-agent consensus based on the opinion dynamics introduced by Krause is studied. A new proof of convergence is given with all agents in the same cluster holding the same opinion (represented by a real number). Lower bounds on the inter-cluster distances at a stable equilibrium are derived. In [33], multi-agent consensus is considered for an active leader-tracking problem under variable interconnection topology. The effects of delays on multi-agent consensus have been considered in [77].
The paper [78] provides a survey of formation control of multi-agent systems. The existing results are categorized into position-, displacement-, and distance-based control. The finite-time formation control for nonlinear multi-agent systems is investigated in [43]. A small number of agents navigate the whole team based on the global information of the desired formation while the other agents regulate their positions by the local information in a distributed manner. A class of nonlinear consensus protocols is first ensured and then applied to the formation control. In [79], a model-independent coordination strategy is proposed for multi-agent formation control in combination with tracking control for a virtual leader. The authors show that the formation error can be stabilized if the agents can track their respective reference points perfectly or if the tracking errors are bounded. In [80], a decentralized cooperative controller for multi-agent formation control and collision avoidance is developed based on the navigation function formalism. The control law is designed as the gradient of a navigation function whose minimum corresponds to the desired formation. Multi-agent formation control with intermittent information exchange is considered in [81]. Energy-based analysis is utilized to derive stability conditions. The paper [82] investigates rotating consensus and formation control problems of second-order multi-agent systems based on Lyapunov theory. Both theoretical and experimental results are presented in [42] on multi-agent decentralized control that achieves leader–follower formation control and collision avoidance for multiple nonholonomic robots.
In [83], synchronization approach is developed for trajectory tracking of multiple mobile robots while maintaining time-varying formations. In [84], synchronization algorithms are designed in a leader–follower cooperative tracking control problem where the agents are modeled as identical general linear systems on a digraph containing a spanning tree. The control framework includes full-state feedback control, observer design, and dynamic output feedback control. In [54], a distributed control scheme is adopted for robust output regulation in a multi-agent system where both the reference inputs and disturbances are generated by an exosystem. In [55], the output regulation problem is extended to multi-agent systems where a group of subsystems cannot access the exogenous signal. In [85], output consensus algorithms are developed for heterogeneous agents with parametric uncertainties. The multi-agent output synchronization problem is also studied in [86] where the coupling among the agents is nonlinear and there are communication delays. In [87], a general result for the robust output regulation problem has been studied for linear uncertain multi-agent systems. In [88], finite-time synchronization is proposed for a class of second-order nonlinear homogenous multi-agent systems with a leader–follower architecture. A finite-time convergent observer and an observer-based finite-time output feedback controller are developed to achieve the goal.
In [89], distributed tracking control is developed for linear multi-agent systems and a leader whose control input is nonzero, bounded, and not available to any follower. The paper [90] considers multi-agent tracking of a high-dimensional active leader, whose state not only keeps changing but also may not be measured. A neighbor-based local state-estimator and controller is developed for each autonomous following agent. A collision-free target-tracking problem of multi-agent robot system is considered in [91], where a cost function using a semi-cooperative Stackelberg equilibrium point component with weights tuned by a proportional-derivative (PD)-like fuzzy controller is formulated. The distributed finite-time tracking control of second-order multi-agent systems is considered in [92]. Observer-based state feedback control algorithms are designed to achieve finite-time tracking in a multi-agent leader-follower system and extended to multiple active leaders. There are also a lot of works focusing on multi-agent target tracking. In [93], the optimal sensor placement and motion coordination strategies for mobile sensor networks are developed in a target-tracking application. Gradient-descent decentralized motion planning algorithms are developed in [94] for multiple cooperating mobile sensor agents for the tracking of dynamic targets. The problem of target tracking and obstacle avoidance for multi-agent systems is considered in [95]. A potential function-based motion control algorithm is proposed to solve the problem where multiple agents cannot effectively track the target while avoiding obstacles at the same time.
The book [96] gives an overview of optimal and adaptive control methods for multi-agent systems. In [56], a distributed subgradient method is developed to solve a multi-agent convex optimization problem where every agent minimizes its own objective function while exchanging information locally with other agents in the network over a time-varying topology. An inverse optimality-based distributed cooperative control law is designed in [97] to guarantee consensus and global optimality of multi-agent systems, where the communication graph topology interplays with the agent dynamics. The work [98] applies stochastic optimal control theory to multi-agent systems, where the agent dynamics evolve with Wiener noise. The goal is to minimize some cost function of different agent–target combinations so that decentralized agents are distributed optimally over a number of targets. An optimal control framework for persistent monitoring using multi-agent systems is developed in [99] to design cooperative motion control laws to minimize an uncertainty metric in a given mission space. The problem leads to hybrid systems analysis, and an infinitesimal perturbation analysis (IPA) is used to obtain an online solution.
Coverage control considers the problem of fully covering a task domain using multi-agent systems. The problem can be solved by either deploying multiple agents to optimal locations in the domain or designing dynamic motion control laws for the agents so as to gradually cover the entire domain. The former solutions entail locational optimization for networked multi-agent systems. Voronoi diagram–based approaches are introduced in [100] to develop decentralized control laws for multiple vehicles for optimal coverage and sensing policies. Gradient descent–based schemes are utilized to drive a vehicle toward the Voronoi centeriod for optimal localization. In [101], the discrete coverage control law is developed and unified with averaging control laws over acyclic digraphs with fixed and controlled-switching topology. In [102], unicycle dynamics are considered and the coverage control algorithms are analyzed with an invariance principle for hybrid systems. The latter solutions focus on the case when the union of the agents' sensor cannot cover the task domain and hence dynamic motion control needs to be designed so that the agents can travel and collaboratively cover the entire domain [103]. A distributed coverage control scheme is developed in [104, 105] for mobile sensor networks, where the sensor has a limited range and is defined by a probabilistic model. A gradient-based control algorithm is developed to maximize the joint detection probabilities of random events taking place. Effective coverage control is developed to dynamically cover a given 2D region using a set of mobile sensor agents [46, 106]. Awareness-based coverage control has been proposed to dynamically cover a task domain based on the level of awareness an agent has with respect to the domain [48]. The paper [107] extends the awareness coverage control by defining a density function that characterizes the importance of each point in the domain and the desired awareness coverage level as a nondecreasing differentiable function of the density distribution. In [108], awareness and persistence coverage control are addressed simultaneously so that the mission domain can be covered periodically while the desired awareness is satisfied.
Passivity-based control approaches have also been developed to guarantee the stability of multi-agent systems [109]. Passivity is an energy-based method and a stronger system property that implies stability [110, 111]. A system is passive if it does not create energy, that is, the stored energy is less than the supplied energy. The negative feedback interconnection and parallel interconnection of passive systems are still passive. The paper [112] discusses the stabilization and output synchronization for a network of interconnected nonlinear passive agents by characterizing the information exchange structure. In [113], a passivity-based cooperative control is developed for multi-agent systems and the group synchronization is proved with the proposed backstepping controller using the Krasovskii–LaSalle invariance principle. The paper [114] introduces a discrete-time asymptotic multi-unmanned aerial vehicle (UAV) formation control that uses a passivity-based method to ensure c01-math-0001 stability in the presence of overlay network topology with delays and data loss. Passivity-based motion coordination has also been used in [115] for the attitude synchronization of rigid bodies in the leader–follower case with communication delay and temporary communication failures. The work [116] uses the multiple Lyapunov function method for the output synchronization of a class of networked passive agents with switching topology. The concept of stochastic passivity is studied for a team of agents modeled as discrete-time Markovian jump nonlinear systems [117]. Passivity-based approaches have also been widely used in the bilateral teleoperation of robots and multi-agent systems. A good amount of work has utilized the scattering wave transformation and two-port network theory to provide stability of the teleoperation under constant communication delays for velocity tracking. A passifying PD controller is developed in [118] for the bilateral teleoperation of multiple mobile slave agents coupled to a single master robot under constant, bounded communication delays. The paper [119] extends the passivity-based architecture to guarantee state (velocity as well as position) synchronization of master/slave robots without using the wave scattering transformation. Passivity-based control strategies are also utilized for the bilateral teleoperation of multiple UAVs [120].
Extensive results presenting algorithms and control methodologies for multi-agent systems cooperation rely on continuous communication between agents. Continuous actuation and continuous measurement of local states may be restricted by particular hardware limitations. A problem in many scenarios is given by the limited communication bandwidth where neighboring agents are not capable of communicating continuously but only at discrete time instants. Limitations and constraints on inter-agent communication may affect any multi-agent network. Consensus problems, in particular, have been analyzed in the context of noncontinuous actuation and noncontinuous inter-agent communication. Several techniques are devised in order to schedule sensor and actuation updates. The sampled-data (periodic) approach [121–123], and [124] represents a first attempt to address these issues. The implementation of periodic communication represents a simple and practical tool that addresses the continuous communication constraint. However, an important drawback of periodic transmission is that it requires synchronization between the agents in two similar aspects: sampling period and sampling time instants, both of which are difficult to meet in practice. First, most results available require every agent to implement the same sampling period. This may not be achievable in many networks of decentralized agents and it is also difficult to globally redefine new sampling periods. Second, not only the agents need to implement the same sampling periods, but also they need to transmit information all at the same time instants. Under this situation each agent is also required to determine the time instants at which it needs to transmit relevant information to its neighbors. Even when agents can adjust and implement the same sampling periods, they also need to synchronize and transmit information at the same time instants for the corresponding algorithms to guarantee the desired convergence properties. Besides being a difficult task to achieve in a decentralized way, the synchronization of time instants is undesirable because all agents are occupying network resources at the same time instants. In wireless networks, the simultaneous transmission of information by each agent may increase the likelihood of packet dropouts since agents that are supposed to receive information from different sources may not be able to successfully receive and process all information at the same time.
Therefore, event-triggered and self-triggered controls for multi-agent systems have been considered for agents with limited resources to gather information and actuate. The event-triggered schemes allow each agent to only send information across the network intermittently and independently determine the time instants when they need to communicate [57]. The use of event-triggered control techniques for decentralized control and coordination has spurred a new area of research that relaxes previous assumptions and constraints associated with the control of multiple agents. In event-triggered control [125–130], a subsystem monitors its own state and transmits a state measurement to the non-collocated controller only when it is necessary, that is, only when a measure of the local subsystem state error is above a specified threshold. In general, the state error measures the difference between the current state and the last transmitted state value. The controller transmits an update by examining the measurement errors with respect to some state-dependent threshold and hence requires continuous monitoring of state error.