University of Strathclyde
Department of Mechanical and Aerospace Engineering
Verifiable Swarm Engineering
with Limited Communication
by
Giuliano Punzo
A thesis presented in fulfilment of the requirements
for the degree of Doctor of Philosophy
2013
Declaration of authors rights
This thesis is the result of the authors original research. It has been composed by
the author and has not been previously submitted for examination which has led
to the award of a degree. The copyright of this thesis belongs to the author under
the terms of the United Kingdom Copyright Acts as qualified by University of
Strathclyde Regulation 3.50. Due acknowledgement must always be made of the
use of any material contained in, or derived from, this thesis.
Acknowledgements
This thesis contains the work developed over 3 years of intense efforts. These efforts inevitably influenced my life beyond the research job and the one of people
around me. My first and more sincere thank you goes to those people, in particular to those not involved in this thesis directly, for tolerating my behaviour.
Elena, my Fiancée, standing always next to me in the crucial moments of my
PhD, proved how our life together will be bright and reliable in a more convincing way I did in this work about the future of swarming systems.
My family in Italy accepted my permanence in the United Kingdom not without
sorrow, but I hope I managed to convince them that was the best thing for myself.
Either way I owe them gratitude. I also want to thank those who keep on having my name on their close friend list without seeing me more than once per year.
To my supervisor Malcolm Macdonald I acknowledge an insane bravery for having allowed me the freedom of developing my research the way I wanted, even
deviating from what I suspect the original plans were. As he already knows this,
the only thing it is left to say is thank you. Derek Bennet knows how important
he was for the continuous support provided throughout three years of “wandering
in the science”; if he didn’t, then this acknowledgement will make him know.
I cannot forget the long coffee trail that linked my PhD to the one of Phil Karagiannakis, and the great work we managed to produce with the help of Stephan
Weiss, another person who doesn’t deserve to be neglected here. We are all looking forward to see where this trail will take.
A special thanks goes to Gordon Dobie, Charles Macleod and Rahul Summan sentinels and keepers of the secrets of the dungeon of the Electrical and Electronic
ii
Engineering department at Strathclyde University. I have to acknowledge them
and the whole CUE research group for experimental work developed in this thesis.
I would then like to thank Jules Simo for his inventive and original contribution
to our work on swarm fragmentation.
Naomi Leonard in Princeton will probably remember me every now and then
when thinking to the fast manoeuvring or when thinking to take a city break in
Glasgow. I aim to bother her together with George Young and Darren Pais again
with the latest old news about eigenvectors and avian flocks. For the time being
they fully deserve my gratitude and my best wishes.
Finally I want to acknowledge the Advanced Space Concept Laboratory and
the Department of Mechanical and Aerospace Engineering at the University of
Strathclyde that provided me with the support to develop my research.
Abstract
The new paradigms of swarm engineering, distributed architectures and autonomous multi-agent systems, are foreseen to redefine the way many engineering
problems are approached. The affirmation of these new concepts requires the
complete understanding of complex dynamics by the designers. That is, any
system whose concept departs from the monolithic architecture must deliver its
tasks in a predictable way and be controlled in a safe manner, while keeping the
maximum possible autonomy.
This work aims to span the gap between a complete foreseeable behaviour and
system autonomy using precise mathematical descriptions of the dynamics and
control of multi-agent systems. Dynamical system theory, Lyapunov stability,
linear algebra and graph theory are used to rigorously frame the problem and delineate the characteristics of such systems in relation to a number of applications
and performance parameters.
The work first considers multi-agent systems as multi-particle systems in a physics
fashion to draw fundamental results about the robustness to fragmentation when
the individuals do not benefit from all-to-all communications. The exploitation of
limited communications together with artificial potential functions is shown to be
an effective way to shape formations of agents in a range of applications for future
engineering, and in particular this scheme is proved to be effective for space-based
communications through the autonomous deployment of antenna arrays. In this
context, application to robotics is explored through laboratory tests exploiting
wheeled robots with possible applications to structural inspection or planetary
exploration. A stable fractal formation is proved to emerge out of a number of
agents whose interaction network presents a recursive layout whereby relative
motion is driven by artificial potential functions. Finally, the fast manoeuvring
problem is covered together with one of allocating resources in an efficient way
to track an external signal for the benefit of the group as a whole. Through an
algebraic approach, the tracking capabilities are distributed amongst the agents
producing advantages at group level for the tracking of an external signal. This
also translates into fast reaction to threats.
Contents
Contents
v
List of Figures
viii
List of Tables
xii
1 Introduction
1
1.1
Towards Decentralised Architectures . . . . . . . . . . . . . . . .
1.1.1 Biological Aspects . . . . . . . . . . . . . . . . . . . . . .
1.1.2 Technology Aspects . . . . . . . . . . . . . . . . . . . . . .
3
6
7
1.2
Self Organization, Group or Swarm Intelligence . . . . . . . . . .
1.2.1 Concept of Emergence . . . . . . . . . . . . . . . . . . . .
7
8
1.3
1.4
Challenges in modelling, simulating, implementing group Behaviours 9
From Particles Systems to Intelligent Cooperating Agents . . . . . 10
1.4.1 Stochastic Approach . . . . . . . . . . . . . . . . . . . . . 11
1.4.2
1.4.3
Deterministic Approach . . . . . . . . . . . . . . . . . . .
Graph Theoretic Methods . . . . . . . . . . . . . . . . . .
12
13
1.5
1.4.4 Behavioral Rules and Algorithms . . . . . . . . . . . . . .
1.4.5 Other Approaches . . . . . . . . . . . . . . . . . . . . . . .
Aims and Objectives . . . . . . . . . . . . . . . . . . . . . . . . .
15
16
16
1.6
1.7
Thesis Layout and Methodology . . . . . . . . . . . . . . . . . . .
Papers Authored . . . . . . . . . . . . . . . . . . . . . . . . . . .
17
19
2 Control Problem, Theory and Tools
2.1
21
2.2
Dynamical system theory . . . . . . . . . . . . . . . . . . . . . . .
2.1.1 System Energy and Hamiltonian . . . . . . . . . . . . . . .
Control Theory . . . . . . . . . . . . . . . . . . . . . . . . . . . .
21
22
23
2.3
Stability Theory . . . . . . . . . . . . . . . . . . . . . . . . . . . .
24
v
2.3.1
2.4
2.5
2.6
2.7
Linear Stability . . . . . . . . . . . . . . . . . . . . . . . .
26
2.3.2 Nonlinear Stability - Lyapunov stability . . . . . . . . . .
Graph Theoretic Methods . . . . . . . . . . . . . . . . . . . . . .
Nonsmooth Analysis . . . . . . . . . . . . . . . . . . . . . . . . .
27
29
33
Swarm Modelling and Control . . . . . . . . . . . . . . . . . . . .
2.6.1 Artificial Potential Functions . . . . . . . . . . . . . . . .
34
35
2.6.2 Connection Networks . . . . . . . . . . . . . . . . . . . . .
2.6.3 The Combined Effect of APFs and Connection Network . .
Schooling, Flocking and Swarming . . . . . . . . . . . . . . . . . .
39
43
44
2.7.1
2.7.2
Biological Studies . . . . . . . . . . . . . . . . . . . . . . .
Crystalline Patterns . . . . . . . . . . . . . . . . . . . . .
44
46
2.7.3
2.7.4
Continuous Swarming Systems . . . . . . . . . . . . . . . .
Consensus, a Common Problem . . . . . . . . . . . . . . .
49
50
3 The Cohesion of Emerging Patterns and the Emergence of Bot52
tlenecks
3.1
3.2
3.3
The Need for a Cohesive Swarm . . . . . . . . . . . . . . . . . . .
A Minimum Number of Links for Cohesion . . . . . . . . . . . . .
The Limit Case and the Cheeger Constant . . . . . . . . . . . . .
52
54
58
3.4
3.5
Effectiveness of the Model . . . . . . . . . . . . . . . . . . . . . .
Final Considerations . . . . . . . . . . . . . . . . . . . . . . . . .
65
69
4 Fractal Fractionated Architectures and Arrays - Application in
Space
4.1 Fractionated Spacecraft and Formation Flight . . . . . . . . . . .
4.1.1 Artificial Potential Functions . . . . . . . . . . . . . . . .
4.1.2
4.1.3
4.2
4.3
4.4
71
72
74
Adjacency Matrix . . . . . . . . . . . . . . . . . . . . . . .
Distributed Fractal Antennas . . . . . . . . . . . . . . . .
75
78
The Control Law and the Fractal Antenna . . . . . . . . . . . . .
4.2.1 Central Symmetry Emergence . . . . . . . . . . . . . . . .
4.2.2 APF Coefficient Definition . . . . . . . . . . . . . . . . . .
81
81
87
4.2.3
4.2.4
Stability of Control Law . . . . . . . . . . . . . . . . . . .
Fractal Antenna Array Design and Analysis . . . . . . . .
87
89
Simulation Results . . . . . . . . . . . . . . . . . . . . . . . . . .
Final Remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
90
97
5 Pattern Creation - Experimental Results
102
5.1
Real World Issues . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
5.1.1 Navigation . . . . . . . . . . . . . . . . . . . . . . . . . . . 104
5.1.2 Actuators . . . . . . . . . . . . . . . . . . . . . . . . . . . 104
5.2
Swarm Robotic Experiments . . . . . . . . . . . . . . . . . . . . . 106
5.2.1 Testbed and Methodology . . . . . . . . . . . . . . . . . . 106
5.3
5.2.2 Guidance and Control . . . . . . . . . . . . . . . . . . . . 109
5.2.3 Experimental Results . . . . . . . . . . . . . . . . . . . . . 115
Real and Simulated Real World . . . . . . . . . . . . . . . . . . . 119
6 Fast Consensus and Manoeuvring
123
6.1
6.2
The Need for Fast Manoeuvring . . . . . . . . . . . . . . . . . . . 123
Fast manoeuvring out of Network Structure . . . . . . . . . . . . 126
6.3
6.4
6.5
Eigenvectors and Eigenvalues in a Network System . . . . . . . . 127
Distributing Tracking Capabilities . . . . . . . . . . . . . . . . . . 130
A Good Choice with Limited Resources . . . . . . . . . . . . . . . 132
6.6
The Significance of the First Left Eigenvector and the Effects of
Scaling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137
6.6.1 Limited Resources and Small Perturbations . . . . . . . . 139
6.7
6.8
6.9
6.6.2 Fixed Trace, a Misleading Effect . . . . . . . . . . . . . . . 142
Solution of the First Order Linear Dynamics . . . . . . . . . . . . 143
6.7.1
6.7.2
6.7.3
The Regular Lattice Case . . . . . . . . . . . . . . . . . . 144
The Ring Case . . . . . . . . . . . . . . . . . . . . . . . . 145
The Small World Network Case . . . . . . . . . . . . . . . 146
6.7.4
6.7.5
The Random Network Case . . . . . . . . . . . . . . . . . 146
Performance of First Order System . . . . . . . . . . . . . 151
The Second Order and the Nonlinear Dynamics . . . . . . . . . . 152
Fast Manoeuvring - Final Considerations . . . . . . . . . . . . . . 160
7 Conclusions and Future Work
7.1 A Summary of the Topics Covered
7.2
7.3
163
. . . . . . . . . . . . . . . . . 163
Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 164
Future Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 166
Bibliography
170
List of Figures
1.1
1.2
1.3
(a) The game of life as example of cellular automata as reported
in Conway’s original work. (b) Beni’s original concept of cellular
robotic systems as illustrated in its work of 1988 . . . . . . . . . .
Examples of emergent patterns in nature . . . . . . . . . . . . . .
4
9
Illustration for the problem of the seven bridges of Königsberg and
the associated graph . . . . . . . . . . . . . . . . . . . . . . . . .
14
2.1
(a) The control scheme of a monolithic system and (b) the control
25
2.2
scheme of a distributed scheme. . . . . . . . . . . . . . . . . . . .
Time history for a single degree of freedom, second order linear
2.3
2.4
system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Original and modified Morse potentials . . . . . . . . . . . . . . .
Examples of potential functions . . . . . . . . . . . . . . . . . . .
28
38
40
(a) Graph representation of a small world network on 20 nodes
with in (c) the correspondent average path lengths. (b) Scale free
network on 60 nodes with in (d) its node degree distribution. . . .
42
2.6
2.7
Vortex Formation . . . . . . . . . . . . . . . . . . . . . . . . . . .
Distorted Formations . . . . . . . . . . . . . . . . . . . . . . . . .
48
49
3.1
The illustration shows how for k = hN/2i the swarm must be
connected . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Swarm systems in a two-dimensional space relaxing to different
56
2.5
3.2
shapes as the number of agents N increases while the number of
connections allowed per agent is held constant at k = 60. . . . . .
viii
57
3.3
Dumbbell emergence in a three-dimensional case. (a) A 130 agent
swarm in a dumbbell shape due to the number of connections per
agent being limited to 65 and (b) the corresponding adjacency
matrix obtained associating each node to one agent, after having
them sorted from one end to the other of the dumbbell. Dots
represent nonzero entries. . . . . . . . . . . . . . . . . . . . . . . .
3.4
3.5
3.6
3.7
58
Circulation definition . . . . . . . . . . . . . . . . . . . . . . . . .
Adjacency matrices for spontaneously relaxed (a) and idealised
connected (b) 60 agent swarm. . . . . . . . . . . . . . . . . . . . .
61
Adjacency matrices for idealised connected 61 agent swarm. . . .
Comparison of Cheeger constant obtained by the analytic expres-
64
62
sion and the one obtained by making a swarm of agents relax
through pairwise interactions for even number of agents. Average
values for each potential and extreme values obtained over all the
3.8
numerical tests are reported as well. . . . . . . . . . . . . . . . . .
Comparison of Cheeger constant obtained by the analytic expres-
66
sion and the one obtained by making a swarm of agents relax
through pairwise interactions for odd number of agents. Average
values for each potential and extreme values obtained over all the
numerical tests are reported as well. Some of the mamimum values
have been cut off the plot to make the curves more clear. . . . . .
4.1
4.2
4.3
4.4
4.5
4.6
67
Connections between groups of fully connected agents. . . . . . .
Adjacency matrix for an ensemble of 25 agents. The nonzero entries are represented by dots. . . . . . . . . . . . . . . . . . . . . .
76
77
Adjacency matrix for an ensemble of 125 agents. The self-similarity
of the matrix can be noticed. . . . . . . . . . . . . . . . . . . . . .
77
Node degrees as number of links belonging to each node. A selfsimilar scheme can be observed with nodes in central position being
the most connected ones. In this scheme the maximum number of
connections per node is 28. . . . . . . . . . . . . . . . . . . . . . .
Homogeneous square lattice array configuration . . . . . . . . . .
5 agents arranged in a homogeneous formation due to all-to-all
78
79
potential with the same coefficients for all the agents . . . . . . .
82
4.7
4.8
Contours for the potential sensed by agent in (0,0) in the case the
other agents have all the same value of the repulsive potential scale
length Lr (i) and in the case one of the other agents has a repulsive
scale distance Lr ′ < Lr . . . . . . . . . . . . . . . . . . . . . . . .
84
Cross pattern emerging by shrinking the repulsive potential scale
length broadcasted for the agent in the centre. . . . . . . . . . . .
85
4.9 First three stages of growth of the Purina fractal array . . . . . .
4.10 Directivity plots for the first three stages of Purina fractal arrays
shown in Figure 4.9, with an assumed direction of the main beam
90
towards broadside (θ = 0) . . . . . . . . . . . . . . . . . . . . . . 91
4.11 Groups of agents controlled at each step throughout the duty cycle. 95
4.12 Formation deployment in GEO. Snapshots taken at (i) t=0 s, (ii)
t=60 s, (iii) t=1 hr, (iv) t=24 hr . . . . . . . . . . . . . . . . . . 97
4.13 Errors in design positioning after 1 day from release of the formation 98
5.1
5.2
Wheeled robots used for swarming test. Not to scale . . . . . . . 107
One of the robots used in the tests. . . . . . . . . . . . . . . . . . 108
5.3
5.4
Testbed architecture . . . . . . . . . . . . . . . . . . . . . . . . . 109
The testbed arena where tests are performed, with Vicon T160
cameras in red. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110
5.5
A formation of 5 agents arranged in (a) pentagon shape due to
all-to-all potential with the same coefficients for all the robots and
(b) in a cross shape due to one agent being less “repulsive” with
respect to the others. Attitude is not considered. Formation is
defined just by the middle point of the wheel axle . . . . . . . . . 112
5.6
Trajectories followed by the 5 agents while arranging in a pentagon
formation first (stars) and then in a cross formation (circles). . . . 118
5.7
A formation of 5 agents arranged in (a) pentagon formation and
then (b) in cross formation. The pictures were extracted from the
video taken by one of the Vicon cameras. . . . . . . . . . . . . . . 119
5.8
Trajectories followed by the 5 agents while arranging in a cross
formation and rotating about its centre. . . . . . . . . . . . . . . 120
A formation of 5 agents arranged in (a) four point star and (b)
5.9
in a three point star that excludes an agent with a failure. The
pictures were extracted from the video taken by one of the Vicon
cameras. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
6.1
Eigenvalues of a negated Laplacian matrix according to the Ger-
6.2
shgorin circle theorem. . . . . . . . . . . . . . . . . . . . . . . . . 128
Sketch of the lattice network. Each node in the inner section has
3 outgoing and 3 incoming edges. Nodes at the left hand side have
more outgoing than incoming edges, while nodes at the right hand
side end have more incoming than outgoing edges. Note that an
6.3
outgoing edge implies that the node where the edge is originated
“observes” the node where the edge ends. . . . . . . . . . . . . . . 145
Consensus in a lattice, performance comparison. . . . . . . . . . . 147
6.4
6.5
Consensus in a ring, performance comparison. . . . . . . . . . . . 148
Consensus in a Small World network, performance comparison. . . 149
6.6
6.7
6.8
Consensus in a random network, performance comparison. . . . . 150
Manoeuvring in three-dimensional space. . . . . . . . . . . . . . . 158
Error time history with respect to the desired heading. . . . . . . 159
6.9 Detrimental effect of inter-agent forces in manoeuvring. . . . . . . 160
6.10 Detrimental effect of inter-agent forces on the error time history
with respect to the desired heading. . . . . . . . . . . . . . . . . . 161
List of Tables
2.1
2.2
Conditions for Lyapunov asymptotic stability. . . . . . . . . . . .
Examples of connected and disconnected graphs. . . . . . . . . . .
29
32
4.1
Numerical values of coefficients used in numerical simulations. . .
96
5.1
Numerical values of coefficients used in numerical simulations re-
5.2
ferred to a cross formation with a arm of 700 mm. . . . . . . . . . 115
Results of the static relative positioning tests for a cross formation
with a 700 mm arm. Tolerance and error are referred to the design
inter-agent distance . . . . . . . . . . . . . . . . . . . . . . . . . . 116
6.1
Graph 2 example . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
6.2
6.3
Graph 1 example . . . . . . . . . . . . . . . . . . . . . . . . . . . 136
First eigenvalues of the system matrix (L + D) where, L is the network graph Laplacian and D is a diagonal matrix whose nonzero
entries are either the first left eigenvector (F. L. E.), a uniform
vector or a vector numerically optimised to maximise the corre-
6.4
6.5
spondent eigenvalue. All of these have unitary L1 norm. . . . . . 151
Coefficients used in the APF for the three dimensional manoeuvring simulations. . . . . . . . . . . . . . . . . . . . . . . . . . . . 157
Desired velocity vectors for the swarm to produce the changes in
heading. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 157
xii
Chapter 1
Introduction
This chapter deals with the different aspects of aggregation of individuals. It
separates the case of groups of individuals from structures composed of cooperating elements, and the case of spontaneous aggregation in biological groups from
the emerging aggregation in engineering applications. It presents the concepts
of swarm engineering and swarm intelligence, defines the concepts of emergence,
illustrates some emerging behaviours and stresses their advantages and the way
they can be possibly exploited in engineering or other scientific disciplines.
Definitions
The engineering systems which are the subject of this dissertation are, in turn,
composed of engineering systems themselves. This characteristic makes the multiagent systems susceptible to confusion. Terms such as units are used to indicate
parts of a more complex device, yet they may be autonomous on their own and,
in most cases, they are engineered devices themselves. In order to avoid misunderstandings due to terminology as much as possible, a glossary of possibly
misleading terms is provided as an aid.
- The term part refers generically to elementary entities, or groups of them,
which belong to a set of interest.
- A system is a set to which several parts belong.
- A component is a part of a system. Components and can be either functional
and dependent on one another, or autonomous and independent.
1
CHAPTER 1. INTRODUCTION
2
- A unit is an individual entity that does not constitute a system on its own,
at least not one of our interest, but is a part of a system composed of
interacting components.
- A multi-agent system is an engineering system composed of a number of
autonomous, yet interacting, units.
- An agent is a unit that is capable of autonomous behaviour.
- The term vehicle is used in place of agent when the attention is put on the
mobility characteristics of it.
- The term swarm is used to indicate a multi-agent system characterised by
relative motion of its agents. Despite its biological connotation, it is used
to address either biological and non biological groups.
- The term ensemble is used as synonym of group of units. Only if this
units are autonomous, hence are agents, ensemble is a synonym of swarm.
Throughout this dissertation the term ensemble does not have any statistical connotation.
- The term architecture is used instead of system to stress the characteristic
of being designed with particular emphasis on the connections of its parts
and their interplay.
- The term distributed indicates the characteristic of multi-agent systems of
performing some tasks in absence of central coordination.
- The term state indicates a set of time dependent parameters that characterise the units and, hence, the system. In the most common use through
the dissertation it indicates position and velocity, but other parameters,
such as internal energy, may be included in general.
Obviously this glossary does not complete the set of terms that are used in the
following, yet it provides some reference to which this work will remain coherent
in its development.
CHAPTER 1. INTRODUCTION
1.1
3
Towards Decentralised Architectures
The idea of achieving challenging tasks through the cooperation of elementary
units, rather than through a dedicated, centralised system, is relatively new in
engineering. Natural science, however, has been looking at the dynamics arising
from the interaction of a number of elementary parts since the mid-20th century
[1]. This particular class of problems has been analysed through in the fields
of chaos, complexity and collective behaviors within disciplines such as particle
physics, mathematics, biology, sociology and many more. The behavior of such
systems cannot be explained by looking at only the single components, instead
the interplay amongst these single components must be considered [2]. While science is in general interested in modelling those systems belonging to the natural
world, to get some characteristics and predict their behaviour, engineering tackles
the problem from the opposite angle: it models and designs the system such as to
get a desired behaviour. In this sense the design of complex systems starting from
the autonomous units that form them is awkward. Biology and science in general
have been fascinated by some collective behaviours and have studied them for
over a century [3]. Conversely, in engineering, the idea of multiple, autonomous
units designed to cooperate and to achieve a result through their interaction is
more recent. The seminal work of John von Neumann about Cellular Automata
in the pioneering era of computer science can be considered the first step towards
the shaping of a new branch of engineering. Cellular automata (CA) are cells
packed in arrays capable of processing information [4]. Each element, (automaton) is characterised by a finite number of states, specifically “ON” and “OFF”,
and it switches between them depending on the information it receives from the
neighbouring units. The switch happens synchronously at discrete time intervals;
the result is the creation of evolving patterns in the array. The popular Conway’s
Game of Life is simply an application of CA [5]. Cells in an array are turned
on when there is a “birth” event, kept on if they “survive” and turned off when
they “die” as Figure 1.1(a) illustrates. In the figure it can be seen how each cell
is adjacent to 8 others in the array. At each time interval, also called generation,
the “birth” of a new cell happens wherever an empty cell is adjacent to 3 existing
ones, while any living cell dies whenever it is adjacent to less than 2, or more than
3 existing others. The patterns propagate from the initial state corresponding
to the first column to successive states. As the rules for “birth”, “survive” and
“die” events are given, then the outcome is determined univocally by the initial
4
CHAPTER 1. INTRODUCTION
configuration.
CA allowed for a rigorous mathematical description of their dynamics but present
limits in the mainly theoretical framework in which they develop and their self
contained dynamics. The information on the neighbour states is processed by the
Automaton and returns its state which is in turn another piece of information
for the other neighbouring CA. This limitation was pointed out and overcome, at
least in a theoretical framework, by Gerardo Beni in 1988 [6]. Beni introduced
the concept of “Cellular Robotic System” (CRS). CRS are different from CA as
the units, although always arranged on a grid, are able to move and operate physically on the surrounding environment. In Figure 1.1(b) the grid arrangement is
visible together with the capability of the CRS to operate on the environment by
moving objects in the grid.
The leap forward provided by CRS with respect to CA can be summarised in the
possibility of processing information and matter as well. This is the characterising feature of robotic systems and, with its multiple agent architecture, the CRS
can be considered the first example of “swarm robotics”.
The term “swarm” in this context was first proposed at “Il Ciocco” conference
while attempting to find a name that could refer to the CRS just presented at
(a)
(b)
Figure 1.1: (a) The game of life as example of cellular automata as reported in [5]. (b)
Beni’s original concept of cellular robotic systems as illustrated in its work of 1988[6].
The picture refers to the case of moving objects in the field.
CHAPTER 1. INTRODUCTION
5
the same event in an appealing way. An immediate link to something popular
such as the biological swarms was considered a key for the success of that new
idea as reported by Beni himself [7]. Indeed, the concept of agents designed to
achieve a global task through local interactions only has several similarities to biological swarms. The number of agents involved is too large for an approach that
considers each of them singularly, and too small for applying statistical methods.
Moreover, all the agents are quasi-identical and autonomous so that a decentralised control scheme emerges naturally. Finally, a consequence of the shift
from Automata to Robotic Agents is the possibility of exploiting “stigmergy”.
Stigmergy is the capability of an agent of passing and processing information by
altering the surrounding environment. It is well know how ants exploit this by
releasing pheromone along the trails from the nest to the food sources. This way
the colony is able to trace the trail back to the food source [8]. CRS are agents
with robotic characteristics, that is to say, they are able to process matter and
hence to operate on the environment. Due to this, CRS are potentially able to
exploit stigmergy. Although the name “Cellular Robots” was used by Fukuda [9]
a few months before the seminal paper of Beni, yet Beni was the first to spot
the potential of exploiting the collective behaviour and hence to start the swarm
robotics. One year before the CRS was presented to the scientific community,
Craig Reynolds, in a context of computer graphics, presented an algorithm to
reproduce the flocking behaviour of birds [10]. The algorithm was based on three
simple rules applied locally for each virtual bird, he called “boid ”. These are:
Separation: each boid shall keep a minimum distance from any of its neighbours;
Cohesion: each boid shall stay within a given distance from any of its neighbours;
Alignment: each boid shall align its velocity to that of its neighbours. The neighbours of each boid were its flock mates within a given sensing distance. As simple
as it was, Reynolds’ flock model, not only was able to reproduce qualitatively the
behaviour of birds flocking, but also was a source of inspiration for the scientific
community that in a few months would have been introduced to swarm robotics.
Initially, the local interactions in groups of swarming robots were defined exploiting heuristic rules, in the same way as Reynolds’ rules operate. Nonetheless, the
engineering of such a new concept could not remain abstract from the certainty
of behaviour for long. By its own nature, engineering needs to rely on certain
rules and its product must deliver what it is designed for, reliably and in safe
CHAPTER 1. INTRODUCTION
6
conditions, regardless the level of autonomy it is provided with. In this respect
a system where the components are designed not to directly deliver the task,
while the totality does so by exhibiting a behaviour different from the one of the
components, presents several hazards for which verification is needed. From the
contrast between the level of autonomy of a swarm robotic system and the need
for predictable behaviour of an engineered product, Alan Winfiled et al. developed the most modern concept of Swarm Engineering [11]. In their work Winfield
et al. point out how verification of a swarm system, from the engineering point
of view, should include verifying both the single component level and the group
level behaviours for the liveliness and safety. These characteristics correspond to
the property of exhibiting a desirable behaviour and avoiding any undesirable behaviour. Verification of the behaviour should then rely on mathematical proofs.
Hence the “Direct Lyapunov Method” is identified as a way to prove stability of
the control system at agent level and statistical methods to model the evolution
of the swarm and hence prove that it is stable for each configuration it is likely
to take.
In a swarm where internal interactions are based on neighbouring, and where,
by definition, internal local interactions produce effects on the state at global
level, the evolution of the neighbouring must be considered ahead of any stability
analysis. On the other hand, predicting all possible neighbouring states is not
feasible from an engineering point of view for a system that processes information,
decides and performs in real time. The system is then to be designed such that,
whatever the relative states of the single agents are, it results in a stable and
predictable behaviour that exhibits the characteristics of liveliness and safety.
This is the starting point of this dissertation.
1.1.1
Biological Aspects
Social aggregation is a characteristic found in many living organisms in nature.
Wolves developed strategies for hunting in packs that have been found to be based
upon stigmergy [12]. Even elementary organisms get benefit from aggregation:
bacteria, that are normally found in colonies, benefit from this property in a
number of situations such as foraging and increasing their resistance to antibiotics.
In doing this bacteria are able to process and transfer information amongst them
not only in a diffusive way, but establishing an efficient network of connections
CHAPTER 1. INTRODUCTION
7
that also in some cases presents features of hierarchical organization [13]. Finally,
popular behaviours that are worth mentioning are those exhibited by fireflies and
ants. Fireflies synchronise their flashing in order to make the signal stronger for
potential mating partners [14], while ants construct social interactions exploiting
stigmergy to locate and optimise the foraging path [8].
1.1.2
Technology Aspects
Multi-agent systems are becoming more and more popular for intelligence and
military applications, see for example [15], mainly because of their flexibility in
adapting to fast changing scenarios and robustness against failures or attacks
[16]. Integrated systems that include space, aerial and ground based segments
have been suggested as well for search and rescue or military applications although
these are far from being realised [17]. More advanced projects are ongoing for
tasks such as assisting human operators, firefighters in the case studied in [18], or
structural or machinery inspection as analysed in [19]. One of the few operative
examples of multi-agent system is the Onyx parachute system [20] that is able
to operate a number of parachutes to land multiple payloads launched from a
cargo aircraft using swarming algorithms for collision avoidance and precision
landing. Very popular projects, now in their test phases (see for example [21]
and the SARTRE project [22]), concern the self-driving vehicles which can be
considered as an application of multi-agent systems. These vehicles do not need
to be designed to collaborate actively, but operate simply considering each other
as non collaborative obstacles. In this sense even just the introduction of an
autonomous vehicle in the normal traffic makes it comparable to a multi-agent
system.
1.2
Self Organization, Group or Swarm Intelligence
The definition of the concept of intelligence is beyond the scope of this thesis, nor
is of interest of that part of literature that this work refers to. Beni [7] pointed
out that one of the characterising aspects of an intelligent behaviour is producing
an improbable outcome, something ordered, hardly achievable simply by chance.
The other is the impossibility of predicting it, i.e. it must be something original,
CHAPTER 1. INTRODUCTION
8
different from the output of any non-intelligent machine such as a printer, that
will always output something more tidy than ink stains but yet it will always
print what it is commanded to print. Nevertheless, according to Beni, an intelligent swarm can be formed out of non-intelligent machines.
Strogatz [23] pointed out how ordered behaviours and in particular synchronization are emergent phenomena in nature for which just small amount or even no
intelligence is needed. Mechanical coupled metronomes were observed to sync
as well as people clapping in a theatre do, although not previously trained for
that. Of course intelligence, in the most common meaning of the word, is not a
characteristic of the metronomes. It can be argued that not only a certain kind of
intelligence, often identified as swarm intelligence, generates emerging behaviour,
but that the swarm intelligence can be considered an emergent characteristic
itself.
1.2.1
Concept of Emergence
The word “emergence” was used already in this introduction. It refers to the
process that leads to self organization of a system in its macroscopic properties
through micro-scale interactions. According to Goldstein, it can be defined as:
“the arising of novel and coherent structures, patterns and properties during the
process of self-organization in complex systems” [24]. The patterns arising in the
Conway’s Game of Life and the corresponding behaviour in the Beni’s Cellular
Robotic Systems are examples of emerging behaviours. Indeed, as for swarm intelligence, emergent behaviours do not need processing to take place. Snow crystals
form out of interactions of water molecules and a number of configurations are
possible depending on the thermodynamic characteristics of the surrounding environment [25], yet no water molecule does anything but acquiring a minimum
energy state. Fish schools are often found forming swirling vortices as result of
keeping the same relative positions and orientation with their local neighbours
while swimming [26]. Moreover, Parrish et al. noted that particular configurations, such as whorls, funnel or tori, can quite often be found in nature as results
of abiotic interactions as well; popular examples include galaxies, hurricanes and
tornadoes. Some cases are illustrated in Figure 1.2.1. In many respects, the
natural world relies on emergent behaviours and its dynamics, at different scales,
resemble very closely a natural cellular automata. One of the most striking examples is the case of the self-regulation of gas exchange in plants, consequence of
CHAPTER 1. INTRODUCTION
9
Figure 1.2: Examples of emergent patterns in nature: a satellite image of an hurricane
in the Atlantic Ocean (NASA), a fish school forming a vortex (National Geographic)
and a snowflake from [25].
a distributed calculation perfectly reproducible by a CA [27].
1.3
Challenges in modelling, simulating, implementing group Behaviours
The wide variety of group behaviours observable in nature and achievable in engineered swarming systems increases the difficulty of studying these systems with
a rigorous scientific approach. On one hand the multitude pushes the analyst
towards finding a common ground on which a critical analysis can be carried out,
on the other the features of the swarming agents so deeply characterise the system that it is impossible to disregard them. The motion constraints, the sensing
capabilities and the environmental conditions make, for instance, a flock of birds
exhibit dynamics different from those of a school of fish; similarly a swarm of
wheeled robots will have completely different dynamics from one of Unmanned
Aerial Vehicles (UAVs). A common approach to modelling the basic features of
most swarming system is considering them as particle systems (see for example
[28–32]). In this case the agents are considered as physical particles moving in a
force field that can be generated by the reciprocal interaction, the surrounding
environment, or both. The evolution of the system is then described through
the dynamical system theory. Considering agents as particles, statistical methods are applicable and particularly suitable for large swarms [33, 34]. In a very
similar physical-like fashion, some researchers exploited entropy measurements to
capture the emerging ordered state [35, 36] that has occasionally created some
confusion within the literature as to the physical meaning of the entropy [35, 37].
These approaches often do not consider any constraints on the agents, yet they
can be claimed to be valid modelling strategies for particular applications such
CHAPTER 1. INTRODUCTION
10
as underwater or space vehicles [38], or to analyse emerging behaviors on their
own (see for example [30, 32]).
When the number of agents is high, modelling them as particles, all driven by the
same force field, is unrealistic. While particles exchange forces, agents exchange
information and produce forces as control actions. As different from forces, information need dedicated communication channels, which are limited in number.
To account for that, the presence of a communication network is considered in
this dissertation consistently with the relevant literature in the field of network
modelling. In this framework graph theoretic methods are used to study networked systems through rigorous mathematical tools (see for example [39–42]).
When peculiar aspects of the agents are considered, such as particular interaction
networks or motion constraints, the modelling looses its universality. It is then
important to find a balance between looking at a wide horizon and focussing on
few specific cases.
In this dissertation the principal methodologies found within the literature will
be reviewed, however the approach taken goes in the direction of preserving the
generality of the system rather than deeply characterise it. Laboratory tests are
then used to confirm part of the theoretic outcomes.
1.4
From Particles Systems to Intelligent Cooperating Agents
The term swarm is found in several contexts. It can refer to the most popular
biological meaning or to particle systems animated by internal motion, or indicating a multi-agent engineering system. This variety of fields has produced a
number of approaches to the swarm modelling and control problem, each capturing different aspects. While statistical methods apply better when swarms are
composed of a very large number of units, they are often considered not effective from a design point of view. Here the various ways of approaching swarm
behaviours are reviewed considering their significance in different fields. The
number of assumptions and the level of detail that make a model more peculiar
to a given phenomenon must be balanced against the extent of its applicability.
CHAPTER 1. INTRODUCTION
11
Hence, modelling self propelled particles, able to move freely in the space, has
more comprehensive features than a model that considers unit actuation capabilities and environmental constraints, but the latest returns more valuable results
when a specific application is targeted.
1.4.1
Stochastic Approach
Statistical physics has been extensively used to model the performance of mobile
agents in large groups. Considering an agent as a self-propelled particle, it is
possible to study the evolution of the particle system state on a statistic basis.
Despite the appealing characteristics and the wide range of possibilities opened
by the stochastic methods applied to swarm systems, in the present thesis these
methodologies are not considered as a completely deterministic approach is instead preferred. The statistical physics approach is more valid as more particles
are considered.
Schweitzer [33] considers some emergent behaviours in particle swarms and relates
them to system thermodynamic properties. This way, rather complex behaviour
can be traced back and linked to some global properties of the system, such as
energy, and their rate of change. The presence of noise is considered as well and
the system state variables are described in terms of their statistical distribution.
Ebeling et al [34] [43] consider the state of a particle swarm subject to thermal
noise in terms of its variable distribution in statistical terms. The velocity distribution across the swarm and its mean square displacement is derived. In this, as
well as in similar works, results from numerical simulations match the predictions
made by means of statistical physics tools when the number of particles in the
swarm is reasonably high to provide the data with statistical significance.
Martinoli [44] considers a different approach by modelling the evolution of a number of coordinated robots with the task of gathering some targets. The robots’
neighbourhood is defined on a statistical base. A series of stochastic events are
triggered, each depending on the success of the previous one. The first event
consists in assigning each robot a random position which also corresponds to assigning a neighbourhood. This includes the target position and the positions of
the other robots in the field, with respect to the first one. If in its position the
CHAPTER 1. INTRODUCTION
12
robot finds the object to catch, then the second event is triggered, that is the
gathering and so on. By showing good matching between the performance of real
gathering robots and the outcome produced by the series of stochastic events,
Martinoli argued that problems such as gathering were easily tractable on a statistical base. This is to say that considering the odds, from time to time, that a
certain event useful towards the accomplishment of the gathering task happens,
is sufficient to predict the success or the failure of the task and in which timescale.
This approach rules out, in theory, some simulation campaigns based on physical
modelling of robots and environment that usually require several repetitions to
account for all possible variables.
An original approach to control problems with a stochastic perspective has been
considered by linking a certain behaviour of a robot in a group of agents to a probability distribution influenced by the external environment. That is, each agent
has a probability distribution associated to all possible behaviour it can pick,
with the expected outcome corresponding to the most appropriate behaviour in
response to the external stimulus as in [45]. This way the possibility of an agent
not picking the correct behaviour is accounted on statistical base.
For a more extensive review of the stochastic methods specifically applied to
robotics the reader can refer to [46] where extended information is provided.
1.4.2
Deterministic Approach
An engineering approach to the swarm problems requires reliable and verifiable
behaviour. This means that proving a given system exhibits some global behaviours, either described through the probability distribution of its states or
through the probability certain events will occur, is not always considered satisfactory from the verification point of view. As it was already pointed out in
Section 1.1, the system should respond in an intelligent way, that is, showing both
liveliness and safety characteristics in every circumstance. For this reason many
authors looked at a more deterministic approach, where the group behaviour can
be studied both at global and at agent level. In particular, the group level is addressed by constructing an attractor for a system of differential equations, each
of them describing the dynamics of a single agent.
CHAPTER 1. INTRODUCTION
13
McInnes proposed the Artificial Potential Function (APF) method, which is extensively described in Section 2.6.1, in the context of space formation flying as
an elegant way to coordinate the deployment of several spacecraft in a given formation, where the final arrangement is guaranteed through the minimization of
an energy function [47]. Bennet [38, 48–50] exploited the APF method and the
bifurcation of the equilibrium to produce a pattern transition between two formations in swarms of autonomous mobile agents. The approach chosen in this case
is not confined to a particular vehicle architecture, although the applications to
spacecraft and unmanned aerial vehicles (UAVs) are considered with real world
issues coming into play such as actuator dynamics and saturation. Mabrouk
[51–53] exploited provable vortex behaviour to drive a swarm of robotic agents
capable to overcome some local minima problems, which are peculiar of the APF
approach, by relying on their internal state. Using similar methodologies and a
focus on space applications, Badawy [54, 55] implemented hyperbolic functions
to control vehicles flying in formation and for on-orbit assembly. In a comparable
way, but exploiting more geometrical rather than energetic considerations, Spears
and Gordon proposed the Artificial Physics method for the control of a formation
of autonomous robots [56, 57].
Leonard [58, 59] proposed an approach for the coordination of groups of vehicles
based on Lie algebra and Lyapunov stability. The latter will be addressed in detail in Section 2.3.2 as it is used to draw some of the results of this thesis. These
studies, together with other ones about motion synchronisation (see for example
[60–62]) were fundamental for the deployment of a fleet of underwater vehicles in
Monterey bay, California, USA, for ocean sampling purposes [63–65].
1.4.3
Graph Theoretic Methods
Graph theory is extensively used when dealing with multi-agent systems and
swarms as a powerful tool for their modelling and analysis. Graph Theory studies the mathematical objects called graphs. A graph is a pair G{V, E} where V is
a set of vertices and E is a set of edges. More rigorous definitions and property
descriptions are provided in Chapter 2, while in the remaining of this section a
general overview of the graph theory practical utility is presented.
14
CHAPTER 1. INTRODUCTION
The first example of graph study in the modern sense can be found in the 1736
paper by Leonhard Euler on the seven bridges of Königsberg problem [66, 67].
Euler’s work focussed on the problem of crossing all seven bridges of the city of
(a)
(b)
Figure 1.3: (a) Illustration for the problem of the seven bridges of Königsberg (source
[66]). (b) Graph associated with the problem
Königsberg in a single trip crossing each of them only once. This is represented
in Figure 1.3 where the city areas are modelled as nodes of the graph and the
bridges as edges. This makes the problem tractable in a rigorous way excluding
from it the information about the path to follow within the city areas.
A seminal contribution to the modern Graph Theory was given by Paul Erdos
with his model of random graphs. These particular graphs are constructed from
a group of nodes that are linked to each other in a random way. Depending on
the number of nodes and edges, it is then possible to study the characteristics of
the network from a statistical point of view [68–70]. The works by Erdos were
of fundamental importance for the study of complex networks and, more refined,
later models largely benefited from them. In particular the model by Barabasi
CHAPTER 1. INTRODUCTION
15
and Albert [71] and the model proposed by Watts and Strogatz [72] departed from
the complete randomness of edge structuring. Both the models are widely used
in works concerning multi-agent systems as they provide robustness and short
average paths between nodes compared to lattice networks respectively. Moreover both have been found to be representative of several biological and social
systems and able to capture some complex behaviours emerging from these.
In multi-agent systems graph theory provides a means to track the network of
connections amongst the agents and to extract its properties, such as its connectivity, which determines the stability of the system and its response as it will be
explained in Section 2.4. In particular it is possible to model a multi-agent system
as a graph whose nodes are the single agents and whose edges are the connections between any pair of them. The distance between any two agents can then
be accounted as the number of edges through which they are connected (provided
that such a path exists) and hence it is possible to predict the capability of the
system to express coherent behaviour and its dynamical properties. This kind of
tracking of system characteristics has also been performed in experimental tests
as in [73]. A line graph coupled with inter-agent potentials is at the base of the
control scheme that produces the ”Tetunagi-oni ” game described by Yamaguchi
and Arai [74]. In this a number of agents, linked by pairwise communication in
a line graph, encircle a single agent not belonging to the line graph. Connected
agents produce an always narrower ark closing around the isolated one. This
behaviour emerges out of the connection scheme and the control laws that are
enforced in it. Graph theory can be used, beside to track the interactions in a
group of agents, as control means itself. Controllability in this sense was investigated in [75], while the problem of converging to consensus, that is to a common
behaviour, in absence of external drives, is extensively covered in [76].
1.4.4
Behavioral Rules and Algorithms
A popular approach to control problems in robotics in general is the use of behavioural rules; this approach is commonly found also in swarm robotics. When
behavioural rules are in place, agents are driven by instructions of the kind ”if...
then...”. Although the results obtainable are remarkable in terms of coordination
and effectiveness in task achievement, the performances of the robots cannot be
described in rigorous mathematical terms, hence a proof of the system responding
CHAPTER 1. INTRODUCTION
16
positively to the requirements of safety and liveliness is not obtainable. Reynolds’
work, described in Section 1.1, can be considered a precursor of behavioural rules
applied to swarm robotics. Lewis and Bekey [77] proposed this behavioural control method to be applied to a colony of microrobots for surgical purposes and
supported their claims through numerical simulation on a grid in a discrete time,
which mimic the CA approach described in Section 1.1. Alternatively this kind of
approach can be validated through extensive simulations and testing campaigns
as in [78]. A mathematical validation of a behavioural algorithm is given in [79]
although no proof is provided. Argumentations based on an approximated mathematical model and numerical simulations support the claim of the validity and
effectiveness of the methodology proposed.
1.4.5
Other Approaches
The examples mentioned are not exhaustive of the extensive literature dealing
with multi-agent systems, yet they provide an idea of the general approach to
this kind of problem from an engineering research perspective. More extended
analyses of some of the methodologies and the approaches taken are provided
later, when relevant to the topic covered. For a more general overview about
swarm robotics literature, which is beyond the scopes of this work, the reader
can refer to the reviews in [80–82].
1.5
Aims and Objectives
This work focuses on controlling a swarm of agents in analytically provable ways.
This is achieved in this dissertation through the use of graph theoretic methods
and artificial potential functions. The work aims to develop a new approach
to distributed control of swarms considering the reduction of the communication
links as a way to enhance the control rather than a limitation. This looks towards
the practical implementation of autonomous swarms where the ideal all-to-all
communication network is often infeasible. In particular this thesis investigates:
♦ The influence of limited communication on the onset of fragmentation in
swarms of autonomous agents;
♦ The leveraging of limited communication to shape a formation of autonomous agents;
CHAPTER 1. INTRODUCTION
17
♦ The issues connected with the implementation of artificial potential function
based methods to real robotic agents;
♦ The connection between the sensing network and the fast maneuvering performance in biological swarms together with the possibility of applying it
to engineered agents;
♦ The exploitation of swarming systems to real world applications such as
space-based telecommunication arrays, structural inspection and defence.
Considering the real world implementation, the application of the theoretical control algorithms produced to real robots will also be investigated.
A clear contribution to knowledge and towards the affirmation of swarm systems
is presented here through meeting practical control problems with analytically
provable methodologies.
1.6
Thesis Layout and Methodology
This thesis presents a quite extended first part over the first two chapters aimed
to introduce the current state of research in swarm engineering and control but
also to introduce the different fields which are touched within it. In this first part
the general theoretical background for the methodologies used later is presented,
clarifying aspects such as emergence, provable behaviours and consensus. The results achieved are then presented from Chapter 3 to Chapter 6. Conclusions and
future developments in the last chapter then contain the author’s observations
on the contents presented.
This chapter has outlined the complexity of the problems related to the interaction of many autonomous units. The challenges presented by an always expanding
technology paradigm were presented noting how the balance between an autonomous and a foreseeable behaviour is to be searched by relying on analytically
provable methodologies. This would allow multi-agent systems to be truly autonomous with the resolution of the technology gap that separates us from this
new engineering.
CHAPTER 1. INTRODUCTION
18
The approach based on considering swarming agents as particle systems is discussed. This will be used often in this dissertation but, as different from some
particle physics works, the methodologies adopted here will not make use of
stochastic techniques. Methodologies more widely applied in engineering are introduced as well as graph theory, which will be exploited extensively in this thesis.
The approach used is then detailed in Chapter 2 which presents the mathematical tools that are exploited in this work. Emphasis is put on the application of
control theory as well as on way these tools apply to swarm control. The control
methodologies used are presented in detail and reference to appropriate literature
is made marking the point from which this thesis departs to produce the original
achievements stated in Section 1.5.
In Chapter 3 the problem of cohesion in a swarm with an incomplete communication graph is considered. The first result of this thesis is drawn by reversing the
problem from a graph theoretical perspective. While a swarm is required to be
connected in order to be controllable, here conditions to have it connected at all
time are drawn avoiding the problem of fragmentation and presenting a measure
for the weakness of the swarms with respect to fragmentation.
While Chapter 3 presents valid results with a wide theoretical applicability, but
does not target any particular application, in Chapter 4 the design of the communication graph in a swarm is used to produce a space based architecture aimed
to telecommunications. Exploiting the emergence of a recursive shape and the
resonances that this produces when composed of radiating elements, the concept
of a distributed fractal antenna is presented. This is done within a control theory
framework aiming to produce a formation flying architecture.
Chapter 5 considers the application of the control laws produced onto hardware.
A coherent behaviour emerges out of the actions of autonomous wheeled robots
controlled by the same means used in Chapters 3 and 4, proving their flexibility
in being transferred onto real hardware and adapting to some specific challenges.
Chapter 6 presents the results obtained in terms of consensus achievement and
its consequences on the fast manoeuvring of swarms of mobile vehicles. A theoretical study is presented and supported by numerical simulations that point to
CHAPTER 1. INTRODUCTION
19
the algebraic characteristics of the communication network to improve the swarm
performance in terms of reactivity.
Conclusions and possible future developments for the results presented are finally
detailed in Chapter 7.
1.7
Papers Authored
Giuliano Punzo, Derek J. Bennet, Malcolm Macdonald. Swarm shape
manipulation through connection control. In: Towards Autonomous Robotic
Systems 2010. Lecture Notes in Artificial Intelligence, 6856 (1). Springer.
Giuliano Punzo, Derek J. Bennet, Malcolm Macdonald. Enhancing selfsimilar patterns by asymmetric artificial potential functions in partially
connected swarms. In: 11th Conference Towards Autonomous Robotic Systems 2011. 2011-08-31 - 2011-09-02, Sheffield, UK. Lecture Notes in Artificial Intelligence, 6856 (1). Institution of Engineering and Technology.
ISBN 978-3-642-23231-2.
Giuliano Punzo Fractal patterns in fractionated spacecraft. In: 62nd
International Astronautical Congress 2011, 2011-10-03 - 2011-10-07, Cape
Town, South Africa.
Giuliano Punzo, Derek J. Bennet, Malcolm Macdonald. A fractally fractionated spacecraft. In: 62nd International Astronautical Congress 2011,
2011-10-03 - 2011-10-07, Cape Town, South Africa.
Gareth Pierce, Giuliano Punzo, Gordon Dobie, Rahul Summan, Charles
N. MacLeod, Colin McInnes, James Biggs, Malcolm Macdonald, Derek J.
Bennet. Reconfigurable robotic platforms for structural health monitoring. In: 6th European Workshop on Structural Health Monitoring & 1st
European Conference of the Prognostics and Health Management (PHM)
Society, 2012-07-03 - 2012-07-06, Dresden.
Punzo, Giuliano, Gordon Dobie, Derek J. Bennet, Jonathan Jamieson,
Malcolm Macdonald. Low-cost, multi-agent systems for planetary surface
exploration. In: 63rd International Astronautical Congress, 2012-10-01 2012-10-05, Naples, Italy.
CHAPTER 1. INTRODUCTION
20
Giuliano Punzo, Philippos Karagiannakis, Derek J. Bennet, Malcolm
Macdonald, Stephan Weiss. Enabling and Exploiting Self similar formations. IEEE Transaction on Aerospace and Electronic Systems. 2013 in
press.
Giuliano Punzo, Jules Simo, Derek Bennet, Malcolm Macdonald. Characteristics of Swarms on the Edge of Fragmentation. In preparation for
resubmission to Physical Review E.
Philippos Karagiannakis, Stephan Weiss, Giuliano Punzo, Malcolm Macdonald, Jamie Bowman, Robert Stewart. Impact of Purina Fractal Array
Geometry on Beamforming Performance and Complexity. 21st European
Signal Processing Conference. 2013-09-09 - 2013-09-13, Marrakech, Morocco.
Chapter 2
Control Problem, Theory and
Tools
In this chapter the tools for the design, modelling and control of multi-agent
systems, that will be used in this dissertation, are illustrated. The two main
theoretical pillars upon which the mathematical foundations of this thesis are
developed are dynamical system theory and graph theory. These will be detailed
in the remainder of this chapter. Moreover, these tools will be linked to modelling
and control of swarms, linking them to different behaviours observable in nature
and reproducible through engineering systems.
2.1
Dynamical system theory
Dynamical system theory is at the base of the analytical modelling of multi-agent
systems regardless of whether the agents considered are particles, biological or
engineered units. It provides the tools to deal with N-agent systems regardless
their size and the physical space considered. The system dynamics are modelled
through differential equations, with partial derivatives in the most general case.
Given a system, its state is identified by the value associated to a number of variables, namely the state variables which are associated with its degrees of freedom.
To make this clear, let’s consider an N-particle system moving on a plane, hence
a two dimensional space; consider furthermore that the particles undergo acceleration depending on their positions and speed. This dynamical system has 2 × N
degrees of freedom as each particle can move in the plane and a state composed
of 2 × 2 × N variables as each particle state is characterised by its position and
21
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
22
velocity, which are in turn composed of two components due to the planar motion.
In this dissertation first and second order dynamical systems are considered. A
first order dynamical system can be expressed as
dx
= ẋ = f (x, t)
dt
(2.1)
where, the vector x is composed by the state variables of the system, hence is
called the state vector of the system. The time derivative of a given variable or
function is often indicated as the same variable dotted. This kind of notation is
used as well through this dissertation.
A second order dynamical system is expressed through a vectorial equation or a
system of equations in the form
d2 x
= ẍ = f (x, ẋ, t)
dt2
(2.2)
where, the state vector is in general composed of the vector x and its derivative
ẋ. Given a system of m differential equations, each of n-th order, it is possible
to turn it into a system of n × m differential equations of first order with n × m
variables by defining (n − 1) × m auxiliary variables and associating them to the
time derivatives of the original state variables. For the case of a second order
system this turns into
ẋ = v
(2.3)
v̇ = f (x, t)
(2.4)
The time dependance is included in the above equations for sake of completeness
although the systems studied in the following do not have any explicit dependance
on time. Such systems are said “autonomous” as opposed to the ones where the
time is an explicit variable and are called “non-autonomous”.
2.1.1
System Energy and Hamiltonian
When multi-particle systems are used as tool to model multi-agent systems, thermodynamic analogies are considered. In such a case canonical coordinates are de-
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
23
fined for the phase space. These are the generalised coordinates q and generalised
momenta p and are linked to the Hamiltonian function H by the Hamiltonian
Equations
∂H
∂q
∂H
q̇ =
∂p
ṗ = −
(2.5)
(2.6)
Hamiltonian function H is defined as
H = p · q̇ − L(q, q̇)
(2.7)
where, (·) indicates the scalar product and L is the Lagrangian, which is directly
linked to the energy of the system and defined as
L= T −U
(2.8)
with T and U being the kinetic and the potential energy respectively.
The energy approach above is particularly important for the development and
stability verification of control methodologies such as the artificial potential functions.
2.2
Control Theory
Control theory provides the framework and the methods to design control algorithms, where such algorithms are meant as the formulas and procedures to
determine the control actions. Control actions are executed by devices able to
process information and translate actions on the system to control [83].
In classical monolithic systems, a controller is in charge of determining control
actions of all the degrees of freedom of the system. In distributed systems, the
degrees of freedom of the systems are the sum of the degrees of freedom of all
the agents. In this case the control is distributed as well, that is, each agent
has a controller that operates on the agent’s state only, although it exploits the
information of other agents and the surrounding environment as well.
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
24
In a classic closed loop control scheme, for a monolithic system, the readings of
the sensors that define the state of the system are compared to a reference state
or guidance law; the control action decided is a consequence of the difference
between the actual state of the system and the reference one. In the distributed
control scheme, the agent’s own state is not only checked against the reference but
also the states of other agents contribute to define the action to be taken. Agents
in the system can be seen as a part of the changing environment surrounding the
agent to control [11]. The controller of that agent has to consider the other units
when driving it, determining the evolution of the distributed system towards its
final state. A comparison between the control of a monolithic system and that of
a distributed system is illustrated in Figure 2.1 where, it is shown how the control
actions driving a classic monolithic system are driven by the system state, the
external environment sensed and the reference or guidance law, whereas in a
distributed system the states of the other agents are used as well to define the
control for any unit and hence for the whole system. In many cases the Guidance
and Control functions are not distinguished from one another and referred to
generically as “Control”. Indeed if the system does not aim to any specific state
in an absolute sense, but instead it aims to achieve a target defined on the base
of the agent relative states, then the control provides the guidance information
at each time step and tracks them as well. A complementary case occurs when
the actuators are not considered or are idealised. What then comes from the
comparison of the reference with the actual state is the feedback the system
receives. This is often used as a simplifying assumption as it does not take into
account how the control action is produced. In refined modelling of engineered
systems, the control input is taken by a low level controller that maps it into a
driving signal for the actuators, which are peculiar to the system considered.
2.3
Stability Theory
Checking the stability of dynamical systems means verifying the long time behaviour converges to either periodic orbits in the phase space or to a fixed point.
The stability verification of the control system is carried out considering the high
level control. What matters for the control stability is that the control action commanded is in the direction of decreasing the error without considering whether
25
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
(a)
(b)
Figure 2.1: (a) The control scheme of a monolithic system and (b) the control scheme
of a distributed system.
26
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
the low level controller and the actuators are able to achieve it. In this dissertation the control of swarming systems is verified for both linear and nonlinear
stability. Here some insights about stability verification are given.
2.3.1
Linear Stability
Linear stability for dynamical systems is verified through considering the eigenvalues of the Jacobian matrix. This is defined as the matrix of the partial derivatives
of the system with respect to its variables. For instance considering a first order
system in m variables
dx1
dt
dx2
dt
= f1 (x1 , x2 ...xm )
= f1 (x1 , x2 ...xm )
... ...
dxm
= fm (x1 , x2 ...xm )
dt
(2.9)
with x = (x1 , x2 ...xm ), the Jacobian matrix is defined as
∂f
1 (x1 ,x2 ,...)
∂x
∂f2 (x1 ,x1 2 ,...)
∂(f1 , f2 , ...fm )
∂x1
J=
=
..
∂(x1 , x2 , ...xm )
.
∂fn (x1 ,x2 ,...)
∂x1
∂f1 (x1 ,x2 ,...)
∂x2
∂f2 (x1 ,x2 ,...)
∂x2
..
.
...
...
...
..
.
...
∂f1 (x1 ,x2 ,...)
∂xm
∂f2 (x1 ,x2 ,...)
∂xm
..
.
∂fn (x1 ,x2 ,...)
∂xm
(2.10)
For linear systems the analysis of the Jacobian is sufficient to prove stability.
Stability of nonlinear systems can be analysed too through linearisation but just
to some extent and in local sense only, that is the results hold just in a neighbourhood of the equilibrium point. The linearisation makes the Jacobian a constant
matrix. If the linearisation is performed about a fixed point x0 , the Jacobian
matrix can be calculated in that point as
J0 =
∂(f1 , f2 , ...fm )
∂(x1 , x2 , ...xm )
x=x0
(2.11)
The Hartman-Grobman theorem states that if all the eigenvalues have nonzero
real part (the fixed point is said hyperbolic), then the trajectories of the nonlinear system can be mapped to trajectories of the local linearised system in a
neighbourhood of the equilibrium. The nonlinear stability can then be determined locally through the spectrum of the Jacobian. In particular the following
behaviours can be related to the eigenvalues as exemplified in Figure 2.2 for a
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
27
linear damped oscillator and to a linearised one:
- if all the eigenvalues are purely imaginary, that is the real part is zero, a
linear system exhibits undamped oscillations (Figure 2.2.a); In this case,
nothing can be said about the stability of a nonlinear system.
- if all the eigenvalues have negative real parts, and some have nonzero imaginary parts, the system (both linear and linearised) stabilises according to
an exponential decay that envelopes the oscillating behaviour damping it
out eventually (Figure 2.2.b);
- if at least one eigenvalue has a positive real parts, and some have imaginary
parts the system (both linear and linearised) will exhibit oscillations of
exponentially increasing magnitude (Figure 2.2.c);
- if all the eigenvalues are real and negative the system (both linear and
linearised) is exponential stable, that is it stabilises to the equilibrium according to an exponential decay (Figure 2.2.d);
- if all the eigenvalues are real and at least one is positive the system (both
linear and linearised) is exponential unstable, that is it exhibits an unbounded motion according to an exponential increase (Figure 2.2.e).
It can be concluded that, for linear systems, an asymptotically stable behaviour requires eigenvalues with negative real parts, whereas a stable behaviour simply excludes the presence of eigenvalues with positive real parts.
Moreover if the system dynamics is governed by a matrix whose eigenvalues
have negative real part, the system will exhibit a stable behaviour. This
kind of matrix is also known as Hurwitz matrix. For linearised system,
instead, nothing can be said if at least one eigenvalue has zero real part.
2.3.2
Nonlinear Stability - Lyapunov stability
In this dissertation the stability of nonlinear systems is analysed using the Lyapunov method, although other criteria to assess nonlinear stability exist for which
the reader is addressed to the literature dealing with nonlinear dynamical systems and their stability (see for example [84–86]). Lyapunov method can assess
28
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
4
1
1.5
0.5
1
3
x 10
0
−0.5
2
Amplitude
Amplitude
Amplitude
2.5
0.5
1.5
1
0.5
0
0
−1
0
10
time
−0.5
0
20
(a)
10
time
20
−0.5
0
5
10
time
(b)
15
20
(c)
8
2.5
0.8
Amplitude
Amplitude
1
0.6
0.4
0.2
0
0
5
10
time
(d)
15
20
x 10
2
1.5
1
0.5
0
0
5
10
time
15
20
(e)
Figure 2.2: Time history for a single degree of freedom, second order linear system
in case of (a) purely imaginary eigenvalues, (b) complex eigenvalues with negative
real part, (c) complex eigenvalues with positive real part, (d) a single negative, real
eigenvalue, (e) a single positive, real eigenvalue
stability for nonlinear systems characterised by smooth dynamics, although it has
also been extended to non-smooth dynamics as will be illustrated in Section 2.5.
In his book of 1892 “The General Problem of Stability” [87] Alexandr Lyapunov
defined a method to verify stability of nonlinear dynamical systems that does not
require solving the system. Before describing how this works, some definitions
must be provided; these are mainly given accordingly to [86]. First the equilibrium point has to be defined: Given a dynamical system as the one in Equation
2.1, with initial conditions x(t0 ) = x0 a point x∗ is said to be an equilibrium
point of the system if f (x∗, t) = 0 for all t ≥ 0. Without loss of generalities and
to comply with the most spread formulation in the literature, from this point on
consider x∗ = 0, that is, the origin. Stability in the sense of Lyapunov prescribes
that x = x∗ = 0 is a stable equilibrium point for the system in equation 2.1 with
initial conditions x(t0 ) = x if for all t0 ≥ 0 and ǫ > 0, there exists δ(t0 , ǫ) such
that
|x0 | < δ(t0 , ǫ) → |x(t)| < ǫ
∀t > t0
(2.12)
where x(t) is the solution of Equation 2.1 starting from x0 at t0 and ǫ and δ are
scalars.
The stability is called uniform if δ can be chosen independent of t0 . The origin
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
29
x = 0 is an asymptotically stable equilibrium point for the system in equation 2.1
with initial conditions x(t0 ) = x if it is a stable equilibrium point for the system
and it is attractive, that is for all t0 ≥ 0, there exists δ(t0 ) such that
|x0 | < δ → lim |x(t)| = 0
t→∞
(2.13)
The origin x = 0 is a Globally Asymptotically Stable equilibrium point if it is
stable and limt→∞ |x(t)| = 0 for all x0 ∈ Rn with n the dimension of the state
space. This means that an equilibrium point is Globally Asymptotically Stable if
the system will always evolve towards it for any set of initial conditions considered.
The method then foresees the verification of the stability by finding a scalar
function V (x), defined over the variables’ space of the system, that takes zero
value at the equilibrium and is positive definite everywhere else; moreover its
time derivative is zero at the equilibrium and negative definite elsewhere. The
conditions are summarised in Table 2.1
Table 2.1: Conditions for Lyapunov asymptotic stability.
V (x) > 0
V̇ (x) < 0
V (x) = 0
for x 6= x0
for x = x0
V̇ (x) = 0
The function V is called Lyapunov function and finding it is the challenge associated with the method as no a-priori indication is given about its definition.
Usually the energy of the system can be taken as a candidate for the definition
of a Lyapunov function. The stability can be global or just local depending on
whether the Lyapunov function satisfies the conditions in table 2.1 respectively
in a neighbourhood of x0 or everywhere.
2.4
Graph Theoretic Methods
Graph theory was introduced in Section 1.4. In this Section the terminology,
used later in this dissertation, is defined. Moreover, some algebraic graph prop-
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
30
erties are illustrated for cases that will be discussed later in this dissertation. The
reader can refer to a number of textbooks for a deeper understanding of graph
theory. The textbooks [88–91] have been used as reference.
A graph G{V, E} is a mathematical object defined as the pair composed of a set
of nodes or vertices V and a set of ordered pairs of vertices E that are called
edges.
• A graph is said to be directed if all the edges are ordered pairs of vertices.
A directed graph is also named digraph. A graph is said to be undirected
if all the edges are unordered pairs of vertices. A graph with ordered and
unordered pairs of vertices is said to be mixed, although it can just be
considered as a special case of directed graph.
• Any two vertices i and j are said to be adjacent if there is an oriented edge
from vertex i to vertex j.
• A path is a subset of edges arranged in a sequence such that any two of
them, which are consecutive in the sequence, connect adjacent vertices.
• A node i of a graph is said to be reachable by another node j if a directed
path from node j to node i exists. Node i is said to be globally reachable
if it is reachable by any other node of the graph.
• A graph is said to be connected if and only if it has at least one globally
reachable node.
• A graph is said to be strongly connected if there exists a directed path
between any two nodes of it.
• A directed graph is said to be weakly connected if it can be made connected
by replacing its directed edges with undirected ones.
• A directed graph is said to be complete if and only if any two distinct
vertices of the graph are the end-points of an edge of the graph.
• The degree of a vertex u in an undirected graph is the number of edges
which include u. The in-degree (out-degree) of a vertex u in a directed
graph is the number of edges entering (exiting from) u. A graph is said to
be regular if all the vertices have the same degree.
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
31
In the following only “sensing” graphs are considered, that is the oriented edges
in the graph point to the nodes which are “observed” and have origin in the
nodes that “observe” or “sense” the first ones. The term degree can then be used
instead of out-degree in the context of directed graphs.
The connectivity properties of a graph are related to the algebraic properties of
two fundamental matrices: the Adjacency Matrix and the Laplacian Matrix. The
adjacency matrix of a graph G on N vertices denoted by A(G) is an N ×N matrix
(square matrix of size N), having rows and columns labeled by the vertices of G,
and ij th entry, aij , defined as
(
aij = 1 ⇔ ui , uj ∈ E
aij =
0
otherwise
(2.14)
where, ui , uj ∈ V.
It is important to notice that the way the adjacency matrix is defined implies
that if an agent i observes an agent j, then there is a nonzero entry at the j-th
column of the i-th row of the adjacency matrix. The Laplacian matrix of a graph
on N vertices can be defined on the base of its adjacency matrix as
L= D−A
where, D is a diagonal degree matrix, that is its i-th diagonal element is the
out-degree of the node i, that is the sum along the rows of the adjacency matrix.
It is immediate to verify that L is positive semidefinite and, because of the zero
row sum, it always has at least one zero eigenvalue. It can be proved that a
graph is connected if and only if the zero eigenvalue of the Laplacian matrix is
simple, that is, it has multiplicity equal to 1. When a graph is disconnected it is
common to refer to its connected subsets as components of the graph. When the
graph is connected, hence one only component is present, it can be referred to
as the giant component. Two examples of graphs are reported in Table 2.2 with
their adjacency matrices and Laplacian matrices. It can be seen that an edge
connecting two generic nodes i, j corresponds to a nonzero entry in the i − th row
at the j − th column of the adjacency matrix. The Laplacian matrix is obtained
from considering a negated adjacency matrix where the i − th entry along the
main diagonal corresponds to the outdegree of node i.
32
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
Table 2.2: Examples of connected and disconnected graphs.
✲ ❞1
2❞
(a)
3
❅
■
❅
❅
❅
❅
❄
❅
✲ ❞4
❞
2❞
❅
■
❅
(b)
3 ❞✠
✒
❅ ❞5
0
1
A=
0
0
0
0
0
1
0
0
1
0
0
1
0
0
0
0
0
0
−1 2 −1 0
L=
0
0
1 −1
0 −1 0
1
❞1
❅
❅
❘❞4
❅
0
0
A=
0
0
1
0
0
0
0
1
0
0
0
0
1
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
L= 0
0
0
0
0
−1 −1 −1 −1
0
0
0
0
4
Example 1. In the first example, (a), the graph is connected as node 1 is globally
reachable. If node 3 is removed the graph would be on just 3 nodes, but it would
still be connected. If instead the edge 3,4 is removed, the 4 node graph would be
disconnected as there would be no path from node 3 to node 1, that hence would
not be globally reachable anymore.
Example 2. In the second example, (b), the graph is disconnected, but it is also
weakly connected. This graph is also known as “exploding star”. If all the existing
edges are made non-directed, the graph becomes connected as node 5 becomes
globally reachable. The same result can be obtained by simply reversing all the
edges, such that these all point to node 5. The graph so produced is also known
as “imploding star”, that, as different from the exploding star, is connected.
Adjacency matrix, and Laplacian in turn, can be weighted, that is the entries
can attain values other than just 0 and 1. The entries in this case would represent the strength of an interaction beside its existence between two nodes. The
linearisation of the interactions amongst agents in a networked system leads to
a matrix representation where the system reduces to its Laplacian, or to a matrix proportional to it. Consider for instance a generic networked system with
nonlinear interactions. Consider that each node or agent is characterised by a
one-dimensional state variable xi . The dynamics of a generic agent i then evolves
as consequence of the interactions with the other agents it is connected into the
network, as a function of the difference between their states, as
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
ẋi =
X
j
f (xj − xi )
.
33
(2.15)
Considering the system in equilibrium for its state corresponding to the vector ξ,
said x̃i the state linearisation about ξ, this is
x̃˙i =
K
X
j
∂f
f (ξj − ξi ) +
∂xj
xj =ξj
(xj − ξj − (xi − ξi ))
(2.16)
where, the sum extends to the K nodes node i is connected with. As ξ is an
equilibrium point, Equation 2.16 reduces to
x̃˙i =
K
X
∂f
j
∂xj
xj =ξj
(xj − ξj − (xi − ξi ))
(2.17)
that, in vector form becomes
x̃˙ = −LJ0T (x − ξ)
(2.18)
where, J0T is the Jacobian matrix referred to the equilibrium point as defined in
Section 2.3.1 and the superscript indicates the transpose. The linear stability can
then be studied considering the spectrum of the LJ0T matrix.
2.5
Nonsmooth Analysis
Nonsmooth analysis is a subfield of nonlinear analysis. It studies non differentiable phenomena, framing them within the rigorous boundary of mathematics
or, conversely, expanding the classic nonlinear analysis to non differentiable functions [92]. When considering a time changing graph, or intermittent interactions
in a swarm of agents, discontinuities are found in the time history of the signal
that the agents receive from the other units in the swarm. In the particular case
of the APF method, the derivative of the potential corresponding to acceleration
or velocity commanded to the agents cannot be defined within the classical analysis whenever a discontinuity is found in the right hand side of the equations.
Without the appropriate tools is then impossible to assert that the system is not
going to output an unbounded response at the discontinuity, that is, stability
cannot be proved. In such cases nonsmooth analysis is used to bound the value
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
34
of the derivative to a set, which can then be used to prove system stability. The
first important contribution to the solution of this kind of problems was given
by Filippov [93] who later provided the analytic tools to study equilibria in this
framework through using smooth Lyapunov functions [94]. The extension to nonsmooth Lyapunov functions was provided by Shevitz and Paden [95].
In this dissertation nonsmooth analysis is considered only as a tool for future
developments, however, it is important to address its role in the study of swarms
and multi-agent problems. In recent papers Tanner [96, 97] studied the performance of groups of mobile agents linked in a network based on proximity. As
the agents are able to interact with neighbours only within a given radius and
the relative motion of the agents changes the local neighbour in a continuous
fashion, the network of interaction changes dynamically. In this case nonlinear attractive-repulsive interactions govern the motion of the agents that can be
proved to align their velocity along a common direction, hence producing a stable
behaviour, by means of the nonsmooth analysis. This is pursued by proving that
the generalised derivative at the discontinuity is such that the system does not
destabilise because of the continuous jumps the agents experience in the control
input. The requirement such systems have to be able to prove their stability is
that the graph representing the interaction network remains connected over all
the configurations the network takes throughout the switches. This hypothesis is
fundamental to guarantee that the system evolves to the consensus.
It is the case to note that being continuously connected is a strong assumption
as the connectivity of the graph of interaction cannot be guaranteed by simply
relying on the interactions within the local neighbours.
2.6
Swarm Modelling and Control
Mathematical handling of swarming problems requires the definition of interactions between the agents from an analytical point of view. This is done for the
interactions in terms of signals and actions exchanged between pairs of communicating agents, and the network that spans the swarm, which is mainly responsible
for its cohesion. In this section, the method of the artificial potential functions
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
35
is explained with reference to the Morse potential. Also there is the need to
consider the role of the network of interactions in the swarm and the control
issues depending on the choice of both the network and the Artificial Potential
Functions.
2.6.1
Artificial Potential Functions
Artificial Potential Functions (APFs) are a popular control means for multi-agent
systems, although it is also quite often used to produce a time variant reference
guide rather than a control output. In APF methods, given a system modelled
as a dynamical system ẋ = f (x, t), a scalar potential is defined as a function of
the system state, namely U = U(x). The control output ẋc is then defined by
the negative gradient of the potential as
ẋc (t) = −∇U(x)
(2.19)
where ∇ indicates the gradient. For a generic function ϕ(χ1 , χ2 ...χn ) in the
generic n variables χ1 , χ2 ...χn the gradient is defined as the vector field defined
by the partial derivatives of ϕ, that is
∇(ϕ) =
∂ϕ
∂ϕ
υ1 +
υ2 + ...
∂χ1
∂χ2
+
∂ϕ
υn
∂χn
(2.20)
with υ1 , υ2 , ...υn being the unit vectors forming the canonical base of the ndimensional space. The control output is zero in case of stationary points in
the potential. The system is therefore driven towards the minimization of the
potential that corresponds to the target state. In order to ensure stability, the
potential function has to be convex about the equilibrium point. The APFs can
either be dependent upon the state of the system in absolute sense, that is, the
position and velocity of each agent with respect to a given pair of parameters, or
in relative sense, that is the potential for each agent is defined on the base of its
state compared to that of one or some of the others. In the second case APFs are
sometimes referred to as based on pairwise interactions, or more briefly pairwise
APFs, which is the case of the systems analysed in this dissertation.
APFs can be used to define the guidance law for an engineering system. In the
case of multi-agent mobile systems, they can define the velocity field according to
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
36
Equation 2.19. The control system has to track this velocity field by producing
the accelerations through the actuators. When the APFs are used to model
dynamical systems this can be done as
ẍ(t) = ẍc (t) = −∇U(x)
.
(2.21)
Equation 2.21 can then be integrated twice to get the time history for a given set
of initial conditions on position and velocity. In this second case the APFs can
be used to model particle systems. When used to control engineered agents some
issues arise. By their own nature, physical systems which undergo the action of
the potential only are conservative, that is the energy provided by the potential
field to the system is transferred to the system as kinetic energy and returned
to the potential field when the system slows down without leakages. Hence the
APFs alone would not be suitable in general to drive a system from an initial
state to a final state characterised by a different level of total energy. For this
reason the APFs are often complimented by an orientation function that, in the
general case, pumps in or out virtual energy to the system. The Hamiltonian
representation of the system, with the notation introduced in Section 2.1.1, is
then
ṗ = −
∂H
∂H
− g(H)
∂q
∂p
(2.22)
where, g(H) is some function in general able to reduce or increase the energy
level of the system. If g(H) is an always positive function, then the system will
always reduce its energy level. Conversely, if it is an always negative function,
the system will increase its energy level. This can be easily seen by computing
the time derivative of the Hamiltonian function as
X ∂H 2
dH
= −g(H)
dt
∂pi
i
(2.23)
In the simplest case of viscous-like effects g(H) is often taken constant and the
system will eventually damp out all its kinetic energy relaxing to a minimum of
its potential energy.
Morse Potential
Morse potential is named after Philip M. Morse who defined it in 1929 [98] for
the purpose of modelling the vibrational structure of diatomic molecules. It is a
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
37
scalar potential composed of the sum of two exponential functions and its first
formulation included explicitly the position and the value of the minimum for the
potential associated to the bond between the atoms of a diatomic molecule. It is
E(r) = De−2a(r−r0 ) − 2De−a(r−r0 )
(2.24)
where, D and a are constants that define the magnitude of the potential as well
as its attraction and repulsion range. The potential has a finite value at zero
distance and vanishes asymptotically for r → ∞. In this formulation the potential attains a unique minimum of −D at a distance r = r0 . This is illustrated
in Figure 2.3.a where, the attractive and repulsive contributions to the potential
are shown as well as their sum.
When modelling particle and swarming systems, a more convenient formulation is
commonly used for pairwise interactions between two generic particles or agents
i and j, see for example [28, 32]. This is
UijM orse = −C a e−
|xi −xj |
La
+ C r e−
|xi −xj |
Lr
(2.25)
where, C a and C r are constants that tune the magnitude of the potential while
La and Lr regulate the range of attraction and repulsion. This formulation uses
vector notation for the positions of two generic agents i and j with | · | indicating
here and thereafter the euclidean norm when applied to vectors, the determinant
when applied to matrices and the absolute value when applied to scalars. The
attractive and repulsive parts are clearly identified by a the superscripts a and r
on the coefficients. This version of the potential presents a unique minimum at
La Lr
ln
xi − xj = r
L − La
C a Lr
C r La
(2.26)
where the potential attains the value of
M orse
Umin
= −C a
C a Lr
C r La
r
LaL−L
r
+ Cr
C a Lr
C r La
a
LaL−L
r
(2.27)
Figure 2.3.b shows the attractive and repulsive functions that together form the
Morse potential. These can be easily compared to the ones in Figure 2.3.a which
refer to the original formulation of the Morse potential.
38
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
10
1
Morse potential
Attractive potential
Repulsive potential
0.5
U
E(r)
5
0
0
−0.5
−5
−10
0
Morse potential
Attractive potential
Repulsive potential
1
2
r
3
4
(a)
−1
0
1
2
xi−xj
3
4
(b)
Figure 2.3: Original and modified Morse potentials. (a) A plot of the Morse potential
as it was published in [98], according to Equation 2.24 with D = a = r0 = 1. The two
exponential functions that shape the potential are shown as well and (b) a plot of the
Morse potential according to Equation 2.25 with C a = C r = La = 1 and Lr = 0.5.
Lennard-Jones and other Potentials
The use of APFs is not limited to the Morse potential. In particular, when
studying molecular dynamics a popular potential is that of Lennard-Jones. It
was originally introduced to model the interaction between neutral atoms in gas
molecules in correlation with the study of the relationship between the gas viscosity and the temperature [99].
Distinct from the Morse potential, the Lennard-Jones potential presents a singularity at the origin, that is, it produces an unbound force over particles at very
close distance. The Lennard Jones potential sensed by a particle i in the vicinity
of a particle j is
λ2
λ1
−
(2.28)
UijLJ =
n
|xi − xj |
|xi − xj |m
where, λ1 and λ2 are free parameters that can be used to set a desired separation distance between particles, n and m are exponents that, in the original
formulation and quite often in literature, take value 6 and 12 respectively; however, these can be tuned to better design the interactions between pairs of agents.
In literature it is easy to find examples of potential functions other than Morse and
Lennard-Jones. Badawy, McInnes and Bennet (see for example [49, 55]) exploited
the property of hyperbolic potentials that have almost constant gradient far from
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
39
equilibrium. The hyperbolic potential U hyp is described by the equation
Uijhyp = Ch [(|xi − xj | − d)2 + 1]0.5
(2.29)
where, d is the desired distance between the agents and Ch is a scaling parameter. This potential produces a control law where saturation problems can be more
easily tackled by scaling the actuator outputs between zero (at equilibrium) and
the maximum gradient achieved by the potential. Moreover the quasi-linearity
of the potential far from equilibrium allows for superposition of the effects easily
tractable in analytic modelling.
Some authors, see for example [59, 97, 100], proposed artificial potentials that
present a discontinuity in the gradient at a designed cut-off distance, that is a
distance within which two agents interact, while they do not anymore if further
apart. In this case the control input experiences a jump when two agents enter each-others range. This should be accounted for in any stability analysis
and tackled through non-smooth analysis. Three examples of potential functions
quoted in this sections are reported in Figure 2.4 where some characteristics can
be noted. The singularities of the Lennard-Jones potential visible in the first
two panels are responsible for unbounded control actions at close distance. The
linear trend of the hyperbolic potential far from equilibrium corresponds to a
bounded derivative, which translates into a constant acceleration commanded to
the agents when far from their design arrangement. Finally, in the rightmost
panels, the non-smooth blending of the potential at a cut-off distance produces
a clear discontinuity in the derivative, that translates to a sudden change from
a continuous to a null control action. Lastly, it is worth noting how, even in
cases where the potential does not present a cut-off distance, some of the functions used in literature present a vanishing gradient at large distances making
the interaction faint and, practically, ineffective beyond a given radius. This is
the case of Lennard-Jones potentials and Morse potential for instance which are
often claimed to affect the agents relative neighbouring only.
2.6.2
Connection Networks
Beside the interactions between pairs of agents, the other fundamental constituent
of distributed control is the network of connections which link all the agents. All-
40
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
3
60
2.5
40
−0.4
U(r)
−0.2
U(r)
U(r)
0
2
20
1.5
0
−0.6
0
1
r
2
1
0
3
−3
3
1
2
r
3
4
−20
0
5
1
r
2
3
−3
x 10
1
x 10
5
4
0.5
1
3
0
f(r)
2
f(r)
f(r)
2
0
1
0
−0.5
−1
−1
−2
−2
0
1
(a)
r
2
3
−1
0
1
2
r
3
4
5
−3
0
0.5
(b)
1
1.5
2
2.5
3
r
(c)
Figure 2.4: Examples of potential functions in the first line and their gradient, below,
in one dimension. (a) Lennard-Jones potential; (b) Hyperbolic potential; (c) Potential
presenting a gradient discontinuity as found in [100]
to-all connections are commonly used in works motivated by the investigation
of particle systems. Attraction and repulsion are experienced by each particle
because of all the other particles in the system: the corresponding graph is hence
complete and equilibrium conditions for a single particle depends on the states
of all the others. In works considering the engineering aspects of multi-agent
systems, the infeasibility of all-to-all communications is often considered. Proper
networks are hence introduced which can be classified in two main categories
based on their graph, that is Symmetric and Asymmetric. In the symmetric case
an edge between two generic nodes i and j implies the existence of another edge
between node j and node i that is oriented in the opposite direction. In this case
the graph is non-directed and its adjacency matrix is symmetric. The symmetry
of the network does not imply symmetry of interactions. In a symmetric network
the existence of a generic non-oriented edge i − j does not imply that the action
of node i on node j is the same of the action produced by node j over node i.
On the other hand, non-symmetric networks exclude the possibility of symmetric
interactions as in this case the corresponding graph is directed.
In the case of agents interacting only in case of proximity, that is within a certain
distance, these are often modelled considering pairwise forces which are null beyond a given range; the network arises then as consequence of this. If the range
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
41
is the same for all the agents, the network that is formed is symmetric but there
is no guarantee that it is or stays connected as the swarm evolves.
Particular networks have been proposed to link autonomous agents in coherent
swarms while enhancing the overall performances by exploiting some of the network characteristics. In particular the models of Scale Free networks by Barabasi
and Albert [71] and the Small World network by Watts and Strogatz [72], already
introduced in Section 1.4.3, deserve further discussion for the attention they received in the literature.
Scale free networks are formed by connecting nodes to an initial cluster, with
each new node having a probability of being connected to an existing node proportional to the number of connections the latter has. The network emerging out
of this process has a node degree distribution (number of nodes featuring a given
number of connections) that presents a small portion of nodes, also named hubs,
with high number of connections while most nodes feature just few connections.
This arrangement produces advantages in terms of controllability as controlling
just the hubs it is possible to control the whole network [101].
The benefits of small world networks in multi-agent systems have been stressed
with an emphasis comparable to scale free networks. Small world networks are
developed by rewiring randomly a small amount of links in a regular lattice. This
is a network where each node is linked to a number of neighbouring nodes and
neighbouring property is determined by the length of the path between the nodes
in the graph. By rewiring a small number of them, shortcuts are created so that
two nodes separated by a long path in the lattice find themselves much closer after
the rewiring [72]. The creation of a small world network, out of a regular lattice,
enhances faster information routing in the network. This has been exploited in
a number of applications which feature the presence of a network, including the
controlled deployment of sensor networks and consensus problems, see for example
[102–104]. In Figure 2.5 the appearance and main features of small world and
scale free networks are illustrated. While the network can significantly change
the performance of a multi-agent system, its connectivity is a critical issue for
the stability of the distributed control. This mostly affects networks based on
proximity, including the case in which interactions present a cut-off distance, and
42
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
(a)
(b)
3
25
20
2
Nr of Nodes
Average Path Length
2.5
1.5
1
10
5
0.5
0
0
15
5
10
Node
(c)
15
20
0
1
2
3
4
Nr of Edges
5
6
(d)
Figure 2.5: (a) Graph representation of a small world network on 20 nodes. (b) Graph
representation of a scale free network on 60 nodes. (c) The comparison between the
average path lengths for the small world network, in red, and the ones of regular lattice,
in yellow, show the characteristics of small world networks of lowering the average path
length of each node with respect to any other in comparison with a regular lattice with
the same number of edges. (d) The node degree distribution of the scale free network
in (b) is typical of this kind of network as characterised by few nodes with many edges
while most of nodes have just few edges.
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
43
dynamically changing networks in general whenever a control on the connectivity
is not considered. The chance of having isolated components in the graph is often
ruled out by hypothesis in literature dealing with stability of distributed control,
but it is a practical issue to account for when actual designs are considered.
2.6.3
The Combined Effect of APFs and Connection Network
The dynamics of swarms are influenced by both the connection network and the
pairwise relations that govern inter-agent interactions in a different way. They
are both responsible for system failures either for what concerns fragmentation of
the swarm, or instabilities in the control respectively. On the other hand, the use
of the connection network and the control law operating along its links can boost
the performance of a multi-agent system. Badawy proposed the exploitation of
the network to route the action of the artificial potential functions towards the
agents to joint in a self-assembling structure by modifying the adjacency matrix
of the graph [54]. This way a number of structural configurations are possible by
changing the adjacency matrix and hence defining which are the agents to lock
together.
An alternative was taken by the Dynamical Control System Laboratory in Princeton. This considers the concept of tensegrity structures [100, 105] for shaping
arbitrary formations of vehicles. Tensegrity structures are composed of a network of linking elements that meet in the nodes of the structure. The position of
the nodes in the structure is uniquely determined by the force in the structural
elements connecting them. In turn this force is determined for each element depending on a rigidity constant and on the base of its length. Depending on the
difference between the rest length and the actual length of the elements, these
ones can either be “strings”, if they exert a pulling force, or “struts” if they exert
a pushing force. This way a multi-agent system can be viewed as a set of masses
interconnected by springs and its equilibrium in a given formation can be found
by assigning “spring constants” accordingly. Also in this case there is no need
to connect all the agents in a complete graph and the stability can be analysed
considering the energy of the system. Indeed the symmetry of the network and
the interactions in it allow for stability proof following a Lyapunov approach and
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
44
relying on the energy of the system as shown in Sections 2.3.2 and 2.6.1. Tensegrity formations, however, exploit the precise tuning of the virtual forces amongst
the agents in the formation. This requires either off-line computations before the
deployment or distributed, heavy, on-line computations. Moreover, because of
the unique choice of the spring constants for a given equilibrium, the failure of
one agent would lead to sensible distortions in the formation.
The limitations highlighted by tensegrity structures are overcome in this dissertation by exploiting some asymmetries in the interactions amongst agents. Moreover the hierarchical arrangement of the agents discussed in Chapter 4 returns
more flexibility in the agent arrangements, as it will be shown later on.
2.7
Schooling, Flocking and Swarming
Schooling, flocking and swarming are three emergent group behaviours found in
nature. Although it is recognised that some individuals are more influential than
others, there is no superior intelligence that coordinates the motion in animal
groups. Since Beni [6], the engineering systems designed taking inspiration out of
such phenomena have been named Swarms, although the biological connotation of
swarming originally only referred to the behaviour observed in bee colonies when
a new queen leaves the hive to form a new one. In this section these terms will
be explained, considering their importance for modelling animal behaviours that
are inspirational for engineering design of multi-agent systems. The conditions
that distinguish a schooling behaviour from a flocking one in terms of engineering
design parameters are extrapolated to then be used to model non-biological systems. Importance is given to the problem of cohesion that, although natural in
biological system, might result of difficult implementation when the agents rely
only on a limited number of communication channels.
2.7.1
Biological Studies
Schooling refers to be behaviour often found in large groups of fish which swim
keeping fixed relative positions and orientations. Flocking instead refers to a
group motion in which individuals, although remaining in close proximity, do
not keep the same positions with respect to one another; this is often associated
to avian groups. Both these behaviours act to better protect the group from
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
45
predators by increasing the capabilities to quickly become aware of an attack
and confuse the predators, making it difficult to isolate a single target. This way
animals, which are smaller and weaker than the predator on their own, increase
their possibility of surviving.
In schools and in flocks, individuals keep track of a fixed number of nearest
neighbours [106, 107]. This is understandable as senses are limited as well as the
amount of resources individuals can invest in social behaviours. It was claimed,
for instance, that the number of neighbours considered by each bird in a flock is
on average between six and seven and this is not dependent upon the distance
between the flock members [106]. This result was found through the analysis
of footage of large starling (Sturnus vulgaris) flocks (see [108, 109]). The same
set of videos provided evidence for a number of behavioural models and a way
to validate their numerical implementation. It was found that the velocities of
the birds in a flock are correlated in a scale-free fashion, that is the correlation
does not depend on the size of the flock [110]. Again it was possible to obtain
important data about the reaction time and the manoeuvre speed of starlings
under predator attack [111], providing evidence in support of the Trafalgar effect
phenomenon [112] and the Chorus line hypothesis [113]. These phenomena will
be better detailed in Chapter 6 together with the importance for their engineering
spin off.
Schooling behaviour has also received noticeable attention, see for example [114,
115]. Biologists attentively studied the mechanism producing the emergence of
tidy patterns. They found maintenance of alignment and lateral distance are
mainly due to the sensing organs schooling fish are provided with. In particular,
tidiness in schooling is possible due to the presence of the so-called “lateral line”,
a string of pressure sensors that, operating in differential fashion, allow the sensing of movement. This is used independently of visual sensing, helping towards
schooling and fast manoeuvring in response to lateral threats [116–119].
As mentioned at the beginning of this section, swarming has a meaning associated to the animal world as well as flocking and schooling. When swarming
bees move in search of a new hive location, the group dynamics is more similar
to flocking than to schooling being the cohesion maintained without necessarily
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
46
keeping alignment and relative separation. Several studies have been conducted
to investigate the emergent behaviours that arise in colonies and the hierarchical
structure that appears to be a result of genetic development of individuals. Individuals swarming in search of a nest site are bigger than their swarm mates and
communicate with visual signals produced by an oscillatory movement. The final
decision for the nest site becomes a “democratic” process in which all the swarm
gives credit to the dancing scouts in the site proximity [120, 121]. The swarming
behaviour is far more complex than a chaotic flight in close proximity and bees
are inspirational for engineers in the design of complex systems [122].
2.7.2
Crystalline Patterns
The control of relative positions within a set of agents is at the base of advances in
the technology for autonomous formation flights and self-driving cars. In schooling behaviour as well as in formation flying of some species, such as geese or
ducks, keeping fixed positions in a given way is useful towards the reduction of
the energy the individuals have to supply to the system. Engineering swarming
systems get advantages from arrangement in fixed relative positions for a number
of cooperative tasks. For example, surveying missions or, more generally, remote
sensing tasks can be carried out by several autonomous agents in place of just
one if these keep fixed relative positions. At the same time agents that keep fixed
relative positions can rely on the navigation capabilities of just some of them in
the position of leaders, leaving the followers more resources to utilise in mission
related tasks. The task allocation is something observed in the natural world as
well, with migratory events generally led by few informed individuals in the group.
Distributed control of swarms in crystalline formation can be achieved by defining
a virtual energy function depending on relative positions that attains a minimum
in correspondence of the desired distance between any two agents. This can be
done for both first order controllers and second order ones, that is, either producing the desired velocity or acceleration components as output of the controller.
From an energy point of view, a swarm whose agents stabilise in fixed relative
positions achieves a minimum, that is a static, stable equilibrium, that corresponds to a given level of energy. This is in general different from the one the
swarm is provided with before the control operates. Controllers based on APFs
leverage the level of total energy by adding or subtracting some virtual potential
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
47
energy when an unwanted quantity of real kinetic energy is respectively missing
or over-abundant into the swarm.
In order to achieve a crystalline pattern the controller has to damp out the velocity
components that tend to make agents drift away from a given pattern or collide
against each other. Damping out the kinetic energy does not imply achieving
a static swarm, which is cancelling the motion of the group while staying in a
pattern. In [32] self propelled particles arrange in a rotating vortex driven by a
pairwise potential and a steering function that produce the equations of motion
dxi
= vi
dt
dvi
m
= −∇Ui − ui
dt
(2.30)
,
(2.31)
for the i−th particle. In Equation 2.31 U is a Morse potential of the kind reported
in Equation 2.25 while ui is a steering function in the form
ui = Co
X
j
(vij · x̂ij )exp(
|xij |
)x̂ij
lo
(2.32)
where, {·} represent the scalar product, x̂ represent the unitary vector in a given
direction, Co and lo are scalar constants and the double subscript indicates the
difference between the variables associated with the subscripts i and j. Under
the effect of the above potential and steering functions, particles then arrange
in a crystalline formation rotating about its centre. This is shown in Figure 2.6
where the time history of the swarm’s total energy is reported as well and it can
be seen that it is an always decreasing function as expected for this particular
model. Excluding external drives or agents inputting energy into the system, in
case of fixed relative positions the total energy stabilises to a minimum as all
kinetic components are damped completely or in part. Results quoted so far in
this section refer to systems where agents are either able to track the positions of
all the swarm members or, in case of particles, they are subject to the potential
generated by all the other particles in the swarm. In this framework it is possible
to deal with the system in terms of total energy and global characteristics such
as linear and angular momentum with a physics approach. When the number of
connections each agent can keep, that is the number of agents it can keep track
of, is limited, several issues arise. Figure 2.7 depicts the arrangement obtained
48
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
−330
0.4
−340
0.3
Total Energy
0.2
0.1
0
−0.1
−0.2
−350
−360
−370
−380
−0.3
−0.4
−0.5
−390
−0.4
−0.3
−0.2
−0.1
0
(a)
0.1
0.2
0.3
0.4
0.5
5
10
Time
15
20
(b)
Figure 2.6: Vortex formation. (a) A crystalline arrangement emerging out of the interactions of 40 agents interacting with APFs and (b) the time history of system total
energy, which is an always decreasing function as per [32].
with the same control law that produced the results in [32] previously shown,
but, in this case, the number of connections per agent is limited and the connections are done producing a small world network and a scale free one (see Section
2.6.2). Clear distortions in the pattern are visible: these are due to the missing
communication links between agents which are physically close but do not have
awareness of each other.
In Chapter 4 reciprocal and symmetric interactions will be shown to be sufficient
to keep on considering the system in a physics oriented framework, with the possibility of achieving useful pattern for some particular space applications. The
physics approach fails, or has to be heavily corrected, when the connections are
not static, that is the case, for instance, of distance dependent interactions. It is
still possible to prove that, as long as the system is cohesive, the energy will eventually decrease and stabilise to a minimum level, as per [96, 97], but this requires
a nonsmooth analysis approach. To avoid this, the reciprocal influence can be
considered a continuous and vanishing function of the distance, rather than a discontinuous one presenting a cut-off distance. This was done, for example, in [123].
Finally, the achievement of fixed relative positions can be put in relation with the
organizational entropy [36]. It was observed that organizational entropy undergoes a phase transition when the free parameters, governing the system’s Hamiltonian, are tuned to pass from a chaotic behaviour to a stabilised one characterised
49
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
0.8
0.8
0.6
0.6
0.4
0.4
0.2
Y
Y
0.2
0
0
−0.2
−0.2
−0.4
−0.4
−0.6
−0.8
−0.6
−0.8
−0.6
−0.4
−0.2
0
0.2
0.4
0.6
0.8
−1
−0.5
0
X
(a)
0.5
1
X
(b)
Figure 2.7: Distorted formations. The vortex arrangement produced in [32] with 40
agents either (a) fails to form or (b) presents dramatic distortions. In (a) a scale
free network is implemented in which agents connect in turn to the ones which are
already connected. In (b) the connections are restricted for each agent to its 8 nearest
neighbours and some of this are rewired randomly. The probability of rewiring is set
to 0.1.
by the agents in fixed relative positions. Similar results were observed by Vicsek
et al. [30] who also accounted for the role of the noise in the transition between
more equilibrium configurations.
2.7.3
Continuous Swarming Systems
In Section 2.7.2 the parallels between the natural behaviour of schooling and the
motion of vehicles, keeping fixed relative positions, was presented. Another recurrent parallel with the natural world is the one between the flocking and swarming
behaviours in birds and bees and the possible application for flight in close proximity but without keeping relative positions. This is most often meant by the
word “swarming”. Such a system is rarely found in literature as the flocking behaviour is inefficient in terms of energy and agents, when provided with a stable
dynamics, will evolve towards a common velocity with no fluctuation. From an
engineering and physical point of view, this minimises the energy dispersions. It
emerges as result of a process that, through mutual influence, drives the agents
towards the so-called “consensus achievement”, which will be detailed in the next
section.
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
50
It is in theory possible to produce continuous fluctuations about the common
velocity by just inserting noise in each agent dynamics or getting rid of the velocity
alignment term, that is ui in Equation 2.31. In practical terms, this would not
find justification in any engineering application where shuffling agent positions
requires resources and, as such, has to be justified by a necessity. In birds flocking
this would correspond to the need of confusing predators or the attempt of some
birds to pass alerting messages by flying against the main direction of the flock.
The latter behaviour can be seen as an information diffusion. The difference
in the behaviour of few birds and the rest of the flock can be considered to be
caused by the information gradient across the flock. A bird that deviates from
the main stream does so alerting the others and producing a manoeuvre wave
with consequent position shuffle. This was observed in the context of the Chorus
Line hypothesis [113] and is the basis on which the results on Chapter 6 will be
constructed.
2.7.4
Consensus, a Common Problem
Regardless of whether the system is similar to a flock of birds or to a school of
fish, consensus problems remain in both cases as convergence to a given response
must be guaranteed. Couzin studied extensively the achievement of a common
will in large groups of animals, where individuals manage to share information
despite uncertainty, see for example [124–127]. Biologists are not interested into
consensus the same way engineers are, their interest is often focussed in linking
corporal characteristics of the species to their behaviours in a group. On the
other hand, they offer a number of hints to engineers who look for the defining
characteristics the single agents have to possess to achieve consensus. A trivial
example is the number of group members each individual tracks in a biological
group to ensure cohesion and continuous information passage: knowing that this
number is usually much smaller than the group size made engineers concentrate
on the problem of achieving an always connected communication network rather
than having a large number of connections per agents.
A stable control algorithm drives the agents of a connected swarm to converge
towards a common, stable behaviour. Consensus is hence a form of stability: in
particular it guarantees that a stable behaviour for a group is achieved whereby
stability of the control law for the single elements is provided.
CHAPTER 2. CONTROL PROBLEM, THEORY AND TOOLS
51
The consensus problem is strictly related to the connectivity network of interactions. If in a swarm there are groups of agents which are isolated, in the sense
that they might have visibility of some neighbours but none of these belongs to
the rest of the swarm, these cannot participate to the group dynamics. Hence an
isolation in terms of network will translate into physical isolation when the group
moves. Indeed a single, globally reachable node in the graph is needed to ensure
the achievement of a consensus as proved for example in [39] and [128]. This
stability and consensus achievement characteristic was extended to asynchronous
swarms under the hypothesis of connectedness by Gazi [129], taking into account
also possible delays in the communication channels.
The challenge of achieving consensus quickly in a swarm, with obvious benefits
for the reactivity of the system, has been attentively investigated. As was discussed in Section 2.6.2, particular networks can enhance the performance in terms
of consensus achievement speed. Olfati-Saber and Murray [130], while pointing
out the centrality of connectivity to reach consensus, related a particular class of
directed graphs to the property of achieving consensus quickly considering also
the effect of delays in communications. The peculiar case of small world networks
has been considered to this end as well, for example in [103, 104]. A review of
the most popular consensus seeking strategies in networked systems was recently
produced by Olfati-Saber, Fax and Murray in [76].
In this dissertation the problem of quick consensus achievement will be analysed
in Chapter 6 in terms of the design of a distributed controller, where the resources
needed for its implementation are provided by the agents on the basis of their
position in the network.
Chapter 3
The Cohesion of Emerging
Patterns and the Emergence of
Bottlenecks
The possibility of shaping a swarm by means of the network of interactions is
explored here for the case of a simple reduction in the number of connections
each agent can establish. A particle system approach is adopted where agents
move in a three-dimensional space by means of pairwise Morse potential and leak
energy because of a viscous damping term. The main results of this chapter have
been reported in [131, 132].
3.1
The Need for a Cohesive Swarm
Within the context of swarming systems it is clear how coherent behaviour of the
swarm depends upon reciprocal interactions amongst the units. These interactions are numerically restricted in biological systems and must be similarly in any
practical implementation of a large scale swarm. As illustrated in the preceding
chapters in these cases, the determinability of the behaviour of the swarm becomes
highly complex. It is thus important to determine conditions in which the behaviour of the swarm becomes more or less coherent due to restricted interactions.
The effect of limited interactions has been addressed in the literature with stability analyses that rely on the swarm staying connected and numerical simulations
in support of the behaviours [61, 96]. At the same time some works have ad52
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
53
dressed the problem of keeping the swarm connected by producing control inputs
in the direction of increasing the connectivity of the graph of interactions, see for
example [133, 134]. However, it is fundamental to assess what characteristics the
swarm needs at agent level in order to establish a communication network that
stays connected as the swarm evolves, even without an active control on the connectivity. Associating agents to nodes of a graph, it is possible to track mutual
interactions as edges of the graph and thus conclude characteristics of the system.
Consider agents that interact according to pairwise potentials. In this chapter,
three potentials will be used to extend the validity of the results obtained beyond
the particular form of interaction of the agents. The Morse potential was already
introduced in Section 2.6.1 and it is repeated for sake of completeness as,
UijM orse
|xij |
|xij |
r
= −C exp − a + C exp − r
L
L
a
(2.25)
where, xij is the relative position vector of an agent i with respect to an agent j.
C a , C r represent the strength of the potential while La and Lr govern the range
of the potentials. The condition La > Lr for C a = C r provide the potential with
a convex shape, which in turn makes its stationary point a stable equilibrium
point, as shown in [28, 32].
The hyperbolic potential was introduced in Section 2.6.1 as,
Uijhyp = Ch [(|xi − xj | − d)2 − 1]0.5
.
(2.29)
where, Ch is a scaling constant and d is a desired inter-agent distance. Finally,
the third potential considered is a simple quadratic potential in the form
Uijq = (|xij | − d)2
(3.1)
where, d is again the desired distance between any two agents. The quadratic
potential provides an inter-agent attraction force increasing linearly as the distance between two connected agents increases beyond d and decreasing in the
same fashion when the distance shrinks below the desired distance.
The equations of motions for the i−th agent of the swarm, regardless the potential
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
54
used, are expressed as,
dxi
= vi
dt
dvi
m
= −∇Ui − σvi
dt
(3.2)
(3.3)
where, σvi introduces a velocity dependent damping term in the dynamics, with
σ being a positive constant. Throughout the analysis presented in this chapter,
σ takes the value of 0.7 as this guarantees a convenient relaxation time in the
P
numerical simulations. Ui = j (aij Uij ), with aij being the entry of the adjacency matrix that takes values “1” or “0” depending respectively on whether the
agents are interacting or not. The presence of a viscous damping guarantees the
achievement of a crystalline formation as per the physical system illustrated in
Section 2.6.1 for an always positive g(H) function. When this happens, as it will
be clear in the following, the connection network doesn’t undergo any further
change, as it is established on the base of inter-agent distances.
3.2
A Minimum Number of Links for Cohesion
The limit to the number of connections per agent necessary to ensure cohesion can
be deduced for some cases depending on the rule agents follow to establish connections. Consider a swarm of N agents whose number of interactions is strictly
limited. Assume a simple connection rule: suppose that an agent can sense the
potential of at most k other agents, and so it does by sensing the k closest ones in
terms of physical positions. As seen in Section 2.7.1, this is consistent with both
biological studies and the performances expected from an engineered multi-agent
system where a limit to the number of communication channels is to be considered. In the remainder of this chapter the connection scheme produced by the
existence of this limit is referred to as the k Nearest Neighbours Rule (k − NNR).
When representing the connection network into a sensing graph, the result is a
directed graph as it can be understood easily through the following considerations. Two generic agents, call them A and B, are each other neighbours in the
graph, if both of them have less than k other agents closer in distance to each of
them. If this is the case, then an undirected link is established between the two
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
55
agents. Conversely, if agent A has k or more neighbours closer in distance, then
agent B will not be included amongst the neighbours of A; hence, although there
might still be a directed edge from B to A, there will be no edge from A to B,
so the link is directed. Each node of the graph then has k outgoing edges, that
is it senses the action of the closest k neighbours, while it can have a number of
incoming edges, that is, it can be sensed by up to the total number of agents in
the graph, excluding itself, i.e. N − 1.
A general consideration is that, for a graph to be connected, the total number
of interactions, hence of connections within the swarm, must be at least N − 1;
this is, amongst others, the case of a line graph or a star graph. In the particular
case presented, when the connections depend upon the relative positions, and
the positions of the agents depend in turn upon the forces exchanged through
the interactions, the pairwise potential presented tends to cluster the interacting
agents. This prevents the spontaneous formation of such edge-saving connected
graphs.
When each agent can keep a number of connections greater or equal to the total
number of agents in the swarm, that is k ≥ N − 1 excluding self connections,
the graph is connected, and in particular it is complete. As more agents join the
system, N increases and eventually becomes greater than k. At this point the
graph of the interactions is no longer complete, yet it can still be connected. It
is straightforward to understand that k = hN/2i (where h·i rounds down to the
nearest integer) for each agent is a sufficient, though not necessary, condition for
the graph to be connected. Hence, the swarm will remain cohesive. Further, in a
dispersed swarm there will be at least one subgroup composed of n agents, with
n ≤ hN/2i = k. As the number of connections per agent is greater than the
number of agents in the subgroup, there must be k − n + 1 edges for each node
in this group connecting to some of the other N − n nodes. As connected agents
gather together, driven to relaxation by the inter-agent forces, this will produce
a cohesive group. This circumstance is illustrated in Figure 3.1 where 8 agents
establish 4 connections each: if the agents found themselves split into isolated
subgroups, at least in one subgroup there would not be enough agents to complete
all the connections. This does not hold anymore when k < hN/2i. For the case
k = hN/2i two clustered, but yet connected, groups arise. This is shown in Figure
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
56
Figure 3.1: The illustration shows how for k = hN/2i the swarm must be connected.
Each of the agents in the subgroup on the left must complete its 4 connections by
joining the group on the right.
3.2 for a planar case where the emergence of a bottlenecked swarm is evident when
the number of agents increases, while the number of interactions per agent is held
constant. As the connections that an agent does not establish in its own cluster
are established always on the base of closeness, these will be with some agents on
the closest region of the other cluster. This gives rise to the dumbbell shape where
the agents in the central, narrower part bond the two clusters together and, by
symmetry, have the same number of connections to both sides of the dumbbell.
Meanwhile, they are sensed by all N agents. Despite the fact that each agent
still keeps a constant number of connection, the bottleneck becomes problematic
when information have to travel across the two halves of the dumbbell. While the
maximum path length is al small as 2 (two hops are sufficient to move from any
node to any other), all the paths that put in communication any two agents not
directly connected pass through the central agents which may experience high
traffic volumes leading to possible chocking problems and lags in the information
transfer. In Figure 3.3 the final arrangement after relaxation of a 130 agent swarm
with 65 connections per agent is shown together with a graphical representation of
the corresponding adjacency matrix for a three-dimensional case. The dumbbell
shape is reflected in the adjacency matrix of the graph as well if each node is
associated to an agent and the nodes are labelled from one-end of the dumbbell
to the other.
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
(N = 60)
(N = 80)
(N = 90)
(N = 100)
(N = 110)
(N = 120)
(N = 121)
(N = 122)
57
Figure 3.2: Swarm systems in a two-dimensional space relaxing to different shapes as
the number of agents N increases while the number of connections allowed per agent
is held constant at k = 60.
58
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
0
20
row index i
40
60
80
100
120
0
20
(a)
40
60
80
column index j
100
120
(b)
Figure 3.3: Dumbbell emergence in a three-dimensional case. (a) A 130 agent swarm in
a dumbbell shape due to the number of connections per agent being limited to 65 and
(b) the corresponding adjacency matrix obtained associating each node to one agent,
after having them sorted from one end to the other of the dumbbell. Dots represent
nonzero entries.
3.3
The Limit Case and the Cheeger Constant
Numerical simulations and logical arguments presented in the previous Section
show dramatically how hN/2i is a critical value for the number of connections per
node. In the following, the bottleneck characteristics of the system for the critical
case of k = hN/2i are studied considering the Cheeger constant of the graph. In
graph theory the Cheeger constant or Isoperimetric number is a numerical value
associated to a connected graph that identifies the groups of vertex which are
more poorly connected to the rest of the graph. A small value of the Cheeger
constant translates to having a group of vertices in the graph that are reachable
through a smaller number of paths compared to the total number of paths that
link the other vertices in the graph. In order to define the Cheeger constant, the
definitions provided by Fan Chung in [135] are used. Consider a graph G{V, E},
according to the symbols defined is Section 1.4.3. Let S be any subset of V in
the graph and S is its complement, while ∂S is the subset of all edges connecting
a node in S to one in S, that is
∂S = {{i, j} ∈ E(G) : i ∈ S and j ∈ S}
.
(3.4)
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
59
For any subset S the volume of S is defined as,
vol(S) =
X
dx
(3.5)
x∈S
where, dx is the degree of the generic node x, equal to the sum of the weights of
all the edges exiting from x in case of G being a weighted graph. Relying on the
above definitions, the Cheeger constant referred to a subset S ∈ G is defined in
[135] as,
hG (S) =
|∂S|
min{vol(S), vol(S)}
.
(3.6)
For an undirected graph, the Cheeger constant is the minimum value achieved
by Equation 3.6 over all possible partitions S of G, that is
h(G) = min hG (S)
S⊂V
.
(3.7)
When the Isoperimetric number is small, there is at least a partition S of the
graph that is hardly reachable, which means there is a bottleneck in the network
of paths that connect the nodes of the graph.
When the graph G is direct, the sum of edge weights cannot be used to calculate
the Cheeger constant that has to be obtained, for an oriented graph G, as [136],
h(G) = inf
S⊂V
F (∂S)
min{F (S), F (S)}
(3.8)
where, F = [fij ] is a function known as “circulation” that takes a nonzero value
on each directed edge of the graph. F can be defined on the basis of the probability distribution matrix P = [pij ] and its dominant left eigenvector, φ. This is
identified, here and thereafter, as the eigenvector corresponding to the rightmost
eigenvalue on the complex plane. Circulation can then be defined as,
fijφ = φi pij
(3.9)
where, i and j are indexes corresponding to generic nodes in the graph while φi
is the i − th entry of the dominant left eigenvector. P = pij is a matrix whose
generic entry i, j gives the probability of moving from vertex i to vertex j based
on the number of links departing from i, accounted through the entries of the
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
adjacency matrix, that is
aij
pij = P
j aij
.
60
(3.10)
Equation 3.9 does not directly provide information on the behaviour of the
Cheeger constant as a function of the number of nodes. However, other circulations can be used to define the Isoperimetric number as long as they satisfy
the condition,
X
i
i→j
fij =
X
fjw
(3.11)
w
j→w
for the generic nodes i, j and w, that is the flow in one node of the graph is null.
Equation 3.11 is to be interpreted as balance between the flow entering in node j
from all the neighbouring nodes accounted by the indices i and flow leaving the
node towards all the neighbours accounted by the indices j. For the matrix F this
translates in having the sum over the rows equal to the sum over the columns,
that is the sum of the elements of the first row is equal to the sum of the elements
of the first column, the sum of the elements of the second row is equal to the sum
of the elements of the second column and so on. Therefore, instead of using a
circulation based on the dominant left eigenvector as in [136], a new circulation
is derived. This leads to an analytic expression for the Cheeger constant, which
is linked directly to the number of agents in the swarm being dependent on the
number of nodes in the graph. The new circulation can be defined by constant
values associated to the edges of the graph grouped as follows. Two edges will be
associated to the same value of the circulation if the nodes to which they point
have the same number of incoming connections, that is are observed by the same
number of agents. This implies that if a node has s incoming edges, then all s
will be associated with the same value of the circulation, that is also the same
for all edges entering nodes with a total of s incoming edges. This is illustrated
by the sketch in Figure 3.4.
This new definition for the circulation requires that the number of nonzero elements in each column of the adjacency matrix is known. This corresponds to
considering known the number of incoming edges for each node. For this reason
the adjacency matrix is idealised as composed of two blocks plus some linking
rows. To be consistent with the connection rule, each row of the idealised adjacency matrix shall still present hN/2i nonzero entries. The cases of even and odd
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
61
Figure 3.4: Circulation definition. The edges of the same colour will be associated with
the same value of the circulation as they all enter in node with the same number of
entering edges. This circulation is composed of just 2 values, one represented by the
colour blue for edges entering nodes with just one incoming edge, and the other by the
colour red, for edges entering nodes with a total of 3 incoming edges.
numbers of nodes can be detailed as follows.
Even Number of Nodes
Two agents, by symmetry, will find their position in the centre of the dumbbell,
as shown in Figure 3.2 for N = 120. The agents in the centre are sensed by
both the clusters and their columns in the adjacency matrix do not have any zero
entries, except along the main diagonal. This idealised approximation is shown in
Figure 3.5 for a graph composed of 60 nodes compared to the one resulting from
a spontaneous relaxation of 60 agents following the dynamics earlier described,
with initial conditions randomly distributed in a unit cube.
The idealisation of the adjacency matrix in Figure 3.5.b represents the connection structure that produces the smallest possible bottleneck compliant with the
physical restriction of having at least 2 central communicating agents for N even.
This leads the corresponding Cheeger constant to be the minimum achievable,
although, given the spontaneous relaxation of the agents, a spatial configuration
that gives rise to this minimum value is not guaranteed to occur as shown in
Figure 3.5.a. Being defined only for existing edges, the circulation used has a
matrix representation with the same nonzero entries as the adjacency matrix.
The actual value for each entry is found by requiring that entries belonging to
columns with the same number of nonzero elements take the same value. This,
together with the condition expressed by Equation 3.11, that is a zero net flow
at the node, provides a number of relations that are sufficient to fully define the
circulation for the matrix in Figure 3.5.b. This circulation takes four possible
62
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
values, as there are four different column lengths in the matrix; call these a, b, c
and d. Four equations are hence derived by requiring the sums along the rows
and the columns of the circulation matrix to be equal, as per Equation 3.11, one
for each column length as
N
2
N
2
− 2 a + b + N4 − 1 c + 2d
− 1 a + N4 − 1 c + 2d
− 1 a + b + N4 − 2 c + 2d
(N − 1) d = b + 2 N4 − 1 c + d
−2 a =
−1 b =
N
c
=
2
N
4
N
4
N
4
(3.12)
where, again N is the number of agents/nodes in the graph. To allow a solution
other than the zero solution, a is considered known and a ∈ ℜ+ . The solution of
the system in Equation 3.12 can be expressed as function of a as,
b =
c =
d =
N −2
a
N
N
4
a−b−2d
N
−1
4
N−2
− N +N
2
N +2
=
2−N
N
+N
−2
4
N
4
2−N + N
2
N
2+N
−1
(3.13)
a
a
0
0
10
10
20
20
row index i
row index i
The Cheeger constant can thus be derived from its definition of Equation 3.7 as
30
30
40
40
50
50
60
0
10
20
30
40
column index j
(a)
50
60
60
0
10
20
30
40
column index j
50
60
(b)
Figure 3.5: Adjacency matrices for spontaneously relaxed (a) and idealised connected
(b) 60 agent swarm.
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
63
follows,
h(G) = h(N) =
N
4
−1
N
2
−2 a+
N
d
2
N
2
+ N4 − 1 c
− 1 b + N4 − 1 N2 c + (N − 1) d
.
(3.14)
This, after long but straightforward algebraic manipulation, can be reduced to
h(G) = h(N) =
2N (N 2 − 4N + 6)
N 4 − 2N 3 − 6N 2 + 20N + 8
.
(3.15)
Equation 3.15 is obtained in the hypothesis of N being a multiple of 4, otherwise
more complicated expressions are to be defined that take into account the shifting
of the rows of one position depending on the value of N. In this case Equation
3.14 should be used considering the minimum achievable for each configuration
rounding up or down the values of N/4 that are used is the equation.
Equation 3.14 can be proved to be a decreasing function of N. It is composed of
a numerator which is a first degree polynomial in N and a denominator which is
a second degree polynomial in N. It follows that for N positive and integer this
is an always decreasing function. In particular, for the same reason, h(N) tends
to zero as N approaches infinity, that is,
lim h(N) = 0
N →∞
(3.16)
The Cheeger constant so defined does not depend on any of the coefficients as
they are all proportional to the variable a, which eventually cancels out by substituting Equation 3.13 into Equation 3.14, that makes all the terms proportional
to a.
Odd Number of Nodes
The same procedure can be followed for an odd number of nodes. The minimum
number of bridging units, in this case, is one, with the two halves of the dumbbell
comprising hN/2i = N 2−1 agents as shown in Figure 3.2 for N = 121. In this case,
a circulation can be composed of as few as three values, as three is the number
of different column length found in the adjacency matrix describing the minimal
link arrangement. When hN/2i is odd (e.g. N = 31, 51 etc...) the positions of
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
64
the nonzero entries of the central row of can be shifted one left or right of one cell.
This is understandable considering the fact that when hN/2i is odd the central
agent cannot split its connections equally between the two halves. The following
derivations are performed considering hN/2i being even as and the central agents
splitting evenly its connections between the two halves as in the case shown in
Figure 3.6 for a swarm of 61 agents. Under these condition, hN/4i = N 4−1 holds.
Following the same derivation described in Section 3.3, in the hypothesis just
0
10
row index i
20
30
40
50
60
0
10
20
30
40
column index j
50
60
Figure 3.6: Adjacency matrices for idealised connected 61 agent swarm.
introduced, the values of the circulation associated to the adjacency matrix in
Figure 3.6 are obtained as,
N
N
N
h 2 i − 1 a = h 4 ia + h 4 ib + c
= h N4 ia + h N4 ib + c
h N2 ib
(N − 1) c = h N2 ib.
(3.17)
Considering a known and a ∈ ℜ+ as for the case of N even, the solution of the
system in Equation 3.17 is then,
b =
c =
ia
hN
4
hN i
2
hN
i−h N
i− N−1
2
4
hN
i
2
N −1
hN
ia
4
(3.18)
hN i
2
i−h N
i− N−1
hN
2
4
.
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
Considering hN/2i =
N −1
2
and hN/4i =
(
b =
c =
N −1
4
N −1
N −3
N −1
2(N −3)
65
these further simplify as
.
(3.19)
The Cheeger constant can thus be derived from its definition of Equation 3.7 as
follows,
h N2 ic
h(G) = h(N) = N
.
(3.20)
h 4 i h N2 i − 1 a + h N4 ih N4 ib
In the hypothesis hN/2i being even, Equation 3.20 can be algebraically manipulated to obtain,
N −1
(3.21)
h(G) = h(N) = 2
N − 4N + 5
As before, it can be easily verified that
lim h(N) = 0
N →∞
(3.22)
being the denominator a polynomial of higher degree of the numerator. Also in
this case the Cheeger constant tends to zero as the number of agents increases
towards infinity.
3.4
Effectiveness of the Model
The results obtained in the previous section are checked against numerical simulations of particle swarms relaxing under the action of the potential functions
described in Section 3.1. A comparison is made between the above analytically
determined Cheeger constant as a function of only the number of nodes, and the
Cheeger constant obtained through the spontaneous relaxation of agents driven
by pairwise potential, that is to say creating a spontaneous network of links based
on the k − NNR. With reference to Equations 2.25, 2.29 and 3.1, the parameters used to shape the potentials are C a = C r = La = Ch = 1, Lr = 0.2. The
desired distance in the hyperbolic and quadratic potentials d is set to 0.1 in order
to obtain a separation of the same order of magnitude of the one obtained with
the Morse potential. Finally, reference to Equations 3.2 and 3.3, the velocity
dependent damping constant sigma is set equal to 0.7 to guarantee a convenient
relaxation time in the simulations, while the mass of each agent is set unitary.
66
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
The simulations are performed providing the agents with random initial positions
and velocities in the unit sphere; the simulations are run for 35 simulated seconds
and the state and the integration is performed using an explicit Euler scheme
with a time step of 10−3 seconds. The values of the Cheeger constants obtained
by averaging 100 tests are presented in Figure 3.7 for an even number of agents.
A comparison for an odd number of agents is also presented here and shown in
Figure 3.8. The tests are compared to analytic detemined Cheeger constant also
for odd values of hN/2i by developing the adaquate expression needed through
the same procedure described in Section 3.3.
0.45
Average Morse potential
Average quadratic potential
Average hyperbolic potential
Upper bound Morse potential
Upper bound quadratic potential
Upper bound hyperbolic potential
Lower bound Morse potential
Lower quadratic Morse potential
Lower bound hyperbolic potential
Analytic lower bound prediction
0.4
0.35
h(G)
0.3
0.25
0.2
0.15
0.1
0.05
0
20
40
60
80
100
120
140
Nr of Agents
Figure 3.7: Comparison of Cheeger constant obtained by the analytic expression and
the one obtained by making a swarm of agents relax through pairwise interactions for
even number of agents. Average values for each potential and extreme values obtained
over all the numerical tests are reported as well.
It is demonstrated that the Cheeger constant obtained from the adjacency matrix
made up of two blocks and one or two linking rows tracks closely the numerical
data, although being always below, or at most equal, to the least value. This
result holds independently from the kind of potential used but differences are
appreciable in the clustering produced by the difference in the interactions. In
67
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
0.1
Analytic lower bound prediction
Average Morse potential
Average quadratic potential
Average hyperbolic potential
Upper bound Morse potential
Upper bound quadratic potential
Upper bound hyperbolic potential
Lower bound Morse potential
Lower bound quadratic potential
Lower bound hyperbolic potential
0.09
0.08
0.07
h(G)
0.06
0.05
0.04
0.03
0.02
0.01
0
20
40
60
80
100
120
140
Nr of Agents
Figure 3.8: Comparison of Cheeger constant obtained by the analytic expression and
the one obtained by making a swarm of agents relax through pairwise interactions for
odd number of agents. Average values for each potential and extreme values obtained
over all the numerical tests are reported as well. Some of the mamimum values have
been cut off the plot to make the curves more clear.
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
68
particular, it can be noticed how the Morse potential guarantees an higher degree
of robustness compared to the other two as it provides an arrangement with more
links across the two halves of the dumbbell. The comparison between quadratic
and hyperbolic potentials reveals how for these two interactions, a more or less
robust behaviour is dependent on the number of agents. Overall, the analytic
curve, which has been proven to converge to zero as the swarm size increases
towards infinite, provides a lower bound for the prediction of the Cheeger constant and can be easily determined for very large swarms, where calculation of
isoperimetric number using eigenvalues becomes problematic.
Results show that, on average, the graph structure the swarm creates keeps a
number of crossing links between the two clusters, in excess of those strictly
needed to ensure cohesion. However, while the actual value of the Cheeger constant depends on the final configuration of the swarm and the form of inter-agent
potential, its lower bound only depends on the number of agents in the swarm,
and, although representing a limit case, it is sometimes achieved by the Cheeger
constant. Moreover, as the lower bound of the Cheeger constant is a decreasing
function of the number of agents, it can be argued that the links between the two
halves of the swarm are an always smaller minority of the total. That is, when
swarms grow larger, the number of links that help towards cohesion reduces as a
function of the total swarm size if compared to the total number of interactions.
Thus as the Cheeger constant is a measure of the bottleneck characteristics of a
graph, the results show how a swarm of agents that interact on the base of the
k − NNR, with k = hN/2i, tends to become more and more bottlenecked as the
number of agents increases. It can be concluded that an increase in the number
of agents is not entirely compensated by an increase in the number of cross-links
between the two clusters of the dumbbell that the system eventually relaxes into.
The Cheeger constant can then be bound from below just by knowing the number of agents and by considering the approximation of the network configuration
produced by the k − NNR. Thinking towards the engineering of multi-agent systems, in the case of a very large number of agents, possible consequences arising
from the behaviour described are even more incisive. Emergence of a bottleneck
restricts sensing and information flow, hence updates of the system’s state, which
agents need for cohesion, is delayed, which in-turn directly results in fragmentation into sub-groups even in the case of k close to, but still greater than, hN/2i.
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
69
A systematic study would be required to better understand the effects of the
Cheeger constant reduction on the engineering of multi-agent systems. This is
not addressed here and represents a possible direction for future developments.
3.5
Final Considerations
The butterfly or dumbbell shape emerging here, as a result of decreasing the
connections per agent to hN/2i, is not common in biological swarms, but it is of
great relevance in engineering systems. The simple reduction in the number of
available connections leads to a shaping of the formation in a predictable way.
This is different from the distortions in the crystalline patterns observable when
a particular network is imposed as shown in Figure 2.7. The shaping towards
a more elongated formation starts appearing already when the number of connections is larger than hN/2i, however, it is just at k = hN/2i that the limit
case presents itself. This produces the narrowest bottleneck achievable for spontaneously relaxing agents driven by inter-agent forces. With a further reduction
of the number of connections per agents, the swarm cannot guarantee cohesion
without actively controlling it. The dumbbell shape is a predictable outcome of
the connection reduction. Beyond the physical shape, the risk of bottlenecks in
communications has to be considered. In an APF approach, as the one just discussed, the shape is the most visible consequence of the connection reduction, but
even more alarming is the reduction in the Cheeger constant which indicates how
a small number of agents become responsible for the cohesion of the whole group.
These can be expected to experience a large volume of information flow to allow
communications between the two halves of the group. This in turn sets the design parameter for the communication capabilities of the agents, or alternatively,
elects as central agents those ones which can guarantee the best performance in
terms of data relaying.
When N increases, k = hN/2i becomes a large number of connections to keep for
single agents, however this is a condition to guarantee cohesion without actively
controlling it. From the design point of view a trade-off should be made to understand whether a multi-agent system would benefit more from a active cohesion
control as in [133] or, as in this case, from a cohesiveness guaranteed regardless
the control.
CHAPTER 3. EMERGING PATTERNS AND BOTTLENECKS
70
While the cohesion is not dependent upon the control law chosen, the shape,
under some respect, is so. The different behaviour of the Cheeger constant with
different potential witness how a change of agent interactions may reflect into
slight changes in the final arrangement. In particular, More potential seems
to cluster agent in the proximity of the bottleneck, more than the quadratic
and hyperbolic potentials. This is confirmend by the higher values achieved on
average by the Cheeger constant, providing a beneficial effect towards keeping
system safe from fragmentation. Similar difference can be expected when the
tuning parameters of the potential ar tuned within the stability region outlined
in Section 3.1, although this was not shown in this chapter. However, despite the
fact different potentials return slightly different results, the physical limit of the
Cheeger constant is confirmed beyond the particular potential used, which is the
main result reported in this chapter.
Chapter 4
Fractal Fractionated
Architectures and Arrays Application in Space
In Chapter 3 it was shown how topographic characteristics of the network reflect
the final arrangement of the agent crystalline formation. The possibility of shaping a crystalline formation through changing the network of interactions can be
exploited in the design of formation flying architectures. In particular, the design
of the final arrangement of a number of agents can be worked out starting from
the arrangement of agents in smaller groups, and then, instead of increasing the
group size, the number of groups can be increased. The same control laws that
drive the arrangement of the agents in a group can be used to drive the arrangement of the groups in larger formations, with the object of the relative positioning
control shifted from agents to groups. This technique can be used iteratively to
produce always larger formations with a well-structured communication network.
The use of the APF method allows for stable behaviour, which is analytically
provable. In this chapter the control of a formation of satellites in Earth orbit is
designed to deploy a fractionated antenna. The self-similar characteristics of the
control law and of the connection network provide a self-similar arrangement of
the spacecraft and hence of the antenna elements carried by them. This enhances
the performance of the antenna array in terms of directivity. The main results of
this chapter have been published or presented in [137–140].
71
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
4.1
72
Fractionated Spacecraft and Formation Flight
A fractionated spacecraft is composed of several independent modules that, flying in close proximity, interact with each other as the subsystems of classic,
monolithic spacecraft do. Having the functional units hosted on board several
platforms, rather than a single one, provides an increased degree of flexibility
and task sharing. A single subsystem can be used in this sense by more than
one spacecraft. Although not strictly necessary, fractionated spacecraft concepts
often make use of formation flying to rely on precise positioning of the different
spacecraft parts beyond the property of being in a close range to allow interactions.
Formation flight techniques are gaining popularity for space science, remote sensing and telecommunication applications [141–145]. Formation flying concepts
proposed so far have been based on a relatively low number of cooperating spacecraft, as in the case of Lisa, Proba-3 or StarLight missions [146–148]. As the
number of spacecraft increases, system complexity increases as well, with significant consequences for the operations and control aspects. Nonetheless, the
exploitation of a fractionated architecture and formation flight with an increased
number of elements, which maintains an acceptable level of system complexity,
can be pursued through the control of autonomous and independent agents as a
single group entity [142, 149].
Formation flying techniques in space have been proposed in the area of communications or for applications where electromagnetic radiations have a primary role,
see for example the Olfar project [150]. The possibility of producing complex
patterns using spacecraft in a reliable way will enable the potential of grouping
a number of antenna elements into a cooperative structure. This has long been
known and applied in antenna array theory [151, 152] and the move to spacebased architectures appears extremely promising.
The key point in the exploitation of formation flying techniques for the deployment of an antenna array is the performance of a homogeneous pattern of array
elements can be matched or surpassed by fractal geometries as per [153] and
[154]. Fractal geometries as defined by [153] can be considered self-similar structures propagated from a core initiator through a number of stages of growth by
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
73
an identical generator. Application of fractal geometries in antenna array design
has mainly focussed on single structures, that is to say one device housing the
antenna array. In this chapter a new scenario is considered where, each satellite
houses an antenna that contributes to form the fractal pattern. Hence, the problem turns into producing a fractal pattern from a formation of spacecraft, which
provide a platform for a number of array elements able to exploit the fractal pattern characteristics.
From a control point of view this can be realised through Artificial Potential
Functions. The way to obtain complex formations through APF, while maintaining a high-degree of reliability and analytically provable characteristics, can be
revealed through the design of a limited connection network. As seen in Section
2.7, network characteristics reflect on the final pattern deployed through APF
acting along its edges. In particular when the connection network presents selfsimilarity characteristics, i.e. the same network structure repeats for nodes and
groups of nodes, this impacts not only on the final formation but also on the
stability and robustness properties, which can be analysed in the same fashion
for single spacecraft or groups of those. The overall control architecture result is
scalable and with a good degree of fault tolerance. The most inner nodes belong
to more than one subgroup of spacecraft and, as such, are more strongly embedded into the network structure and more resistant to connection loss. The most
peripheral nodes just contribute marginally to the formation shaping and a fault
at this level is not catastrophical for the formation. This will be clearly explained
later on in this chapter.
From the array point of view self-similarity and sparseness, that is the low number
of elements per unit area, lead to a number of benefits: similar performance in
operation across a number of frequencies becomes possible due to the repetitive
nature of the array pattern; additionally array performance degrades gracefully
with element failure and finally equivalent performance can be achieved for a
fraction of the number of elements used in square lattice arranged arrays [155].
An ensemble of N spacecraft is considered divided in subgroups of n agents
such that N = nk for some integer k > 0. It is assumed that each spacecraft
carries an element of the array where the pair spacecraft-array element will be
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
74
named as agent. Spacecraft and array element will be instead used when referring
to these parts in any of the agents. The agents are connected according to a
non-directed graph described by the adjacency matrix A of dimension N. The
spacecraft are controlled through APFs that act only along the edges of the graph
with a pairwise scheme. There is no global position or orientation of the agent
formation, but within the formation, relative positions are considered for agents
and groups of agents while relative orientation is considered for groups of agents
only. This implies that the single array elements are assumed to be correctly
pointed or, which is what is done here, they are assumed as isotropic sources.
In this section it is shown how it is possible to obtain a self-similar formation
starting from mutually interacting agents and how the array performances can
be analysed for such a system. Artificial potential function characteristics and
communication graph topology are described as well. The fundamental concept
of applying fractal geometries to the design of antenna arrays using a self-scaling
method is described. Only planar configurations are considered but the same
argumentations can also apply to linear as well as 3D formations.
4.1.1
Artificial Potential Functions
The spacecraft are controlled through artificial potential functions operating
along the edges of a communication network. The APFs operate on a pairwise
basis, that is they do not depend on position or velocity of the agents but only
on their positions relative to the other spacecraft which they are connected with;
in particular the Morse potential is used. This was already introduced in Section
2.6.1 and used in Chapter 3, but here, differently from what done previously, the
tuning parameters are not the same for all the agents: they depend on the pair
of agents they refer to. The parameters C a , C r , La and Lr are redefined and
associated to a pair of agents. For the generic pair of agents ij the attractive and
repulsive components of the potential are respectively defined as,
Uija
Uijr
|xij |
=
exp − a
Lij
|xij |
r
= Cij exp − r
Lij
−Cija
(4.1)
(4.2)
where, Cija , Cijr , Laij and Lrij are constants whose values shape potentials sensed by
agent i because of interaction with agent j; xij is the relative position vector of
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
75
agent i with respect to agent j. The control law is completed by a virtual viscouslike damping in the form σvi . This control law together with the hypothesis of
no external disturbances and idealised sensing and actuation capabilities results
in the following equations of motion,
ẋi = vi
(4.3)
mv̇i = −∇Uia − ∇Uir − σvi
(4.4)
where,
Uia =
X
(aij Uija ) Uir =
j
X
(aij Uijr )
(4.5)
j
and aij is the entry of the adjacency matrix as defined in Section 4.1.2. This approach for controlling spacecraft formations was already undertaken in literature,
see for example [38, 47, 50, 55]
4.1.2
Adjacency Matrix
As reported in Section 4.1.1 agents communicate through a network of links. In
general in a network system studied through graph theory, an adjacency matrix
is a matrix which presents a nonzero entry in the ij location whenever there is a
directed edge from node i to node j, which corresponds to a communication link
between the two agents represented by nodes i and j. Moreover, the matrix is not
weighted, that means its elements only take 0 or 1 values; the “weight”, i.e. the
strength of the interactions, is provided by the APF used. The adjacency matrix
here proposed is symmetric, hence the graph is not directed but this does not
imply that the virtual interactions amongst the agents are symmetric. Along the
diagonal the edges belonging to fully connected n-agent groups form a block-like
matrix. In each group of n2 agents the groups of n agents are directly linked
through 2 linking agents (4 directed edges). These account for both position and
orientation as shown in Figure 4.1. Extending this scheme to more agents, it can
be defined that for any group of nk agents for any integer k > 0 the connections
with groups of equal number of agents are ensured through 2nk−1 agents. The
adjacency matrices for 25 and 125 agents are given in Figures 4.2 and 4.3. In
Figure 4.4 the node degree is reported for the adjacency matrix of dimension 125,
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
76
Figure 4.1: Connections between groups of fully connected agents. Four groups of
four agents each keep relative distances and orientations by means of links, in red,
connecting agents of different groups. As positions of connected agents can be adjusted
and kept, this results into the positions and relative orientation of the groups being
uniquely determined.
that is the number of links each node is connected to. Nodes are sorted from the
central to the peripheral ones.
The network is designed such that the peripheral nodes are weaker than the
central ones. This means that loss of control of one node due to loss of link is
more likely for nodes that belong to peripheral region of the formation, hence
they do not act as a bridge between large portions of the ensemble. This implies
that the loss of some links is more likely to produce the disconnection of a smaller
and peripheral portion of the network than of a large portion. Each node is in
any case at least connected to n − 1 other nodes. When the number of generators
increases, those groups which were end-points for the previous generator become
embedded and more firmly bonded into the larger pattern. This ensures that in
the most critical scenario the loss of at least n−1 links is needed for fragmentation
to occur.
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
77
0
row index j
5
10
15
20
25
0
5
10
15
column index i
20
25
Figure 4.2: Adjacency matrix for an ensemble of 25 agents. The nonzero entries are
represented by dots.
0
20
row index j
40
60
80
100
120
0
20
40
60
80
column index i
100
120
Figure 4.3: Adjacency matrix for an ensemble of 125 agents. The self-similarity of the
matrix can be noticed.
78
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
30
25
Degree
20
15
10
5
0
0
20
40
60
80
100
120
Node Index
Figure 4.4: Node degrees as number of links belonging to each node. A self-similar
scheme can be observed with nodes in central position being the most connected ones.
In this scheme the maximum number of connections per node is 28.
4.1.3
Distributed Fractal Antennas
The advantage of grouping individual antenna elements into arrays consists in
the possibility to produce a highly directional radiation pattern, that is the resulting antenna has a very narrow beam and concentrates the radiated energy in
a given direction minimizing its dispersion elsewhere. This can be quantified by
the directivity, which is a measure of how efficient an antenna is at radiating its
energy in a desired direction. Using a group of smaller distributed antenna elements to steer the main beam, especially in the context of space communications,
offers a number of benefits including cost reduction and risk mitigation. Fractal
electrodynamics studies the arrangement of these radiating elements into fractal
patterns. It is defined as the combined study of fractal geometries with electromagnetic theory and provides methods for the theoretical analysis and synthesis
of fractal antenna arrays. When an array presents radiating elements arranged
into a fractal geometry, then it is termed as a fractal array. It provides the
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
79
Figure 4.5: Homogeneous square lattice array configuration
opportunity to merge the benefits offered by periodic and random arrays, while
mitigating some of the side effects.
In deterministic fractals the pattern can be achieved by switching elements of
a fully populated (square lattice) array on/off until the desired fractal pattern
emerges. Following this procedure, the thinned generating sub-array can be
copied, scaled and translated to produce the final array. Deterministic fractal
arrays created in this manner can be viewed as arrays of arrays due to the recursive nature of the development procedure.
The effects of several antenna elements combined in an array are mainly visible in the overall radiation pattern, which changes from the one of the single
elements. These changes can be modelled using the so called array factor : it
quantifies the effect of combining radiating elements in an array without the element specific radiation pattern being taken into account. Figure 4.5 shows a
symmetric planar array with uniformly spaced elements, dx and dy apart in the
x− and y−directions, where each element is considered an isotropic source; spacing of elements is commonly measured in fractions of wavelengths. In terms of
mathematical modelling the array factor Γ in a given direction, identified by the
azimuthal angle φ and the elevation angle θ, enters the directivity in the sense
that it can be used to quantify the amount of energy radiated in a certain direction, which is then to be compared with the total energy radiated. Directivity is
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
hence:
Γ2P (θ, φ)
DP (θ, φ) =
1
4π
R2π Rπ
0 0
80
(4.6)
Γ2P (ϑ, ϕ) sin(ϑ) dϑ dϕ
where, the array factor in the (θ, φ) direction is compared against its integral. The
integral sweeps an hemispherical surface because of the planar configuration of
the array. The subscript P refers to the level of growth of the fractal as ΓP attains
different values depending on the stage of growth of the fractal. In particular,
defining δ as the growth or expansion factor, that is the number of elements on
the side of a square lattice needed to produce the initiator, the basic structure of
the fractal, then the array factor is
ΓP (u) =
P
Y
Γfrac (δ p−1u) ,
(4.7)
p=1
where, u = (θ, φ). δ controls how much the array grows with each application of
the generating sub-array and corresponds to the number of elements on the side
of the symmetric planar array before the thinning.
The frequency at which the array operates scales with the fractal configuration
starting from the frequency f0 at which the individual elements are designed to
operate
fP = f0 δ −P
.
(4.8)
Conventional symmetric arrays are typically designed to operate at a chosen
frequency f0 in the sense that they are expected to focus most of their energy in
the main lobe rather than radiate their energy evenly in all directions. Fractal
arrays are known to exhibit multiband characteristics as demonstrated by [153]
and [154], which concluded that infinite fractal arrays have the same array factor
at any infinite number of bands. For bandlimited realisation similar properties
only exist for as many bands as growth stages, i.e. P . Fractal distributed arrays
present a number of appealing features for space architecture to produce their
deployment. The use of some characteristics of APF allows a Purina fractal
arrangement to be produced in a relatively simple way. While fractal arrangement
of array elements have been investigated in the context of fractal electrodynamics,
the following sections are aimed to illustrate the innovative integration of this
fractal shaped antenna in a formation flying architecture controlled through the
APF method. A more detailed analysis of the array characteristics emerging out
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
81
of the fractal deployment is contained in the work the author developed with
Philippos Karagiannakis et al. [137].
4.2
The Control Law and the Fractal Antenna
In this section the characteristics of the control technique used to drive an ensemble of agents towards the formation of a fractal pattern and the issues related
to the design of a fractal shaped antenna array are considered. It is first shown
how an asymmetry in attraction-repulsion potential leads necessarily to a central
symmetry configuration. It is then shown how the APF coefficients are calculated
in order to get the desired distance between agents. Analysis of the control law is
completed by considering the nonlinear stability characteristics. The characteristics of the fractal antenna deployed through the satellite formation are finally
illustrated for the case of a Purina fractal antenna array [153]. In the remainder
of this chapter just the case of an initiator of 5 elements is considered, that is,
with reference to Section 4.1, n = 5.
4.2.1
Central Symmetry Emergence
Central symmetry emerges at initiator level by means of an asymmetry between
the interactions of an agent with the group. This is obtained through a different
value of the Lrij parameter along the directed edges connecting the agent to the
other 4 in the initiator structure. This is herein explained by finding the conditions that make the artificial potential derivatives null along two orthogonal axes
centred on the agent considered. Considering a spatial arrangement of n agents
the first derivative of the artificial potential sensed by the generic i − th agent
can be calculated as in Equations 4.9 and 4.10.
n
Cija
|xi − xj |
exp −
−
Laij
Laij
n
|xi − xj |
∂Ui X Cija
exp −
−
=
∂yi
Laij
Laij
j=1
∂Ui X
=
∂xi
j=1
Cijr
|xi − xj |
xi − xj
exp −
(4.9)
r
r
Lij
Lij
|xi − xj |
Cijr
|xi − xj |
yi − yj
exp −
(4.10)
r
r
Lij
Lij
|xi − xj |
Excluding the trivial case for Lrij = Laij and Cijr = Cija , Equations 4.9 and 4.10
can be made null while satisfying the stability conditions in [32] Lrij < Laij . From
here on, only changes in the Lrij parameter are considered, where i, j refers to
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
82
their indexing within the 5 agent group, while Laij , Cija and Cijr are considered
independent from the pair of agents, that is, these take the same value for every
index i, j, hence the indices will be dropped.
Consider the planar formation in Figure 4.6, which will be used extensively
throughout this chapter. Agent 1, by symmetry, is in equilibrium for the y
component of the potential.
Figure 4.6: 5 agents arranged in a homogeneous formation due to all-to-all potential
with the same coefficients for all the agents
The equilibrium for the y-axis holds for all possible distances d either in case
Lrij = Lr for all ij, that is, it takes the same values along all the edges, or
in the case agent 1 has a different repulsive scale distance. In the reference
frame chosen any point along the x-axis has null y derivative of the potential
because of the symmetry of the formation about the x-axis. Equilibrium along
the x-axis for agent 1 does not lead to an explicit expression for the equilibrium
distance, nonetheless the x derivative of the potential referring to any agent can
be calculated. Due to the homogeneity of the configuration, any agent can be
taken to analyse the artificial potential field. In particular for agent 1
∂U1
∂x1
where,
Ca
d
d2
= 2 a exp − a (cos α) + exp − a (cos β)
L
L
L
pentagon
r
d
d2
C
−2 r ′ exp − r ′ (cos α) + exp − r ′ (cos β)
L
L
L
(4.11)
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
d
d2 =
2
s
1
tan α +
cos α
2
+ 1 = kd
83
(4.12)
can be determined, where k > 1. This is considered as an initial equilibrium scenario for some equilibrium distance d and for Lr = Lr ′ that is the same repulsive
scale distance sensed by all the agents. In this scenario, Equation 4.11 must return zero but if Lr 6= Lr ′ and in particular Lr ′ < Lr the separation distance must
shrink, that is the equilibrium distance reduces as the scale separation distance
shrinks. This can be verified by differentiating Equation 4.11 with respect to Lr ′ .
This returns
dU1
dx1
Cr
d
kd
= 2 r ′2 exp − r ′ (cos α) + exp − r ′ (cos β)
L
L
L
r
r
r
C
d
kd
C
C
−2 r ′
exp − r ′ (cos α) + ′2 exp − r ′ (cos β)
.
L
Lr ′2
L
Lr
L
pentagon
dLr ′
(4.13)
Equation 4.13 can be shown to be negative, as given in Equation 4.15, that
is verifying that a reduction of Lr ′ produces an acceleration on agent 1 in the
direction of positive x-axis, hence a reduction of its equilibrium distance:
Cr
d
kd
2 r ′2 exp − r ′ (cos α) + exp − r ′ (cos β)
L
L
L
r
r
r
C
C
d
kd
C
exp − r ′ (cos α) + r ′2 exp − r ′ (cos β) < 0
−2 r ′
L
L
L
Lr ′2
L
(4.14)
d
d
kd
kd
∴ 1 − r′
exp − r ′ (cos α) + 1 − r ′
exp − r ′ (cos β) < 0
.
L
L
L
L
(4.15)
′
This is always verified for d > Lr . The sufficient condition d > Lr ′ can be
obtained by a wide choice of system parameters and understood by inspecting
the equilibrium distance for the simple case of two agents. This is obtained by
summing up and setting equal to zero the derivatives of Equations 4.1 and 4.2
for |xij | = d, and then solving for d
C a Lr
La Lr
ln r a > Lr
d= r
a
L −L
C L
.
(4.16)
84
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
y
y
5
5
5
-5
10
x
x
5
-5
-5
10
-5
(i)
(ii)
Figure 4.7: Contours for the potential sensed by agent in (0,0) in the case the other
agents have all the same value of the repulsive potential scale length Lr (i) and in the
case one of the other agents has a repulsive scale distance Lr′ < Lr
In particular, for C a = C r this is verified as long as La 6= Lr , but as stability
imposes La > Lr , to make the potential function convex about equilibrium, it can
be concluded that this is always verified in this condition and possible to achieve
for other choices of C a and C r parameters.
The other agents in the group considered tend to keep the same relative distance
with respect to agent 1. This produces the new equilibrium configuration that
sees the agent with reduced separation distance finding its equilibrium position
in the centre of the 5-agent group while fulfilling also equilibrium conditions for
the other agents. A contour plot of the potential which agent 1 senses is shown in
Figure 4.7 for both equilibrium and non-equilibrium parameter choices. The same
could be stated for parameter C r as Equation 4.11 is linear in C r . Here parameter Lr ′ is used to impose the central symmetry configuration over the pentagon
one, while parameter C r is used to produce the desired inter-agent distance only.
The cross configuration generated by the asymmetry in the potential repulsive
scale length is sketched in Figure 4.8 where the reference frame is rotated by 45
degrees compared to Figure 4.6.
Considering that interactions amongst agents are only along the edges of the adjacency matrix, a representation of the repulsive and attractive scale parameter,
as well as of the other coefficients influencing Equations 4.1 and 4.2, can be given
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
85
Figure 4.8: Cross pattern emerging by shrinking the repulsive potential scale length
broadcasted for the agent in the centre.
in terms of matrix that has the same structure as the adjacency matrix described
in Section 4.1.2. An extract from repulsive distance matrix is reported in the
Equation 4.17
0 Lr
r′
L
0
r′
L Lr
r′
L Lr
Lr ′ Lr
r′
L2 0
0 Lr
3
0
0
Lr Lr Lr
Lr Lr Lr
Lr2 0
0 Lr3
0
0
0 Lr Lr
Lr 0 Lr
Lr Lr 0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0 Lr Lr
Lr ′ 0 Lr
0
0
..
.
0
Lr ′ Lr
0
...
..
.
,
(4.17)
where, zeros are in the same positions in the adjacency matrix in Figure 4.2 and
4.3, and where the coefficients regulating the interactions among nodes that are
centres of two different 5-agent groups are denoted by Lr2 . Finally Lr3 is used to
indicate the value along the edges connecting peripheral agents across different
5-agent groups. Hence, coefficients Lr , La , C r and C a can be arranged in square
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
86
matrices of dimension N as they are referred to the edge and take a different
value depending on which agents the edge connects.
It is important to notice how the pentagon arrangement is not guaranteed by the
condition Lr ′ = Lr , that is, the cross configuration is an equilibrium configuration
also in the case Lr ′ = Lr . Conversely, having Lr ′ 6= Lr excludes an equilibrium
configuration in the shape of a pentagon. It can be concluded that the pentagon
arrangement, representing a local minimum configuration, is excluded by the
choice of a reduced Lr ′ . As the cross configuration in Figure 4.8 can be obtained
for both the choices of Lr ′ considered, the reduction of the Lr parameter produces
the exclusion of one of the two configurations, hence, it can be seen as a method
for escaping a local minimum configuration. A consequence of this is once the
cross configuration is achieved through the reduced Lr ′ , returning this parameter
to its original value Lr does not produce any change on the formation. This is because, due to the symmetry, the central agent is in equilibrium whatever choice of
Lr parameter is done as long as it guarantees a stable minimum of the potential.
The symmetry of the arrangement translates to two pairs of equal and opposite
terms for the sums in Equations 4.9 and 4.10 making both equations trivially null.
Equilibrium conditions for the surrounding agents according to the scheme of
Figure 4.8 is only determined by Equation 4.9 as the y-component is null by
symmetry. Equilibrium distance d as reported in Figure 4.8 is found by solving
for d the Equation
Cr
Lr
exp
√ !!
√
2d
−2d
+ exp
+ 2 exp
=
r
L
Lr
√
Ca
−d
−2d
2 exp
exp
+
exp
+
La
La
La
−d
Lr
√ !!
− 2d
La
. (4.18)
Equation 4.18 is obtained by expanding Equation 4.9 for the present case; however, this is not analytically solvable. On the other hand, equilibrium can be
found for a given d, by tuning C a and C r parameters. This is better explained
in Section 4.2.2.
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
4.2.2
87
APF Coefficient Definition
The coefficients of the APF acting along the edges of the graph are calculated such
as to set the desired distance amongst the spacecraft. Just the C r coefficient is
calculated as function of the others which are set arbitrarily as long as suitable to
produce an equilibrium configuration according to the criteria discussed in Section
4.2.1. The change of C r parameter only or, more precisely, the change in the ratio
C r /C a is sufficient to modify the position of the minimum of the potential, hence
the design distance, for the APF used. In particular, an interaction between two
spacecraft belonging to two different n-agent groups is considered, with a design
distance dd ; C r coefficient can hence be calculated by manipulating Equation 4.16
as
Cr
Lr
La − Lr
= a exp dd a r
Ca
L
L L
.
(4.19)
Equation 4.19 can be reversed to calculate the equilibrium distance once the
coefficients are set. When more than 2 agents are involved, an analytic expression
for the equilibrium distance cannot be defined, but given a desired distance, it
is always possible to get an expression for the value of the ratio C r /C a that
produces that separation. In particular for a fully connected group of 5 agents
C r /C a ratio can be calculated equating to zero the gradient of the potential for
the formation according to the scheme in Figure 4.8. As the y-component is
trivially null, C r /C a can be calculated considering just the x-component of the
gradient in Equation 4.9. It is then possible to solve for the C r /C a ratio and
obtain
√
√
2dd
dd
2
exp
− L2da d
+
exp
−
+
exp
−
La
La
L
C
√
= a
Ca
L exp − dd + exp − 2dd + √2 exp − 2dd
Lr
Lr
Lr
r
r
.
(4.20)
This tuning method can be extended to the other links of the adjacency matrix;
by defining the coefficients in this way the desired self-similar pattern is produced.
4.2.3
Stability of Control Law
Asymptotic stability can be proved by adopting the procedure introduced in [32].
Consider the time derivative of the energy as sum of artificial potential and real
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
kinetic energy:
dEt
dKt dUt
=
+
dt
dt
dt
where
Ut =
,
88
(4.21)
1 XX
aij Uij
2 i j
(4.22)
is the total potential energy per unit mass with
Uij = Uija + Uijr
and
Kt =
,
(4.23)
1X
1X
Ki =
(vi · vi )
2 i
2 i
(4.24)
is the total kinetic energy per unit mass. Expanding 4.21,
∂Kt
dEt X
=
∇Ut · vi +
· v̇i
dt
∂v
i
i
(4.25)
where, the operator ∇(·) is the one defined in Equation 2.20. Substituting Equation 4.4 and Equation 4.23 into Equation 4.25
dEt X
dKt
=
∇Ut · vi +
· (−∇Ui − σvi )
dt
dv
i
i
∴
dEt X
=
(∇Ut · vi − ∇Ui · vi ) − σ|vi |2
dt
i
(4.26)
.
(4.27)
As the potential depends upon pairwise interactions, xi derivative is not null for
both the Uij and Uji potentials that form the total potential Ut . If the agents
interacted in a symmetric way, this would cancel out with the gradient ∇Ui , but
as the sum of the potential derivatives upon any agent includes asymmetric terms,
this does not occur. Nevertheless the difference between the gradients can always
be damped by the artificial viscous damping. Hence, it can be concluded that
∃σ>0:
X
i
(∇Ut · vi − ∇Ui · vi ) − σ|vi |2 ≤ 0
.
(4.28)
This is enabled by the property that the Morse potential and its gradients are
bounded functions.
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
89
As total energy time derivative can be made a negative semi-defined function, this
can be compared to a Lyapunov-like function whose derivative is always proved
to be negative and zero at equilibrium, corresponding to null speed. Thus, the
system will leak energy and stabilise eventually into a static formation which
corresponds to a local minimum of the total energy.
4.2.4
Fractal Antenna Array Design and Analysis
The fractal geometry of the formation provides the possibility of reducing the
number of elements of a lattice array without significantly diminishing the performance of the array. In a formation flying architecture this is a noteworthy
factor as, especially in close proximity, the challenges in position-keeping increase
with the number of agents, as the presence of each agent in the group introduces
position constraints for the others. Basing antenna array formations on fractal
geometries provides not only the potential to reduce the number of elements but
also offers the possibility to operate across a range of frequencies and the selfreplicating nature of fractal patterns extends to their performance characteristics.
Following the analysis methodology described in Section 4.1.3, a thinned planar
array based on the Vicsek or Purina fractal can be considered; this fractal has
the following simple sub-array at growth scale one (P = 1):
1 0 1
.
S1 = 0 1 0
1 0 1
(4.29)
The array fractal pattern SP at an arbitrary growth scale P ∈ N, P ≥ 2 is given
by
SP = S1 ⊗ SP −1
,
(4.30)
with ⊗ denoting the Kronecker product. Hence, the growth factor δ is 3 and the
array will increase three times its linear size for each stage of growth.
This first stage of growth starts off with a 9-element generating sub-array of
uniformly spaced elements; the center and corner four elements are switched on
(“1”), while the remaining elements are switched off (“0”) or can be considered
completely removed. The Purina fractal array is generated recursively by repeat-
90
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
edly applying this sub-array over P -scales of growth. This is done by replacing
each “1” or switched on element by a generating sub-array and replacing each
“0” by an all zero array the size of the generating sub-array. The result of the
process is illustrated in Figure 4.9.
(a)
(b)
(c)
Figure 4.9: First three stages of growth of the Purina fractal array for (a) P = 1, (b)
P = 2, and (c) P = 3. Credits P. Karagiannakis (from [137]).
The theory developed by Werner [153] about fractal arrays describes in details
the above steps illustrating how the combination of the antenna elements has
the potential to alter the radiation characteristics of an ensemble of antennas and
results in a steerable and highly directive beam. Providing an in-depth analysis of
the radiation pattern and the electrodynamics properties of this fractal antenna
array is beyond the targets of this thesis. However it is important to understand
what the advantages of such a configuration are. This can be done by plotting
Equation 4.6 for the arrangements shown in Figure 4.9. This is done in Figure
4.10 where it is evident how as the stages of growth increase, the beam is more
concentrated, that is, a greater percentage of the radiated energy is concentrated
into the prescribed direction, resulting in a better signal strength for arrays with
the same total area.
4.3
Simulation Results
The control method illustrated in Section 4.2 is used to simulate a possible operative scenario in which a spacecraft formation is used to form a distributed array
in Earth orbit. A geostationary orbit is chosen to simulate the dynamics although
the application is not specifically aimed at terrestrial telecommunications. Deployment of a fractal antenna array is simulated where the system is composed
91
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
5
0
−10
−15
P
|D (θ,φ) | / [dB]
−5
10log
10
−20
−25
−30
P=1
P=2
P=3
−35
−40
−80
−60
−40
−20
0
20
40
60
80
o
θ/[ ]
Figure 4.10: Directivity plots for the first three stages of Purina fractal arrays shown
in Figure 4.9, with an assumed direction of the main beam towards broadside (θ = 0)
of 125 radiating elements.
The requirements of the system drive selection of actuators and dictate to a
certain degree choices regarding agent selection and separation. The method of
control and the possibilities offered by reducing the size of individual radiating
elements, while maintaining an overall large aperture, drive towards the selection
of a satellite in the size range of pico- or nano-satellite suitable for a separation in
the order of 1m. This is the separation chosen as the inter-spacecraft distance is
still small enough to control motion through mutually exchanged electromagnetic
forces and far apart enough to allow for relatively coarse accuracy, in particular
at the release from a carrier spacecraft or launcher.
The 125 unitary mass agents reproduce the shape of a Purina fractal at a growth
stage of P = 3; they are deployed in 25 groups of 5-agent subgroups which
is the elementary unit of the formation (N = 125, n = 5). The dynamics of
the spacecraft formation is based on Clohessy-Wilthshire (CW) [156] linearised
equations in an orbiting reference frame. This is so oriented:
92
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
• the x-axis is tangent to the orbit and parallel to orbital velocity vector;
• the y-axis is parallel to angular momentum vector;
• the z-axis is orthogonal to the first two and pointing towards the Earth
centre.
CW equations in the reference frame chosen are
ẍ = −2ν ż
ÿ = −ν 2 y
z̈ = −2ν ẋ − 3ν 2 z
(4.31)
where, ν is the orbital frequency.
Initial conditions were set such as each spacecraft had an initial position randomly
picked within a sphere centred on its final position and radius equal to 1.5 times
the distance to its nearest neighbour to account for possible initially swapped positions between near agents; initial relative velocities are null. This corresponds
to assuming that a carrier spacecraft or launcher releases the agents with coarse
accuracy i.e. not completely randomly. Attitude for the single spacecraft is not
considered while overall formation attitude control for rotation around x and y
axis is guaranteed by positioning control through a parabolic potential that flatten the formation on the x − y plane. Sensors are idealised, that is, the exact
position of each agent is known without delay by all the agents it is linked with.
In terms of the actuators, although these are not modeled, some characteristics
related to the possible use of electromagnetic forces are considered. In particular,
actuators of the kind proposed in [157] and [158] are considered. In, particular,
in [157] the effectiveness of Coulomb actuators is assessed. The magnitude of
the Coulomb forces attainable by two spacecraft in a geostationary orbit, lies
between 10−3 and 10−5 newtons for separation distances between 10 and 50 meters. By inspection of Equations 4.31, considering that the orbital frequency ν
has an order of magnitude of 10−5 for the geostationary orbit chosen, a unitary
ż (corresponding to the drift of one agent out of the plane of the formation at
the speed of 1 ms−1 , that is extremely high velocity) would fall within the actuation capabilities of the system. The actuators would then be able to correct the
error in this case that represents the most critical scenario to face in a realistic
setting. In fact, the z acceleration is the z direction is of the same magnitude
93
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
of the x component while the y component is 5 orders of magnitude smaller while .
Electromagnetic actuators, in particular the ones based on Coulomb forces, cannot be used all at the same time to avoid interferences, a duty cycle is set up and
the ensemble is split into a number of groups so that any two groups which are
active at the same time are relatively far apart. This allows interferences to be
neglected. Each group is controlled across a time period of the duty cycle. Over
the whole duty cycle each group of agents is controlled for the same amount of
time. As a consequence, agents belonging to more than one group (e.g. linking
agents between groups) are controlled for longer. The frequency of the duty cycle
needs to be high enough to prevent spacecraft drifting away between control periods. This can be bounded from below by considering a linearised version of the
control law and computing the frequency of the associated harmonic oscillator.
Considering the APF only, the x-component of the acceleration provided to the
generic agent i, that is ẍi by the control can be linearised about equilibrium as
mẍ˜i =
X Cija
j
Cijr
−dij
−dij
exp
− r exp
−
Laij
Laij
Lij
Lrij
)
"
#
Cijr
Cija
−dij
−dij
− r 2 exp
(xi − dij )
exp
Laij
Lrij
Laij 2
Lij
, (4.32)
where, it was assumed that the equilibrium position is at a distance d from the
neighbouring agents and that these agents are fixed in their positions. The ycomponent of the acceleration is not considered for symmetry reasons. The sum
is extended to all the neighbouring agents acting along one axis. As an example,
considering the central agent of Figure 4.8, this means that only 2 agents contribute to its oscillatory motion along the orthogonal axes.
Equation 4.32 is in the form of a linearised harmonic oscillator perturbed by a
constant acceleration. The frequency associated with this system is:
v
uX a
Cij
Cijr
u
−dij
−dij
ωi = t
− r 2 exp
exp
Laij
Lrij
Laij 2
Lij
j
.
The frequency chosen for the control to operate shall not be smaller than
(4.33)
94
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
sup ωi
(4.34)
i
obtained by considering all the sets of values defining the control of the groups.
For the case under consideration the whole duty cycle lasts 2 seconds, which corresponds to a control frequency of 0.5 Hertz. The 125 spacecraft are considered
as belonging to 9 groups, these are reported in the following list while Figure 4.11
identifies them by highlighting their position in the formation.
• the 5 5-agent groups at the centre of 25-agent groups (Figure 4.11.a);
• the 5 5-agent groups at the top of 25-agent groups (Figure 4.11.b);
• the 5 5-agent groups at the bottom of 25-agent groups (Figure 4.11.c);
• the 5 5-agent groups at the left of 25-agent groups (Figure 4.11.d);
• the 5 5-agent groups at the right of 25-agent groups (Figure 4.11.e);
• the agents linking the centres of the 5-agent groups in 25 agent groups
(Figure 4.11.f);
• the agents bonding the 5-agent side by side in the 25-agent groups (Figure
4.11.g);
• the agents bonding the centres of the 25-agent groups (Figure 4.11.h);
• the agents bonding the sides of the 25-agent groups (Figure 4.11.i).
The connections between each group (consisting of 25 agents) are ensured by
pairs of agents instead of groups of agents. This allows a reduction of the computational efforts for each agent and a reduction of the computational resources
needed for the simulation. On the other hand, this reduces the control power and
slows down the deployment of the formation. Table 4.1 reports the values of the
coefficients used.
The agent at the centre of the formation (say agent 1) is the only one linked to
the centre of reference frame by a quadratic potential in the form Uc = ζ|x1 |2
with ζ as a weighting parameter set to 0.1. The control force is provided through
the negative gradient of Uc as before plus a dissipative term in the form −σ ẋi
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
(i)
Figure 4.11: Groups of agents controlled at each step throughout the duty cycle.
95
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
96
Table 4.1: Numerical values of coefficients used in numerical simulations.
Fully connected groups (f.c.g.)
Centres of f.c.g.
Peripheral between adjacent f.c.g.
Centres of 25-agent groups
Peripheral of 25-agent groups
σ = 0.1 for all the agents
Ca
4
Cr
La
3.94722 2
Lr
1
Lr’=0.5
1
0.99596 4.5 4
Lr’=2
0.8925 1
2
0.5
2500
2505.3 10 9.9
Lr’=4.5
69.96 70.0
3
2.9
with σ also set to 0.1. This is to provide a kind of orbit tracking capability
or, in practical terms, the possibility to stay anchored to centre of the reference
frame. The value of 0.1 is chosen in order to make the orbit tracking effective
without superimposing disturbing dynamics over the inter-agent APFs. Also,
this suggests that the task of tracking the orbit can potentially be carried out
by only a single agent, while the others just track their relative position with
respect to the central agent. This is not necessarily the one in the centre as it is
done here for simplicity. The control law is applied for just x and y axis of the
orbital reference frame with control on z-axis performed through the gradient of
a simple parabolic potential plus a dissipative term in the form z̈i = −∇Uzi −σ żi ,
for i = 1...N, with Uzi = ζ|zi|2 where, ζ and σ have the same roles and values as
used previously, that flatten the formation on the plane z = 0. Four snapshots
from the deployment are reported in Figure 4.12. It can be noticed that after
one day the deployment presents some distortions in particular from peripheral
groups.
Finally in Figure 4.13 errors on the designed relative positions after one day
are plotted. More precisely the error is considered as the difference between
the actual distance of each spacecraft from the centre of the formation and the
ideal design distance; this is then plotted as a percentage of the desired spacing.
It can be seen that the greatest error is lower than 5%. Both snapshots and
error evaluations are considered after a maximum of 24 hours; this is sufficient to
show the self arranging capabilities of the control technique. After a further 24
hours the magnitude of the maximum error is halved. Theoretically a complete
97
10
10
5
5
Y [m]
Y [m]
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
0
0
−5
−5
−10
−10
−15
−10
−5
0
5
10
15
−15
−10
−5
X [m]
10
10
5
5
0
10
15
5
10
15
0
−5
−5
−10
−10
−10
5
(ii)
Y [m]
Y [m]
(i)
−15
0
X [m]
−5
0
5
10
15
−15
−10
X [m]
(iii)
−5
0
X [m]
(iv)
Figure 4.12: Formation deployment in GEO. Snapshots taken at (i) t=0 s, (ii) t=60 s,
(iii) t=1 hr, (iv) t=24 hr
relaxation with no positioning errors is possible but only after an infinite period
of time due to the viscous-like damping. On the other hand, the error after finite
time can be improved by either optimising the damping constant or replacing
the viscous damping with more sophisticated means of virtual energy dissipation.
However, both these improvements are beyond the scope of the study presented
here althugh represent possible development directions.
4.4
Final Remarks
The idea of meeting needs for highly directional wideband antenna arrays through
a space based fractionated architecture is constructed around the possibility of
locating a number of spacecraft, each carrying an antenna element, according to
98
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
0.6
1
0.5
0.8
∆ Y [m]
∆ X [m]
0.4
0.6
0.3
0.4
0.2
0.2
0
0
0.1
20
40
60
AGENT
80
100
120
0
0
20
40
(a)
60
AGENT
80
100
120
(b)
5
DISTANCE IN %
4
3
2
1
0
0
20
40
60
AGENT
80
100
120
(c)
Figure 4.13: Errors in design positioning after 1 day from release of the formation.
Distances are computed with respect to the agent at the centre of the formation (a)
along the x axis, (b) along the y axis and (c) in percentage of the design distance from
it.
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
99
a precise fractal scheme. This improves overall antenna performance and capabilities while using a reduced number of elements. In turn the possibility of using
small spacecraft enables the formation of a wideband fractionated antenna, but
requires accurate spacing between the elements. Orientation is not considered
here for single agents as they are assumed to be isotropic sources. Thus, in the
case of an antenna array as described above, the relative agent positions within
the whole array is the key requirement as this influences the performance of the
array. Hence considering just coarse attitude control for single agents, a description of the system characteristics in a global sense is possible as long as relative
positions are precisely known. Utilising this knowledge, directivity through array
phasing is achievable: at group-level for compensation of global attitude errors
and at agent-level to accommodate misalignment of the single elements.
From a control point of view, the need for precise close formation flying can be
tackled through using reliable techniques and implementing these on relatively
small agents. In this respect APFs are particularly suited to the task as their stability characteristics are analytically provable, hence they do not need extensive
montecarlo test campaigns to validate their behaviour. Moreover, APFs allow for
highly nonlinear control through quite straightforward computation due to their
smoothness. As the amount of information needed is just the relative position
of a number of neighbours, the connection network presented here has the double advantage of shaping the formation on one side and reducing the number of
connections on the other. These combined characteristics make small spacecraft,
even with reduced computation capabilities, able to carry out the task of arranging into a formation through exclusively inter-agent interaction in a decentralised
way.
APFs provide collision avoidance capabilities to the agents directly connected into
the network. However, some agents might be found to fly, at least in some phases,
close to others which they are not connected with. This means that these agents
would not sense each other as they are unaware of their relative positions. The
problem can be tackled considering avoidance manoeuvres to operate just in case
of close proximity, revealed through any sensor scanning of the local neighbours.
Any avoidance action should then be properly designed so as not to introduce
persistent instability in the control of the agents already linked through the net-
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
100
work. Moreover, as the communication architecture presented requires each agent
to keep a relatively small number of communication channels open, temporary
links can be established between spacecraft in close proximity to negotiate on a
case-by-case basis a suitable avoidance manoeuvre.
The concept of self-similar arrangement is quite general and can be exported from
formation flying for the exploitation of electromagnetic phenomena, to other applications. However, in this chapter the space flight case was illustrated, hence
the dynamics of the formation in the orbit environment imposes to consider specific orbit parameters and suitable actuators, at least in the definition of the
simulation scenario.
A geostationary orbit was considered although agents are not specifically targeted
at telecommunication purposes. The choice not to detail the actuator modelling
in depth was made to mainly concentrate on the control aspects and the benefits the distributed architecture can get out of the hierarchical communication
scheme. Actuator characteristics were considered only in part. Although the response of the actuators was not included, their choice took into account the close
proximity scenario. The use of inter-agent electromagnetic forces was proposed
rather than thrusters as it avoids possible plume impingement problems considered the close proximity at which the agents operate. Moreover the APF methods
drive the system through an oscillatory stage before the achievement of the equilibrium configuration during which residual energy (both virtual potential and
real kinetic) is dissipated. This translates into fuel wasting when considering the
use of thrusters. The introduction of a duty cycle in the control operation is a
consequence of the choice of actuators. Also the advantage of having actuators
that mimic the virtual inter-agent action of the artificial potential makes the performances of control techniques more evident as less dependent on the actuators.
The duty cycle just applies to inter-agent forces, which are supposed of Coulomb
type [157], while no hypothesis is done on forces that act along z-axis or that bond
the centre of the formation to the origin of the reference frame. For what concerns
z-axis, the use of Lorentz forces as in [158] might be considered, although their
effectiveness is to be investigated further in relation to the magnetic environment.
CHAPTER 4. FRACTAL FRACTIONATED ARCHITECTURES
101
The communication network was intended in the first place for control purposes
only, but the need for task assignment in the fractionated architecture as well as
array phasing can be carried out through the same architecture. In particular,
the system inherits a structured hierarchical network, where the ranking of the
agents depends on the number of links they are connected to. This does not imply that the resulting architecture is centralised, but allows the task assignment
to be carried out on the basis of the hierarchy of the agents. For instance the
guidance for the whole formation can be carried out by a number of spacecraft
within the group which communicate in an all-to-all scheme in order to share the
computational efforts (e.g. the centres of the 25 agent groups shown in Figure
4.11.h), and then passed to another module able to compare this to the navigation
to eventually generate a control input for the whole formation. This is different
from the guidance, navigation and control (GNC) functions that each spacecraft
carries out: while each spacecraft should find its position in a distributed architecture, the whole system follows a guidance law that enables the mission task
achievement. It is worthwhile stressing how the position of each agent is not
pre-determined in a strict sense. The links of each agent are pre-assigned, but
this does not prevent agents, or groups of agents belonging to the same level, to
swap their positions.
Finally, despite the planarity of the formation, nothing prevents to produce the
emergence of a central symmetry and the building up of several hierarchical levels
in a self-similar fashion also in 3D formations. This can be done considering an
initiator composed of a different number of agents as well. An analysis on the
ground of fractal electrodynamics can be carried out in three dimensions using
the same methodology described in this chapter.
Chapter 5
Pattern Creation - Experimental
Results
In Chapter 3 and Chapter 4 the formation shaping and control for swarms of
agents was analysed from a theoretical point of view, demonstrating the soundness
and the benefits of techniques based on the APF methods, with advantages from
modification and design of the communication network. Practical aspects such as
latencies in the process, accuracy of the navigation sensors, actuator saturation
and dynamics, or possible constraints due to their nature, were not taken into
account so far. In Section 4.4 it was pointed out how the choice of actuators may
heavily influence the dynamics of the system, but for the purpose of the control
technique analysis carried out in Chapter 4, a deeper analysis of the “real world ”
issues was avoided. In this chapter, the implementation on real hardware of the
control technique proposed in Chapter 4 provides insight into some real world
issues that are either difficult to model through simulations, or requiring a number
of assumptions that diminish the validity of the modelling. The implementation
of innovative technologies on hardware is a fundamental step when climbing the
Technology Readiness Level (TRL) ladder, from mathematical modelling up to
the point the technology is made available for the final users. The main results
of this chapter have been reported in [159, 160].
102
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
5.1
103
Real World Issues
When departing from theoretical models and looking into specific applications,
APF methods are widely used to define the guidance law, that is the velocity
field, rather than the forces to produce through the actuators to drive the system
to its desired state; this is done, for example, in [161]. APFs are indeed suitable
for defining the desired, time varying state of the system, and less appropriate for
driving it there. The oscillatory motion produced about the equilibrium is characteristic of this control method and, although the virtual damping presented
in Chapter 4 can be improved by a certain amount, using the APFs to define
the accelerations of the agents returns improvable relaxation times and energy
exploitation. Another fundamental reason that prevents the application of the
APF to the definition of the required action from the actuators is the lack of full
actuation in a real system. The APF method was applied in the previous chapter
to produce a force dependent on the positions of a number of agents. This cannot
be generally done in real systems: primarily because actuators cannot produce
an action in any direction at any time. Constraints apply to actuator operations,
which are due to their nature and to the physical characteristics of the system
and the environment. Consider specifically the ground mobility case: wheels are
the most widely used actuators and they are extremely limiting if compared to
the assumption of free translation along any direction which was made in Chapter 4. Excluding particular and complex designs, wheeled agents are prevented
from producing pure lateral shifting. Curvature radius also has to be taken into
account for any manoeuvre. As different from the idealised free space case, APF
cannot be simply used to produce the motion of the vehicles as a direct response
to the potential gradient, which may be in a direction perpendicular to the heading of the vehicle, precluded in the motion of the wheeled system.
APFs are very effective in designing the guidance law, that is where the vehicle
should move and with which velocity. However, a controller different from the
APF that produced the guidance law has to be implemented. This maps the error
generated by the difference between the actual state and the desired one produced
by the APFs into a command for the actuators at the agent-level. Although this
second controller could exploit APFs as well, the advantages of the APF method
would be greatly diminished when used to control a single agent rather than a
swarm in distributed fashion. For this reason, in the experiments presented in
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
104
Section 5.2, the low level control, that is the control of the actuators, is performed
through a quasi-linear function limiting the use of the artificial potential to the
definition of the target position to achieve.
5.1.1
Navigation
Navigation is the determination of position and direction of motion of an object.
In a swarm, navigation must also provide information about the relative position
of an agent with respect to the others so to provide a meaningful input to the
guidance and control systems which drive the agent motion relying on relative
position and velocity.
In numerical simulations of swarms, agents are virtual objects for whose motion
all information can be known. In the real world swarming agents must be able to
estimate their own position and velocity and the ones of their neighbours. This
is a demanding task for the agents, and, as it happens also for the control, is the
object of specific studies oriented either to the definition of the state of a single
agent or to its state relatively to the other agents of the swarm.
A detailed analysis of the navigation problems is beyond the scope of this dissertation, hence, the problem of navigation is not faced directly. The results
reported in this chapter are obtained using wheeled robots and the navigation
problems, which cannot be avoided as in numerical simulations, are overcome
by tracking the agents through external sensors and then controlling them in
distributed fashion by exploiting this information.
5.1.2
Actuators
The problems connected with actuators were partially anticipated at the beginning of this section. They can be summarised into 3 main concepts:
- Actuators have their own dynamics that reflects onto the swarm behaviour
mainly through delaying and limiting the motion within the swarm;
- Actuators are subject to problems such as saturation, deadbands and locking: control system should be robust enough to compensate for these;
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
105
- Actuators interface with the external environment and often exploit its
feature to produce forces and torques on the system.
Actuators dynamics can be seen in terms of rising time from the moment the
input is fed to the achievement of the level of actuation requested, provided that
this is feasible. When the input is decided on the base of a time varying guidance
law, reaction time of the actuators should be small enough to keep up with it.
This is indeed the case of swarming systems where at each instant, desired states
for each agent depend upon the continuously changing states of the others. Failing to keep up may result in the loss of convergence to consensus, that is agents
do not manage to coordinate their action and produce a coherent behaviour.
Saturation, deadband and locking are undesirable states which are present in all
actuators. Avoiding operative conditions close to, or with the possibility of evolving into, those states relies on the controller characteristics. Trivially, the control system should not command the actuators to output within their deadband
(which would result in null output) or close to the maximum output achievable
(which would leave the system with no control margin for contingency manoeuvres, beside stressing consistently the actuators).
Finally, most actuators use the external environment to input forces and torques
onto the system. For example, wheels use a solid surface and the grip with it
to roll while propellers use a fluid flow to accelerate vehicles. The interaction
with the external environment is difficult to model as very little is usually known
about the environment where the actuators are operating. The design of the control relies on a partial knowledge of its characteristics: the torque to the wheels in
a road vehicle should be, for instance, commanded considering the static friction
provided by the ground and the characteristics of the tire.
There are then characteristics that are peculiar of particular actuators. For a
wheeled vehicle, control system shall not command, for instance, out-of-plane
motions or lateral shifting. By the same argument, the autonomous controller of
a fixed wing aircraft should not command a climbing angle beyond the stalling
point of the vehicle. It has instead to work out the best way to minimize the
error between the actual state and the desired one by relying on the physical
characteristics of the system and its actuators. All these aspects need to be con-
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
106
sidered carefully when passing from a numerical simulation to the implementation
of swarm control onto real robotic agents.
5.2
Swarm Robotic Experiments
In this section tests performed on a number of robots are described. First the
adaptation of the APF control method to the task is presented with a description
of the experimental apparatus. Tests of formation achievement, the switching of
the formation from a pentagon to a cross shape, and the case of a cross formation
rotating about its centre are then described.
5.2.1
Testbed and Methodology
Hardware setup
The robots used were designed at the Centre for Ultrasonic Engineering and
manufactured within the facilities of the University of Strathclyde. These are
shown in Figure 5.1. They are composed of two differentially driven motors
mounted on an aluminium chassis, a ball bearing is used as third support. They
are approximately 175 × 124 × 80 mm with a wheel diameter of 54 mm. The
robots are controlled by an embedded Linux computer (720MHz ARM CortexA8 with 512MB of RAM) and powered by a Lithium Polymer battery (11.1V,
2Ahr, providing 4 hour run time).
Guidance and control functions were performed centrally on a host computer
using a C# Graphical User Interface (GUI) that was interfaced with the robot’s
Software development kit (SDK). The SDK interfaces with the robots using WiFi.
The robots were tracked by a 6 camera VICON T160 [162] positioning system,
which provides 1 millimetre, 6 degree of freedom accuracy at 100 Hz over a 3.8×4
m2 area. Individual robots are identified through a unique pattern of reflective
targets (14 mm spheres) affixed to the chassis, these targets are visible in Figure
5.2. In these experiments the robots were driven on the floor of the cell, so
only X,Y positions and yaw angle were required to fully characterise the robot
position. A more detailed description of the robotic agents used can be found in
[163, 164].
Control actions were defined on the basis of the VICON positional estimate,
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
Figure 5.1: Wheeled robots used for swarming test. Not to scale
107
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
108
Figure 5.2: One of the robots used in the tests.
which was passed directly to the robots SDK. The computer emulated distributed
control: inputs were produced for each agent in turn on the basis of its position
relative to the others. This ensured that each robot was controlled individually
hence any group behaviours emerge out of singles’ actions. Virtual robots can
be added to the test. These are simulated agents that, through the GUI, can be
made part of the experiment increasing the number of agents considered. This
feature was not used for the tests presented in the following. A scheme of how
the testbed works is presented in Figure 5.3 while the VICON [162] positioning
cell is sketched in Figure 5.4.
Software
The guidance and control laws were coded using C# and, as previously explained,
run on a dual core 2.5 GHz, 2 GB RAM, Windows XP desktop computer rather
than on the single robots. The results are fed to the guidance law determined
by the Artificial Potential Functions that in turn was fed to the control of the
wheel speed as it will be explained in the next section. This is done in turn for
each object robot considered by the GUI, that automatically labels the robots
consecutively from 1 to 5. Robot state, as measured by VICON, was exported to
file for post-processing and visualisation in MATLAB R .
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
109
Figure 5.3: Testbed architecture. The Vicon cameras track the robots in the test area
and send their positions and attitude to the centralised computer that can also include
virtual agents. The control is computed and input to the robots’ wheels are passed
through the wireless network.
5.2.2
Guidance and Control
The argumentations illustrated in Section 5.1 drive to the conclusion that the
guidance must be separated from the control when including experiments with
real hardware rather than just simulations.
The Guidance Functions
Morse artificial potential functions, already used in this thesis, produce the desired velocity for the system ẋdi = (ẋdi , ẏid ). As before, these are composed of
exponential functions that provide an attractive and repulsive component defined
as,
|xij |
=
exp − a
Lij
|xij |
r
r
Uij = Cij exp − r
Lij
Uija
−Cija
(4.1)
(4.2)
with constants Cija , Cijr , Laij and Lrij shaping the formation and its size. Differently
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
110
Figure 5.4: The testbed arena where tests are performed, with Vicon T160 cameras in
red.
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
111
from what done in Chapters 3 and 4, this time APF gradients are used to obtain
the desired velocity for any robot i as,
ẋdi = −∇Uia − ∇Uir
(5.1)
where,
∇(·) =
d(·)
dxi
Uia =
X
j
(Uija ) Uir =
X
(Uijr )
.
(5.2)
j
The order of the dynamic equation drops from second to first and, compared
to Equation 4.4. In Equation 5.1 the virtual viscous damping is missing as the
first order dynamics does not produce overshoots to damp out. Up to this point,
relative position was used to produce the guidance law. The velocity field can
be modified considering global translational and rotational terms. Here just a
rotational one, Hi , is considered. It is defined as
Hi = H(xi ) = kr K(xi − X)
(5.3)
where, X is the position of the centre of mass of the formation; K is a square
matrix of dimension 2 that produces a rotation θ of the velocity vector. The
value θ = π2 K is considered here to produce a velocity component in a direction
orthogonal to the position vector of each robot with respect to the centre of mass
of the formation, that is producing a pure rotation of the swarm. The matrix K
is hence defined as
"
# "
#
cos(θ) − sin(θ)
0 −1
K=
=
.
(5.4)
sin(θ) cos(θ)
1 0
Matrix K produces a component of the desired velocity orthogonal to the position
vector xi − X of each agent with respect to the centre of mass of the formation.
A general form can include a parameter kr that tunes the magnitude of the
tangential velocity. For the experiments presented in the proceeding, this was
taken as unitary. Additionally the rotational term returns a general form of the
guidance law as,
ẋdi = −∇Uia − ∇Uir + Hi
.
(5.5)
The process that leads a group of 5 agents to achieve a pentagon rather than a
cross shape has already been described in Section 4.2.1. The group of 5 robots,
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
(a)
112
(b)
Figure 5.5: A formation of 5 agents arranged in (a) pentagon shape due to all-to-all
potential with the same coefficients for all the robots and (b) in a cross shape due to
one agent being less “repulsive” with respect to the others. Attitude is not considered.
Formation is defined just by the middle point of the wheel axle
controlled through pairwise artificial potential, can arrange in two different cluster configurations that correspond to stable equilibria. Also here, by tuning one
single parameter Lrij along the directed edges connecting one single agent to the
other 4, the pentagon formation is no longer an equilibrium one. The formation
is pictured in Figure 5.5 and it can be analysed considering the gradient of the
artificial potential sensed by any of the robots as already done in Equations 4.9
and 4.10, for the initiator of the fractal formation presented in Chapter 4.
For what concerns the stability of the control law, satisfying the condition Lrij <
Laij guarantees the presence of a minimum (i.e. convexity property) at the equilibrium [32]. Null derivative can be obtained as a function of Cijr /Cija ratio. For
agent 1, as done already in Chapter 4, exploiting the convenient reference frame
of Figure 5.5(a), dropping the indices, the x-derivative becomes
∂U1
=
∂x1 pentagon
Ca
d
d2
= 2 a exp − a (cos α) + exp − a (cos β)
L
L
L
r
d
d2
C
−2 r exp − r (cos α) + exp − r (cos β)
L
L
L
where,
(5.6)
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
d
d2 =
2
s
1
tan α +
cos α
2
+ 1 = kd
.
113
(5.7)
If the repulsive scale distance considered just for agent 1 shrinks to a value Lr ′
such that Lr ′ < Lr the pentagon configuration is not an equilibrium one anymore.
Indeed, the derivative of Equation 5.6 with respect to Lr ′ can be proved to be
always negative for d > Lr ′ , as per Equations 4.13-4.15, or otherwise can be made
always negative by setting correctly the other parameters of the equation. Again
a reduction of the repulsive scale length produces an increase in velocity along
the positive x − axis. When this is done for just one robot, it will collapse in
the centre of the formation, which will rearrange in the cross shape. Moreover,
as Equation 5.6 is linear with respect to C r , as pointed out previously, the ratio
C r /C a can be used to scale the physical size of the formation.
The Control Functions
The low level controller uses the desired velocity output of APF to provide the
rotational speed values for the wheels. This was developed considering the design
of the wheeled robots used.
The desired velocity vector components are mapped to the wheel mean speed
(WMS) and the wheel speed difference (WSD). These are calculated, for a generic
robot i, as function of the magnitude of the artificial potential derivative and of
the error in heading,
W MSi =
|ẋdi |
Sr Υ i
|∇Ui∗ |
W SDi =
∆θi
Sr
π
(5.8)
where, Sr is the maximum speed value allowed by the motor in the non-dimensional
range [-100; 100], ∆θ is the error in heading measured as difference between the
desired heading and the actual one,
∆θi = θid − θi
θid = tan−1
ẏid
ẋdi
.
(5.9)
while θi is provided by the tracking system. U ∗ is the value of the artificial
potential measured at a desired stand-off distance. The function Υi damps the
magnitude of the speed commanded to the wheel in case of large heading error.
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
It is defined as
Υi = Υ(∆θi ) = cos
∆θi
2
.
114
(5.10)
This function, positive semi-definite in the range [−π/2, π/2], weights the error
in heading more than the ones in position. It ensures that the correct heading is
acquired before producing a high WMS which might be demanded by the APF in
case for large errors in position that correspond to large absolute value of artificial
potential.
The actual command to the speed is 2 element vector where the first element is
the left wheel speed and the second one is the right wheel speed. It is defined as
wi =
"
W MSi − W SDi /2
W MSi + W SDi /2
#
.
As the motors used do not output any torque in the interval ]-13;13[ a deadband
was set imposing a rotational speed corresponding to level 13 for any input in the
interval with exclusion of zero. Saturation of the motors can only be partially
tackled by the scaling parameter U ∗ in Equation 5.2.2 that can be considered
valid just for the case of 2 robots. When more robots are considered, it potential
|∇U ∗ | is heuristically defined as
|∇Ui∗ |
= (N − 1)
Ca
exp
La
−d∗
La
Cr
− r exp
Li
−d∗
Lri
(5.11)
where, just the index of Lr parameter is considered being the other parameters
the same for all the agents. The constant d∗ = 500mm is a separation distance
between two generic robots, considered isolated, and N is the number of robots.
d∗ is chosen considering twice the maximum distance between axle centre (around
which the robot rotates) and the farthest point of the robot chassis from this
one. Finally the sensitivity to positioning errors is defined as the error on the
magnitude of the artificial potential gradient for each agent. Once again in the
case of 2 or 3 robots it is relatively easy to scale up the threshold on the potential
gradient below which the group of robots is considered in a formation, that is,
when the potential gradient magnitude is directly related to the error in relative
positioning. When there are more than 3 robots, because any possible equilibrium
does not produce the same separation between any two robots the computation of
the threshold becomes more difficult. For these reasons the tolerance to relative
115
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
positioning errors, that maps into tolerance on the APF gradient magnitude was
heuristically scaled to account for a variable number of robots. The value of the
threshold was taken in terms of potential gradient as ∇U(dd + dtoll ) where, dd
is the design distance and δd is the required precision, or the tolerance. The
associated potentia is
∆Uitoll
N −1
=
N +1
Ca
exp
La
−(dd + dtoll )
La
Cr
− r exp
Li
−(dd + dtoll )
Lri
.
(5.12)
The achievement of the desired spacing is obtained by solving Equation 4.18 for
Cijr /Cija as a function of the other parameters, which are assigned, with the distance amongst the robots considered as a design parameter. The same values
were used for parameter Laij , for all (i, j) to calculate Cijr /Cija . Lrij parameter
was instead assumed to be the same for all the interactions amongst the robots,
except for the interactions sensed by the robot taking the central position.
As the testbed labels the robots consecutively (i.e. from 1 to 5), agent number
1 was given a reduced value for Lrij , that is Lr ′ . Values used, calculated using
Equation 4.20 for a cross arm design size of 700 mm, are reported in Table 5.1.
Table 5.1: Numerical values of coefficients used in numerical simulations referred to a
cross formation with a arm of 700 mm.
Ca
99.9996
Cr
100
La
700
Lr
698
Lr ′ =69.8
A series of tests were performed to validate the switching and the rotational
behaviours that the system exhibits when the value of Lr parameter sensed by one
agent drops to Lr ′ and when the coupling to the centre of mass of the formation
is enabled through matrix H respectively. Here observations made are reported
with two tests described in detail as examples.
5.2.3
Experimental Results
In this section tests performed are described. The arrangement in a cross formation is tested and outputs for 3 requested positioning distances are illustrated,
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
116
beside the capability of rotating the formation or switching from a pentagon to
a cross formation.
Formation Acquisition and Static Positioning Accuracy
The success in the achievement of a static formation is strongly dependent upon
the level of accuracy in positioning that is requested. For the 700 mm cross case
an accuracy of 20 mm was difficult to obtain. Relaxing the requirement to 30 mm
acquiring a static formations took on average 324 seconds but this figure shrinks
dramatically as soon as the precision requirement is relaxed further.
To test the ability of achieving the cross formation within a prescribed tolerance,
30 tests were performed, 10 tests for each required tolerance on the inter-agent
distance (30, 50 and 100 mm). The robots were given random initial positions
within the testing area and zero initial velocities. As the final configuration is
achieved by a minimization of the inter-agent potential that just depends on relative positions and the final positions with respect to the external reference frame
are not assigned, the accuracy in getting a precise positioning must be mapped
into a threshold value of the potential gradient sensed by each rover. Results
obtained are summarised in Table 5.2.
Table 5.2: Results of the static relative positioning tests for a cross formation with a
700 mm arm. Tolerance and error are referred to the design inter-agent distance
Tolerance
Average Error
Error Variance
Average time taken
30mm
18.3mm
364mm2
324s
50mm
32.7mm
1238mm2
107s
100mm
61.9mm
3965mm2
94s
A critical issue was spotted in the frequency at which the system is updated,
which for 5 agents and the testbed architecture described previously, was 10 Hz.
This is ten times lower than the frequency at which position data are streamed
by the Vicon; hence just one every ten measurements from Vicon was used to
generate the commands to the actuators. As each command keeps on executing
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
117
till the new command is fed in, robots may end up passing from an error in positioning on one side to an error on the other side while trying to correct the first
one. Moreover, the control frequency scales down linearly with the increase in the
number of robots, while the number of local maxima and minima in the artificial
potential field increases. Difficulties in the achievement of very precise positioning are also partially due to the noise in measurements and the misalignment
between the centre of the robots perceived by the tracking system and the actual
centre about which the robot rotates. This range of problems is closely linked to
the testbed architecture which includes external tracking and centralised computation to mimic on-board intelligence. As such, these limitations can be overcome
through improving the testing technology available but do not diminish the validity of the results obtained in terms of pattern formation using the APF method.
Some issues can be spotted in the design of the robots and their suitability for
the use in conjunction with the APF method. The limited mobility due to the
differential drive is a key issue to account for when considering the capabilities of
the system to quickly achieve a precise configuration. It was noted how, despite
the cosine term in the low level control function, the final equilibrium position
was missed several times when the rovers were committed to sharp bends in the
final approach to the equilibrium positions. Actuators other than differentially
driven wheels can be used to improve the design of the robots and make them
more suitable for the APF control method.
The deadband of the actuators, that is, the lower bound on the available spinning
rate of the wheels, was found to contribute significantly, in negative sense, to the
coarse precision in the final approach. The performance of the actuators should
be compliant with the action requested by the controller, which, in turn, may
be improved and optimised to bridge the APF velocity field with the achievable
performance of the actuators.
Formation Switching
In the formation switching tests a group of 5 robots randomly deployed in the
test area at the beginning of the test acquires the pentagon formation described,
then, the Lr parameter is manually switched to Lr ′ through the GUI for the agent
that the system labelled as nr 1 according to the automatic labelling introduced
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
118
400
200
Y [mm]
0
−200
−400
−600
−800
−2800 −2600 −2400 −2200 −2000 −1800 −1600 −1400 −1200 −1000
X [mm]
Figure 5.6: Trajectories followed by the 5 agents while arranging in a pentagon formation first (stars) and then in a cross formation (circles).
in Section 5.2.1. This collapses into the centre forcing the others to adjust their
distance accordingly. The two formations are pictured in Figures 5.6 and 5.7.
Rotating Formation and Failure Tolerance
The rotating formation was tested starting from a random arrangement and both
including the rotation since the beginning of the test, and switching the rotational
term on when the formation was already stabilized in a cross shape. Just the cross
shape was tested although, in theory, nothing would prevent the pentagon formation to undergo rotation by the same means. Results obtained are illustrated in
Figures 5.8 and 5.9. Although the tracks of the robots in the rotating formation
do not overlap exactly, the maximum distance between 2 trajectories is always
below 40 mm.
The loss of one agent was simulated as well in this context. When this happens,
agents are sorted again and the one with Lrij = Lr ′ is identified always as the
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
(a)
119
(b)
Figure 5.7: A formation of 5 agents arranged in (a) pentagon formation and then (b)
in cross formation. The pictures were extracted from the video taken by one of the
Vicon cameras.
corresponding to the first sorted agent. The formation will then rotate about
the agent newly labelled 1 that now presents the reduced value of repulsive scale
distance. The rotation about the central agent is a consequence of the rotation
about the centre of mass of the formation, which coincides with the position of
the agent number 1.
The pattern reconfiguration confirms that the emergence of the central symmetry
is not dependent upon the number of agents involved. It is the equilibrium
configuration that robotic agents achieve to compensate for the asymmetry of
the potential with respect to one agent.
5.3
Real and Simulated Real World
The effective implementation of APF techniques towards autonomous robotics is
to be regarded as a step in the direction of increasing the technology readiness
level (TRL) of this control method. This has been done in this chapter in a all-toall connection scheme, that is, each robot has access to the information about the
state of every other robot and uses this information to compute the desired state
and evolve towards it. However, this is not always possible, in particular when
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
120
200
0
Y [mm]
−200
−400
−600
−800
−1000
−1200
−2800 −2600 −2400 −2200 −2000 −1800 −1600 −1400 −1200 −1000
X [mm]
Figure 5.8: Trajectories followed by the 5 agents while arranging in a cross formation
and rotating about its centre. Each robot is tracked with a different colour. The initial
positions are marked by the diamonds while the final positions by circles. The picture
shows the robustness to the loss of an agent. The robot tracked with the red line stops
operating while the formation is rotating about the robot tracked in blue. The system
then sorts again the agents forming a 3 point star rotating about its centre where the
agent in black finds its place.
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
(a)
121
(b)
Figure 5.9: A formation of 5 agents arranged in (a) four point star and (b) in a three
point star that excludes an agent with a failure. The pictures were extracted from the
video taken by one of the Vicon cameras.
the number of robots increases significantly (see also [105, 165]). Findings related
to incomplete or switching communication graphs should also be verified in an
experimental environment. The main contribution given by the tests presented
is the experimental validation of one particular emergent behaviour, that is the
collapse of a given formation into one of central symmetry by non symmetric
changes in the guidance potential. Testing similar behaviour in a not fully connected system is the next step for this research. This kind of multi-agent system
can find application in structural inspection because of the advantages of having
devices able to perform the task in an autonomous and flexible fashion. These
may include hard to access areas, or areas where access may have consequences
for the plant.
Concerns may arise because of the complex geometries and tough conditions that
are usually found in real world industrial environments, while in the experiments
robots were moving on a plane, hazard-free surface. This is the main difference,
and the step that separates the real world from a simulated real world. As such, to
achieve a further improvement on the TRL ladder, this issue has to be considered.
However, the tests aimed to reproduce emergent behaviours experimentally that
had previously been seen in simulations, or in mathematical models. Thus, they
CHAPTER 5. PATTERN CREATION - EXPERIMENTAL RESULTS
122
consolidate the control techniques developed while keeping them unattached from
any particular operational theatre. Beside this, the test campaign included the
evaluation of the effect that the failure of one robot can have on the whole system,
which is often the case in real operational scenarios.
Chapter 6
Fast Consensus and Manoeuvring
In the two previous chapters the focus was on formations, here the point is remaining cohesive and exploiting the network of connections to boost manoeuvring
performance. The content of this chapter, even more than the others, is inspired
by the biological world. The starting point is represented by concepts well known
amongst naturalists such as the “Chorus line hypothesis” and the “Trafalgar Effect”, previously mentioned in Section 2.7. This chapter turns the problem of
consensus and fast manoeuvring in avian flocks into one of designing an efficient
distributed control scheme for swarms. This, taking advantage of the swarm characteristics, drives the agents to consensus and to manoeuvre quickly in response
to external stimuli.
6.1
The Need for Fast Manoeuvring
The break through of autonomous systems, especially in defence and intelligence
applications, imposes that complex manoeuvres, up to now left to the human
ability, be performed autonomously. This applies to swarms as well and, even
more, to swarms characterised by quick dynamics. As an example the DARPA
F6 project aimed to produce a highly responsive space fractionated architecture,
with the capability of quickly dispersing and regathering [15]. Works confirming
the growing interest in the direction of the F6 project can also be found in the
literature, see for example [166].
In the design of a control architecture able to produce fast responses, nature
can be extremely inspirational. In avian flocks, birds manage to share informa123
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
124
tion about migratory routes by influencing each other direction of flight and at
the same time they defend themselves efficiently from predators. In a flock the
chances of spotting predators increase together with the capabilities of confusing
them when attacking. The quicker the decision - manoeuvre - regathering chain
is triggered and performed, the better the chances for the flock. As inspired by
natural world, swarm engineering aims to turn the spontaneous behaviour of biological swarm, optimised by several years of Darwinian evolution, into design
rules able to return the same or comparable behaviour for engineered swarming
systems. Performance of such systems are strongly dependent on the capabilities
of communicating and sharing tasks. Naturalists observed a number of dynamics
contributing to the fast response in animal groups. Some of them are inspiring
and worth mentioning.
In 1981 Theherne and Foster isolated a dynamic in the group motion of marine
insects able to disorient the approaching predators. They named this dynamic
as the “Trafalgar Effect” [112]. Predators approaching a side of the group produce changes in the locomotory behaviour of the individuals at the periphery.
These ones increase their activity with rapid turnings that disorient the attacker.
This produces more, apparently random, encounters which are responsible for the
information to be transmitted throughout the group. This kind of behaviour triggers an avoidance manoeuvre faster than the approaching speed of the predator
preserving the insects from the attack. Information is then passed on a neighbouring basis, in the sense that, individuals aware of the danger alert neighbours
through their movements.
A few years after the study by Theherne and Foster, Wayne Potts postulated
the even more interesting “Chorus line hypothesis” [113]. Through footage analysis of avian flocks, Potts noticed that the manoeuvres, including those to avoid
predators, are initiated by individuals at the edges and propagated inward with
increased speed. Flock’s members realise of the approaching manoeuvre wave and
start propagating it before the wave reaches them, producing an increase in the
propagation speed. The manoeuvre wave then travels throughout the flock up
to four time faster than the speed achievable just by passing the information on
neighbouring basis. This reduces the chances of the predators to catch any of the
flock’s members. The propagation of manoeuvring waves was later found to be
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
125
unrelated to the flock size. More precisely, it exhibits scale-free characteristics,
whereby the speed of propagation in flocks of larger sizes increases compensating
for the increased sizes and hence requiring approximately the same time for a
manoeuvring wave to be transmitted throughout the flock [110].
Transferring the capabilities of animal groups onto engineered systems raises the
question: how is it possible to ensure that information flows quickly and sufficient
attention is paid to the threats of the external world as well as the targets the system is designed for? The answer has to consider two aspects. The first one is
translating the concept of “paying attention” typical of mindful creatures into
something that can be used by devices with a very low-level of, or no intelligence
at all. The second one is ensuring this task is carried out by the swarm efficiently
as a whole, exploiting its own features.
Flocking birds manoeuvre in response to predator attacks or follow the group
motion initiated elsewhere in the flock; in both cases they respond to an external
stimulus but, as different from monolithic systems, the group response emerges
out of the responses of the single flock members to the stimulus and to the
group motion itself. The idea of “paying attention” to something becomes then
weighting the external drives against the group dynamics. In terms of engineered
swarms, an external drive can easily be seen as a signal to follow, regardless
this signal contains the route to the accomplishment of a task or is generated by
an approaching threat. Providing all agents with the capability of tracking the
external drives would mean not taking advantage of the group structure, that is
having a number of independent agents, each one looking after itself only. On the
other hand, exploiting the connections in the group it is possible to allocate the
tracking duties more efficiently amongst all the swarm members. The amount of
resources each individual has to provide for tracking is then strongly connected
to the position of the agent within the system. This is only partially related to
the physical position of the agents as in the case of marine insects through which
the Trafalgar Effect was postulated; the position of an agent in the swarm has to
be considered as its position in the network of agents. This will be discussed in
the following sections.
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
6.2
126
Fast manoeuvring out of Network Structure
In Section 2.6.2 it was shown how particular networks make quick information
spreading possible, and drive multi-agent systems towards consensus and coherent
responses. When the spatial arrangement of the agents determines the network,
as in the case of nearest neighbours networks, and the network determines the
dynamics of the agents, the swarm is not guaranteed to evolve towards a configuration that boosts the consensus speed or the information spreading such
as the small world network. The lack of control on network arrangement gives
cause to find alternative methods to achieve quick consensus. The problem is
then reversed: given any network of interaction, how can this be exploited to
produce fast responses? In a swarm where all agents have similar characteristics,
the network can be used to rank them according to their influence on the rest of
the swarm. Consider an agent observed by a large number of swarm members.
This agent will be highly influential as its dynamics influences a wide portion of
the ensemble. Ideally this agent should be visible by its local neighbours (other
agents at close distance) and, possibly, by some distant ones. This would push the
swarm towards a small world arrangement and would allow the agent to influence
both its local neighbourhood and more remote regions of the swarm. It would
be clever then to provide this agent with more power to follow an external signal
or the ability to spot threats for the whole group. Indeed, even without having
such an advantageous network of connections, it is always possible to provide the
most observed agents with more tracking capabilities.
Several measurements are available in the literature that aim to define the degree
of centrality of a node in a network. These include, though being not limited to,
Betweenes Centrality, Closeness Centrality, Node Certainty [167, 168]. Although
none of them is detailed in the following, the idea of leveraging a “reciprocal
influence” measurement can be used to improve the swarm reactivity.
In the next section it will be shown how algebraic network characteristics can be
fed back into the dynamics to influence it directly.
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
6.3
127
Eigenvectors and Eigenvalues in a Network
System
In Section 2.4 it was stated how the number of zero eigenvalues of the Laplacian
indicates the number of components in the graph. If zero is a simple eigenvalue,
then the graph is connected. It can be proved that the spectrum of the Laplacian
matrix is positive semi-definite. Consider a complex, square matrix A of order
P
n, with entries aij , for i, j ∈ {1, . . . , n}. Define Ri = j6=i |aij | as the sum of the
absolute values of the non-diagonal entries in the i-th row. For such a matrix
a Gershgorin disc D(aii , Ri ) is defined as the closed disc on the Re − Im plane
centred at aii with radius Ri . Consider then Gershgorin circle theorem:
Theorem 1. Every eigenvalue of A lies within at least one of the Gershgorin
discs D(aii , Ri ).
The proof of the Gershgorin Circle Theorem can be found in many linear algebra
textbooks, see for example [169], hence is here omitted.
As the diagonal elements of the Laplancian are either positive or null, and each
row, excluding the diagonal entry, sums up to the negated diagonal element, all
the eigenvalues of the Laplacian matrix are either positive or null. As discussed
in Section 2.4, the negated Laplacian can be used to model a multi-agent system
in a linear fashion (see for example [76, 103, 104]). In this case the modelling
takes the form
ẋ = −Lx
(6.1)
As long as the zero eigenvalue is a simple eigenvalue, as state in Section 2.4, the
graph is connected and this model guarantees the achievement of a consensus. In
particular, the system converges to a consensus value as the negated Laplacian,
considering the Gershgorin theorem, will have all its eigenvalues in the left part
of the Re − Im plane, which includes the origin. The Re − Im plane for the
eigenvalues of a negated Laplacian is illustrated in Figure 6.1 where, Gershgorin
disks are centred on three diagonal entries of the Laplacian. As the zero eigenvalue always belong to the spectrum, the convergence speed of a linear system
modelled through the negated Laplacian matrix is expressed by the second smallest eigenvalue in magnitude. The larger the magnitude of the second eigenvalue,
the faster the system will converge.
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
128
Figure 6.1: Eigenvalues of a negated Laplacian matrix according to the Gershgorin
circle theorem. Every eigenvalue lies within at least one of the Gershgorin discs; these
are centred on the matrix diagonal entry and have radius equal to the row sum. All
the eigenvalues will have non-positive real part
In a matrix the first left eigenvector, here identified by c, is the one associated
with the largest eigenvalue in magnitude. For the Perron-Frobenius theorem
this coincides with the spectral radius for positive semidefinite matrices [170]. A
negated Laplacian matrix −L can be turned into a positive semi-definite matrix
M by adding its spectral radius ρ(L) to the diagonal entries, i.e.
M = −L + ρ(L)I
(6.2)
where, I is the identity matrix of appropriate dimensions. This way the whole
spectrum is shifted in the positive half of the Re − Im plane and the zero eigenvalue coincides with the spectral radius. Consequently, the left eigenvector corresponding to the zero eigenvalue of the Laplacian becomes the first left eigenvector
of the augmented Laplacian M, without undergoing any change as a constant
diagonal perturbation does not change the eigenvectors. As a consequence of
this, all the properties that apply to the dominant eigenvalue and its eigenvector of positive semidefinite matrices translate for the negated Laplacian for its
zero eigenvalue and the corresponding eigenvectors (left and right) equally. This
includes the properties obtainable as a consequence of the Perron-Frobenius theorem.
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
129
For the Laplacian and the adjacency matrices, the eigenvector associated to the
zero eigenvalue reveals some fundamental characteristics of the network. In particular, it is a measure of centrality for the nodes of the network as it quantifies
the number of paths that pass through a given node compared to all possible
paths. In order to understand this, consider the power method which can be
used to approximate the first left eigenvector. Given a generic matrix A, the
power method provides an approximation of its first eigenvector c, closer to the
actual eigenvector as higher powers are applied to the matrix A. This is multiplied by the uniform vector w(0) = (1, 1, . . . , 1); at the l − th iteration the method
returns
w(l) = (1, 1, , 1)A(l)
.
(6.3)
When A is the adjacency matrix, the first iteration w(1) returns the row sum of
A, that is the outdegree of each node, i.e.
w(1) = (1, 1, , 1)A = (d1 , d2, ...., dn )
.
(6.4)
At the second iteration the i − th entry of vector w(2) corresponds to the sum of
the outdegrees of all the neighbours of node i, that is the number of possible walks
of length 2 departing from node i. Propagating this, it is easy to understand how,
for high l, w(l) provides the number of paths of any length departing from each
node. As the power method provides an estimate of the Perron vector, it can
then be concluded that the i − th entry of vector c is the property of node i to be
the origin of a number of arbitrary long paths in the graph. An important consequence of this is the first left eigenvector ranks the nodes based on their ability to
start a successful information spreading across the network. This characteristic
goes beyond the use done here of the Perron vector, as applications can be easily
targeted in the spread of epidemic diseases, fades or emerging behaviour in fields
other than engineering [171, 172].
When the graph’s Laplacian matrix can be taken as a representation of the associated dynamical system, the characteristic of ranking the most influential nodes
allows use of the first left eigenvector to define the final, steady state of the unperturbed system subject to some initial conditions. This is easily proved considering
that the inner product of the Perron left eigenvector and the state vector is an
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
130
invariant of the system as Equation 6.5 shows:
d(c · x)
= c · ẋ = −cLx = 0
dt
.
(6.5)
Hence, the expression c · x(0) = c · x(t) holds and considering a generic vector
norm k · k, it follows
lim x(t) =
t→∞
c · x(0)
(1, 1, ...1)T
kck
.
(6.6)
Furthermore, the Perron-Frobenius theorem reveals that all the entries of the first
eigenvector are nonnegative, which will be useful in the following sections.
Finally, it is worth noting how the zero row sum of the Laplacian causes the
right eigenvector, call it e, corresponding to the zero eigenvalue, to be a uniform
vector, that is
e ∈ span(1)
6.4
.
Distributing Tracking Capabilities
In a swarm designed to accomplish some mission tasks, the amount of computational resources used for signal pursuing is to be considered relative to the total
amount of resources on which the swarm relies to carry out its mission. Allocating
these resources then becomes central for the accomplishment of the swarm tasks.
The problem of signal pursuing carried out by a group of individuals connected
through a network of sensing links is then analysed here.
When the resources are limited, their allocation amongst the group members
for signal pursuing has to be considered. In particular the performance of fast
convergence in the direction of the stimulus is strongly influenced by the way the
swarm pursues the stimulus. Consider a generic, linear, first order dynamics as
ẋ = Ax + Bu
(6.7)
where, x is the system state, a stacked vector of order n containing single agent
states, A ∈ Rn×n is the system matrix defining the linear system dynamics, that
is how the states affect each other; u is the input vector of order m that contains
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
131
the external stimuli and B ∈ Rn×m is the input matrix that determines how the
inputs affect the system states. In this case the relative importance of the external drives, compared to the group dynamics, is given by the relative weight on
the system of the Ax and the Bu terms. Looking at the same system’s representation, the problem of tracking an external signal efficiently turns then into
designing the input matrix B, defining which state is influenced by which input,
or, thinking about a swarm system, which agent pays attention to which input
and how much it does so compared to the other agents. For the purpose of this
analysis u is constant as it is used to evaluate the response of the system to a
simple perturbation. It is possible to associate u to a step input for the system
and use it to characterise its performance. However, later on in this chapter, u
will change with time becoming piecewise constant. The case of a continuous time
changing reference signal u(t) is not considered in this chapter and considered
towards possible future developments
In multi-agent systems the linear relations are often expressed using the Laplacian matrix, as explained is Section 2.4. The Laplacian L hence will replace
matrix A in Equation 6.7. Matrix B is then to be determined. Let’s assume that
each agent in the swarm is able to track the same signal, although the strength
sensed depends, for each agent, on the amount of resources this allocates to the
tracking task. This is modelled using an input vector u ∈ span {1}, that is,
u = u(1, 1, ....1)T which has dimension n and B ∈ Rn×n . In particular B can
be taken diagonal, that is each state is influenced by one single entry of u, in
this case the entries of B can be organised into an n-dimensional vector b which
contains the signals for each of the n states.
In a multi-agent system where the Laplacian is used to model inter-agent interactions and a driving signal from outside is fed to each agent, the dynamics of
the i-th agent can then be modelled as
ẋi =
X
j∈Ni
(xi − xj ) + bi (u − xi )
(6.8)
where, xi is the state of the i − th agent, u is the reference signal that the
agents pursue with different strength, according to the gains bi . Finally, Ni is
the set of neighbours of agent i in the graph, that is the set comprising of those
132
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
agents which agent i observes. Equation 6.8 considers an unweighted graph. The
external signal u is compared to the actual state as each agent can only see the
external signal as deviation from its actual state. For instance if the external
signal is a heading direction to follow, then the agent can only measure the new
heading starting from the actual one, hence as a deviation from that. Each agent
i is able to track the signal as long as the corresponding bi is nonzero. In the next
section the model in Equation 6.8 will be used to show how bi coefficients can be
chosen to drive the swarm to quick consensus about the external signal, i.e. all
the states will converge to u.
6.5
A Good Choice with Limited Resources
The choice of the bi coefficients in Equation 6.8 determines the amount of effort
each agent makes to track the signal u, as previously explained. Considering the
role of the first left eigenvector in ranking the nodes, based on their reciprocal
influence, a possible choice is leveraging the Perron vector of the Laplacian, i.e.
the one associated with the zero eigenvalue. This is a positive vector as per the
Perror-Frobenius theorem. The magnitude of c is determined by the way the
eigenvectors are scaled, but, when fed into a dynamical system, it has to be compared with the magnitude of the underlying dynamics, that is, the Laplacian. A
large magnitude of c implies a considerable amount of resources allocated globally
to follow an external signal. Although the following arguments apply with any
scaling of the eigenvectors, it is useful at this point to introduce the definition of
L1 vector norm, also known as Frobenius norm. This will be important in Section
6.6.
Definition 1. Given a generic vector v ∈ RN its L1 norm is defined as the sum
of the absolute value of its components, that is,
kvkF = kvk1 =
N
X
i=1
|vi |
.
(6.9)
At this point the model can be written in a more elegant way using vector notation
as
ẋ = −Lx + C(u − x)
(6.10)
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
133
where, u is the signal which is the same for all the agents. C is a diagonal matrix
whose entries are the entries of the first left eigenvector c of the Laplacian matrix
L of the system, Equation 6.10 can be further transformed to a more suitable
version as
ẋ = −(L + C)x + Cu
.
(6.11)
A change of coordinates can be performed to eliminate the term Cu from Equation 6.11. Consider
y = x − (L + C)−1 Cu
.
(6.12)
With u constant in time, the derivative of y is equal to the derivative of x.
Isolating x from Equation 6.12 and substituting in Equation 6.11, this becomes
ẏ = ẋ = −(L + C)(y + (L + C)−1 Cu) + Cu
(6.13)
where, most of terms cancel out turning the equation into
ẏ = −(L + C)y
.
(6.14)
This change of coordinates shifts the origin of the reference frame to the value of
the signal u, that hence is associated with the value zero.
The first left eigenvector allows the most influential nodes to be identified and
allocated with a better signal pursuing capability. Moreover, it allocates no pursuing capabilities to those node that are not observed by any other nodes, and
hence cannot contribute to make the system converge towards the signal. This is
confirmed by the following theorem and lemmas.
Theorem 2. Let L ∈ Rn×n be the Laplacian matrix of a connected graph and let
the matrix C being defined as C = diag{c} with c being the first left eigenvector
of L satisfying cT L = 0. Then the matrix −L̃ = −L − C is Hurwitz.
In order to prove this theorem the following Lemmas are needed.
Lemma 1. Let L be the Laplacian matrix of a connected digraph and let c be
its first left eigenvector corresponding to the zero eigenvalue. Then c has a zero
component ci if and only if node i has only outgoing edges or for each edge j − i
entering node i with nonzero out-degree, cj = 0 holds.
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
134
The lemma implies that all nodes i that correspond to ci = 0 are not globally
reachable.
Proof. Consider the i − th component of the vector cT L
[cT L]i = ci Lii +
X
cj Lji = 0
.
(6.15)
j6=i
By hypothesis the j = i term, that was taken out of the sum, is zero. For the
Perron Froebenius theorem, elements of c are nonnegative, hence for the sum
in Equation 6.15 to be zero, it must be cj = 0 for each j corresponding to an
existing edge j − i. Considering the same equation, the reversed implication
becomes straightforward.
The lemma also implies that if a node is globally reachable, it must have ci 6= 0.
Indeed, looking at Equation 6.15, a globally reachable node must have at least
one incoming edge Lji corresponding to a cj 6= 0, hence the equation must be
satisfied for some nonzero, hence positive, value of ci Lii .
Lemma 2. Let L be the Laplacian matrix of a connected digraph and let c be
its first left eigenvector corresponding to the zero eigenvalue. Then c has a zero
component ci = 0 if and only if node i is not globally reachable.
Proof. The first implication is a consequence of Lemma 1. More precisely, if
ci = 0 Equation 6.15 can be satisfied either for Lji = 0 for all j 6= i, or for
cj = 0 for Lji 6= 0. In the first case the node would not be reachable, i.e.
observed, by any other node; in the second case all the nodes connecting to
node j would have either no incoming connections or all connections belonging to
nodes that correspond to zero components of the eigenvector. As there must be
at least one nonzero component of c, that means that the nodes with cj = 0 are
either not reachable or reachable amongst them, but not globally. For the second
implication, in order to have non-globally reachable nodes, the graph must not
be strongly connected, hence the Laplacian matrix L is reducible and, through
permutations, it is possible to arrange it in the form
L=
"
L0
B
[0] LGR
#
(6.16)
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
135
where, L0 ∈ Rl×l and LGR ∈ Rk×k with n = l + k square matrices whose rows
refer respectively to non-globally reachable and globally reachable nodes. [0] is
a matrix of consistent dimensions whose entries are all zeros and B is a matrix
of consistent dimensions with no relevant characteristics. Note that L0 is not a
Laplacian, while LGR still is a Laplacian, being its row sum equal to zero. The
eigenvalues of L are the union of the spectra of L0 and LGR , and because the
graph is connected L has one single eigenvalue equal to zero that must belong to
the spectrum of LGR as this is a Laplacian itself. Consider to partition the first
left eigenvector accordingly, that is the first l and the second m entries as
c=
h
cN GR cGR
iT
.
(6.17)
As cTN GR L0 = 0 must hold, and zero is not an eigenvalue for L0 , this means that
cN GR = 0.
Lemma 3. Let L be the Laplacian matrix of a connected digraph and let c be
its first left eigenvector corresponding to the zero eigenvalue. If c has a zero
component ci = 0 then the diagonal element of the Laplacian matrix Lii 6= 0.
Proof. The lemma can be proved by contradiction by noting that if both ci = 0
and Lii = 0 then the node i would have no outgoing edges and all its incoming
edges would start from non-globally reachable nodes, hence the node i would
belong to an isolated component and the graph would be disconnected.
The following examples show the importance of the lemmas stated above.
Example 3. Consider the digraph already discussed in table 2.2 and here
reported again in table 6.1 with its Laplacian matrix and its first left eigenvector.
Table 6.1: Graph 2 example
2❡
❅
■
❅
❅
❄
3❡
✲ ❡1
❅
❅
❅
❅
❅
✲ ❡4
0
0
0
0
−1 2 −1 0
L=
0
0
1 −1
0 −1 0
1
1
0
c=
0
0
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
136
Nodes 2, 3 and 4 are not globally reachable (they cannot be reached from node
1), but still they have both incoming and outgoing edges. The second, third and
fourth entries of the first left eigenvector are then zeros.
Example 4. Consider the digraph in table 6.2 with its Laplacian matrix and its
first left eigenvector.
Table 6.2: Graph 1 example
2 ❡✛
❅
■
❅
❅
❄
3❡
❡1
❅
❅
❅
❅
❅
✲ ❡4
1 −1 0
0
0 1 −1 0
L=
0 0
1 −1
0 −1 0
1
0
0.33
c=
0.33
0.33
Reorienting the edge 2-1 makes vertices 2, 3 and 4 globally reachable, while node
1 is no longer globally reachable, nor simply reachable from any other node. This
is reflected into the first left eigenvector which is scaled to 1 considering the L1
norm.
It is now possible to prove theorem 2:
Proof. First suppose L to be reducible and consider its partition as per Equation
6.16. The first l columns of L correspond to those nodes with index i for which
ci = 0. All incoming edges to those nodes must have origin in nodes indexed by
some j for which cj = 0 too. So all the nonzero entries of the first l columns
must be in the top left partition L0 . As the zero eigenvalue is found in LGR , −L0
is Hurwitz. It follows that the matrix −L̃ = −L − C is Hurwitz too because
the zero entries of c sum up the diagonal of L0 and the positive ones make the
matrix −LGR − Ck Hurwitz, where Ck is the diagonal matrix composed of the
last k components of c. The case of L being irreducible, can be considered as a
particular one where l = 0
A consequence of Theorem 2 is shown through the following
Theorem 3. Let L ∈ Rn×n be the Laplacian matrix of a connected graph and let
c be its first left eigenvector, corresponding to the zero eigenvalue. Let the matrix
D ∈ Rn×n be a diagonal matrix whose entries are the elements of the vector d
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
137
and are all nonnegative. Then the matrix −L∗ = −L − D is Hurwitz if and only
if the scalar product hd, ci 6= 0.
Proof. As both d, c are nonnegative, for their scalar product to be nonzero it
must be that di , ci 6= 0 for at least one index i. If this is the case, considering the
matrix L partitioned as in Equation 6.16, the nonzero elements of D would sum
up the lower partition LGR making at least one row sum of it positive. The proof
is completed considering the Theorem III in [173] and here reported as Theorem
4 with the notation used so far for sake of completeness.
Theorem 4. ([173])
Let L ∈ Rn×n such that Lii > 0 and Lij 6 0 for j 6= i. Assume moreover that
P
Lii > j6=i Lij and that L is irreducible. The determinant of L then vanishes if
P
and only if nj=1 Lij = 0 for i = 1, 2, ..., n.
Because the determinant does not vanish, the zero eigenvalue of the Laplacian
disappears and Gershgorin circle theorem ensures that all the eigenvalues of −L∗
have then negative real part.
The result of Theorem 3 can be explained considering that, whenever a driving
signal is fed to one of the globally observable nodes, this will affect the whole
group making it converge towards the signal. It can be concluded that a diagonal
matrix, whose nonzero entries are the elements of the first left eigenvector, has
the property of stabilising the system once summed up to the Laplacian of the
graph.
Up to now this chapter showed that the first left eigenvector of the Laplacian matrix is able to stabilise and drive to consensus a swarm of agents whose dynamics
is modelled by the Laplacian itself. However, the effectiveness of it compared to
other possible choices was not assessed. This assessment is produced in the next
section.
6.6
The Significance of the First Left Eigenvector and the Effects of Scaling
In a number of cases, the first left eigenvector represents the best possible choice
of resource allocation over the nodes of the system for improving fast pursuing
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
138
of an external signal. The gain used to track the signal can be introduced in
the form of a diagonal perturbation of the Laplacian. This is valid when the
perturbation is of small magnitude with respect to the Laplacian. The amount of
resources, which corresponds to the scaling of the perturbation, is a fundamental
parameter. Consider again the system in Equation 6.18
ẏ = −(L + D)y
(6.18)
where, as different from Equation 6.14, matrix D represents a generic diagonal
perturbation to the Laplacian. The spectrum of −(L+D) is shifted by a constant
quantity k if a uniform diagonal perturbation kI is added, where I is the identity
matrix of appropriate dimensions. This was already done in Section 6.3. Assume
now that the matrix −(L + D) is Hurwitz and that the scalar k is positive and
larger than the largest eigenvalue in magnitude of the −(L + D) matrix. This
makes [−L − D + kI] a strictly positive matrix with the smallest eigenvalue
in magnitude of −(L + D), call it λ1 , contributing to the spectral radius of
[−L − D + kI], namely ρ([−L − D + kI]) = λ1 (−L − D) + k. The spectral radius
received more attention in the literature than the smallest eigenvalue considered
so far. In particular, given a generic matrix A ∈ Rn×n , the spectral radius is an
always increasing function of each entry of A, and in particular the expression
∂ρ(A)
ci ej
= ∇ρ(A) = T
∂ij
e c
(6.19)
holds where, e and c are, respectively, the first right and left eigenvectors as
reported in [174]. In [175] an expression is given for the spectrum of a matrix
subject to a small perturbation, that is
λ(A + D) = λ(A) +
cT De
eT c
.
(6.20)
This have important consequences. The first is that the spectral radius is an
always increasing function of the matrix entries. Consider the matrix M = −L +
kI: its spectrum is the spectrum of Laplacian matrix L shifted by the positive
quantity k, and its eigenvectors are exactly the ones of L. This implies that,
as long as a perturbation of small magnitude is introduced, the spectral radius
ρ(M − D) = ρ(−L − D + kI) can be approximated by means of the first left and
right eigenvectors of the Laplacian. This is detailed in the following subsection.
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
6.6.1
139
Limited Resources and Small Perturbations
The magnitude of the diagonal perturbation to the system dynamics can be measured in terms of vector norm. Consider a perturbation D = −C = −diag{c}: in
particular let c and e be scaled such that their L1 norm is equal to some generic
positive scalars γ and ǫ, that is,
kck1 = γ
kek1 = ǫ
X
X
i
i
|ci | = γ
|ei | = ǫ
X
X
ci = γ
i
ei = ǫ
.
i
Furthermore, assume that the system has high coupling gains, that is kLk ≫ kDk
for some norm k·k. Then, for what was previously stated,
P
cT Ce
γ 2 ǫ i c2i
P
ρ(M − C) = ρ(M) − T = ρ(M) −
= ρ(M) − γkck22
e c
γǫ i ci
. (6.21)
This implies that the smallest eigenvalue in magnitude of the perturbed negated
Laplacian is
λ1 (−L − C) = −γkck22
.
(6.22)
Note that here the Euclidean and the Frobenius norm are indicated with the
classical p-norm notations to avoid confusion.
An important consequence is that the first left eigenvector represents a gradient
descending direction for the spectral radius at −L that is for the unperturbed
system. This local property also holds in a neighbourhood of C = [0]. As the
perturbation increases in magnitude, the minimum of the spectral radius is not
said to be obtainable through a diagonal perturbation composed of the first left
eigenvector.
Considering a fixed Frobenius norm for the perturbation, the same argumentations used to find the diagonal matrix that maximizes the spectral radius in [174]
can be adopted to find the minimum of it. Consider the gradient of the spectral
radius in Equation 6.19 and consider the eigenprojection E of the generic matrix
A as defined in [176].
Definition 2. The eigenprojection of a matrix A corresponding to the eigenvalue
0 or, for brevity, the eigenprojection of A, is the projection, i.e. the idempotent
140
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
matrix, Z such that
R(Z) = N (Aν )
and
N (Z) = R(Aν )
where, R indicates the Range and N the Null space. ν is the smallest k such that
rank(Ak+1 ) = rank(Ak )
k ∈ {0, 1, 2....}
.
As the idempotent matrix is uniquely defined, so is the eigenprojection. It can
be stated that the gradient in Equation 6.19 corresponds to the transpose of the
eigenprojection E, that is
E=
ecT
cT e
(6.23)
ceT
.
(6.24)
cT e
The following lemma shows how a diagonal perturbation ED composed of the
∇ρ(A) = E T =
Perron vector attains the minimum of the spectral radius for the matrix (−L −
ED + KI) amongst the diagonal perturbations with fixed Frobenius norm, in the
hypothesis of small magnitude of it.
Lemma 4. Let L be the Laplacian matrix of a directed graph on N nodes. Let
e and c be respectively the right and left Perron vectors associated with the zero
eigenvalue. Let K be a positive scalar greater than the largest eigenvalue in magnitude of L and ED be the diagonal matrix defined as
ED = diag
e ◦ c
cT e
(6.25)
where “◦” indicates the product element by element. Then there exists a t∆ > 0
such that for all 0 < t ≤ t∆ the minimum of the spectral radius of (−L−t∆+KI)
with ∆ belonging to the space of diagonal matrices of unitary Frobenius norm is
achieved for ∆ = ED /kED kF .
Proof. Consider the derivative of the spectral radius of the matrix M = −L + KI
in the direction of ∆ and ED /kED kF , respectively
141
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
ρ (M − t∆) − ρ (M)
t→0
t
T
= −E · ∆ = −trace (E∆)
ρ M − t kEEDDkF − ρ (M)
(M) = lim
t→0
t
E
D
= −kED kF
= −E T ·
kED kF
ρ′∆ (M) = lim
ρ′ED /kED kF
(6.26)
(6.27)
where the inner product between matrices is defined as
P · Q = trace (P Q)
which provides a definition for the Frobenius norm of matrices as kP kF =
Now, for the Cauchy-Schwarz inequality [174]
trace(E∆) = kE T ∆k2F ≤ kEE T kF k∆T ∆kF = kEkF
p
trace (P T P )
(6.28)
that implies
ρ′ED /kED kF (M) ≤ ρ′∆ (M)
.
(6.29)
Due to the definition of ED in Equation 6.25, and due to the fact that M has
the same eigenvectors of L, e is always uniform and the gradient of the spectral
radius depends upon the left eigenvector only.
In the special case of a balanced graph, the Laplacian matrix is line sum symmetric, that is rows and columns sum up to the same values, though not necessary
the same for every row or column. In this special case both the first left and
the first right eigenvectors are uniform. Given a norm of the perturbation vector, the property of achieving the minimum of the spectral radius locally extends
to all possible perturbation magnitudes, as the matrix M − C keeps the same
eigenspace of M regardless of the magnitude of C.
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
6.6.2
142
Fixed Trace, a Misleading Effect
The following theorem and lemma, given and proved in [177], provide a useful
means to find the best control matrix D for the system previously described, in
the space of diagonal matrices with fixed trace. In particular, the theorem provides a way to calculate the matrix D with zero trace which returns the lowest
spectral radius possible and the lemma extends the results to all possible traces.
Note that when elements of D are all nonnegative, the trace corresponds to the
L1 norm of the vector composed of the same nonzero entries of D.
In order to make the following theorem and lemma understandable, the following
notation is needed:
- µ(A; t) is the minimum of the spectral radius for a generic matrix A
subject to a diagonal perturbation with trace t. In particular for zero
trace perturbation, it holds µ(A) , µ(A; 0).
- Dq indicates a diagonal matrix whose nonzero entries are the elements of
vector q.
- The diagonal similarity of a matrix A is defined as A1 = Dq−1 ADq where q
is a positive vector. The diagonal similarity of a nonnegative matrix A is
nonnegative too and the value of µ(A) is not effected by the
transformation.
Theorem 5. Let A be an n-by-n essentially nonnegative matrix, and suppose
that P is a permutation matrix such that P T AP is in Frobenius normal form.
Let A0 be the direct sum of irreducible matrices that is obtained from P T AP by
replacing all entries in off-diagonal blocks with 0’s, and let B be the line sum
symmetric diagonal similarity of A0 . Then µ(A) = µ(A0 ) = (1/n)1T B1, and
µ(A) = min{λmax (A + Dq ) : 1T q = 0} is achieved only by q = P [µ(A)1 − B1].
Lemma 5. Let A be an n-by-n essentially nonnegative matrix and let t be a scalar.
Then µ(A; t) = µ(A) + t/n. Furthermore, if the minimum µ(A) is achieved only
by the vector q with 1T q = 0, then the minimum µ(A; t) is achieved only by a
vector q + t/n1 with 1 = (1, 1, 1....1)T .
The calculation of the matrix B from Theorem 5 is not immediate. It consists in
finding the positive diagonal matrix Dq that makes matrix A0 line sum symmetric.
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
143
This corresponds to finding the solution of the system of equations
Dq−1 Aq − Dq AT δ = 0
(6.30)
where δ is a vector whose elements are the inverse of the elements of q. This can
be solved through iterative procedures such as the Newton-Raphson method.
The above theorem and lemma provide useful way to calculate the best diagonal
perturbation for fixed trace. Unfortunately the fixed trace condition produces,
in many cases, the effect of “fouling” the less influential nodes to allocate more
influence to the most influential ones. As the arithmetic sum of the diagonal
element must be the same, some elements, those corresponding to less observed
nodes, can be taken negative so to allow other elements to achieve larger values,
even beyond the amount which all the elements have to sum up to. This “mathematical trickery” does not match the initial intent of distributing resources in
an efficient way as a negative element, that corresponds to passing a node wrong
information about the driving signal, requires a resource allocation by the node
anyway. It is then the absolute value of the elements, that corresponds to the
actual amount of resources used, to be considered, and this is the case of L1 norm
which the present chapter is mostly concerned on.
6.7
Solution of the First Order Linear Dynamics
Numerical simulation for a network of 20 nodes connected in different ways were
carried out and are reported here. The networks used are a non-symmetric lattice, a periodic lattice or ring, a random network and a Small World network.
The system is of first order, of the kind described by Equation 6.10 and consensus speed is measured by the smallest eigenvalue in magnitude of the dynamical
system matrix L + C or L + D, depending on the kind of diagonal matrix used
to stabilise the system. Performances obtained through a uniform distribution
of tracking capabilities and one corresponding to the first left eigenvector were
compared with the speed of consensus obtained for a capability distribution obtained through numerical optimisation. Due to the continuous nature of the
problem, the optimisation is performed using a gradient-based method implemented through the MATLAB R built-in optimiser fmincon. This performs the
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
144
gradient descending search using one of four algorithms automatically selected
time to time depending on the problem’s peculiarities. Its objective is finding the
diagonal perturbation D that maximises the smallest eigenvalue in magnitude of
the matrix (L + D). The optimisation is constrained to the vectors of unitary L1
norm to produce a meaningful comparison with the first left eigenvector, which
is scaled in the same way.
Each node is seen as a mono dimensional variable “x” whose initial condition is
zero and is pushed towards convergence to 1, being vector u in Equation 6.10
defined as u = (1, 1, ....1)T . The nodes are hence in agreement at the beginning
of the simulation and converge towards the unitary signal in different fashions
depending on how the tracking resources are allocated.
In the following subsections the cases referred to the different networks considered
are detailed. In subsection 6.7.5 a summary of the results is given, which includes
a comparison of the performance achieved for the different networks with the
numerical optimiser and the first left eigenvector. The comparison is made using
the value achieved for the smallest eigenvalue of the perturbed Laplacian.
6.7.1
The Regular Lattice Case
The lattice structure considered is non-symmetric in the sense that a node in
general does not observe and is observed by the same subset of nodes, hence
the corresponding graph adjacency and Laplacian matrices are non-symmetric.
The generic node i “observes” the two in front of it, i + 1 and i + 2 and just
one behind it, i − 1. Node 1 observes just nodes 2 and 3; the second-last node
observe just the last node and the third-last while the last node just observes the
second-last node. The asymmetry in the graph is wanted as a symmetric graph
would produce a uniform first left eigenvector. This case is considered separately.
A graphical representation of the lattice used is reported in Figure 6.2.
This structure of the lattice graph finds justification in the preferential direction
of view found in animal groups. The attention of individuals in large group focalises on a restricted group of individuals found in particular areas within their
sensing range. Here it was chosen to give preference to the forward sensing more
than the backward one within a maximum sensing range of 2 subsequent individuals, hence the asymmetry of the network. Moreover, a lattice is a plausible,
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
145
Figure 6.2: Sketch of the lattice network. Each node in the inner section has 3 outgoing
and 3 incoming edges. Nodes at the left hand side have more outgoing than incoming edges, while nodes at the right hand side end have more incoming than outgoing
edges. Note that an outgoing edge implies that the node where the edge is originated
“observes” the node where the edge ends.
though quite rigid, representation of the sensing links in groups where sensing
range is limited.
The values for the first eigenvalues achieved in the three cases considered for this
graph are λ1 = −0.2994 when the first left eigenvector is used, λ1 = −0.3253 for
the numerically optimised vector and λ1 = −0.05 for the case of uniform resource
allocation. The difference between the performance using the optimised vector
and using the first left eigenvector is very small compared to the consensus speed
attained using a uniform distribution. In the latter case all nodes move rigidly
towards the consensus value, but the motion is much slower than the one obtained
with a non uniform allocation of resources. This is well represented in Figure 6.3.
6.7.2
The Ring Case
The case of a periodic lattice builds upon the aperiodic lattice illustrated in
the previous section. The difference consists in having agents at the extremities
connected so to have the same number of inbound and outbound edges for each
node, regardless its index. The resulting adjacency and Laplacian matrices are
said to be “circulant”. Circulant matrices are balanced matrices and, as well as
symmetric matrices, are a subset of normal matrices. In normal matrices the left
and right eigenspaces present no difference and in particular the first left and right
eigenvectors are the same. Thus, the first eigenvectors of the Laplacian for this
network are both uniform. As shown in Section 6.6.1 the choice that maximise the
speed of consensus, for a given magnitude, is obtained for a uniform distribution
of the weights. The plots in Figure 6.4 illustrate the outcomes of the numerical
simulation for the 20 node, first order, linear system for a ring network.
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
6.7.3
146
The Small World Network Case
The Small World network is often formed with a ring network with a number of
rewired connections. Here the rewiring is kept but the original network is a nonperiodic lattice, as in Section 6.7.1. The rewiring is performed with probability
psw = 0.1. The plots in Figure 6.5 show the dynamics of the 20 node, first order,
linear system for a Small World network.
6.7.4
The Random Network Case
Distinct from the three previous cases of lattice, ring and Small World networks,
a random network does not have an underlying or base structure that is then
modified. Each node connects to the others with a certain probability, here
arbitrarily set to pr = 0.3. The resulting graph might be not connected but
the results presented here do refer to a connected graph. The plots in Figure
6.6 show the dynamics of the 20 node, first order, linear system for a random
network. The resource allocation presented in the figure is an example of how a
numerical optimiser can encounter problems due to local minima. Because of this
the first left eigenvector can surpass the performance of the optimised dynamics.
However it should be noted how this is strongly dependent on the initial guess
provided to the optimiser. Providing the first left eigenvector or a uniform vector
as initial guess would have probably produced a different outcome. In particular,
when the first left eigenvector is taken, the optimiser is likely to return the first
left eigenvector as optimal choice. This is a consequence of the local optimal
conditions guaranteed by this choice.
147
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
0
2
0.5
Optimised vector
Uniform vector
First left eigenvector
4
Resource allocation
Node index
6
8
10
12
14
0.4
0.3
0.2
16
0.1
18
20
0
5
10
Node index
15
0
0
20
(a)
5
10
Node index
15
20
(b)
1
0.9
0.8
0.7
X
0.6
0.5
0.4
0.3
First left eigenvector
Uniform vector
Optimised vector
First left eigenvector average
Optimised vector average
0.2
0.1
0
0
5
10
t [s]
15
20
(c)
Figure 6.3: Consensus in a lattice, performance comparison. (a) The adjacency matrix
of the lattice where the dots represent nonzero entries. (b) Comparison between the
values of the diagonal entries of matrix D for the matrix composed of the first left
eigenvector, a vector obtained through numerical optimisation and one obtained by
distributing evenly the tracking characteristics. All these vectors are scaled to have
a unitary L1 norm. (c) Time evolution of the first order system driven to consensus
about x = 1 by the diagonal perturbation of the Laplacian.
148
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
2
0.05
4
0.045
6
0.04
Resource allocation
Node index
0
8
10
12
14
0.035
0.03
0.025
Optimised vector
Uniform vector
First left eigenvector
0.02
0.015
16
0.01
18
0.005
20
0
5
10
Node index
15
0
0
20
(a)
5
10
Node index
15
20
(b)
0.7
0.6
0.5
X
0.4
0.3
0.2
First left eigenvector
Uniform vector
Optimised vector
0.1
0
0
5
10
t [s]
15
20
(c)
Figure 6.4: Consensus in a ring, performance comparison. (a) The adjacency matrix of
the ring where the dots represent nonzero entries. (b) Comparison between the values of
the diagonal entries of matrix D for the matrix composed of the first left eigenvector,
a vector obtained through numerical optimisation and one obtained by distributing
evenly the tracking characteristics. All these vectors are scaled to have a unitary L1
norm. (c) Time evolution of the first order system driven to consensus about x = 1
by the diagonal perturbation of the Laplacian. As the three distributions coincide the
plots overlap.
149
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
0
2
0.5
Optimised vector
Uniform vector
First left eigenvector
4
Resource allocation
Node index
6
8
10
12
14
0.4
0.3
0.2
16
0.1
18
20
0
5
10
Node index
15
0
0
20
(a)
5
10
Node index
15
20
(b)
1
0.9
0.8
0.7
X
0.6
0.5
0.4
0.3
First left eigenvector
Uniform vector
Optimised vector
First left eigenvector average
Optimised vector average
0.2
0.1
0
0
5
10
t [s]
15
20
(c)
Figure 6.5: Consensus in a Small World network, performance comparison. (a) The
adjacency matrix of the network where the dots represent nonzero entries. (b) Comparison between the values of the diagonal entries of matrix D for the matrix composed of
the first left eigenvector, a vector obtained through numerical optimisation and one obtained by distributing evenly the tracking characteristics. All these vectors are scaled
to have a unitary L1 norm. (c) Time evolution of the first order system driven to
consensus about x = 1 by the diagonal perturbation of the Laplacian.
150
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
0.4
0
2
0.35
Optimised vector
Uniform vector
First left eigenvector
4
0.3
Resource allocation
Node index
6
8
10
12
14
0.25
0.2
0.15
16
0.1
18
0.05
20
0
5
10
Node index
15
0
0
20
(a)
5
10
Node index
15
20
(b)
1
0.9
0.8
0.7
X
0.6
0.5
0.4
0.3
First left eigenvector
Uniform vector
Optimised vector
First left eigenvector average
Optimised vector average
0.2
0.1
0
0
5
10
t [s]
15
20
(c)
Figure 6.6: Consensus in a random network, performance comparison. (a) The adjacency matrix of the network. The dots represent nonzero entries. (b) Comparison
between the values of the diagonal entries of matrix D for the matrix composed of the
first left eigenvector, a vector obtained through numerical optimisation and one obtained by distributing evenly the tracking characteristics. All these vectors are scaled
to have a unitary L1 norm. (c) Time evolution of the first order system driven to
consensus about x = 1 by the diagonal perturbation of the Laplacian.
151
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
6.7.5
Performance of First Order System
The dominant eigenvalue can be taken as a measure of the speed of the system
dynamics. For a system modelled through a Hurwitz matrix this eigenvalue is
the smallest in magnitude. It is then possible to summarise the results obtained
in terms of speed of response for the linear dynamics depending on the diagonal
perturbation added to the Laplacian. This is done in table 6.3 for the cases
analysed within this Section.
Table 6.3: First eigenvalues of the system matrix (L+D) where, L is the network graph
Laplacian and D is a diagonal matrix whose nonzero entries are either the first left
eigenvector (F. L. E.), a uniform vector or a vector numerically optimised to maximise
the correspondent eigenvalue. All of these have unitary L1 norm.
D diagonal values
Lattice
Ring
Small World
Random
F. L. E.
0.2994
0.05
0.2416
0.2430
Uniform
0.05
0.05
0.05
0.05
Optimised
0.3256
0.05
0.2466
0.0823
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
6.8
152
The Second Order and the Nonlinear Dynamics
The dynamics observed for the first order system, where linear interactions represented by the Laplacian of the connected graph are augmented by the first left
eigenvector of it, can be translated for a second order system where nonlinear
relations can be put in place. This has the advantage of attaining a less abstract
representation of the dynamical system composed of a swarm of mobile agents
that manoeuvre in response to external stimuli while staying cohesive. Consider
the second order dynamics:
ẍi = Cid
X
j∈Ni
(xj − xi − d) + Civ
X
j
(ẋj − ẋi ) + Ciz (zi − xi ) + Ciu (ui − ẋi ) (6.31)
where, the symbols are the same used for first order system; moreover Cid , Civ , Ciz ,
Ciu are constants that weight the relative importance of the terms of the equation, zi is a reference for the actual position xi and d is a desired gap between the
positions of any two agents i and j connected in the network. Equation 6.31 can
be used to model the dynamics of a number of agents aligned along a line. The
difference between the positions of any two of them can be positive or negative
depending on one preceding the other or viceversa. The difference xj − xi in the
first term of Equation 6.31 is positive if the position of the j − th agent is further
down the line than agent i. In this case it makes sense to compare it to the
positive scalar d. Equation 6.31 can then be used as it is to model the dynamics
of a number of forward-looking agents along the line. This corresponds to a line
graph, either periodic (a ring) or not. This simple one-dimensional model can
be used to show that some properties of the first order dynamics hold for the
second order dynamics as well, where nonlinearities can be introduced to make
the model better suited to reproduce the motion of a swarm of autonomous agents.
First consider that each agent does not have a preference, or a driving input for
its position in absolute terms, that is Ciz = 0∀i ∈ {1, ...N}. Equation 6.31 can be
written in vector form as
ẍ = −Cd Lx + d − Cv Lẋ + Cu u − ẋ
(6.32)
153
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
where, C d , C v and C u are diagonal matrices and d is a uniform vector stating
the desired distance each agent aims to keep with respect to its neighbours in the
graph. This second order model can be translated into the state space becoming
(
ẇ = y
ẏ = (−C v L − C u )y − Cd Lw + Cu u − Cd Ld
.
(6.33)
A coordinate change can be now operated to get rid of the terms related to the
external input. Define two new variables ξ and θ such that
(
ξ = w+d
θ = y − (Cv L + Cu )−1 Cu u
.
(6.34)
Using this transformation, Equation 6.33 becomes
(
ξ̇ = θ
θ̇ = −(C v L + C u )θ − Cd Lξ
.
(6.35)
Finally Equation 6.35 can be turned into matrix form as
( ) "
#( )
ξ˙
[0]
I
ξ
=
d
v
u
θ̇
−C L −(C L + C )
θ
(6.36)
where, I is the identity matrix of appropriate dimensions. The entries of C d , C v
and C u matrices are to be chosen consistently with the constant (unitary) total
amount of resources the swarm uses to track an external signal and to produce a
coherent swarm behaviour, that is
Cid + Civ + Ciu = 1 ∀i ∈ {1, ...N}
(6.37)
For positive C d , C v and C u the system matrix is not Hurwitz as it keeps the zero
eigenvalue but, for a connected swarm, this is a simple eigenvalue and all the
other eigenvalues have negative real parts. This can be proved in the following
Lemma 6. Let L be the Laplacian matrix of a connected directed graph on N
nodes. Let C d , C v and C u be diagonal matrices with positive entries and in
particular let C u have nonzero elements along the diagonal where the first left
eigenvector, corresponding to the zero eigenvalue of the Laplacian matrix, presents
154
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
nonzero entries. Then all the eigenvalues of matrix
S=
"
[0]
I
−C d L −(C v L + C u )
#
(6.38)
have non-positive real part. Moreover the spectrum of submatrix −(C v L + C u )
belongs to the spectrum of matrix S.
Proof. Consider the characteristic equation for matrix S, that is
det(S − λI) =
−λI
I
d
v
−C L −(C L + C u + λI)
.
(6.39)
As −λI and −C d L are permutable, from [178], the determinant can be calculated
as
det(S − λI) = |λI(C v L + C u + λI) + C d L| = 0
.
(6.40)
Consider to reduce all the matrices, which are not already diagonal, to their
diagonal entries, that is their eigenvalues, and indicate these reduced matrices
with subscript D, then
det(S − λI) = |λI(C v LD + C u + λI) + C d LD | = 0
.
(6.41)
The characteristic equation reduces then to the determinant of a diagonal matrix.
As this is given by the product of the diagonal terms, the eigenvalue problem can
be treated as a scalar one and turned into finding the λs that make a generic
diagonal element null. As there are only diagonal matrices, it is possible to easily
multiply the terms in Equation 6.41 to obtain, for the generic i − th diagonal
element,
λ2i + (Civ Li + Ciu )λi + Cid Li = 0
,
(6.42)
which is a second degree equation. The λ that satisfies it can be found as
λi =
−(Civ Li + Ciu ) ±
p
(Civ Li + Ciu )2 − 4Cid Li
2
.
(6.43)
As Cid Li ≥ 0 the real part of the Equation 6.43 will always be non-positive.
Moreover, the zero eigenvalue belongs to the spectrum of S as it belongs to the
spectrum of C d L.
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
155
The second part of the Lemma can be proved by considering the Schur formula
[178] according to which the determinant det(S − λI) can also be calculated as
det(S − λI) = | − λI − (C v L + C u + λI)C d L|| − C v L − C u − λI| = 0
(6.44)
from which it is clear how the spectrum S includes the eigenvalues of −(C v L +
C u ).
The linear system is then stable and the presence of the zero eigenvalue reflects
the fact that agents do not achieve consensus about a unique position as a consequence of C z = 0. However, they maintain distances with respect to each other.
Extending the model to more dimensions, the design distance d has to be compared with a norm to map the relative positions between pairs of connected agents
into a scalar. As an Euclidean norm is a nonlinear function of the relative positions, the extension beyond the one-dimensional model is made together with
the introduction of nonlinear interactions in the network in the form of artificial
potential functions. As the artificial potential functions represent a viable way to
control relative positions in a highly nonlinear fashion they are suitable for the
needs herein, and can be introduced considering the background outlined in the
previous chapters. The introduction of nonlinearities is presented here although
most results would require deeper analyses to be fully understood. In the following, results based on numerical tests are reported that open a wide horizon for
prospective developments. Once more, the Morse potential is used and the model
takes then form
dx
dt
dv
dt
= v
= −∇U − Lv + D(vdes − v)
(6.45)
where, ∇U is a time dependent vector representing the gradient of the Morse
potential used in Chapters 3, 4 and 5. The numerical values for the coefficients
of the potential are summarised in Table 6.4. The small magnitude of C a and
C r coefficient is chosen not to let the inter-agent distance keeping dominate over
the consensus to a common velocity. The signal to track is introduced into the
equation as vdes while L is the graph Laplacian and D is a diagonal matrix
containing the weights through which each agent tracks the signal. In Figure 6.7
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
156
it is shown how manoeuvring in a three-dimensional space is different depending
on the values of D despite the fact that in both cases D has unitary Frobenius
norm. The network used is generated at the beginning of any new manoeuvre as
a directed graph in which each node observes the 5 nodes which are physically
closer to it. This is similar to the connection rule described in Chapter 3, although
the number of neighbours is significantly smaller, hence the connectivity of the
graph cannot, in principle, be guaranteed. However, this choice is coherent with
the structure of sensing networks in animal groups, as previously explained. As
already verified for the first order case, when the diagonal of D is composed of the
elements of the first left eigenvector, the swarm accelerates towards a reference
direction quicker, covering longer distances in the same time. In Figure 6.7 this
is compared to the case in which the weights are the same for all the agents, that
is, the diagonal entries of D are all the same.
The difference in performance for the change in heading can be better understood
looking at the time history of the errors between the actual and the desired
headings, shown in Figure 6.8. The error is calculated as the instantaneous angle
∆Θ between each agent’s velocity vector and the desired velocity vector. It can be
seen how the error reduces quicker for the swarm where the tracking resources are
allocated according to the first left eigenvector. Moreover, comparing panels (b)
and (d) of Figure 6.8 it can be noticed how, for approximately the same change
in heading, the swarm with uniform distribution reacts differently, increasing
the gap in performance. This is most likely due to the increased inertia effects
in panel (d) respect to panel (b) for the swarm with uniform distribution of the
weights. This, in turn, is caused by an increased modulus of the velocity at which
the change of heading is performed. In fact, the manoeuvre in panel (b) takes
place at lower speed than in panel (d) due to the initial zero velocity condition,
that, as understandable from the shorter distance covered (Figure 6.9), influenced
more the swarm with uniform distribution of the weights. In the last manoeuvre
the influence of the initial condition becomes negligible and the difference in
performance appears more clear.
The heading requested to the agents, after the initial static relaxation, are reported in Table 6.5.
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
157
Table 6.4: Coefficients used in the APF for the three dimensional manoeuvring simulations.
Ca
10−5
Cr
10−5
La
10
Lr
5
Table 6.5: Desired velocity vectors for the swarm to produce the changes in heading.
The table shows the three components of the reference velocity vector which each agent
is supposed to follow in the simulations shown in this section.
1
2
3
4
Vx
Vy
Vz
√
√
√
1
0
3
3
0
√
2
2
3
3
0
√
2
2
√
− 22
−
3
3
−
√
2
2
0
Avoiding the Undesirable Effects of Distance Keeping
The use of nonlinear inter-agent forces for relative position keeping, that is the
APFs in this case, was limited to preserve the group from dispersing. As the
graph is not symmetric, so the inter-agent forces are, and this produces unbalanced drifts into the swarm of the same kind of the ones used in Chapter 4 to
make one agent overcome the potential barrier and change the configuration of
the swarm. This doesn’t always return a stable configuration as in Chapter 4,
and the unbalanced interactions may produce an effect counteracting the desired
convergence towards a common velocity.
Increasing by a factor 102 the magnitude of the APFs, that is having C r = C a =
10−3 , with all the other parameters unchanged, introduces fluctuations in the velocity that reflect on the final trajectory and on the speed of manoeuvre. In Figure
6.9 the final trajectories followed by the swarm with the uniform tracking capability distribution (in blue) and the one with a tracking capability distribution
reflecting the first left eigenvector are shown. It can be seen how dramatically the
difference builds up. In Figure 6.10 instead the comparisons of the heading errors
158
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
Uniform vector
First left eigenvector
10
Z [m]
5
0
−5
10
40
0
20
0
Y [m]
−10
−20
X [m]
Figure 6.7: Manoeuvring in three-dimensional space. Agents track a step changing
external signal that provides the vectorial reference velocity. The agents sensing the
signal weighted by the first left eigenvector (in red) achieve the velocity target sooner
than the ones sensing the signal uniformly (in blue), hence cover more distance.
are shown. The set of reference velocity used here is the same previously used and
reported in Table 6.5. It can be noted how a uniform distribution of the weights
produces bounces in the error time history. These are due to the effect of the
asymmetries in the graph that generate asymmetric distance-keeping forces, and
when the effect of these become predominant over the consensus to a common
velocity, the consensus achievement slows down and the error increases again.
When the consensus is sought through the first left eigenvector, the detrimental
effect of cohesion forces looks to be minimised and a more coherent convergence
to consensus emerge. This can be in part due to the fact that the first left eigenvector awards the nodes that are mostly observed with more tracking capability
than the nodes that are mostly observers. As the observers tend to keep a given
relative distance from the observed, the distribution provided by the first left
eigenvector makes the observed node to drive the swarm relying on the distance
keeping of the observers to follow up on the motion. However, these are just
preliminary studies whose outcome look promising for the future expansion of
the research branch here outlined.
159
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
60
150
Uniform vector
First left eigenvector
100
40
∆ θ [deg]
∆ θ [deg]
Uniform vector
First left eigenvector
50
50
30
20
10
0
20
25
30
t [s]
35
0
40
40
45
(a)
50
t [s]
55
60
(b)
60
100
Uniform vector
First left eigenvector
Uniform vector
First left eigenvector
50
80
∆ θ [deg]
∆ θ [deg]
40
60
40
30
20
20
0
60
10
65
70
t [s]
(c)
75
80
0
80
85
90
t [s]
95
100
(d)
Figure 6.8: Error time history with respect to the desired heading. Before starting the
manoeuvres the agents relax for the first 20 seconds aiming to a null velocity. (a) The
first manoeuvre is a straight acceleration that sets the agents in motion aiming to get
unitary speed; (b) the first change in heading is performed with the swarm weighting
the signal using the first left eigenvector (in red) surpassing the swarm were sensing is
uniform distributed (in blue) in achieving the new heading; (c) a similar difference can
be observed for the second change in direction; (d) the gap in performance increases
in the last change in direction. Although the angular gap is similar to the one in case
(a), the increased effects of the inertia penalise the swarm characterised by a uniform
distribution more than the other.
160
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
Uniform vector
First left eigenvector
5
Z [m]
0
−5
−10
10
0
5
−10
Y [m]
−20
−5
10
15
0
X [m]
Figure 6.9: Detrimental effect of inter-agent forces in manoeuvring. Agents track a
step changing external signal that provides the vectorial reference velocity. The agents
sensing the signal weighted by the first left eigenvector (in red) are less affected by
the contrasting behaviour of keeping relative distances than the agents using a uniform
weight distribution (in blue).
6.9
Fast Manoeuvring - Final Considerations
The first left eigenvector can be used to create a diagonal control matrix that
stabilises a multi-agent system with dynamics governed by the Laplacian matrix. In the case of a small magnitude of the control matrix with respect to
the coupling strength, the first left eigenvector represents the optimal choice for
maximising the magnitude of the system smallest eigenvalue, that is ensuring fast
convergence towards an external input signal. This finds evidence into the local
maximisation of the magnitude of the first eigenvalue of the system, that is the
smallest in magnitude for a system with dynamics described by a Hurwitz matrix.
In the case where the connection graph of the system is balanced, the first left
eigenvector represents the best possible choice of allocating resources amongst
the system’s nodes for tracking the external signal, regardless its magnitude. In
this case the first left eigenvector belongs to span{1}.
161
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
200
100
Uniform vector
First left eigenvector
Uniform vector
First left eigenvector
50
∆ θ [deg]
∆ θ [deg]
150
100
50
0
20
0
−50
−100
25
30
t [s]
35
−150
40
40
45
(a)
55
60
(b)
100
200
Uniform vector
First left eigenvector
Uniform vector
First left eigenvector
150
80
100
50
∆ θ [deg]
∆ θ [deg]
50
t [s]
0
−50
−100
60
40
20
−150
−200
60
65
70
t [s]
(c)
75
80
0
80
85
90
t [s]
95
100
(d)
Figure 6.10: Detrimental effect of inter-agent forces on the error time history with
respect to the desired heading. Before starting the manoeuvres the agents relax for the
first 20 seconds aiming to a null velocity. (a) The effect of the inter-agent forces can
be noticed since the first acceleration: as the swarm with uniform distributed tracking
gets close to the target velocity, some agents experience a braking effect that can be
visualised as reversing the direction of motion of 180 ◦ C; (b) the first change in heading
highlights the difference in behaviour between the two swarms; (c) a similar difference
can be observed for the second change in direction; (d) the swarm driven to consensus
through uniform weight distribution seems to produce a quick minimisation of the error
but, it bounces back towards larger errors before achieving the target heading.
CHAPTER 6. FAST CONSENSUS AND MANOEUVRING
162
When the system is not balanced nor the control matrix is of small magnitude,
the first left eigenvector cannot be said to drive the best choice possible for tracking resource allocation. It still leads to a minimum of the first eigenvalue of the
system though just in local sense. The global minimum can be analytically found
for the set of diagonal control matrices with fixed trace although this has less
applicability to a problem of resource allocation including negative entries in the
control gains. Through linking the scaling of the eigenvector to the amount of
resources available for pursuing a given external signal, it is possible to build up
a meaningful model of resources allocation in distributed architectures.
The fundamental problem of calculating the first left eigenvector in a distributed
fashion, that is referred to a network for which each node does not have full visibility, can be overcome by using a distributed estimate as proposed by Qu et
al. [179]. However, it is beyond the objectives of this dissertation to assess the
performance of the swarm when distributed calculations are performed.
Relying on the graph algebraic properties to improve the reactivity of a swarm is
advantageous with respect to a numerical optimisation for the definition of the
amount of control each node has to provide. Local minima can be avoided and
results are close to the optimal ones, if not optimal.
The extension towards second order nonlinear dynamics introduces a number of
challenges to account for. This is just partially done here, outlining the possibilities to expand the research in this direction.
It is important to stress how, relying on the communication network rather than
on the agents own dynamics, this methodology is widely applicable across a number of distributed systems, which are not restricted to swarms of mobile agents.
Information networks may get benefits through a quicker diffusion of new data,
power grids can potentially be monitored against cascade failures in a more efficient way or computer virus infections spreading across the internet can be
detected and targeted quicker by providing just some nodes with the required
capabilities.
Chapter 7
Conclusions and Future Work
7.1
A Summary of the Topics Covered
In this Thesis a general introduction to a rather complex field such as swarm
engineering was first provided through Chapter 1. The approach based on considering agents as particles was discussed in as well as this method is widely used
throughout the Thesis and allows for obtaining results of wide applicability. The
first chapter also stated also the targets this thesis aimed and introduced its development.
The mathematical tools used and their application to the specific problem of distributed control were presented in Chapter 2 with a review of the previous works
in the area of swarming featuring the elements common to this thesis presented
in Section 2.7. This was done by connecting the present work to the inspiring
biological fields. The way group behaviours found in nature can be achieved with
the techniques presented in the previous chapters is introduced.
Chapter 3 leads to the first of the results of this Thesis. The problem of fragmentation in particle swarms, which are more in general representative of agent
swarms, was considered by finding the conditions that guarantee the swarm to
be connected. Reversing the popular approach which requires a swarm to be
connected to guarantee its stability, in Chapter 3 conditions that lead to swarm
fragmentation are highlighted and the Cheeger constant is introduced as metric
for the swarm fragility. A technique requiring minimal computational efforts to
have a measure of the swarm fragility was finally presented.
163
CHAPTER 7. CONCLUSIONS AND FUTURE WORK
164
Chapter 3 also introduces the possibility of formation shaping offered by the
reduction of links. These are more extensively exploited in Chapter 4 where
a formation flying concept is designed around this emerging characteristic of
the swarming system. It was shown how, by exploiting asymmetric potential
functions and a particular communication network, the deployment of a fractal
arrangement of agents can emerge out of the swarm. This arrangement finds
application in the electromagnetic field as antenna array, which is provided with
enhanced radiating properties by its peculiar shape.
The asymmetry in the artificial potential functions was then exploited again in a
test campaign aimed to validate the emerging behaviour on real hardware. This
is described in Chapter 5 where, the performance of a group of wheeled robots
are evaluated when controlled through the artificial potential functions aforementioned. This clearly goes in the direction of increasing the level of confidence in
swarming systems towards real world applications.
Lastly, a noteworthy point covered in this Thesis in the area of consensus achievement was discussed in Chapter 6. Manoeuvring performance of a swarm of mobile
agents are analysed considering the amount of resources the swarm can allocate
to tracking an external signal or staying in watch for threats. It was proved
how a uniform distribution of resources is able to provide the swarm with good
performance only in few cases. The results of Chapter 6 indicate a criterium for
selection of the agents mainly deputed to signal following in a swarm depending
on their position in the network. These results, inspired by flocking behaviours of
birds, are here applied to the manoeuvres in swarms of mobile agents. However
the extension of these results into more general area of networked systems was
suggested considering the capital importance of the reaction speed to external
inputs for interconnected systems.
7.2
Conclusions
This dissertation has investigated the problems of modelling and controlling a
swarm of agents in an effective and reliable way departing from the all-to-all
CHAPTER 7. CONCLUSIONS AND FUTURE WORK
165
communication schemes and leveraging the communication network in a swarm
to produce emergent behaviours. This has been done in a rigorous analytic way
with the scope of addressing real problems of distributed control with effective
and reliable tools. In particular this dissertation has provided the following novel
contributions to knowledge.
♦ The problem of fragmentation in large groups of agents with reduced communication capabilities was linked to the formation of a bottleneck in the
communication network. This reflects onto the final spatial arrangement
for swarms controlled through APFs. A fundamental limit to the number
of links each agent has to keep was identified and connected to the flow of a
generic quantity across the network. The decreasing trend of the flow with
the increase in the number of agents highlights how in large engineering
multi agent systems the loss of cohesion can happen also when agents are
supposed to keep an high number of connections.
♦ The possibilities disclosed in the area of formation shaping by the clever
design of the communication network were explored and the application to
space-based telecommunication arrays discussed. An original contribution
to knowledge was provided in the exploitation of a fractal shaped satellite
formation for the deployment of a distributed array. Calculations show how
such a telecommunication architecture surpassed in performance both large
monolithic antennas and distributed, regularly spaced, array lattices.
♦ The artificial potential function method was used to validate an emergent
segregation behaviour and explore the implications of applying APF methods to real robotic agents. Although experimental tests match validations,
these highlight a strong dependance of the performance on the particular
actuators used.
♦ The fast manoeuvring performance in swarms were linked with the capability to cleverly exploit the connection network characteristics and allocate
limited tracking resources across the swarm members. Quick consensus and
consequent fast manoeuvring are achieved without the need for numerical
optimisation.
These four topics cover open problems in swarm engineering and provide original
results for the development of this discipline.
CHAPTER 7. CONCLUSIONS AND FUTURE WORK
7.3
166
Future Work
Multi-agent systems, and swarms in particular, have to overcome a certain resistance to their affirmation which is mainly related to their autonomy. Having
one single autonomous vehicle interacting with the real world means achieving
a very high level of confidence in the autonomous controls. This is penetrating
into the technological paradigm nowadays. Having a swarm of vehicles mutually interacting and then, as a whole, interacting with the real world, in a fully
autonomous way, is probably something that will be hardly seen in the near future. Nevertheless, both the always increasing interests in this technology raised
within the research environments, and the evolution of our society towards an
always more interconnected world suggest this is the direction our technological
paradigm will develop. Hence the need, for the scientific community, to increase
the knowledge about the multi-agent systems getting ready for the time in which
our technological level and social structure will be ready to welcome them. One
of the paths that leads to this increase of technological awareness starts from this
Thesis and develops throughout a number of steps which can be summarised as
follows.
A fundamental step to take consists in the increasing of the TRL of swarming systems by constructing and testing fully autonomous swarms of robots,
UAV, (micro) spacecraft and so on. Initially this has to be done in an isolated environment. Swarm systems are being tested nowadays in laboratories but few of them are ready lo leave this ”simulated real environment”
to then operate in the real world. Before this, swarm systems need to be
tested in the absence of external tracking and far from human presence and
intervention. This can be achieved by employing these systems in tasks
that do not require the presence of humans in the neighbourhood. Autonomous mining, structural inspections, the exploration of remote regions, or
space applications, as the ones suggested in this work, can be a valuable
validation for technologies that, at some point will join our everyday life.
The research about autonomy can progress through considering the possibility of evolving the control laws here presented with time varying potentials.
This should include more than one emergent behaviour and the switching
from one to the other in an autonomous way as well. Here the introduction
CHAPTER 7. CONCLUSIONS AND FUTURE WORK
167
of an asymmetry in the artificial potential was proved to produce a switch
from one configuration to the other. In the future having systems that are
able to switch their configuration depending on the external stimuli, but
in the same predictable and reliable way as it was done here, would significantly increase the level of autonomy. This of course is to be seen in a
context even wider than formation shaping.
A refinement to the theoretical results of this Thesis would be the introduction of noise in the driving signal and the capability of the swarm of filtering
out the disturbance through sharing their understanding of the signal. To
the author’s knowledge this development direction is being pursued in several research institutions and represents another steps towards meeting the
challenges of the real world.
A further step into the refining of the theoretical models presented may
include the problems related with the discontinuities when the network
changes. This would shift the problem into the field of nonsmooth analysis.
The results presented here mainly exploited the heterogeneity of the interactions, that was used to rank and group the agents. Future swarming system
are likely to include agents heterogeneous by their physical characteristics.
Wheeled vehicles may interact with aerial ones and also include computing
nodes with no capability of processing matter but just information as it was
in the early Cellular Automata.
The concept of distributed fractal antenna presented here requires almost
continuous actuation to contrast the orbital dynamics and force the agents
in frozen relative positions. Nevertheless if the emergent behaviour can be
blended with the natural motion given by the orbital dynamics, a more
refined system can be produced. Moreover the array produced can be improved significatively by considering three-dimensional formations. This
would imply a deeper integration of the swarm dynamics with the operational scenario towards a more defined, although less open-ended, applica-
CHAPTER 7. CONCLUSIONS AND FUTURE WORK
168
tion.
Although the number makes swarming systems intrinsically safe from mission critical point of view, their interactions with humans has to be faced at
some point. This will produce a shift from mission-critical to safety-critical
in the design of swarming systems. Hence research has to be addressed in
the direction of human-machine interaction, with the machine being something different from what it is nowadays meant. However, what the machine
should be is not fully defined at present.
The technology that produced self-driving cars looks to be ready for the public in
the next future. Although this is not exactly an example of swarm engineering,
yet it witness its feasibility or, at least, the will to produce the shift to the next
technological paradigm. From fast flying cars visionary foresaw to self driving
vehicles on the doorstep the paradigm has somehow already shifted from “speed”
to “control”. Drawing physical devices out of the paradigm and making them
available to anyone is likely to produce the next social shift.
Bibliography
[1] W. Weaver. Science and complexity. American scientist, (36):536–44, 1948.
[2] A. Ilachinski. Land Warfare and Complexity, Part I: Mathematical Background and Technical Sourcebook. AD-a362 620. Center for naval analyses,
Alexandria, VA, 1996.
[3] I. Couzin. Collective minds. Nature, 445(7129):715, 2007.
[4] J. Von Neumann. The general and logical theory of automata. Cerebral
mechanisms in behavior, pages 1–41, 1951.
[5] M. Gardner. The fantastic combinations of John Conway’s new solitaire
game “life”. Scientific American, 223:120–123, October 1970.
[6] G. Beni. The concept of cellular robotic system. In Intelligent Control,
1988. Proceedings., IEEE International Symposium on, pages 57 –62, aug
1988.
[7] G. Beni. From swarm intelligence to swarm robotics. In Proceedings of
the 2004 international conference on Swarm Robotics, SAB’04, pages 1–9,
Berlin, Heidelberg, 2005. Springer-Verlag.
[8] R. Beckers, S. Goss, J. L. Deneubourg, and J.M. Pasteels. Colony size,
communication and ant foraging strategy. Psyche, (96):239–256, 1989.
[9] T. Fukuda and S. Nakagawa. Approach to the dynamically reconfigurable
robotic system. Journal of Intelligent and Robotic Systems, 1:55–72, 1988.
[10] C. W. Reynolds. Flocks, herds and schools: A distributed behavioral model.
SIGGRAPH Comput. Graph., 21(4):25–34, August 1987.
170
171
BIBLIOGRAPHY
[11] A. F. T. Winfield, C. J. Harper, and J. Nembrini. Towards dependable
swarms and a new discipline of swarm engineering. volume 3342, pages 126
– 142, 2005.
[12] K. A. J. White, J. D. Murray, and M. A. Lewis. Wolf-deer interactions: A
mathematical model. Proceedings of the Royal Society of London. Series B:
Biological Sciences, 263(1368):299–305, 1996.
[13] E. Ben-Jacob. Bacterial self-organization: co-enhancement of complexification and adaptability in a dynamic environment. Philosophical Transactions: Mathematical, Physical and Engineering Sciences, 361(1807):1283–
1312, 2003.
[14] A. Moiseff and J. Copeland. Firefly Synchrony: A Behavioral Strategy to
Minimize Visual Clutter. Science, 329(5988):181+, July 2010.
[15] DARPA. Broad agency announcement system f6, 2010.
[16] O. Brown, P. Eremenko, and B. A. Hamilton. Fractionated spacearchitectures: A vision for responsive space. In Proceedings of the 4th Responsive
Space Conference. AiAA, 2006.
[17] H. Kitano. Robocup rescue: a grand challenge for multi-agent systems. In
MultiAgent Systems, 2000. Proceedings. Fourth International Conference
on, pages 5 –12, 2000.
[18] J. Saez-Pons, L. Alboul, J. Penders, and L. Nomdedeu. Multi-robot team
formation control in the GUARDIANS project. Industrial Robot: An International Journal, 37(4):372–383, 2010.
[19] N. Correll and A. Martinoli. A challenging application in swarm robotics:
The autonomous inspection of complex engineered structures. Bulletin of
the Swiss Society for Automatic Control, pages 15–19, 2007.
[20] Atair
Aerospace.
Atair
-
onyx
parachute
system.
http://www.atair.com/onyx/. Accessed 25 March, 2013.
[21] C. Urmson and W. Whittaker. Self-driving cars and the urban challenge.
Intelligent Systems, IEEE, 23(2):66 –68, march-april 2008.
BIBLIOGRAPHY
172
[22] C. Bergenhem, E. Hedin, and D. Skarin. Vehicle-to-vehicle communication for a platooning system. Procedia - Social and Behavioral Sciences,
48(0):1222 – 1233, 2012.
[23] S. H. Strogatz. Sync: The Emerging Science of Spontaneous Order. Hyperion, 1st edition, 2003.
[24] J Goldstein. Emergence as a Construct: History and Issues. Emergence:
Complexity and Organization, (1):1–9, 1999.
[25] J. Gravner and D. Griffeath. Modeling snow-crystal growth: a threedimensional approach. Physical Review E, 75:1–18, 2009.
[26] J.K. Parrish, S.V. Viscido, and D. Grnbaum. Self-organized fish schools:
an examination of emergent properties. Biology Bulletin, 202(3):206–305,
june 2002.
[27] D. Peak, J. D. West, S. M. Messinger, and K. A. Mott. Evidence for complex, collective dynamics and emergent, distributed computation in plants.
Proceedings of the National Academy of Sciences of the United States of
America, 101(4):918–922, January 2004.
[28] M. R. D’Orsogna, Y. L. Chuang, A. L. Bertozzi, and L. S. Chayes. SelfPropelled Particles with Soft-Core Interactions: Patterns, Stability, and
Collapse. Physical Review Letters, 96(10), 2006.
[29] H. Levine, W. Rappel, and I. Cohen. Self-organization in systems of selfpropelled particles. Phys. Rev. E, 63:017101, Dec 2000.
[30] T. Vicsek, A. Czirók, E. BenJacob, I. Cohen, and O. Shochet. Novel type
of phase transition in a system of self-driven particles. Phys. Rev. Lett.,
75:1226–1229, Aug 1995.
[31] V. Gazi and K.M. Passino. Stability analysis of swarms. Automatic Control,
IEEE Transactions on, 48(4):692 – 697, april 2003.
[32] C. R. McInnes. Vortex formation in swarms of interacting particles. Phys.
Rev. E, 75:032904, Mar 2007.
BIBLIOGRAPHY
173
[33] F. Schweitzer, W. Ebeling, and B. Tilch. Statistical mechanics of canonicaldissipative systems and applications to swarm dynamics. Phys. Rev. E,
64:021110, Jul 2001.
[34] Werner E. Nonequilibrium statistical mechanics of swarms of driven particles. Physica A: Statistical Mechanics and its Applications, 314(14):92 –
96, 2002.
[35] H. Van Dyke Parunak and S. Brueckner. Entropy and self-organization in
multi-agent systems. In Proceedings of the fifth international conference on
Autonomous agents, AGENTS ’01, pages 124–130, New York, NY, USA,
2001. ACM.
[36] W. A. Wright, R. E. Smith, M. Danek, and P. Greenway. A generalisable measure of self-organisation and emergence. In Proceedings of the
International Conference on Artificial Neural Networks, ICANN ’01, pages
857–864, London, UK, 2001. Springer-Verlag.
[37] F. L. Lambert. Shuffled cards, messy desks, and disorderly dorm rooms examples of entropy increase? nonsense! Journal of Chemical Education,
76(10):1385, 1999.
[38] D. J. Bennet. Pattern Formation in Swarming Systems using Bifurcating
Potential Fields. PhD thesis, 2010.
[39] W. Ren and R.W. Beard. Consensus seeking in multiagent systems under dynamically changing interaction topologies. Automatic Control, IEEE
Transactions on, 50(5):655 – 661, may 2005.
[40] R. P. Wiegand, M. A. Potter, D. A. Sofge, and W. M. Spears. A generalized graph-based method for engineering swarm solutions to multiagent
problems. In Proceedings of the 9th international conference on Parallel
Problem Solving from Nature, PPSN’06, pages 741–750, Berlin, Heidelberg,
2006. Springer-Verlag.
[41] A. Jadbabaie, J. Lin, and A.S. Morse. Coordination of groups of mobile
autonomous agents using nearest neighbor rules. Automatic Control, IEEE
Transactions on, 48(6):988 – 1001, june 2003.
174
BIBLIOGRAPHY
[42] L. Moreau. Stability of multiagent systems with time-dependent communication links. Automatic Control, IEEE Transactions on, 50(2):169 – 182,
feb. 2005.
[43] W. Ebeling and U. Erdmann.
Nonequilibrium statistical mechanics of
swarms of driven particles. 8(4):23–30, 2003.
[44] A. Martinoli, A.J. Ijspeert, and F. Mondada. Understanding collective
aggregation mechanisms: From probabilistic modelling to experiments with
real robots. Robotics and Autonomous Systems, 29(1):51 – 63, 1999.
[45] E. Sahin, T. H. Labella, V. Trianni, J. Deneubourg, P. Rasse, D. Floreano,
L. Gambardella, F. Mondada, S. Nolfi, and M. Dorigo. Swarm-bot: Pattern
formation in a swarm of self-assembling mobile robots. volume 4, pages 145
– 150, Yasmine Hammamet, Tunisia, 2002.
[46] K. Lerman, A. Martinoli, and A. Galstyan. A review of probabilistic macroscopic models for swarm robotic systems. volume 3342, pages 143 – 152,
Santa Monica, CA, United states, 2005.
[47] C. R. McInnes. Autonomous ring formation for a planar constellation of
satellites. Journal of Guidance, Control, and Dynamics, 18(5):1215 – 1217,
1995.
[48] D. J. Bennet and C. R. McInnes. Verifiable control of a swarm of unmanned
aerial vehicles. Proceedings of the Institution of Mechanical Engineers, Part
G: Journal of Aerospace Engineering, 223(7):939–953, 2009.
[49] D. J. Bennet and C.R. McInnes. Spacecraft formation flying using bifurcating potential fields. In Proceeding of the 59th International Astronautical
Congress, Glasgow, Scotland, UK, 2008.
[50] D. J. Bennet and C. R. McInnes. Pattern transition in spacecraft formation
flying via the artificial potential field method and bifurcation theory. In 3rd
International Symposium on Formation Flying, Missions and Technologies,
European Space Agency, (Special Publication) ESA SP, ESA-ESTEC, The
Netherlands, April 2008. European Space Agency.
BIBLIOGRAPHY
175
[51] M. H. Mabrouk and C. R. McInnes. Nonlinear stability of vortex formation
in swarms of interacting particles. Physical Review E (Statistical, Nonlinear,
and Soft Matter Physics), 78(1), 2008.
[52] M.H. Mabrouk and C.R. McInnes. Wall following to escape local minima
for swarms of agents using internal states and emergent behaviour. London,
England, UK, 2008.
[53] M. M. Mabrouk, C. W. Murray, K. Johnstone, and C.R. McInnes. Internal
agent states: experiments using the swarm leader concept.
[54] A. Badawy and C. McInnes. On-orbit assembly using superquadric potential fields. Journal of Guidance, Control and Dynamics, 31(1):30–43,
2008.
[55] A. Badawy and C.R. McInnes. Small spacecraft formation-flying using
potential functions. Acta Astronautica, 65(11-12):1783–1788, 2009.
[56] D.F. Gordon, W.M. Spears, O. Sokolsky, and L. Insup. Distributed spatial
control, global monitoring and steering of mobile agents. In Information
Intelligence and Systems, 1999. Proceedings. 1999 International Conference
on, pages 681 –688, 1999.
[57] W.M. Spears, R. Heil, D.F. Spears, and D. Zarzhitsky. Physicomimetics
for mobile robot formations. In Autonomous Agents and Multiagent Systems, 2004. AAMAS 2004. Proceedings of the Third International Joint
Conference on, pages 1528 –1529, july 2004.
[58] T. R. Smith, H. Hanmann, and N. E. Leonard. Orientation control of multiple underwater vehicles with symmetry-breaking potentials. In 40th IEEE
Conference on Decision and Control (CDC), December 4, 2001 - December
7, 2001, volume 5 of Proceedings of the IEEE Conference on Decision and
Control, pages 4598–4603. Institute of Electrical and Electronics Engineers
Inc., 2001.
[59] N. E. Leonard and E. Fiorelli. Virtual leaders, artificial potentials and
coordinated control of groups. In Decision and Control, 2001. Proceedings
of the 40th IEEE Conference on, volume 3, pages 2968–2973 vol.3.
BIBLIOGRAPHY
176
[60] R. Sepulchre, D. A. Paley, and N. E. Leonard. Stabilization of planar collective motion: All-to-all communication. IEEE Transactions on Automatic
Control, 52(Compendex):811–824, 2007.
[61] R. Sepulchre, D. A. Paley, and N. E. Leonard. Stabilization of planar collective motion with limited communication. IEEE Transactions on Automatic
Control, 53(Compendex):706–719, 2008.
[62] S. Nair and N. E. Leonard. Stable synchronization of rigid body networks.
Networks and heterogeneous Media, 2(4):597–626, 2007.
[63] E. Fiorelli, N. E. Leonard, P. Bhatta, D. Paley, R. Bachmayer, and D. M.
Fratantoni. Multi-auv control and adaptive sampling in monterey bay. 2004
IEEE/OES Autonomous Underwater Vehicles, pages 134 – 147, 2004.
[64] N. E. Leonard, D. A. Paley, F. Lekien, R. Sepulchre, D. M. Fratantoni,
and R. E. Davis. Collective motion, sensor networks, and ocean sampling.
Proceedings of the IEEE, 95(Compendex):48–74, 2007.
[65] N. E. Leonard, D. A. Paley, R. E. Davis, D. M. Fratantoni, F. Lekien, and
F. Zhang. Coordinated control of an underwater glider fleet in an adaptive
ocean sampling field experiment in monterey bay. Journal of Field Robotics,
27(6):718–740, 2010.
[66] L. Euler. Solutio problematis ad geometriam situs pertinentis. Commentarii
academiae scientiarum Petropolitanae, 8:128–140, 1741.
[67] N. Biggs, E. Lloyd, and R. Wilson. Graph Theory, 1736-1936. Oxford
University Press, 1986.
[68] P. Erds and A. Rnyi. On random graphs, i. Publicationes Mathematicae
(Debrecen), 6:290–297, 1959.
[69] P. Erdos and A. Renyi. On the evolution of random graphs. Publ. Math.
Inst. Hung. Acad. Sci, 5:17–61, 1960.
[70] P. Erdos and A. Rnyi. On the strength of connectedness of a random graph.
Acta Mathematica Hungarica, 12(1):261–267, 1961.
[71] A. Barabasi and R. Albert. Emergence of scaling in random networks.
Science, 286(5439):509–512, 1999.
BIBLIOGRAPHY
177
[72] D. J. Watts and S. H. Strogatz. Collective dynamics of ‘small-world’ networks. Nature, 393(6684):440–442, 1998.
[73] N. E. Leonard, G. Young, K. Hochgraf, D. Swain, A. Trippe, W. Chen, and
S. Marshall. In the dance studio: Analysis of human flocking. In American
Control Conference (ACC), 2012, pages 4333 –4338, june 2012.
[74] H. Yamaguchi and T. Arai. Distributed and autonomous control method for
generating shape of multiple mobile robot group. In Intelligent Robots and
Systems ’94. ’Advanced Robotic Systems and the Real World’, IROS ’94.
Proceedings of the IEEE/RSJ/GI International Conference on, volume 2,
pages 800 –807 vol.2, sep 1994.
[75] A. Rahmani, M. Ji, M. Mesbahi, and M. Egerstedt. Controllability of multiagent systems from a graph-theoretic perspective. SIAM J. Control Optim.,
48(1):162–186, February 2009.
[76] R. Olfati-Saber, J.A. Fax, and R.M. Murray. Consensus and cooperation in
networked multi-agent systems. Proceedings of the IEEE, 95(1):215 –233,
jan. 2007.
[77] M.A. Lewis and G.A. Bekey. The behavioral self-organization of nanorobots
using local rules. In Intelligent Robots and Systems, 1992., Proceedings of
the 1992 lEEE/RSJ International Conference on, volume 2, pages 1333
–1338, jul 1992.
[78] T. Balch and R.C. Arkin. Behavior-based formation control for multirobot
teams. Robotics and Automation, IEEE Transactions on, 14(6):926 –939,
dec 1998.
[79] L. Blazovics, K. Csorba, B. Forstner, and C. Hassan. Target tracking and
surrounding with swarm robots. In Engineering of Computer Based Systems (ECBS), 2012 IEEE 19th International Conference and Workshops
on, pages 135 –141, april 2012.
[80] L. Bayindir and E. Sahin. A review of studies in swarm robotics. Turkish
Journal of Electrical Engineering and Computer Sciences, 15(2):115 – 147,
2007.
BIBLIOGRAPHY
178
[81] Y. Mohan and S.G. Ponnambalam. An extensive review of research in
swarm robotics. pages 140 – 145, Coimbatore, India, 2009.
[82] E. Bahceci, Soysal O., and E. Sahin. A review: Pattern formation and
adaptation in multi-robot systems. Technical report, Robotics Institute,
Carnegie Mellon University, Pittsburgh, PA, Oct 2003.
[83] Z. Bubnicki. Modern Control Theory. Springer, Berlin Heidelberg, 2006.
[84] D. P. Atherton. Stability of Nonlinear Systems. Research Study Press - John
Wiley & Sons Ltd, 8 Willian Way, Letchworth, Herts SG6 2HG, England,
1981.
[85] V Lakshmikantham, S. Leela, and A. A. Martynyuk. Stability Analysis of
Nonlinear Systems. Marcel Dekker, Inc., 270 Madison Avenue, New York,
New York 10016, 1989.
[86] S. Sastry. Nonlinear Systems: Analysis, Stability and Control. SpringerVerlag, New York Berlin Heldelberg, 1999.
[87] A. M. Lyapunov. The general problem of the stability of motion. International Journal of Control, 55:531–534, 1992.
[88] B. Carre. Graphs and networks. Clarendon Press, 1979.
[89] R. Gould. Graph theory. Benjamin/Cummings Pub. Co., Menlo Park,
Calif., 1988.
[90] J. A. Bondy and U. S. Murty. Graph Theory. Graduate texts in Mathematics, 244. Springer, New York, 2008.
[91] N. Biggs. Algebraic Graph Theory. Cambridge University Press, 2nd edition,
1993.
[92] F. H. Clarke. Nonsmooth analysis and control theory. Graduate texts in
mathematics 178. Springer, New York, 1998.
[93] A. Filippov. Differential Equations with Discontinuous Right-Hand Side.
American Mathematica Society Translations, (2):191–231, 1964.
[94] A. Filippov. Differential Equations with Discontinuous Right-Hand Side.
Kluver, Boston, MA, 1988.
BIBLIOGRAPHY
179
[95] D. Shevitz and B. Paden. Lyapunov stability theory of nonsmooth systems.
In Proceedings of the 1993 ASME Winter Annual Meeting, November 28,
1993 - December 3, 1993, volume 53 of American Society of Mechanical Engineers, Dynamic Systems and Control Division (Publication) DSC, pages
53–60. Publ by ASME.
[96] H. G. Tanner, A. Jadbabaie, and G. J. Pappas. Flocking in fixed and
switching networks. Automatic Control, IEEE Transactions on, 52(5):863–
868, 2007.
[97] H. G. Tanner, A. Jadbabaie, and G. J. Pappas. Stable flocking of mobile
agents - part ii: Dynamic topology. 42nd Ieee Conference on Decision and
Control, Vols 1-6, Proceedings, pages 2016–2021 6583, 2003.
[98] P. M. Morse. Diatomic molecules according to the wave mechanics. ii.
vibrational levels. Phys. Rev., 34(1):57–64, 07 1929.
[99] J. E. Jones. On the Determination of Molecular Fields. II. From the Equation of State of a Gas. Royal Society of London Proceedings Series A,
106:463–477, October 1924.
[100] B. Nabet and N. E. Leonard. Shape control of a multi-agent system using
tensegrity structures. In 3rd IFAC Workshop on Lagrangian and Hamiltonian Methods in Nonlinear Control, LHMNLC 2006, July 19, 2006 - July
21, 2006, volume 366 LNCIS of Lecture Notes in Control and Information
Sciences, pages 329–339. Springer Verlag.
[101] X. F. Wang and G. Chen. Pinning control of scale-free dynamical networks.
Physica A: Statistical Mechanics and its Applications, 310(3-4):521–531,
2002.
[102] K.A. Hawick and H.A. James. Small-world effects in wireless agent sensor networks. International Journal of Wireless and Mobile Computing,
4(3):155 – 164, 2010.
[103] R. Olfati-Saber. Ultrafast consensus in small-world networks. volume 4,
pages 2371 – 2378, Portland, OR, United States, 2005.
[104] F. Yu, D. Lv, and X. Liu. Optimal distributed consensus in small world
network. In 25th Chinese Control Conference, CCC 2006, August 7, 2006 -
BIBLIOGRAPHY
180
August 11, 2006, 2006 Chinese Control Conference Proceedings, CCC 2006,
pages 333–336. Inst. of Elec. and Elec. Eng. Computer Society.
[105] D. Pais, M. Cao, and N. E. Leonard. Formation shape and orientation
control using projected collinear tensegrity structures. In 2009 American
Control Conference, ACC 2009, June 10, 2009 - June 12, 2009, Proceedings
of the American Control Conference, pages 610–615. Institute of Electrical
and Electronics Engineers Inc.
[106] M. Ballerini, N. Cabibbo, R. Candelier, A. Cavagna, E. Cisbani, I. Giardina, V. Lecomte, A. Orlandi, G. Parisi, A. Procaccini, M. Viale, and
V. Zdravkovic. Interaction ruling animal collective behavior depends on
topological rather than metric distance: Evidence from a field study. Proceedings of the National Academy of Sciences, 105(4):1232–1237, 2008.
[107] J. K. Parrish, S. V. Viscido, and D. Grnbaum. Self-organized fish schools:
An examination of emergent properties. Biol. Bull, 202:296–305, 2002.
[108] A. Cavagna, I. Giardina, A. Orlandi, G. Parisi, and A. Procaccini. The
starflag handbook on collective animal behaviour: 2. three-dimensional
analysis. Animal Behaviour, 76(1):237–248, 2008.
[109] A. Cavagna, I. Giardina, A. Orlandi, G. Parisi, A. Procaccini, M. Viale,
and V. Zdravkovic. The starflag handbook on collective animal behaviour:
1. empirical methods. Animal Behaviour, 76(1):217–236, 2008.
[110] A. Cavagna, A. Cimarelli, I. Giardina, G. Parisi, R. Santagati, F. Stefanini,
and M. Viale. Scale-free correlations in starling flocks. Proceedings of the
National Academy of Sciences, 107(26):11865–11870, 2010.
[111] A. Procaccini, A. Orlandi, A. Cavagna, I. Giardina, F. Zoratto, D. Santucci,
F. Chiarotti, C. K. Hemelrijk, E. Alleva, G. Parisi, and C. Carere. Propagating waves in starling, sturnus vulgaris, flocks under predation. Animal
Behaviour, 82(4):759–765, 2011.
[112] J. E. Treherne and W. A. Foster. Group transmission of predator avoidance
behaviour in a marine insect: The trafalgar effect. Animal Behaviour,
29(3):911–917, 1981.
BIBLIOGRAPHY
181
[113] W. K. Potts. The chorus-line hypothesis of manoeuvre coordination in
avian flocks. Nature, 309(5966):344–345, 1984.
[114] B. L. Partridge. The structure and function of fish schools. Sci. Amer.,
246(6):114–123, 1982.
[115] T. J. Pitcher. Some ecological consequences of fish school volumes. Freshwater Biology, 10(6):539–544, 1980.
[116] TJ Pitcher, BL Partridge, and CS Wardle. A blind fish can school. Science,
194(4268):963–965, 1976.
[117] J. R. Hunter. Communication of velocity changes in jack mackerel (trachurus symmetricus) schools. Animal Behaviour, 17(Part 3):507–514, 1969.
[118] J. K. Parrish. Re-examining the selfish herd: are central fish safer? Animal
Behaviour, 38(6):1048–1053, 1989.
[119] A. E. Magurran and A. Higham. Information transfer across fish shoals
under predator threat. Ethology, 78(2):153–158, 1988.
[120] P. K. Visscher and S. Camazine. Collective decisions and cognition in bees.
Nature, 397(6718):400–400, 1999. 10.1038/17047.
[121] M. R. Myerscough. Dancing for a decision: a matrix model for nest-site
choice by honeybees. Proc Biol Sci, 270(1515):577–82, 2003. Myerscough,
Mary R England Proceedings. Biological sciences / The Royal Society Proc
Biol Sci. 2003 Mar 22;270(1515):577-82.
[122] P. Lucic and D. Teodorovic. Transportation modeling: an artificial life
approach. In Tools with Artificial Intelligence, 2002. (ICTAI 2002). Proceedings. 14th IEEE International Conference on, pages 216–223, 2002.
[123] F. Cucker and S. Smale. Emergent behavior in flocks. Automatic Control,
IEEE Transactions on, 52(5):852–862, 2007.
[124] I. D. Couzin, J. Krause, N. R. Franks, and S. A. Levin. Effective leadership
and decision-making in animal groups on the move. Nature, 433(7025):513–
516, 2005. 10.1038/nature03236.
BIBLIOGRAPHY
182
[125] I. D. Couzin. Collective cognition in animal groups. Trends in Cognitive
Sciences, 13(1):36–43, 2009.
[126] N. O. Handegard, K. M. Boswell, C. C. Ioannou, S. P. Leblanc, D. B.
Tjstheim, and I. D. Couzin. The dynamics of coordinated group hunting
and collective information transfer among schooling prey. Current biology
: CB, 22(13):1213–1217, 2012.
[127] B. Nabet, N. E. Leonard, I. D. Couzin, and S. A. Levin. Dynamics of
decision making in animal group motion. Journal of Nonlinear Science,
19(Compendex):399–435, 2009.
[128] L. Zhiyun, B. Francis, and M. Maggiore. Necessary and sufficient graphical
conditions for formation control of unicycles. Automatic Control, IEEE
Transactions on, 50(1):121–127, 2005.
[129] V. Gazi. Stability of an asynchronous swarm with time-dependent communication links. IEEE Transactions on Systems, Man, and Cybernetics, Part
B: Cybernetics, 38(Compendex):267–274, 2008.
[130] R. Olfati-Saber and R. M. Murray. Consensus problems in networks of
agents with switching topology and time-delays. Automatic Control, IEEE
Transactions on, 49(9):1520–1533, 2004.
[131] G. Punzo, J. Simo, D. J. Bennet, and M. Macdonald. Characteristics of
swarms on the edge of fragmentation. under review in Physical Review E.
[132] G. Punzo, D. J. Bennet, and M. Macdonald. Swarm shape manipulation through connection control. In 12th Conference Towards Autonomous
Robotic Systems 2010, August 2010.
[133] E. Stump, A. Jadbabaie, and V. Kumar. Connectivity management in
mobile robot teams. In Robotics and Automation, 2008. ICRA 2008. IEEE
International Conference on, pages 1525–1530, May 2008.
[134] M.M. Zavlanos and G.J. Pappas. Controlling connectivity of dynamic
graphs. In Decision and Control, 2005 and 2005 European Control Conference. CDC-ECC ’05. 44th IEEE Conference on, pages 6388–6393, December 2005.
BIBLIOGRAPHY
183
[135] F. Chung. Spectral Graph Theory (CBMS Regional Conference Series in
Mathematics, No. 92). American Mathematical Society, Providence, RI,
02940-6248, USA, 1997.
[136] F. Chung. Laplacians and the cheeger inequality for directed graphs. Annals
of Combinatorics, 9(1):1–19, 2005.
[137] G. Punzo, P. Karagiannakis, D. J. Bennet, M. Macdonald, and S. Weiss.
Enabling and exploiting self-similar central symmetry formations. IEEE
Transaction on Aerospace and Electronic Systems, 2013 in press.
[138] G. Punzo, D. J. Bennet, and M. Macdonald. A fractally fractionated spacecraft. In 62nd International Astronautical Congress 2011, pages Article
IAC–11–D1.4.11, October 2011.
[139] G. Punzo. Fractal patterns in fractionated spacecraft. In 62nd International
Astronautical Congress 2011, pages Article IAC–11–E2.2.6, October 2011.
[140] G. Punzo, D. J. Bennet, and M. Macdonald. Enhancing self-similar patterns by asymmetric artificial potential functions in partially connected
swarms. In Towards Autonomous Robotic Systems, volume 6856 of Lecture
Notes in Artificial Intelligence. Institution of Engineering and Technology,
September 2011.
[141] J. Leitner. Formation flying: The future of remote sensing from space.
Technical report, NASA, Goddard Space Flight Center, MD, United States,
2004.
[142] E. A. Barnes and B. E. Bishop. Utilizing navigational capability to control
cooperating robotic swarms in reconnaissance-based operations. Proceedings of the 40th Southeastern Symposium on System Theory, pages 338–342,
March 2008.
[143] S Schwartz. Cross-scale - multi-scale coupling in space plasmas. Technical
report, ESA, December 2009.
[144] M. Bester, M. Lewis, B. Roberts, J. McDonald, D. Pease, J. Thorsness,
S. Frey, D. Cosgrove, and D. Rummel. Themis operations. Space Science
Reviews, 141(1-4):91–115, December 2008.
BIBLIOGRAPHY
184
[145] P. R. Lawson, O. P. Lay, S. R. Martin, C. A. Beichman, K. J. Johnston,
W. C. Danchi, R. O. Gappinger, S. L. Hunyadi, A. Ksendzov, B. Mennesson, R. D. Peters, D. P. Scharf, E. Serabyn, and S. C. Unwin. Terrestrial planet finder interferometer 2006-2007 progress and plans - art. no.
669308. Techniques and Instumentation for Detection of Exoplanets III,
6693:69308–69308, September 2007.
[146] K. Danzmann. LISA - an ESA cornerstone mission for the detection and
observation of gravitational waves. Advances in Space Research, 32(7):1233
– 1242, 2003.
[147] F. Sjoberg, A. Karlsson, and B. Jakobsson. PROBA-3: A formation flying
demonstration mission on the verge to be realised. Number 654 SP, April
2008.
[148] G. Blackwood, O. Lay, A. Deininger, W.and Ahmed, R. Duren, C. Noecker,
and B. Barden. The starlight mission: A formation-flying stellar interferometer. Proceedings of SPIE - The International Society for Optical Engineering, 4852(2):463 – 480, February 2002.
[149] C. R. McInnes. Velocity field path-planning for single and multiple unmanned aerial vehicles. Aeronautical Journal, 107(1073):419–426, July
2003.
[150] M. J. Bentum, C. J. M. Verhoeven, and A. J. Boonstra. Olfar - orbiting
low frequency antennas for radio astronomy. In Proceedings of the ProRISC 2009, Annual Workshop on Circuits, Systems and Signal Processing,
Veldhoven, pages 1–6, November 2009.
[151] C. Balanis. Antenna theory: analysis and design. New York: Wiley, 1997.
[152] D. H. Johnson. Array signal processing:concepts and techniques. Englewood
Cliffs, NJ : P T R Prentice Hall, 1993.
[153] D. H. Werner and R. Mittra. The theory and design of fractal antenna
arrays. In Frontiers in Electromagnetics, chapter 3, pages 94–203. WileyIEEE Press, 1 edition, 2000.
BIBLIOGRAPHY
185
[154] C. Puente-Baliarda and R. Pous. Fractal design of multiband and low sidelobe arrays. Antennas and Propagation, IEEE Transactions on, 44(5):730–
739, May 1996.
[155] A. Maffett. Array factors with nonuniform spacing parameter. Antennas
and Propagation, IRE Transactions on, 10(2):131 –136, March 1962.
[156] H. Schaub and J.L. Junkins. Analytical mechanics of space systems. American Institute of Aeronautics and Astronautics, Reston, Va., 2003.
[157] L.B. King, G.G. Parker, S. Deshmukh, and J. Chong. Study of interspacecraft coulomb forces and implications for formation flying. Journal of
Propulsion and Power, 19(3):497 – 505, May/June 2003.
[158] M. A. Peck, B. Streetman, Saaj C. M., and V. Lappas. Spacececraft formation flying using lorentz forces. Journal of the British Interplanetary
Society, 60:263–267, 2007.
[159] G. Punzo, G. Dobie, D. J. Bennet, J. Jamieson, and M. Macdonald. Lowcost, multi-agent systems for planetary surface exploration. In 63rd International Astronautical Congress, October 2012.
[160] S. Pierce, G. Punzo, G. Dobie, R. Summan, C. N. MacLeod, C. McInnes,
J. Biggs, M. Macdonald, and D. J. Bennet. Reconfigurable robotic platforms
for structural health monitoring. In 6th European Workshop on Structural
Health Monitoring & 1st European Conference of the Prognostics and Health
Management (PHM) Society, July 2012.
[161] D. Izzo and L. Pettazzi. Autonomous and distributed motion planning for
satellite swarm. Journal of Guidance, Control, and Dynamics, 30(2):449–
459, March/April 2007.
[162] Vicon t160 website, http://www.vicon.com/products/t160.html.
[163] G.I. Dobie. Ultrasonic Sensor Platforms for Non-Destructive Testing. PhD
thesis, University of Strathclyde, 2010.
[164] S.G. Pierce, k. Worden, R. Summan, G. Dobie, and J.J. Hensman. Towards
implementation of reconfigurable robotic strategies for structural health
monitoring. In 5th European Workshop on Structural Health Monitoring
EWSHM, pages 610 –615, Sorrento, Italy, July 2010.
BIBLIOGRAPHY
186
[165] Y. Liu, J. Slotine, and A. Barabasi. Controllability of complex networks.
Nature, 473(7346):167–173, 2011. 10.1038/nature10011.
[166] S. Nag and L. Summerer. Behaviour based, autonomous and distributed
scatter manoeuvres for satellite swarms. Acta Astronautica, 82(1):95–109,
2013.
[167] L. C. Freeman. A Set of Measures of Centrality Based on Betweenness.
Sociometry, 40(1):35–41, March 1977.
[168] I. Poulakakis, L. Scardovi, and N.E. Leonard. Node certainty in collective
decision making. In Decision and Control (CDC), 2012 IEEE 51st Annual
Conference on, pages 4648–4653, 2012.
[169] R. S. Varga. Matrix Iterative Analysis. Springer-Verlag, Berlin, 2009.
[170] D. Serre. Matrices: theory and applications. Graduate texts in mathematics;
v. 216. Springer, 2010.
[171] Z. Qu, C. Li, and F. Lewis. Cooperative control based on distributed estimation of network connectivity. In American Control Conference (ACC),
2011, pages 3441–3446.
[172] K. Klemm, M. Serrano, V. M. Eguı́luz, and M. S. Miguel. A measure of
individual role in collective dynamics. Scientific Reports, 2, February 2012.
[173] O. Taussky. A recurring theorem on determinants. The American Mathematical Monthly, 56(10):pp. 672–676, 1949.
[174] L. Han, M. Neumann, and M. Tsatsomeros. Spectral radii of fixed frobenius
norm perturbations of nonnegative matrices. SIAM J. Matrix Anal. Appl.,
21(1):79–92, October 1999.
[175] G. W. Stewart. Introduction to matrix computations [by] G. W. Stewart.
Academic Press New York, 1973.
[176] R. P. Agaev and P. Yu. Chebotarev. On determining the eigenprojection
and components of a matrix. Autom. Remote Control, 63(10):1537–1545,
October 2002.
BIBLIOGRAPHY
187
[177] C. R. Johnson, D. P. Stanford, D. D. Olesky, and P. van den Driessche.
Dominant eigenvalues under trace-preserving diagonal perturbations. Linear Algebra and Its Applications, 212-213:415–435, 1994.
[178] F.R. Gantmacher. Matrix theory, Volume 1. Chelsea Publishing Company,
New York, NY, 1960.
[179] Z. Qu, J. Wang, and R. A. Hull. Cooperative Control of Dynamical Systems With Application to Autonomous Vehicles. Automatic Control, IEEE
Transactions on, 53(4):894–911, 2008.