by an angle when it tails another drone.
In the literature there are two design methods to
develop collective behaviours in swarm systems:
behaviour-based design and automatic design
(Brambilla, 2013). The former implies the
developers to implement, study, and improve the
behaviour of each single individual until the desired
collective behaviour is achieved. This is the
approach adopted in (Cimino, 2015b). The latter is
usually used to reduce the effort of the developers.
Automatic design methods can be furtherly divided
in two categories: reinforcement learning and
evolutionary robotics (Brambilla, 2013). The first
implies a definition at the individual level of positive
and repulsive reinforce to give reward to the
individual. In general, it is usually hard for the
developer to decompose the collective output of the
swarm in individual rewards. Evolutionary robotics
implies evolutionary techniques inspired by the
Darwinian principle of selection and evolution.
Generally in these methods each swarm consists of
individuals with the same behaviour. A population
of swarms is then computed, where each population
member has a particular behaviour. A simulation is
made for each member and a fitness function is
computed. Then through a mutation and crossover
procedure a new generation is computed. This
process iteratively repeats improving the
performance of the swarm population.
2 RELATED WORK
DE has been used in several domains for
optimization and parameterization tasks (Das, 2011).
As an example, in (Nikolos, 2005) the authors used a
classical DE variant, namely DE/1/rand/bin, to
coordinate multiple drones navigating from a known
initial position to a predetermined target location.
Here, DE is set up with N=50, F=1.05 and CR=0.85.
The algorithm was defined to terminate in 200
generations, but it usually converges in 30 iterations.
Our problem sensibly differs, because the target
position is unknown, and our approach is
independent of the initial position.
In (Chakraborty, 2008) the authors confront DE
and Particle Swarm Optimization (PSO) for co-
operative distributed multi-robot path planning
problem. As for (Nikolos and Brintaki, 2005) initial
position of the robots and final position are known.
Here, both centralized and decentralized
formulations are proposed. In the centralized
approach, DE minimizes the distance for the next
step of each robot. In this case all information of the
position of each robot, the next position, and the
possible collision are provided to DE. In the
decentralized formulation, each robot runs DE for
itself considering the information of neighbour
robots. Authors conclude that the decentralized
approach needs less time in comparison to the
centralized one; moreover the performance is
comparable to PSO. In our approach, we consider to
use DE offline to find a proper and general purpose
parameter tuning for the swarms. Moreover, in our
formulation drones have a limited computing
capability, and then an online execution of DE is not
feasible.
In (Cruz-Alvarez, 2013) DE/1/rand/bin is used
with F=0.7, CR=1.0, N=120 for 250 generations, and
another variant called DE/1/best/bin is used with
F=0.8, CR=1.0, N=150 for 200 generations to tune
the behaviour of a robot in wall-following task. Here
it seems that DE/1/best/bin is able to find a slightly
better solution than DE/1/rand/bin. However,
authors used different parameters settings (F, N and
number of generations) for each variant, thus a
comparative analysis is difficult. In our approach we
focus on DE/1/rand/bin variant and evaluate several
combinations of CR and F.
3 SWARM BEHAVIORAL MODEL
In this section we improve the swarm algorithm of
(Cimino, 2015b). We refer to the time unit as a tick,
i.e., an update cycle of both the environment and the
drones. Each drone is equipped with: (a) wireless
communication device for sending and receiving
information from a ground station; (b) self-location
capability, e.g. based on global position system
(GPS); (c) a sensor to detect a target in proximity of
the drone; (d) processor with limited computing
capability; (e) a sensor to detect obstacles.
The environment and the pheromone dynamics
We consider a predefined area that contains a set
of targets to be identified. The environment is
modelled by a digital grid corresponding to the
physical area. The grid has C
2
cells, each identified
by (x, y) coordinates with x, y ∈ {1,…,C}. The
actual size of the area and the granulation of the grid
depend on the domain application. Figure 1 shows
Pheromone dynamics in an urban scenario. Here, the
intensity of the pheromone is represented as a dark
colour, and each target is represented by an “X”. A
darker gradation means higher pheromone intensity.
At the beginning, the pheromone is in one cell at its
maximum intensity, and then it diffuses to