2009; Piotrowski et al., 2009), there is still a number
of challenges to tackle. First, to the best of our knowl-
edge, most of previous studies do not encompass both
the distribution and the replication aspects. Second,
the proposed solutions are usually not tested experi-
mentally at large-scale, while we stress how real de-
ployment is of primary importance when developing
WSN applications.
In this paper, we propose a low complexity dis-
tributed data replication mechanism to increase the
resilience and storage capacity of a WSN-based ob-
servation system against node failure and local mem-
ory shortage. We evaluate our approach through ex-
perimental tests conducted on the SensLab real plat-
form (SensLAB, 2008). We show how careful config-
uration of key system parameters has a positive effect
on the performance. The portfolio of parameters in-
cludes: (i) the node transmit power, which should be
reduced to save energy; (ii) the communication over-
head, i.e., the amount of control messages exchanged
by the nodes; (iii) the number of deployed nodes; and
(iv) the data redundancy, i.e., the number of copies to
be stored per sensed data unit.
This paper is organized as follows. Section 2 is
dedicated to related works. In Section 3, we moti-
vate and detail the design of our greedy mechanism
for distributed data replication. Section 4 is devoted to
the large scale experiments performed on the SensLab
testbed, by considering the aforementioned scenarios.
At last, Section 5 concludes the paper.
2 RELATED WORK
Various schemes to efficiently store and process
sensed data in WSNs have been proposed in the past
years (Hongping and Kangling, 2010).
In distributed WSN storage schemes, nodes coop-
erate to efficiently distribute data across them. Pre-
vious studies focused on data centric storage (Rat-
nasamy et al., 2003; Madden et al., 2005; Awad et al.,
2009). According to a centric storage approach, data
collected in some WSN regions are stored at super
nodes that are responsible for these regions and/or ac-
cording to the type of data. Hash values are used to
distinguish data types and the corresponding storage
locations. Even if this approach is based on node co-
operation, it is not fully distributed, since super nodes
store all the contents generated by the others.
In a fully distributed data storage approach, all
nodes participate in sensing and storing in the same
way. All nodes, first, store their sensor readings lo-
cally and, once their local memories have filled up,
delegate other nodes to store them. A first signifi-
cant contribution in this direction is given by Data
Farms (Omotayo et al., 2007). The authors propose
a fully data distributed storage mechanism with pe-
riodical data retrieval. They derive a cost model to
measure energy consumption and show how a care-
ful selection of nodes offering storage, called “donor
nodes”, optimizes the system capacity at the price of
slightly higher transmission costs. They assume the
network is built in a tree topology and each sensor
node knows the return path to a sink node, who pe-
riodically retrieves data. Another study is proposed
in EnviroStore (Luo et al., 2007). The authors focus
on data redistribution when the remaining storage of
a sensor node exceeds a given threshold, by means
of load balancing. They use a proactive mechanism,
where each node maintains locally a memory table
containing the status of the memory of its neighbors.
Furthermore, mobile nodes (called
mules
) are used to
carry data from an overloadedarea to an offloaded one
and to take data to a sink node. The major lack of the
abovestudies pertains to the absence of large scale ex-
periments to evaluate the system performance, which
is a key contribution of our work.
Data replication consists in adding redundancy to
the system by copying data at several donor nodes
(within the WSN) to mitigate the risk of node fail-
ure. It has been widely studied for WSNs and several
works are available in the literature (Neumann et al.,
2009; Hamed Azimi et al., 2010). A scoring func-
tion for suitably choosing a replicator node is pro-
posed in (Neumann et al., 2009). The function is in-
fluenced by critical parameters such as the number of
desired replicas, the remaining energy of a replicator
node and the energy of the neighbors of the replicator
node. In TinyDSM (Piotrowski et al., 2009), a reac-
tive replication approach is discussed. Data burst are
broadcasted by a source node and then handled inde-
pendently by each neighbor, which decides to store a
copy of the burst or not. Unlike the above approaches,
we propose a replication-based distributed data stor-
age mechanism with lower complexity, since the re-
sponsibility of finding donor nodes is not centralized
at the source node but handed over to consecutive
donor nodes in a recursive manner.
The main contributions of this paper are the fol-
lowing. First, we encompass, with a fully distributed
mechanism, both data replication and distributed stor-
age. Second, we conduct large scale experiments on
the Senslab real testbed in order to evaluate our solu-
tion. Even if the Senslab platform is suitable for test-
ing WSN-based applications (Ducrocq et al., 2010),
to the best of our knowledge, no previous studies with
results collected in Senslab have been published.
SENSORNETS 2012 - International Conference on Sensor Networks
16