data, touristic and cultural data, cartographic maps,
mostly in the form of comma-separated values or
geospatial vector data in the case of geographic in-
formation. To the best of our knowledge, the datasets
made available at the portal have static character (that
is, they will not reflect changes occurred after the re-
lease date) and are of two kinds, namely: aggregated
and anonymized data (e.g. number of children for
each school in the region); and identification data of
public entities (e.g. names and addresses of restau-
rants in the region). We believe this is a general is-
sue affecting the whole current generation of Open
Data. Indeed no data concerning individual people
may been released, because of obvious privacy rea-
sons. On the other hand, data of public entities (e.g.
list of restaurants) are already available on the web,
although possibly not in an organized form. Thus, the
only new thing in the current wave of Open Data is
the massive release of aggregated anonymized data,
which can be of great interest for statistical reasons
but of limited use for building services. In our opin-
ion, the lack of data concerning individual people,
along with the static nature of the datasets, are weak-
ness points of the current generation Open Data; with-
out personal data and without “freshness”, it is indeed
impossible to build useful services tailored to the ac-
tual needs of a given individual at a given time
4
.
At this point, we wondered about what would be
a useful complement to the Open Data idea. Personal
and sensitive data cannot be released without access
control and online permission by the individual. If we
need data that are always “fresh”, we need to abandon
the concept of dataset being released once for ever. It
would then make much more sense to leave the data
where they are, namely, the back ends of the numer-
ous websites, and let the webservers export them to
the web using some sort of accepted standard at the
front end. As we will discuss in this position paper,
we do not need to devise anything really new: it is
just a matter of leveraging existing technologies and
standards, and then the massive amount of personal
and even sensitive data concerning individuals would
become open, thus available for a new wave of in-
novative and personalized services, yet preserving an
acceptable degree of privacy.
The key point is that there now exist standard tech-
nologies for online authorization, by which an indi-
vidual can exert access control over personal data re-
gardless of the physical location where the data are
actually stored and managed; it is only necessary that
4
It is not by chance, that the Apps For Italy developer
contest (http:/www.apps4italy.org), calling for interesting
Open Data applications, has shifted the currently open sub-
mission deadline from February 10th to April 30th, 2012.
the manager of data (a public administration dealing
with citizen data, but also a business company manag-
ing client data) conforms to these authorization tech-
nologies and APIs. Another point is that there now
exist mature technologies for representing and export-
ing data items (e.g. records of a back end database).
A third point is that there would be no need to change
the internal organization of data at the back ends, so
no need for coordination among the many entities that
currently manage our personal data. Just glue these
pieces together, and get what we call the “Open Data
for the masses”.
Last but not least, with these technologies each in-
dividual would regain ownership on personal data, af-
ter decades in which “the owner” could only be the
same as “the manager”; this is very important for a
true ecosystem of online services to grow, free from
the monopolistic control of data managers improperly
acting as owners.
This position paper is structured as follows. Sec-
tion 2 briefly summarizes the current standard ingre-
dients of the Web 2.0. Section 3 describes the new
emerging standards and technologies for data exporta-
tion, authorization and access control; access control
by user permission, based on cryptographic creden-
tials, is the ingredient the lack of which whould make
it impossible to safely export personal data to the web.
Finally, in Section 5 the paper concludes suggesting
what we consider the open points to be addressed by
near-term research in order for the proposed approach
to become effective.
2 WEB 2.0 TECHNOLOGIES
Data Formats, Web API. In order to go over the
HTML page building block, new and more generic
languages to model structured data, independently
from their future usage, such as XML and JSON, have
been standardised. Using HTTP as a communication
protocol, each website can nowadays export its data
and services by offering its own set of web APIs, ac-
cessible to other software developers.
SOA, REST. The interaction of different applica-
tions, possibly running on heterogeneous architec-
tures, has led at the beginning of the 2000 to the de-
velopment of the so-called Service-Oriented Architec-
ture (SOA), Amazon Web Services
5
being the first
case of success with the introduction of its cloud com-
puting platform in which e-commerce applications
are built by means of separate cooperating services.
5
http://aws.amazon.com
WEBIST2012-8thInternationalConferenceonWebInformationSystemsandTechnologies
202