query will greatly assist a geospatial analyst’s effi-
ciency, accuracy and productivity. Being able to or-
chestrate geospatial datasets and WPS to achieve a
task will be adaptable to other fields as well. Integrat-
ing Semantic Web concepts and technologies into var-
ious non-geospatial datasets and Web Services will al-
low for automatic orchestration and hence, increased
productivity. As an example, Kauppinen (Kauppinen
et al., 2014) has documented how semantic technolo-
gies are being integrated within Brazilian Amazon
Rainforest data, that has led to increased productiv-
ity and efficiency in that area.
To advance automated and intelligent orchestra-
tion, certain features and specifications must be added
to the current WPS standard to allow for machine in-
terpretation of these Web Services, to be able to ef-
fectively and reliably determine which Web Services
are appropriate for completing a given task. Critical
information that paves the way for automatic orches-
tration are currently not defined in the WPS specifica-
tion and this research aims to set an example for the
addition of metadata and functions that allow for au-
tomatic orchestration as the next logical step for WPS.
This paper explores automated orchestration
methods of web services and data from multiple, dis-
parate sources; in contrast to the current widespread
method of supplying all the data and services to the
end user and leaving it to them to manually analyse
and process the vast amounts of varying kinds of data,
and determine what processing needs to be executed.
Natural language processors and ontologies are pro-
posed to build the required Artificial Intelligence to
automatically chain together the resources to produce
useful output for the end user.
2 BACKGROUND
In the last decade, the Web has been moving to-
wards Service-Oriented Computing architecture sup-
porting automated use (Ameller et al., 2015) (Huhns
and Singh, 2005). This architecture aims to build a
network of interoperable and collaborative applica-
tions, independent of platform, called services (Pa-
pazoglou and Georgakopoulos, 2003) (Pugliese and
Tiezzi, 2012). The geospatial world is also mov-
ing away from the traditional desktop application
paradigm to processing and accessing data on-the-fly
from the Web using Web Processing Services, as out-
lined by Granell et al. (Granell et al., 2012). As Web
Service technology has matured in recent years, an
increasing amount of geospatial content and process-
ing capabilities are available online as Web Services
(Zhao et al., 2012). These Web Services enable inter-
operable, distributed, and collaborative geoprocessing
to significantly enhance the abilities of users to col-
lect, analyze and derive geospatial data, information,
and knowledge over the Internet (Zhao et al., 2012).
Current geospatial workflows and processes rely
on manual human intervention in searching for the
relevant and/or required datasets and Web Services
(Granell et al., 2012). These workflows also require
human analysis of the output at each stage of process-
ing, and manual determination of which Web Process-
ing Service to use next on the data to achieve the final
required output (Granell et al., 2012). This has been
observed to lead to inefficiencies in the accuracy and
currency of the data as we are relying on a human user
to search for these ill-exposed datasets and Web Ser-
vices. For example, a human user will tend to have a
bias towards a dataset or Web Service that he/she has
used before, regardless of the currency of the data or
the frequency of updating of the data, a phenomenon
known as the Mere-repeated-exposure paradigm (Za-
jonc, 2001). Geospatial Web Services and datasets
that may be vital in contributing to the final result
may also be left unexposed due to current search tech-
nologies not being able to expose Web Services suffi-
ciently. The way that Web Services are searched for is
by functional and non-functional requirements as well
as interactive behaviour (Wu, 2015a), which require
more than simple keyword matching, as per current
search algorithms.
2.1 The Semantic Web (Web 3.0)
The Semantic Web aims to create a web of infor-
mation that is machine-readable and not just human-
readable (Berners-Lee et al., 2001). This allows ma-
chines to automatically find, combine and act upon
information found on the Web (Pulido et al., 2006).
The objective of the Semantic Web is accom-
plished by integrating semantic content into web
pages that helps describe the contents and context of
the data in the form of metadata (data about data)
(Handschuh and Staab, 2003). This greatly improves
the quality of the data so that a machine is able to un-
derstand what the data is for, what it can be used for
and what other things are linked to it (Harth, 2004)
(Berners-Lee et al., 2001). This allows the machine
to process and use the data, instead of the current
paradigm of relying on a human to interpret, process
and understand data.
Ontologies are a core component of the Semantic
Web, and are required by machines to be able to in-
telligently reason and infer data (Pulido et al., 2006).
An ontology is a set of data elements within a domain
that are linked together to denote the types, proper-
GISTAM 2016 - 2nd International Conference on Geographical Information Systems Theory, Applications and Management
72