Author:
Ingo Simonis
Affiliation:
Open Geospatial Consortium and OGC, United Kingdom
Keyword(s):
Big Data, Standards, Spatial Data Infrastructure, OGC, Cloud, Container, Docker.
Abstract:
There is a growing number of easily accessible Big Data repositories hosted on cloud infrastructures that offer additional sets of cloud-based products such as compute, storage, database, or analytics services. The Sentinel-2 earth observation satellite data available via Amazon S3 is a good example of a petabyte-sized data repository in a rich cloud environment. The combination of hosted data and co-located cloud services is a key enabler for efficient Big Data processing. When the transport of large amounts of data is not feasible or cost efficient, processes need to be shipped and executed as closely as possible to the actual data. This paper describes standardization efforts to build an architecture featuring high levels of interoperability for provisioning, registration, deployment, and execution of arbitrary applications in cloud environments. Based on virtualization mechanisms and containerization technology, the standardized approach allows to pack any type of application or
multi-application based workflow into a container that can be dynamically deployed on any type of cloud environment. Consumers can discover these containers, provide the necessary parameterization and execute them online even easier than on their local machines, because no software installation, data download, or complex configuration is necessary.
(More)