two major bottlenecks of current methods by using
a divisive approach to reduce effects of Gaussian Pro-
cesses’ cubic runtime complexity as well as employ-
ing a purposive strategy to generate fewer candidate
models.
Our performance evaluation has revealed that
GPMs resulting from the proposed TAGR algorithm
deliver similar model quality in comparison to those
models produced by state-of-the-art algorithms. In
addition, runtime of the retrieval process is reduced
significantly especially for larger time series, where
we achieve a speed-up factor of approximately 500
with regards to existing methods such as CKS, ABCD
and SKC.
As for future work, we consider global approxi-
mations an opportunity for further optimizing our ap-
proach. We therefore plan to address the opportuni-
ties of low-rank approximations such as the Nystr
¨
om
method (Hensman et al., 2013) in our future work.
In addition, we plan to develop GPM retrieval algo-
rithms for big data processing frameworks in order to
scale GPM retrieval to very large and even multidi-
mensional datasets.
REFERENCES
Abrahamsen and Petter (1997). A review of gaussian ran-
dom fields and correlation functions: Technical report
917.
Aminikhanghahi, S. and Cook, D. J. (2017). A survey
of methods for time series change point detection.
Knowl. Inf. Syst., 51(2):339–367.
Berns, F. and Beecks, C. (2020a). Automatic gaussian pro-
cess model retrieval for big data. In CIKM. ACM.
Berns, F. and Beecks, C. (2020b). Towards large-scale gaus-
sian process models for efficient bayesian machine
learning. In DATA, pages 275–282. SciTePress.
Berns, F., Schmidt, K. W., Grass, A., and Beecks, C. (2019).
A new approach for efficient structure discovery in iot.
In BigData, pages 4152–4156. IEEE.
Cheng, C. and Boots, B. (2017). Variational inference for
gaussian process models with linear complexity. In
NIPS, pages 5184–5194.
Chollet, F. (2018). Deep learning with Python. Manning
Publications Co, Shelter Island New York.
Csat
´
o, L. and Opper, M. (2000). Sparse representation for
gaussian process models. In NIPS, pages 444–450.
MIT Press.
Duvenaud, D., Lloyd, J. R., Grosse, R. B., Tenenbaum,
J. B., and Ghahramani, Z. (2013). Structure dis-
covery in nonparametric regression through composi-
tional kernel search. In ICML (3), volume 28 of JMLR
Workshop and Conference Proceedings, pages 1166–
1174. JMLR.org.
Gittens, A. and Mahoney, M. W. (2016). Revisiting the nys-
trom method for improved large-scale machine learn-
ing. J. Mach. Learn. Res., 17:117:1–117:65.
Hebrail, G. and Berard, A. (2012). Individual house-
hold electric power consumption data set.
https://archive.ics.uci.edu/ml/datasets/individual+
household+electric+power+consumption. Accessed:
09/01/2020.
Hensman, J., Fusi, N., and Lawrence, N. D. (2013). Gaus-
sian processes for big data. In UAI. AUAI Press.
Hong, T., Pinson, P., and Fan, S. (2014). Global energy
forecasting competition 2012. International Journal
of Forecasting, 30(2):357–363.
Iliev, A. I., Kyurkchiev, N., and Markov, S. (2017). On the
approximation of the step function by some sigmoid
functions. Math. Comput. Simul., 133:223–234.
Kim, H. and Teh, Y. W. (2018). Scaling up the automatic
statistician: Scalable structure discovery using gaus-
sian processes. In AISTATS, volume 84 of Proceed-
ings of Machine Learning Research, pages 575–584.
PMLR.
Li, S. C. and Marlin, B. M. (2016). A scalable end-to-end
gaussian process adapter for irregularly sampled time
series classification. In NIPS, pages 1804–1812.
Liu, H., Ong, Y., Shen, X., and Cai, J. (2020). When gaus-
sian process meets big data: A review of scalable gps.
IEEE Transactions on Neural Networks and Learning
Systems, pages 1–19.
Lloyd, J. R., Duvenaud, D., Grosse, R. B., Tenenbaum,
J. B., and Ghahramani, Z. (2014). Automatic con-
struction and natural-language description of nonpara-
metric regression models. In AAAI, pages 1242–1250.
AAAI Press.
Low, K. H., Yu, J., Chen, J., and Jaillet, P. (2015). Paral-
lel gaussian process regression for big data: Low-rank
representation meets markov approximation. In AAAI,
pages 2821–2827. AAAI Press.
Malkomes, G., Schaff, C., and Garnett, R. (2016). Bayesian
optimization for automated model selection. In NIPS,
pages 2892–2900.
Max Planck Institute for Biogeochemistry (2019). Weather
Station Beutenberg / Weather Station Saaleaue: Jena
Weather Data Analysis. https://www.bgc-jena.mpg.
de/wetter/. Accessed: 09/01/2020.
Park, C. and Apley, D. W. (2018). Patchwork kriging
for large-scale gaussian process regression. J. Mach.
Learn. Res., 19:7:1–7:43.
Rasmussen, C. E. and Williams, C. K. I. (2006). Gaussian
processes for machine learning. Adaptive computa-
tion and machine learning. MIT Press.
Resende, M. G. and Ribeiro, C. C. (2016). Optimization by
GRASP: Greedy Randomized Adaptive Search Proce-
dures. Springer New York, New York, NY.
Rivera, R. and Burnaev, E. (2017). Forecasting of com-
mercial sales with large scale gaussian processes. In
ICDM Workshops, pages 625–634. IEEE Computer
Society.
Roberts, S., Osborne, M., Ebden, M., Reece, S., Gibson, N.,
and Aigrain, S. (2013). Gaussian processes for time-
series modelling. Philosophical transactions. Series
Large-scale Retrieval of Bayesian Machine Learning Models for Time Series Data via Gaussian Processes
79