
tance are marked with colored boxes in the reduced
time series. Value curves are extracted and displayed
in separate diagrams for all classes (see Figure 6b).
The question remains whether these patterns in
the reduced time series are also present in the orig-
inal MTS. Therefore, as an example, the original
segments of the identified pattern 2 are investigated
(see Figure 6c). Positive and negative correlations are
evident between sensor values, mainly between gy-
roscope data in x and z direction and accelerometer
data in x and z direction. Patterns are found between
variables of gyroscope data in x direction, accelerom-
eter data in x, and z direction, indicating a link be-
tween correlation and patterns. However, some sen-
sors have little to no correlation to other sensors, such
as speed and accelerometer data in y direction. A cor-
relation between sensor values is necessary for DR,
especially for PCA methods, which use correlation as
a primary factor. Patterns 0 and 2 show similar re-
sults, indicating that patterns found for reduced time
series also occur in the MTS. However, higher vari-
ances or noise among pattern instances occur in a sub-
set of the dimensions, making them sub-dimensional
patterns. Utilizing upstream DR for TSPR results in a
time saving of 89 % compared to multivariate cases.
5 CONCLUSION & OUTLOOK
This paper compares various methods for dimen-
sion reduction in the context of unsupervised pat-
tern recognition. As a result, Autoencoder, Func-
tional Principle Component Analysis (FPCA), and
Factor Analysis (FA) produce dimensionally reduced
data with the least loss of ∆F1
a
using a synthetic
dataset. Furthermore, dimension reduction with FA
and FPCA yields a runtime advantage of up to 90 %
over a non-reduced pattern search while losing only
18 % of ∆F1
a
. This result is validated with the real-
world dataset Commercial Vehicles Sensor Dataset.
However, the speed benefits must be weighed against
potential loss in accuracy and tested in advance, es-
pecially in the case of sub-dimensional pattern recog-
nition. Future studies can explore alternative pattern
algorithms, increase dataset diversity, investigate sub-
dimensional pattern recognition and variable pattern
lengths, and consider the temporal offset of patterns.
REFERENCES
AB, S. S. Commercial vehicles data set. Accessed: 2023-
04-11.
Alaee, S., Kamgar, K., and Keogh, E. (2020). Matrix profile
xxii: Exact discovery of time series motifs under dtw.
Proceedings - IEEE International Conference on Data
Mining, ICDM, 2020-November:900–905.
Bascol, K., Emonet, R., Fromont, E., and Odobez, J. M.
(2016). Unsupervised interpretable pattern discovery
in time series using autoencoders. Lecture Notes in
Computer Science (including subseries Lecture Notes
in Artificial Intelligence and Lecture Notes in Bioin-
formatics), 10029 LNCS:427–438.
Berndt, D. J. and Clifford, J. (1994). Using dynamic time
warping to find patterns in time series. In Proceed-
ings of the 3rd International Conference on Knowl-
edge Discovery and Data Mining, AAAIWS’94, page
359–370. AAAI Press.
Cattell, R. B. (1965). A biometrics invited paper. factor
analysis: An introduction to essentials ii. the role of
factor analysis in research. Biometrics, 21(2):405–
435.
Chan, N. H. (2007). Time series: applications to finance.
Statistica, 62.
Chaudhuri, P. (2006). Functional data analysis by j. o. ram-
say; b. w. silverman. Sankhy
¯
a: The Indian Journal of
Statistics (2003-2007), 68:364–365.
Dodge, Y. (2008). The concise encyclopedia of statistics.
Springer Science & Business Media.
Hochreiter, S. and Schmidhuber, J. (1997). Long short-term
memory. Neural computation, 9:1735–80.
Hotelling, H. (1933). Analysis of a complex of statistical
variables into principal components. Journal of Edu-
cational Psychology, 24:498–520.
Kam, H. J., Sung, J. O., and Park, R. W. (2010). Predic-
tion of daily patient numbers for a regional emergency
medical center using time series analysis. Healthcare
informatics research, 16(3):158–165.
Law, S. M. (2019). STUMPY: A Powerful and Scalable
Python Library for Time Series Data Mining. The
Journal of Open Source Software, 4(39):1504.
Mairal, J., Bach, F., Ponce, J., and Sapiro, G. (2009). Online
dictionary learning for sparse coding. In Proceedings
of the 26th Annual International Conference on Ma-
chine Learning, ICML ’09, page 689–696, New York,
NY, USA. Association for Computing Machinery.
McInnes, L., Healy, J., Saul, N., and Großberger, L. (2018).
Umap: Uniform manifold approximation and projec-
tion. Journal of Open Source Software, 3(29):861.
M
¨
orchen, F. (2006). Time series knowlegde mining. G
¨
orich
und Weiersh
¨
auser.
Noering, F. K. D. (2022). Unsupervised pattern discovery
in automotive time series.
Pearson, K. (1905). The problem of the random walk. Na-
ture 1905 72:1865, 72:294–294.
Petersen, P. and Sax, E. (2022). A fully automated method-
ology for the selection and extraction of energy-
relevant features for the energy consumption of bat-
tery electric vehicles. SN Computer Science, 3(5):342.
Rumelhart, D. E., Hinton, G. E., and Williams, R. J. (1986).
Learning representations by back-propagating errors.
Nature, 323:533–536.
Comparison of Dimension Reduction Methods for Multivariate Time Series Pattern Recognition
815