ble of generalising over the training data. To that end
we regress from the proportion of the training data
to the test data and evaluate how the prediction error
changes by altering this proportion. The prediction is
performed using simple least-square regression over
the kernel induced feature space. In Figure 6, the
results are shown using different sizes of the train-
ing data. As shown, the path kernel performs signif-
icantly better compared to the global alignment ker-
nel and the results improve with the size of the train-
ing dataset. Interestingly, the global alignment ker-
nel produces very different results dependent on the
size of the training dataset indicating that it is severely
over-fitting the data.
5 10 15 20 25 30 35 40 45 50
training examples
0
10
20
30
40
50
RMS Error
Figure 6: The above figure depicts the RMS error when
predicting the phase shift from a noisy sine waveform by
a regression over the feature space induced by the kernels.
The red bars correspond to the global alignment kernel and
the green bars to the path kernel. The y-axis shows the error
in percentage of phase, while the x-axis indicates the size of
the training dataset.
5 CONCLUSIONS
In this paper, we have presented a novel kernel for en-
coding sequences. Our kernel reflects and encodes all
possible alignments between two sequences by asso-
ciating a cost to each. This cost encodes a preference
towards specific paths. The kernel is applicable to
any kind of symbolic or numerical data as it requires
only the existence of a kernel between symbols. We
have presented both qualitative and quantitative ex-
periments exemplifying the benefits of the path kernel
compared to competing methods. We show that the
proposed method significantly improves results both
with respect to discrimination and generalisation es-
pecially in noisy scenarios. The computational cost
associated with the kernel is considerably lower than
competing methods, making it applicable to data-sets
that could previously not be investigated using ker-
nels.
Our experimental results indicate that the kernel
we propose is positive semi-definite. In future we
intend to investigate proving this property. Further-
more, in this paper, we have chosen a very simplis-
tic dataset in order to evaluate our kernel. Given
our encouraging results, we are currently working on
applying our kernel to more challenging real-world
datasets.
REFERENCES
Bahlmann, C., Haasdonk, B., and Burkhardt, H. (2002).
Online handwriting recognition with support vector
machines - a kernel approach. In 8th International
Workshop on Frontiers in Handwriting Recognition.
Berlinet, A. and Thomas-Agnan, C. (2004). Reproducing
kernel Hilbert spaces in probability and.
Buhmann, M. D. and Martin, D. (2003). Radial basis func-
tions: theory and implementations.
Chang, C. C. and Lin, C. J. (2001). LIBSVM: a library for
support vector machines.
Cristianini, N. and Shawe-Taylor, J. (2006). An introduction
to support Vector Machines: and other kernel-based
learning methods.
Cuturi, M. (2010). Fast Global Alignment Kernels. In In-
ternational Conference on Machine Learning.
Cuturi, M., Vert, J.-P., Birkenes, O., and Matsui, T. (2007).
A Kernel for Time Series Based on Global Aligh-
ments. In IEEE International Conference on Acous-
tics, Speech and Signal Processing, pages 413–416.
Gudmundsson, S., Runarsson, T. P., and Sigurdsson, S.
(2008). Support vector machines and dynamic time
warping for time series. IEEE International Joint
Conference on Neural Networks, pages 2772–2776.
Haasdonk, B. (2005). Feature space interpretation of SVMs
with indefinite kernels. IEEE Transactions on Pattern
Analysis and Machine Intelligence, 27(4):482–492.
Haasdonk, B. and Burkhardt, H. (2007). Invariant kernel
functions for pattern analysis and machine learning.
Machine learning, 68(1):35–61.
Haussler, D. (1999). Convolution kernels on discrete struc-
tures. Technical report.
Leslie, C. and Kuang, R. (2004). Fast String Kernels using
Inexact Matching for Protein Sequences. The Journal
of Machine Learning Research, 5:1435–1455.
Li, M. and Zhu, Y. (2006). Image classification via LZ78
based string kernel: a comparative study. Advances in
knowledge discovery and data mining.
Lodhi, H., Saunders, C., Shawe-Taylor, J., Cristianini, N.,
and Watkins, C. (2002). Text classification using
string kernels. The Journal of Machine Learning Re-
search, 2:419–444.
Luo, G., Bergstr
¨
om, N., Ek, C. H., and Kragic, D. (2011).
Representing actions with Kernels. In IEEE/RSJ In-
ternational Conference on Intelligent Robots and Sys-
tems, pages 2028–2035.
ICPRAM2013-InternationalConferenceonPatternRecognitionApplicationsandMethods
56