number of hyperslabs p = 8, the weight coefficient
ω = 0.1250 and the kernel parameter γ = 1.
The obtained results lead to the following obser-
vations:
• The four algorithms presented in this experiment
have a fast performance, taking less than 1 minute
in the estimation with 1000 training samples. As
is showed in Table I, the KSCP achieves the mini-
mum mean square error, the KAP and the KLMS-
SCAL1 achieves more velocity in execution but
with a higher mean square error. It is important
to highlight that the settings of the T
1
and T
2
pa-
rameters are essential for the execution time and
the accuracy. The abnormality threshold parame-
ter T
1
can be adjusted for achieving more velocity
and the redundancy threshold parameter T
2
must
be adequately limited to guarantee high accuracy.
• The learning curve figure showed that in the Ker-
nel Affine Projection algorithm, the mean square
error begins to grow exponentially in the iteration
number 400; the KLMS-SCAL1 takes several ite-
rations to stabilize and the EXKRLS and KSCP
reach zero and remain stable from the first 50 ite-
rations for the considered situation.
5 CONCLUSIONS
This paper presents our proposed algorithm KSCP
for adaptive nonlinear estimation. The KSCP impro-
ves the estimation performance based on 3 aspects:
First, an effective dictionary control is established,
guaranteeing an exhaustive selection of the most im-
portant data for the estimation, and low computational
complexity, basing not on heuristics but on a strong
mathematical foundation approach with statistical and
probabilistic techniques. Second, we take advantage
of the Kernel Least Mean Square and the Surprise Cri-
terion combining them to reduce the complexity of
the calculations of variance prediction and error pre-
diction of the incoming data. Third, for the high accu-
racy goal, we use the Parallel Hyperslab Projection
Along Affine Subspace.
With all of these ideas, we achieve a fast conver-
gence, high accuracy, small size of dictionary and fast
performance algorithm, which is demonstrated by the
computational experiment and the comparison with
some important and recognized algorithms like Ker-
nel Affine Projection with Coherence Criterion, Ker-
nel Least Mean Square with Coherence-Sparsification
criterion and Adaptive L1-norm regularization and
Extended Recursive Least Squares.
REFERENCES
Aronszajn, N. (1950). Theory of reproducing kernels.
Transactions of the American mathematical society,
68(3):337–404.
Engel, Y., Mannor, S., and Meir, R. (2004). The kernel
recursive least-squares algorithm. IEEE Transactions
on Signal Processing, 52(8):2275–2285.
Gao, W., Chen, J., Richard, C., Huang, J., and Flamary, R.
(2013). Kernel lms algorithm with forward-backward
splitting for dictionary learning. In 2013 IEEE Inter-
national Conference on Acoustics, Speech and Signal
Processing, pages 5735–5739.
Haykin, S., Sayed, A. H., Zeidler, J. R., Yee, P., and Wei,
P. C. (1997). Adaptive tracking of linear time-variant
systems by extended rls algorithms. IEEE Transacti-
ons on Signal Processing, 45(5):1118–1128.
Kivinen, J., Smola, A. J., and Williamson, R. C. (2004).
Online learning with kernels. IEEE Transactions on
Signal Processing, 52(8):2165–2176.
Liu, W., Park, I., and Principe, J. C. (2009a). An infor-
mation theoretic approach of designing sparse kernel
adaptive filters. IEEE Transactions on Neural Net-
works, 20(12):1950–1961.
Liu, W., Park, I., Wang, Y., and Principe, J. C. (2009b). Ex-
tended kernel recursive least squares algorithm. IEEE
Transactions on Signal Processing, 57(10):3801–
3814.
Liu, W., Pokharel, P. P., and Principe, J. C. (2008). The ker-
nel least-mean-square algorithm. IEEE Transactions
on Signal Processing, 56(2):543–554.
Liu, W., Principe, J., and Haykin, S. (2010). Kernel Adap-
tive Filtering: A comprehensive Introduction. Wiley,
1 edition.
Ozeki, K. (2016). Theory of affine projection algorithms for
adaptive filtering. Springer.
Platt, J. (1991). A resource-allocating network for function
interpolation. Neural computation, 3(2):213–225.
Richard, C., Bermudez, J. C. M., and Honeine, P. (2009).
Online prediction of time series data with kernels.
IEEE Transactions on Signal Processing, 57(3):1058–
1067.
Said
´
e, C., Lengell
´
e, R., Honeine, P., Richard, C., and Ach-
kar, R. (2012). Dictionary adaptation for online pre-
diction of time series data with kernels. In 2012 IEEE
Statistical Signal Processing Workshop (SSP), pages
604–607.
Scholkopf, B. and Smola, A. (2001). Learning with Kernels.
Cambridge. MIT Press, 1 edition.
Slavakis, K., Theodoridis, S., and Yamada, I. (2008). On-
line kernel-based classification using adaptive pro-
jection algorithms. IEEE Transactions on Signal Pro-
cessing, 56(7):2781–2796.
Takizawa, M. and Yukawa, M. (2013). An efficient data-
reusing kernel adaptive filtering algorithm based on
parallel hyperslab projection along affine subspaces.
In 2013 IEEE International Conference on Acoustics,
Speech and Signal Processing, pages 3557–3561.
Nonlinear Adaptive Estimation by Kernel Least Mean Square with Surprise Criterion and Parallel Hyperslab Projection along Affine
Subspaces Algorithm
367