impact of outliers on classifier accuracy) in a
classical Levenberg-Marquadt learning algorithm.
The proposed learning algorithm is tested and
compared with three other ones on a simulation
example. The impact of the choice of
misclassification costs and of the presence of
outliers is investigated. The results show that the
conjoint use of cost-sensitive weight and robust
criterion improves the classifier accuracy.
In our future works, this approach will be tested
on other benchmark datasets in order to confirm the
results. The impact of imbalanced repartition in the
dataset will be also investigated. Last this approach
will be extended to the multiclass case.
REFERENCES
Aström, K.J., 1980. Maximum likelihood and prediction
error methods. Automatica, 16, 551-574.
Bloch G., Theilliol D., Thomas P., 1994. Robust
identification of non linear SISO systems with neural
networks. System Identification (SYSID'94), a
Postprint Volume from the IFAC Symp., Copenhagen,
Denmark, July 4-6, M. Blanke, T. Söderström (Eds.),
Pergamon, 1995, Vol. 3, pp. 1417-1422.
Bloch G., Thomas P., Theilliol D., 1997. Accommodation
to outliers in identification of non linear SISO systems
with neural networks. Neurocomputing, 14, 85-99.
Barnett V. Lewis T., 1994, Outliers in statistical data,
John Wiley, ISBN 0-471-93094-6, Chichester.
Castro C.L., Braga A.P., 2013. Novel cost-sensitivity
approach to improve the multilayer perceptron
performance on imbalanced data. IEEE Trans. On
Neural Networks and Learning Systems, 24, 6, 888-
899.
Cateni S., Colla V., Vannucci M., 2008. Outlier Detection
Methods for Industrial Applications, Advances in
Robotics, Automation and Control, Jesus Aramburo
and Antonio Ramirez Trevino (Ed.), ISBN: 978-953-
7619-16-9, InTech, Available from:
http://www.intechopen.com/books/advances_in_roboti
cs_automation_and_control/outlier_detection_method
s_for_industrial_applications
Chen D.S., Jain R.C., 1991. A robust back propagation
learning algorithm for function approximation. Proc.
Third Int. Workshop on Artificial Intelligence and
Statistics, Fort Lauderdale, FL, 218-239.
Cybenko G., 1989. Approximation by superposition of a
sigmoïdal function. Math. Control Systems Signals 2,
303-314.
Demuth H., Beale P. 1994. Neural networks toolbox user's
guide V2.0. The MathWorks, Inc.
Domingos P., 1999. MetaCost: A general method for
making classifiers cost sensitive. Proc. of the 5
th
Int.
Conf. on Knowledge Discovery and Data Mining, 78-
150.
Drummond C., Holte R.C., 2000. Exploiting the cost
(in)sensitivity of decision tree splitting criteria. Proc.
of the 17
th
Int. Conf. on Machine Learning, 239-246.
Fan W., Stolfo S.J., Zhang J., Chan P.K., 1999. AdaCost:
Misclassification cost-sensitive boosting. Proc. of Int.
Conf. on Machine Learning, pp. 97-105.
Frénay B., Verleysen M., 2014. Classification in the
presence of label noise: a survey. IEEE trans. On
Neural Networks and Learning Systems, 25, 845-869.
Funahashi K., 1989. On the approximate realisation of
continuous mapping by neural networks. Neural
Networks 2, 183-192.
Garcia R.A.V., Marqués A.I., Sanchez J.S., Antonio-
Velasquez J.A., 2013. Making accurate credit risk
predictions with cost-sensitive MLP neural networks
in Management Intelligent Systems, Advances in
Intelligent Systems and Computing, 220, 1-8.
Geibel, Peter, Brefeld, Ulf, and Wysotzki, Fritz.
Perceptron and svm learning with generalized cost
models. Intelligent Data Analysis, 8:439–455, 2004
Hand, D, Mannila, H., Smyth, P., 2001. Principles of data
mining. The MIT press, Cambridge
Hawkins, D., 1980. Identification of Outliers, Chapman
and Hall, London.
Huber P.J., 1964. Robust estimation of a location
parameter. Ann. Math. Stat., 35, 73-101.
Kotsiantis, S.B., 2007. Supervised machine learning: a
review of classification techniques. Informatica, 31,
249–268.
Liano K., 1996. Robust error for supervised neural
network learning with outliers. IEEE Trans. on Neural
Networks, 7, 246-250.
Lin Y., Lee Y., Wahba G., 2000. Support vector machines
for classification in nonstandard situations. Technical
Repport,
http://roma.stat.wisc.edu/sites/default/files/tr1016.pdf.
Ljung L., 1987. System identification: theory for the user.
Prentice-Hall, Englewood Cliffs.
Manwani N., Sastry P.S. 2013. Noise tolerance under risk
minimization. IEEE Trans. Cybern., 43, 1146–1151.
Margineantu D., 2002. Class probability estimation and
cost-sensitive classification decision. Proc. of the 13
th
European Conference on Machine Learning, 270-281.
Moore D.S., McCabe G.P., 1999. Introduction to the
Practice of Statistics. Freeman & Company.
Nguyen D., Widrow B., 1990. Improving the learning
speed of 2-layer neural networks by choosing initial
values of the adaptive weights. Proc. of the Int. J.
Conf. on Neural Networks IJCNN'90, 3, 21-26.
Puthenpura S., Sinha N.K., 1990. A robust recursive
identification method. Control-Theory and Advanced
Technology 6: 683-695.
Raudys S., Raudis A., 2010. Pairwise costs in multiclass
perceptrons. IEEE Tans. On Pattern Analysis and
Machine Intelligence, 32, 7, 1324-1328.
Sàez J., Galar M., Luengo J., Herrera F. 2014. Analyzing
the presence of noise in multi-class problems:
Alleviating its influence with the one-vs-one
decomposition, Knowl. and Information Systems, 38,
179–206.
NCTA 2015 - 7th International Conference on Neural Computation Theory and Applications
112