the missing data problem using processing function
included in the algorithm. The pruning process has a
great effect on the decision tree that improved the
results of prediction after applying it on the tree by
reducing the error of prediction from 3.2% to 2.2%.
The learning rate 0.00001 leads to the best
training results in ANN and 20000 iterations was
enough to run the training that the error is decreased
very slowly after 10000 iterations. Changing the
output to binary offered improvement on the
classification and prediction process using ANN that
reduced the error of prediction the state of bending
protein from 12.2% to 3.7%. Removing the missing
data reduced the number of training examples from
2800 example to 1947 examples made the network
faster in training phase and improved the error of
prediction from 3.7% to 3.1%. Finding the mean of
data set for each class as applied in section 6.3.3 was
not helpful in any way but also have a poor results in
prediction, also rearranging the input values between
0 and 1 shows no improvement on the results or the
speed of learning. Increasing the number of neurons
more than 20 neurons effects the results of
prediction in a bad way that decrease the speed of
learning and increase the error of prediction.
In this study it was obvious that the dataset
quality is not good, it have a lot of missing values
and there are a shortage of examples in class 2 and
class 3, but the comparison in this data
circumstances shows that C4.5 decision tree is more
reliable method to use but In future it is worth trying
to run these experiments again using another data
sets in many aspects of medical diagnosis domain
that have better quality examples with less numbers
of missing values and rich examples in each class
attributes and then run the comparison again to see
which method will overcome in the field of medical
diagnoses domain.
REFERENCES
Leiva, H., 2002. A multi-relational decision tree learning
algorithm ,Msc, Iowa State University Ames.
Thomas, D., Hild, H., and Ghulum Bakiri, G., 1995. A
Comparison of ID3 and Backpropagation for English
Text-to-Speech Mapping Machine Learning, Kluwer
Academic Publishers, Boston.
Berkman, S., Lubomir, H., Ping, C., Chuan, Z., Wei Jun,
W,. 2005. Comparison of decision tree classifiers with
neural network and linear discriminant analysis
classifiers for computer-aided diagnosis: a Monte
Carlo simulation study, Medical Imaging: Image
Processing, Volume 5747.
Bagnall, A., Cawley, G., 2000. Learning Classifier
Systems for Data Mining: A Comparison of XCS with
Other Classifiers for the Forest Cover Data Set,
University of East Anglia, England.
The UCI KDD archive. Irvine, University of California,
Department of Information and Computer Science,
http://kdd.ics.uci.edu. Last access September 2007.
Mitra, S., Tinkuacharya ., 2003, Data mining multimedia,
soft computing, and Bioinformatics, John Wiley &
Sons, Inc.
Ye, N., 2003. Hand book of data mining, Arizona State
University, Lawrence Erlbaum Associates, Inc, New
Jersey.
Kantardzic, M., 2003. Data Mining: Concepts, Models,
Methods, and Algorithms, John Wiley & Sons.Inc.
Larose, D., 2005, Discovering Knowledge in data, an
introduction to data mining, John Wiley & Sons.Inc.
Michael, A., Berry, S., 2004. Data mining Techniques,
John Wiley & Sons.Inc, 2nd edition.
Paplinski, A., 2004. Basic structures and properties of
Artificial Neural Networks.retrivedfrom:
lsc.fie.umich.mx/~juan/Materias/ANN/Papers/basic
structures-and-properties.pdf, last access June 2006.
Zurada, J., 1992. Introduction to Artificial Neural
Systems, West publishing Company, Singapore.
A COMPARISON STUDY OF TWO KNOWLEDGE ACQUISITION TECHNIQUES APPLIED TO THYROID
MEDICAL DIAGNOSIS DOMAIN
365