Table 4: Table comparing the mean training time for
AURA k-NN using time-series data compared to AURA
k-NN with CFS on the same data. The mean was
calculated over five runs.
AURA
ts
AURA
CFSts
Training time (secs)
0.30
0.17
7 CONCLUSIONS
In this paper, we have introduced a unified
framework for attribute selection and prediction.
Classification is also available in the framework
(Hodge and Austin, 2005); (Krishnan et al., 2010).
Previously, we demonstrated two attribute
selection approaches in AURA (Hodge et al., 2006).
We have now added the multivariate CFS selector
which is based on entropy. No attribute selector
excels on all data or all tasks so we need a range of
selectors to select the best for each task. We showed
that CFS improved the prediction accuracy of the
AURA k-NN on the real world task of bus journey
prediction. We demonstrated that using attribute
selection to reduce the dimensionality reduces the
training time allowing larger data to be processed.
The AURA framework described is flexible and
easily extended to other attribute selection
algorithms. Ultimately, we will provide a parallel
and distributed data mining framework for attribute
selection, classification and prediction drawing on
our previous work on parallel (Weeks et al., 2002)
and distributed AURA (Jackson et al., 2004).
REFERENCES
Austin, J., 1995. Distributed Associative Memories for
High Speed Symbolic Reasoning. In IJCAI: Working
Notes of Workshop on Connectionist-Symbolic
Integration, pp. 87-93.
Bishop, C., 1995. Neural networks for pattern recognition,
Oxford University Press, Oxford, UK.
Box, G., Jenkins, G., 1970. Time series analysis:
Forecasting and control, San Francisco: Holden-Day.
Cover T., Hart P, 1967. Nearest neighbor pattern
classification. IEEE Transactions on Information
Theory 13(1): 21–27.
Engle, R., 1982. Autoregressive Conditional
Heteroscedasticity with Estimates of the Variance of
UK Inflation. Econom., 50: 987-1008.
Fayyad, U., Irani K., 1993. Multi-Interval Discretization
of Continuous-Valued Attributes for Classification
Learning. In Procs International Joint Conference on
Artificial Intelligence, pp. 1022-1029.
Guyon, I. et al., 2002. Gene selection for cancer
classification using support vector machines. Mach.
Learn., 46(1): 389-422
Hall, M., 1998. Correlation-based Feature Subset
Selection for Machine Learning. Ph.D. Thesis,
University of Waikato, New Zealand.
Hall, M., Smith, L., 1997. Feature subset selection: a
correlation based filter approach. In, International
Conference on Neural Information Processing and
Intelligent Information Systems, pp. 855-858.
Hodge, V., 2011. Outlier and Anomaly Detection: A
Survey of Outlier and Anomaly Detection Methods.
LAMBERT Academic Publishing, ISBN: 978-3-8465-
4822-6.
Hodge, V., Austin, J., 2005. A Binary Neural k-Nearest
Neighbour Technique. Knowl. Inf. Syst. (KAIS), 8(3):
276-292, Springer-Verlag London Ltd.
Hodge, V. et al., 2004. A High Performance k-NN
Approach Using Binary Neural Networks. Neural
Netw., 17(3): 441-458, Elsevier Science.
Hodge, V. et al., 2006. A Binary Neural Decision Table
Classifier. NeuroComputing, 69(16-18): 1850-1859,
Elsevier Science.
Hodge, V. et al., 2011. Short-Term Traffic Prediction
Using a Binary Neural Network. 43rd Annual UTSG
Conference, Open University, UK, January 5-7.
Jackson, T. et al., 2004. Distributed Health Monitoring for
Aero-Engines on the Grid: DAME. In Procs of IEEE
Aerospace, Montana, USA, March 6-13.
Kohavi, R. John, G., 1997. Wrappers for Feature Subset
Selection. In Artif. Intell. J., Special Issue on
Relevance, 97(1-2): 273-324
Krishnan, R. et al., 2010. On Identifying Spatial Traffic
Patterns using Advanced Pattern Matching
Techniques. In Procs of Transportation Research
Board (TRB) 89th Annual Meeting, Washington, D.C.
Labeeuw, W. et al., 2009. Prediction of Congested Traffic
on the Critical Density Point Using Machine Learning
and Decentralised Collaborating Cameras.
Portuguese
Conference on Artificial Intelligence, pp. 15-26.
Liu, H., Setiono, R., 1995. Chi2: Feature selection and
discretization of numeric attributes. In Procs IEEE 7th
International Conference on Tools with Artificial
Intelligence, pp. 338-391.
Quinlan, J., 1986. Induction of Decision Trees. Mach,
Learn., 1: 81-106.
Vlahogianni, E. et al., 2005. Optimized and meta-
optimized neural networks for short-term traffic flow
prediction: A genetic approach, Transp. Res. Part C:
Emerging Technologies, 13(3) (2005): 211-234.
Weeks, M. et al., 2002. A Hardware Accelerated Novel IR
System. In Procs 10th Euromicro Workshop (PDP-
2002), Gran Canaria, Jan. 9–11, 2002.
Zhou, P. et al., 1999. High Performance k-NN Classifier
Using a Binary Correlation Matrix Memory. In Procs
Advances in Neural Information Processing Systems
Vol. II, MIT Press, Cambridge, MA, USA, 713-719.
ABinaryNeuralNetworkFrameworkforAttributeSelectionandPrediction
515