els that meaningfully impact the overall game score.
While it lacks the generality of game independent ap-
proaches, it shows that minimal expert involvement
can enable even simple game playing agents to per-
form very well.
Because in this paper we focused on the feature
extraction, we investigated the benefits on a limited
set of agents only. Future work might revolve around
applying our findings to a broader range of agents
and games. Another interesting addition could be to
further automate the manual feature selection. The
methods presented in this paper often reduce the pos-
sible byte candidates to just a handful and it might be
feasible to reduce the number even more. If nothing
else, this work should demonstrate that feature quality
can have a very meaningful impact in ALE. Taking
this into account, it would be interesting to investigate
how fully automatic dimensionality reduction methods
can influence game playing performance for different
agents.
Overall, we have shown that enabling descriptive
features to build knowledge-based agents is a very
promising route. It yields agents that are not only
comprehensible but that are also able to outperform
state-of-the-art solutions in difficult situations.
REFERENCES
Bellemare, M. G., Naddaf, Y., Veness, J., and Bowling, M.
(2013). The arcade learning environment: An evalua-
tion platform for general agents. Journal of Artificial
Intelligence Research, 47:253–279.
Bishop, C. M. (2006). Pattern recognition and machine
learning. Springer.
Boser, B. E., Guyon, I. M., and Vapnik, V. N. (1992). A
training algorithm for optimal margin classifiers. In
Proceedings of the fifth annual workshop on Computa-
tional learning theory, pages 144–152. ACM.
Breiman, L. (2001). Random forests. Machine learning,
45(1):5–32.
Chang, C.-C. and Lin, C.-J. (2011). Libsvm: A library
for support vector machines. ACM Trans. Intell. Syst.
Technol., 2(3):27:1–27:27.
Coates, A. and Ng, A. Y. (2012). Learning feature represen-
tations with k-means. In Neural Networks: Tricks of
the Trade, pages 561–580. Springer.
Fernández-Delgado, M., Cernadas, E., Barro, S., and
Amorim, D. (2014). Do we need hundreds of clas-
sifiers to solve real world classification problems? The
Journal of Machine Learning Research, 15(1):3133–
3181.
Fix, E. and Hodges Jr, J. L. (1951). Discriminatory analysis-
nonparametric discrimination: consistency properties.
Technical report, DTIC Document.
Friedman, J., Hastie, T., and Tibshirani, R. (2001). The
elements of statistical learning, volume 1. Springer
series in statistics Springer, Berlin.
Hausknecht, M., Khandelwal, P., Miikkulainen, R., and
Stone, P. (2012). Hyperneat-ggp: A hyperneat-based
atari general game player. In Proceedings of the four-
teenth international conference on Genetic and evolu-
tionary computation conference, pages 217–224.
Ho, T. K. (1998). The random subspace method for construct-
ing decision forests. IEEE Transactions on Pattern
Analysis and Machine Intelligence, 20(8):832–844.
Lipovetzky, N., Ramirez, M., and Geffner, H. (2015). Classi-
cal planning with simulators: Results on the atari video
games. Proc. IJCAI 2015.
Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A.,
Antonoglou, I., Wierstra, D., and Riedmiller, M. (2013).
Playing atari with deep reinforcement learning. NIPS
Deep Learning Workshop 2013.
Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A. A., Ve-
ness, J., Bellemare, M. G., Graves, A., Riedmiller, M.,
Fidjeland, A. K., Ostrovski, G., Petersen, S., Beat-
tie, C., Sadik, A., Antonoglou, I., King, H., Kumaran,
D., Wierstra, D., Legg, S., and Hassabis, D. (2015).
Human-level control through deep reinforcement learn-
ing. Nature, 518(7540):529–533.
Naddaf, Y. (2010). Game-independent ai agents for playing
atari 2600 console games. University of Alberta.
Nair, A., Srinivasan, P., and Blackwell, S. (2015). Mas-
sively parallel methods for deep reinforcement learn-
ing. arXiv preprint arXiv:1507.04296.
Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V.,
Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P.,
Weiss, R., Dubourg, V., et al. (2011). Scikit-learn:
Machine learning in python. The Journal of Machine
Learning Research, 12:2825–2830.
Sculley, D. et al. (2011). Results from a semi-supervised
feature learning competition. NIPS 2011 Workshop on
Deep Learning and Unsupervised Feature Learning.
Vapnik, V. and Chapelle, O. (2000). Bounds on error expecta-
tion for support vector machines. Neural computation,
12(9):2013–2036.
ICAART 2017 - 9th International Conference on Agents and Artificial Intelligence
238