subset selection into the neuroevolutionary process of
NEAT. Instead of evolving networks with the full set
of features as NEAT does, IFSE-NEAT initializes net-
works with one feature. IFSE-NEAT then iteratively
adds features to the current best network that con-
tributes most to its performance improvement while
evolving the weights and topology of that network.
Prior to this work, feature selection for reinforce-
ment learning has focused on linear value function ap-
proximation (Kolter and Ng, 2009; Parr et al., 2008)
and model-based RL algorithms (Kroon and White-
son, 2009). For neuroevolution algorithms such as
NEAT, only random search has been explored (White-
son et al., 2005). In this light we can see that IFSE-
NEAT is a novel approach in feature selection for RL.
Our experimental study has shownseveral promis-
ing results for IFSE-NEAT.We find that the algorithm
is nearly unaffected in its ability to select relevant fea-
tures as the number of irrelevant features grows very
large. This, in turn, allows for a better policy to be
derived than NEAT and FS-NEAT. Additionally, by
using only a few relevant features we are able to learn
a good policy while limiting model complexity.
The rest of the paper is organized as follows: Sec-
tion 2 introduces the NEAT algorithm and the pro-
posed IFSE-NEAT algorithm. Section 3 describes the
experimental setup. Section 4 presents and discusses
experimental results. Finally, Section 5 concludes this
work and identifies some future research directions.
2 APPROACH
2.1 NEAT
Neural networks (NNs) are efficient function approx-
imators that can model complex functions to an arbi-
trary accuracy. The drawbacks of using NNs in RL
domains have been that NN design was a difficult
manual process and training was a supervised learn-
ing process. Neuroevolutionary approaches, which
utilize genetic algorithms to automate the process of
training and/or designing NNs, eliminate these draw-
backs allowing NNs to be easily applied to RL do-
mains. NeuroEvolution of Augmenting Topologies
(NEAT) is a novel RL framework based on neuroevo-
lution. By evolving both the network topology and
weights of the connections between network nodes,
NEAT solved typical RL benchmark problems sev-
eral times faster than competing RL algorithms with
significantly less system resources (Stanley and Mi-
ikkulainen, 2002).
However, one limiting issue with NEAT is that it
assumes that all features provided by the environment
are relevant and necessary, and attempts to incorpo-
rate all the features into its solution networks. The
extraneous features will unnecessarily complicate the
networks and severely slow the rate at which NEAT is
able to derive an effectivepolicy. In the following sec-
tion we describe a new algorithm based upon NEAT
that builds a small set of required features while learn-
ing an effective policy.
2.2 Incremental Feature Selection
Embedded in NEAT
To deal with the exponential search space, we adopt
sequential forward search (SFS), an efficient search
strategy which has proven effective in finding near-
optimal subsets in supervised feature selection. Start-
ing from an empty set, SFS iteratively adds one fea-
ture at a time to the current best set until a desired
number of features k are selected. Since in each of
the k iterations, it goes through all N features outside
of the current best set, the time complexity of SFS is
O(kN). Although SFS does not guarantee the optimal
solution, it is capable of selecting relevant features
while keeping irrelevant or redundant features out of
the final subset. The method is particularly suitable
for high-dimensional problems where a large portion
of the features are irrelevant or redundant.
Algorithm 1 provides a basic overview about how
IFSE-NEAT functions and is able to select a mini-
mal set of features. IFSE-NEAT incrementally adds
features to a NN that we call the BACKBONE. The
BACKBONE network utilizes the best discovered fea-
ture set and represents the current best derived policy.
It is persistent through additions of new features to
the feature set and it is what makes IFSE-NEAT an
embedded algorithm as opposed to a straightforward
wrapper algorithm.
Initially, the BACKBONE network consists of
only the output nodes (line 5). Then, for each of
the individual features available, F
q
, a NN is gener-
ated by connecting a single input node to every output
node (line 13). In parallel, or independently, a popu-
lation of networks based upon this single-input base
network is generated. Each network in the popula-
tion share the topology of the base network, but have
randomly generated weights on the edges joining the
nodes. The population of NNs are then evolved via
the standard NEAT algorithm for L generations (lines
17-19). At the end of the NEAT process, the cham-
pion of each population (the network representing the
best policy) is identified. The champions (each corre-
sponding to a candidate feature F
q
) are then compared
against one another to decide the BEST NETWORK
and BEST FEATURE (lines 22-25). It is our hypoth-
esis that the best performing network, BEST NET-
ICAART 2011 - 3rd International Conference on Agents and Artificial Intelligence
264