Efficient Evolution of Neural Network Topologies
Kenneth
0.
Stanley and Risto Miikkulainen
Department of Computer Sciences
The University of Texas at Austin
Austin, TX 78712
kstanley,
risto@cs.
utexas.edu
Abstract
Neuroevolution, i.e. evolving artificial neural networks
with genetic algorithms, has been highly effective in rein-
forcement learning
tasks,
particularly those with hidden
state information. An important question in neuraevolu-
tion is how to gain an advantage from evolving neural net-
work topologies along with weights. We present a method,
NeuroEvolution of Augmenting Topologies (NEAT) that
outperforms the best fixed-topology methods on a challeng-
ing benchmark reinforcement learning
task.
We claim that
the increased efficiency is due to
(1)
employing a principled
method of crossover of different topologies,
(2)
protecting
structural innovation using speciation, and
(3)
incremen-
tally growing from minimal structure. We test this claim
through a series of ablation studies that demonstrate that
each component is necessary to the system
as
a whole and
to each other. What results is significantly
faster
learning.
NEAT is also an important contribution to GAS because
it shows how it is possible for evolution to both optimize
and
complexify
solutions simultaneously, making it possible
to
evolve increasingly complex solutions over time, thereby
strengthening the analogy with biological evolution.
I. INTRODUCTION
Neuroevolution (NE), the artificial evolution of neural net-
works using genetic algorithms, has shown great promise
in reinforcement learning tasks. NE outperforms standard
reinforcement learning methods in many benchmark tasks
[6,
10,
11
3.
Neural networks are a good class of decision
making systems to evolve because they are capable of repre-
senting solutions to many different kinds of problems, and the
mapping from genotype to phenotype is generally efficient.
NE is particularly well suited to reinforcement learning tasks
because NE does not require supervision.
A major question in NE is how to gain an advantage from
evolving topology in addition to connection weights. On one
hand, evolving topology might overcomplicate the search. On
the other, it can also save time by finding the right number of
hidden neurons
for
a particular problem automatically [7].
A previous study showed that fixed-topology NE can out-
perform a topology-evolving system on the benchmark dou-
ble pole balancing task
[6].
This finding is important because
pole balancing has been a benchmark task in NE and rein-
forcement learning for over
30
years
[l,
6,
7,
91,
and
dou-
ble pole balancing is challenging to even the best of modem
methods. Doing well at this important benchmark suggests
that a method will do well in other tasks as well. Whether
Topology and Weight Evolving Artificial Neural Networks
(TWEANNs) can enhance the performance
of
NE remains
an open question.
In this article, we aim to show that evolving topology can
indeed increase performance. We present a new
TWEANN,
NeuroEvolution of Augmenting Topologies (NEAT), that sig-
nificantly outperforms the fixed-topology NE method that
currently takes the fewest evaluations
on
the double pole
balancing task. We identify three major challenges for
TWEA”s and present solutions
to
each of them:
(1)
Is
there
a genetic representation that allows disparate topologies to
crossover in a meaningful way?
Our
solution is to use his-
torical markings to line up genes with the same origin. (2)
How can topological innovation that needs a few generations
to optimize be protected
so
that it does not disappear from the
population prematurely?
Our
solution is to separate each in-
novation into a different species.
(3)
How can topologies
be
minimized
throughout evolution
without the need for a spe-
cially contrived fitness function that measures complexity?
Our
solution is to start from a minimal structure and grow
only when necessary. This paper establishes that each of our
solutions is necessary by showing that
NE
performance sig-
nificantly declines with the ablation
of
any of the major
so-
lution components. Working together in NEAT these compo-
nents constitute a promising new approach to difficult rein-
forcement learning tasks.
We begin by describing the NEAT method, including re-
sults showing that NEAT is significantly faster than other NE
methods on the hardest pole balancing benchmark. We then
present ablation studies designed to explain NEAT’S perfor-
mance in terms of its components.
11. NEUROEVOLUTION
OF
AUGMENTING
A.
Genetic
Encoding
NEAT is designed specifically to address the three challenges
raised in the introduction. Each genome includes a list of
connection genes,
each of which refers
to
two
node genes
be-
ing connected (figure
l).
Each connection gene specifies the
in-node, the out-node, the weight of the connection, whether
TOPOLOGIES (NEAT)
0-7803-7282-4/02/$10.00 02002
IEEE
1757