TY - GEN
T1 - Evolving large-scale neural networks for vision-based reinforcement learning
AU - Koutník, Jan
AU - Cuccu, Giuseppe
AU - Schmidhuber, Jürgen
AU - Gomez, Faustino
N1 - Generated from Scopus record by KAUST IRTS on 2022-09-14
PY - 2013/9/2
Y1 - 2013/9/2
N2 - The idea of using evolutionary computation to train artificial neural networks, or neuroevolution (NE), for reinforcement learning (RL) tasks has now been around for over 20 years. However, as RL tasks become more challenging, the networks required become larger, so do their genomes. But, scaling NE to large nets (i.e. tens of thousands of weights) is infeasible using direct encodings that map genes one-to-one to network components. In this paper, we scale-up our "compressed" network encoding where network weight matrices are represented indirectly as a set of Fourier-type coefficients, to tasks that require very-large networks due to the high-dimensionality of their input space. The approach is demonstrated successfully on two reinforcement learning tasks in which the control networks receive visual input: (1) a vision-based version of the octopus control task requiring networks with over 3 thousand weights, and (2) a version of the TORCS driving game where networks with over 1 million weights are evolved to drive a car around a track using video images from the driver's perspective. Copyright © 2013 ACM.
AB - The idea of using evolutionary computation to train artificial neural networks, or neuroevolution (NE), for reinforcement learning (RL) tasks has now been around for over 20 years. However, as RL tasks become more challenging, the networks required become larger, so do their genomes. But, scaling NE to large nets (i.e. tens of thousands of weights) is infeasible using direct encodings that map genes one-to-one to network components. In this paper, we scale-up our "compressed" network encoding where network weight matrices are represented indirectly as a set of Fourier-type coefficients, to tasks that require very-large networks due to the high-dimensionality of their input space. The approach is demonstrated successfully on two reinforcement learning tasks in which the control networks receive visual input: (1) a vision-based version of the octopus control task requiring networks with over 3 thousand weights, and (2) a version of the TORCS driving game where networks with over 1 million weights are evolved to drive a car around a track using video images from the driver's perspective. Copyright © 2013 ACM.
UR - http://dl.acm.org/citation.cfm?doid=2463372.2463509
UR - http://www.scopus.com/inward/record.url?scp=84883060087&partnerID=8YFLogxK
U2 - 10.1145/2463372.2463509
DO - 10.1145/2463372.2463509
M3 - Conference contribution
SN - 9781450319638
SP - 1061
EP - 1068
BT - GECCO 2013 - Proceedings of the 2013 Genetic and Evolutionary Computation Conference
ER -