TY - GEN
T1 - Co-evolving recurrent neurons learn deep memory POMDPs
AU - Gomez, Faustino J.
AU - Schmidhuber, Jürgen
N1 - Generated from Scopus record by KAUST IRTS on 2022-09-14
PY - 2005/12/1
Y1 - 2005/12/1
N2 - Recurrent neural networks are theoretically capable of learning complex temporal sequences, but training them through gradient-descent is too slow and unstable for practical use in reinforcement learning environments. Neuroevolution, the evolution of artificial neural networks using genetic algorithms, can potentially solve real-world reinforcement learning tasks that require deep use of memory, i.e. memory spanning hundreds or thousands of inputs, by searching the space of recurrent neural networks directly. In this paper, we introduce a new neuroevolution algorithm called Hierarchical Enforced SubPopulations that simultaneously evolves networks at two levels of granularity: full networks and network components or neurons. We demonstrate the method in two POMDP tasks that involve temporal dependencies of up to thousands of time-steps, and show that it is faster and simpler than the current best conventional reinforcement learning system on these tasks. Copyright 2005 ACM.
AB - Recurrent neural networks are theoretically capable of learning complex temporal sequences, but training them through gradient-descent is too slow and unstable for practical use in reinforcement learning environments. Neuroevolution, the evolution of artificial neural networks using genetic algorithms, can potentially solve real-world reinforcement learning tasks that require deep use of memory, i.e. memory spanning hundreds or thousands of inputs, by searching the space of recurrent neural networks directly. In this paper, we introduce a new neuroevolution algorithm called Hierarchical Enforced SubPopulations that simultaneously evolves networks at two levels of granularity: full networks and network components or neurons. We demonstrate the method in two POMDP tasks that involve temporal dependencies of up to thousands of time-steps, and show that it is faster and simpler than the current best conventional reinforcement learning system on these tasks. Copyright 2005 ACM.
UR - http://portal.acm.org/citation.cfm?doid=1068009.1068092
UR - http://www.scopus.com/inward/record.url?scp=32444448207&partnerID=8YFLogxK
U2 - 10.1145/1068009.1068092
DO - 10.1145/1068009.1068092
M3 - Conference contribution
SN - 1595930108
SP - 491
EP - 498
BT - GECCO 2005 - Genetic and Evolutionary Computation Conference
ER -