TY - GEN
T1 - Schema networks
T2 - 34th International Conference on Machine Learning, ICML 2017
AU - Kansky, Ken
AU - Silver, Tom
AU - Mély, David A.
AU - Eldawy, Mohamed
AU - Lázaro-Gredilla, Miguel
AU - Lou, Xinghua
AU - Dorfman, Nimrod
AU - Sidor, Szymon
AU - Phoenix, Scott
AU - George, Dileep
N1 - Publisher Copyright:
© Copyright 2017 by the author(s).
PY - 2017
Y1 - 2017
N2 - The recent adaptation of deep neural network-based methods to reinforcement learning and planning domains has yielded remarkable progress on individual tasks. Nonetheless, progress on task-to-task transfer remains limited. In pursuit of efficient and robust generalization, we introduce the Schema Network, an object-oriented generative physics simulator capable of disentangling multiple causes of events and reasoning backward through causes to achieve goals. The richly structured architecture of the Schema Network can learn the dynamics of an environment directly from data. We compare Schema Networks with Asynchronous Advantage Actor-Critic and Progressive Networks on a suite of Breakout variations, reporting results on training efficiency and zero-shot generalization, consistently demonstrating faster, more robust learning and better transfer. We argue that generalizing from limited data and learning causal relationships are essential abilities on the path toward generally intelligent systems.
AB - The recent adaptation of deep neural network-based methods to reinforcement learning and planning domains has yielded remarkable progress on individual tasks. Nonetheless, progress on task-to-task transfer remains limited. In pursuit of efficient and robust generalization, we introduce the Schema Network, an object-oriented generative physics simulator capable of disentangling multiple causes of events and reasoning backward through causes to achieve goals. The richly structured architecture of the Schema Network can learn the dynamics of an environment directly from data. We compare Schema Networks with Asynchronous Advantage Actor-Critic and Progressive Networks on a suite of Breakout variations, reporting results on training efficiency and zero-shot generalization, consistently demonstrating faster, more robust learning and better transfer. We argue that generalizing from limited data and learning causal relationships are essential abilities on the path toward generally intelligent systems.
UR - https://www.scopus.com/pages/publications/85048412662
UR - https://www.scopus.com/inward/citedby.url?scp=85048412662&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85048412662
T3 - 34th International Conference on Machine Learning, ICML 2017
SP - 2879
EP - 2889
BT - 34th International Conference on Machine Learning, ICML 2017
PB - International Machine Learning Society (IMLS)
Y2 - 6 August 2017 through 11 August 2017
ER -