TY - GEN
T1 - Dynamic deep neural networks
T2 - 32nd AAAI Conference on Artificial Intelligence, AAAI 2018
AU - Liu, Lanlan
AU - Deng, Jia
N1 - Publisher Copyright:
Copyright © 2018, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.
PY - 2018
Y1 - 2018
N2 - We introduce Dynamic Deep Neural Networks (D 2 NN), a new type of feed-forward deep neural network that allows selective execution. Given an input, only a subset of D 2 NN neurons are executed, and the particular subset is determined by the D 2 NN itself. By pruning unnecessary computation depending on input, D 2 NNs provide a way to improve computational efficiency. To achieve dynamic selective execution, a D 2 NN augments a feed-forward deep neural network (directed acyclic graph of differentiable modules) with controller modules. Each controller module is a sub-network whose output is a decision that controls whether other modules can execute. A D 2 NN is trained end to end. Both regular and controller modules in a D 2 NN are learnable and are jointly trained to optimize both accuracy and efficiency. Such training is achieved by integrating backpropagation with reinforcement learning. With extensive experiments of various D 2 NN architectures on image classification tasks, we demonstrate that D 2 NNs are general and flexible, and can effectively optimize accuracy-efficiency trade-offs.
AB - We introduce Dynamic Deep Neural Networks (D 2 NN), a new type of feed-forward deep neural network that allows selective execution. Given an input, only a subset of D 2 NN neurons are executed, and the particular subset is determined by the D 2 NN itself. By pruning unnecessary computation depending on input, D 2 NNs provide a way to improve computational efficiency. To achieve dynamic selective execution, a D 2 NN augments a feed-forward deep neural network (directed acyclic graph of differentiable modules) with controller modules. Each controller module is a sub-network whose output is a decision that controls whether other modules can execute. A D 2 NN is trained end to end. Both regular and controller modules in a D 2 NN are learnable and are jointly trained to optimize both accuracy and efficiency. Such training is achieved by integrating backpropagation with reinforcement learning. With extensive experiments of various D 2 NN architectures on image classification tasks, we demonstrate that D 2 NNs are general and flexible, and can effectively optimize accuracy-efficiency trade-offs.
UR - http://www.scopus.com/inward/record.url?scp=85056780727&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85056780727&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85056780727
T3 - 32nd AAAI Conference on Artificial Intelligence, AAAI 2018
SP - 3675
EP - 3682
BT - 32nd AAAI Conference on Artificial Intelligence, AAAI 2018
PB - AAAI press
Y2 - 2 February 2018 through 7 February 2018
ER -