TY - GEN
T1 - Rethinking Neural Operations for Diverse Tasks
AU - Roberts, Nicholas
AU - Khodak, Mikhail
AU - Dao, Tri
AU - Li, Liam
AU - Re, Christopher
AU - Talwalkar, Ameet
N1 - Publisher Copyright:
© 2021 Neural information processing systems foundation. All rights reserved.
PY - 2021
Y1 - 2021
N2 - An important goal of AutoML is to automate-away the design of neural networks on new tasks in under-explored domains. Motivated by this goal, we study the problem of enabling users to discover the right neural operations given data from their specific domain. We introduce a search space of operations called XD-Operations that mimic the inductive bias of standard multi-channel convolutions while being much more expressive: we prove that it includes many named operations across multiple application areas. Starting with any standard backbone such as ResNet, we show how to transform it into a search space over XD-operations and how to traverse the space using a simple weight-sharing scheme. On a diverse set of tasks- solving PDEs, distance prediction for protein folding, and music modeling-our approach consistently yields models with lower error than baseline networks and often even lower error than expert-designed domain-specific approaches.
AB - An important goal of AutoML is to automate-away the design of neural networks on new tasks in under-explored domains. Motivated by this goal, we study the problem of enabling users to discover the right neural operations given data from their specific domain. We introduce a search space of operations called XD-Operations that mimic the inductive bias of standard multi-channel convolutions while being much more expressive: we prove that it includes many named operations across multiple application areas. Starting with any standard backbone such as ResNet, we show how to transform it into a search space over XD-operations and how to traverse the space using a simple weight-sharing scheme. On a diverse set of tasks- solving PDEs, distance prediction for protein folding, and music modeling-our approach consistently yields models with lower error than baseline networks and often even lower error than expert-designed domain-specific approaches.
UR - http://www.scopus.com/inward/record.url?scp=85129543564&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85129543564&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85129543564
T3 - Advances in Neural Information Processing Systems
SP - 15855
EP - 15869
BT - Advances in Neural Information Processing Systems 34 - 35th Conference on Neural Information Processing Systems, NeurIPS 2021
A2 - Ranzato, Marc'Aurelio
A2 - Beygelzimer, Alina
A2 - Dauphin, Yann
A2 - Liang, Percy S.
A2 - Wortman Vaughan, Jenn
PB - Neural information processing systems foundation
T2 - 35th Conference on Neural Information Processing Systems, NeurIPS 2021
Y2 - 6 December 2021 through 14 December 2021
ER -