TY - JOUR
T1 - HYDRA
T2 - 34th Conference on Neural Information Processing Systems, NeurIPS 2020
AU - Sehwag, Vikash
AU - Wang, Shiqi
AU - Mittal, Prateek
AU - Jana, Suman
N1 - Funding Information:
This work was supported in part by the National Science Foundation under grants CNS-1553437, CNS-1704105, by Qualcomm Innovation Fellowship, by the Army Research Office Young Investigator Prize, by Schmidt DataX Fund, by Princeton E-ffiliates Partnership, by Army Research Laboratory (ARL) Army Artificial Intelligence Institute (A2I2), by Facebook Systems for ML award, by a Google Faculty Fellowship, by a Capital One Research Grant, by an ARL Young Investigator Award, by the Office of Naval Research Young Investigator Award, by a J.P. Morgan Faculty Award, and by two NSF CAREER Awards.
Publisher Copyright:
© 2020 Neural information processing systems foundation. All rights reserved.
PY - 2020
Y1 - 2020
N2 - In safety-critical but computationally resource-constrained applications, deep learning faces two key challenges: lack of robustness against adversarial attacks and large neural network size (often millions of parameters). While the research community has extensively explored the use of robust training and network pruning independently to address one of these challenges, only a few recent works have studied them jointly. However, these works inherit a heuristic pruning strategy that was developed for benign training, which performs poorly when integrated with robust training techniques, including adversarial training and verifiable robust training. To overcome this challenge, we propose to make pruning techniques aware of the robust training objective and let the training objective guide the search for which connections to prune. We realize this insight by formulating the pruning objective as an empirical risk minimization problem which is solved efficiently using SGD. We demonstrate that our approach, titled HYDRA1, achieves compressed networks with state-of-the-art benign and robust accuracy, simultaneously. We demonstrate the success of our approach across CIFAR-10, SVHN, and ImageNet dataset with four robust training techniques: iterative adversarial training, randomized smoothing, MixTrain, and CROWN-IBP. We also demonstrate the existence of highly robust sub-networks within non-robust networks.
AB - In safety-critical but computationally resource-constrained applications, deep learning faces two key challenges: lack of robustness against adversarial attacks and large neural network size (often millions of parameters). While the research community has extensively explored the use of robust training and network pruning independently to address one of these challenges, only a few recent works have studied them jointly. However, these works inherit a heuristic pruning strategy that was developed for benign training, which performs poorly when integrated with robust training techniques, including adversarial training and verifiable robust training. To overcome this challenge, we propose to make pruning techniques aware of the robust training objective and let the training objective guide the search for which connections to prune. We realize this insight by formulating the pruning objective as an empirical risk minimization problem which is solved efficiently using SGD. We demonstrate that our approach, titled HYDRA1, achieves compressed networks with state-of-the-art benign and robust accuracy, simultaneously. We demonstrate the success of our approach across CIFAR-10, SVHN, and ImageNet dataset with four robust training techniques: iterative adversarial training, randomized smoothing, MixTrain, and CROWN-IBP. We also demonstrate the existence of highly robust sub-networks within non-robust networks.
UR - http://www.scopus.com/inward/record.url?scp=85099180835&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85099180835&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85099180835
SN - 1049-5258
VL - 2020-December
JO - Advances in Neural Information Processing Systems
JF - Advances in Neural Information Processing Systems
Y2 - 6 December 2020 through 12 December 2020
ER -