Frank R. Schmidt

Postdoctoral Researcher


Scaling provable adversarial defenses

Eric Wong, Frank R. Schmidt, Jan Hendrik Metzen, J. Zico Kolter
Advances in Neural Information Processing Systems (NIPS) - Dec 2018
Download the publication : 445 KB 
Recent work has developed methods for learning deep network classifiers that are provably robust to norm-bounded adversarial perturbation; however, these methods are currently only possible for relatively small feedforward networks. In this paper, in an effort to scale these approaches to ubstantially larger models, we extend previous work in three main directions. First, we present a technique for extending these training procedures to much more general networks, with skip connections (such as ResNets) and general nonlinearities; the approach is fully modular, and can be implemented automatically (analogous to automatic differentiation). Second, in the specific case of l∞ adversarial perturbations and networks with ReLU nonlinearities, we adopt a nonlinear random projection for training, which scales linearly in the number of hidden units (previous approaches scaled quadratically). Third, we show how to further improve robust error through cascade models. On both MNIST and CIFAR data sets, we train classifiers that improve substantially on the state of the art in provable robust adversarial error bounds: from 5.8% to 3.1% on MNIST (with l∞ perturbations of ε=0.1), and from 80% to 36.4% on CIFAR (with l∞ perturbations of ε=2/255).

For a more recent version please visit: arXiv

Images and movies

teaser.png (11 KB)

BibTex references

  author       = "Wong, Eric and Schmidt, Frank R. and Metzen, Jan Hendrik and Kolter, J. Zico",
  title        = "Scaling provable adversarial defenses",
  booktitle    = "Advances in Neural Information Processing Systems (NIPS)",
  month        = "Dec",
  year         = "2018",
  url          = ""