Robustness to Adversarial Examples

Adversarial example
  • Deep neural networks are vulnerable to small adversarial perturbations. We study a front end signal processing defense that exploits sparsity of natural data.
Front end defense
  • We show that sparsity-based defenses are provably effective for linear classifers, reducing the impact of $\ell_\infty$-bounded attacks by a factor of roughly K/N, where N is the data dimension and K is the sparsity level.

  • We then extend our approach to deep networks using locally linear modeling.

Publications