This is the public, feature-limited version of the conference webpage. After Registration and login please visit the full version.

Most ReLU Networks Suffer from $\ell^2$ Adversarial Perturbations

Amit Daniely, Hadas Shacham

Spotlight presentation: Orals & Spotlights Track 26: Graph/Relational/Theory
on 2020-12-10T08:10:00-08:00 - 2020-12-10T08:20:00-08:00
Poster Session 6 (more posters)
on 2020-12-10T09:00:00-08:00 - 2020-12-10T11:00:00-08:00
Abstract: We consider ReLU networks with random weights, in which the dimension decreases at each layer. We show that for most such networks, most examples $x$ admit an adversarial perturbation at an Euclidean distance of $O\left(\frac{\|x\|}{\sqrt{d}}\right)$, where $d$ is the input dimension. Moreover, this perturbation can be found via gradient flow, as well as gradient descent with sufficiently small steps. This result can be seen as an explanation to the abundance of adversarial examples, and to the fact that they are found via gradient descent.

Preview Video and Chat

To see video, interact with the author and ask questions please use registration and login.