Most ReLU Networks Suffer from $\ell^2$ Adversarial Perturbations
Amit Daniely, Hadas Shacham
Spotlight presentation: Orals & Spotlights Track 26: Graph/Relational/Theory
on Thu, Dec 10th, 2020 @ 16:10 – 16:20 GMT
on Thu, Dec 10th, 2020 @ 16:10 – 16:20 GMT
Poster Session 6 (more posters)
on Thu, Dec 10th, 2020 @ 17:00 – 19:00 GMT
GatherTown: Theory ( Town D0 - Spot B1 )
on Thu, Dec 10th, 2020 @ 17:00 – 19:00 GMT
GatherTown: Theory ( Town D0 - Spot B1 )
Join GatherTown
Only iff poster is crowded, join Zoom . Authors have to start the Zoom call from their Profile page / Presentation History.
Only iff poster is crowded, join Zoom . Authors have to start the Zoom call from their Profile page / Presentation History.
Toggle Abstract Paper (in Proceedings / .pdf)
Abstract: We consider ReLU networks with random weights, in which the dimension decreases at each layer. We show that for most such networks, most examples $x$ admit an adversarial perturbation at an Euclidean distance of $O\left(\frac{\|x\|}{\sqrt{d}}\right)$, where $d$ is the input dimension. Moreover, this perturbation can be found via gradient flow, as well as gradient descent with sufficiently small steps. This result can be seen as an explanation to the abundance of adversarial examples, and to the fact that they are found via gradient descent.