Timezone: »

Approximate Feature Collisions in Neural Nets
Ke Li · Tianhao Zhang · Jitendra Malik

Tue Dec 10 10:45 AM -- 12:45 PM (PST) @ East Exhibition Hall B + C #161

Work on adversarial examples has shown that neural nets are surprisingly sensitive to adversarially chosen changes of small magnitude. In this paper, we show the opposite: neural nets could be surprisingly insensitive to adversarially chosen changes of large magnitude. We observe that this phenomenon can arise from the intrinsic properties of the ReLU activation function. As a result, two very different examples could share the same feature activation and therefore the same classification decision. We refer to this phenomenon as feature collision and the corresponding examples as colliding examples. We find that colliding examples are quite abundant: we empirically demonstrate the existence of polytopes of approximately colliding examples in the neighbourhood of practically any example.

Author Information

Ke Li (UC Berkeley)
Tianhao Zhang (Nanjing University)
Jitendra Malik (University of California at Berkley)

More from the Same Authors