Poster
Exponential Separations in Symmetric Neural Networks
Aaron Zweig · Joan Bruna
Hall J (level 1) #815
Keywords: [ set-based ] [ self-attention ] [ symmetric function ] [ relational network ] [ separation ] [ deepsets ]
Abstract:
In this work we demonstrate a novel separation between symmetric neural network architectures. Specifically, we consider the Relational Network~\parencite{santoro2017simple} architecture as a natural generalization of the DeepSets~\parencite{zaheer2017deep} architecture, and study their representational gap. Under the restriction to analytic activation functions, we construct a symmetric function acting on sets of size with elements in dimension , which can be efficiently approximated by the former architecture, but provably requires width exponential in and for the latter.
Chat is not available.