Timezone: »
The strong lottery ticket hypothesis has highlighted the potential for training deep neural networks by pruning, which has inspired interesting practical and theoretical insights into how neural networks can represent functions. For networks with ReLU activation functions, it has been proven that a target network with depth L can be approximated by the subnetwork of a randomly initialized neural network that has double the target's depth 2L and is wider by a logarithmic factor. We show that a depth L+1 is sufficient. This result indicates that we can expect to find lottery tickets at realistic, commonly used depths while only requiring logarithmic overparametrization. Our novel construction approach applies to a large class of activation functions and is not limited to ReLUs. Code is available on Github (RelationalML/LT-existence).
Author Information
Rebekka Burkholz (CISPA, Helmholtz Center, Saarland Informatics Campus)
More from the Same Authors
-
2022 : Using hierarchical variational autoencoders to incorporate conditional independent priors for paired single-cell multi-omics data integration »
Ping-Han Hsieh · Ru-Xiu Hsiao · Tatiana Belova · Katalin Ferenc · Anthony Mathelier · Rebekka Burkholz · Chien-Yu Chen · Geir Kjetil Sandve · Marieke L Kuijjer -
2022 Spotlight: Most Activation Functions Can Win the Lottery Without Excessive Depth »
Rebekka Burkholz -
2021 Poster: Scaling up Continuous-Time Markov Chains Helps Resolve Underspecification »
Alkis Gotovos · Rebekka Burkholz · John Quackenbush · Stefanie Jegelka -
2019 Poster: Initialization of ReLUs for Dynamical Isometry »
Rebekka Burkholz · Alina Dubatovka