Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Symmetry and Geometry in Neural Representations (NeurReps)

Understanding Optimization Challenges when Encoding to Geometric Structures

Babak Esmaeili · Robin Walters · Heiko Zimmermann · Jan-Willem van de Meent

Keywords: [ Representation Learning ] [ Autoencoders ] [ Lie groups ] [ topological ] [ homeomorphism ] [ isometry ] [ Equivariant ]


Abstract:

Geometric inductive biases such as spatial curvature, factorizability, or equivariance have been shown to enable learning of latent spaces which better reflect the structure of data and perform better on downstream tasks. Training such models, however, can be a challenging task due to the topological constraints imposed by encoding to such structures. In this paper, we theoretically and empirically characterize obstructions to training autoencoders with geometric latent spaces. These include issues such as singularity (e.g. self-intersection), incorrect degree or winding number, and non-isometric homeomorphic embedding. We propose a method, isometric autoencoder, to improve the stability of training and convergence to an isometric mapping in geometric latent spaces. We perform an empirical evaluation of this method over 2 domains, which demonstrates that our approach can better circumvent the identified optimization problems.

Chat is not available.