Poster
Your contrastive learning problem is secretly a distribution alignment problem
Zihao Chen · Chi-Heng Lin · Ran Liu · Jingyun Xiao · Eva Dyer
East Exhibit Hall A-C #2303
Despite the success of contrastive learning (CL) in vision and language, its theoretical foundations and mechanisms for arranging representations remain poorly understood. In this work, we build connections between noise contrastive estimation losses widely used in CL and distribution alignment with entropic optimal transport (OT). This connection allows us to develop a family of different losses and multistep variants for existing CL methods. Intuitively, by using more information from the distribution of latents, our approach allows a more distribution-aware manipulation of the relationships within augmented sample sets. We provide theoretical insights and experimental evidence demonstrating the benefits of our approach for generalized contrastive alignment. Through this framework, it is possible to leverage tools in OT to build unbalanced losses to handle noisy views and customize the representation space by changing the constraints on alignment. By reframing contrastive learning as an alignment problem and leveraging existing optimization tools for OT, our work provides new insights and connections between different self-supervised learning models in addition to new tools that can be more easily adapted to incorporate domain knowledge into learning.
Live content is unavailable. Log in and register to view live content