Skip to yearly menu bar Skip to main content


Poster

Statistical bounds for entropic optimal transport: sample complexity and the central limit theorem

Gonzalo Mena · Jonathan Niles-Weed

East Exhibition Hall B, C #220

Keywords: [ Learning Theory ] [ Theory ] [ Frequentist Statistics ]


Abstract:

We prove several fundamental statistical bounds for entropic OT with the squared Euclidean cost between subgaussian probability measures in arbitrary dimension. First, through a new sample complexity result we establish the rate of convergence of entropic OT for empirical measures. Our analysis improves exponentially on the bound of Genevay et al.~(2019) and extends their work to unbounded measures. Second, we establish a central limit theorem for entropic OT, based on techniques developed by Del Barrio and Loubes~(2019). Previously, such a result was only known for finite metric spaces. As an application of our results, we develop and analyze a new technique for estimating the entropy of a random variable corrupted by gaussian noise.

Live content is unavailable. Log in and register to view live content