Conference item icon

Conference item

Statistical bounds for entropic optimal transport: sample complexity and the central limit theorem

Alternative title:
Conference poster
Abstract:
We prove several fundamental statistical bounds for entropic OT with the squared Euclidean cost between subgaussian probability measures in arbitrary dimension. First, through a new sample complexity result we establish the rate of convergence of entropic OT for empirical measures. Our analysis improves exponentially on the bound of Genevay et al.~(2019) and extends their work to unbounded measures. Second, we establish a central limit theorem for entropic OT, based on techniques developed by Del Barrio and Loubes~(2019). Previously, such a result was only known for finite metric spaces. As an application of our results, we develop and analyze a new technique for estimating the entropy of a random variable corrupted by gaussian noise.
Publication status:
Published
Peer review status:
Peer reviewed

Actions


Access Document


Authors


More by this author
Institution:
University of Oxford
Division:
MPLS
Department:
Statistics
Role:
Author
ORCID:
0000-0003-4432-9679


Publisher:
MIT Press
Host title:
Advances in Neural Information Processing Systems 32 (NIPS 2019)
Publication date:
2019-12-12
Acceptance date:
2019-09-20
Event title:
2019 Advances in Neural Information Processing Systems (32nd NeuIPS)
Event series:
Vancouver, Canada
Event website:
https://nips.cc/Conferences/2019
Event start date:
2019-12-08
Event end date:
2019-12-14
ISSN:
1049-5258


Language:
English
Keywords:
Pubs id:
1136141
Local pid:
pubs:1136141
Deposit date:
2020-10-05

Terms of use



Views and Downloads






If you are the owner of this record, you can report an update to it here: Report update to this record

TO TOP