Learning Generative Models with Sinkhorn Divergences. Genevay, A., Peyré, G., & Cuturi, M. arXiv:1706.00292 [stat], October, 2017. arXiv: 1706.00292Paper abstract bibtex The ability to compare two degenerate probability distributions (i.e. two probability distributions supported on two distinct low-dimensional manifolds living in a much higher-dimensional space) is a crucial problem arising in the estimation of generative models for high-dimensional observations such as those arising in computer vision or natural language. It is known that optimal transport metrics can represent a cure for this problem, since they were specifically designed as an alternative to information divergences to handle such problematic scenarios. Unfortunately, training generative machines using OT raises formidable computational and statistical challenges, because of (i) the computational burden of evaluating OT losses, (ii) the instability and lack of smoothness of these losses, (iii) the difficulty to estimate robustly these losses and their gradients in high dimension. This paper presents the first tractable computational method to train large scale generative models using an optimal transport loss, and tackles these three issues by relying on two key ideas: (a) entropic smoothing, which turns the original OT loss into one that can be computed using Sinkhorn fixed point iterations; (b) algorithmic (automatic) differentiation of these iterations. These two approximations result in a robust and differentiable approximation of the OT loss with streamlined GPU execution. Entropic smoothing generates a family of losses interpolating between Wasserstein (OT) and Maximum Mean Discrepancy (MMD), thus allowing to find a sweet spot leveraging the geometry of OT and the favorable high-dimensional sample complexity of MMD which comes with unbiased gradient estimates. The resulting computational architecture complements nicely standard deep network generative models by a stack of extra layers implementing the loss function.
@article{genevay_learning_2017,
title = {Learning {Generative} {Models} with {Sinkhorn} {Divergences}},
url = {http://arxiv.org/abs/1706.00292},
abstract = {The ability to compare two degenerate probability distributions (i.e. two probability distributions supported on two distinct low-dimensional manifolds living in a much higher-dimensional space) is a crucial problem arising in the estimation of generative models for high-dimensional observations such as those arising in computer vision or natural language. It is known that optimal transport metrics can represent a cure for this problem, since they were specifically designed as an alternative to information divergences to handle such problematic scenarios. Unfortunately, training generative machines using OT raises formidable computational and statistical challenges, because of (i) the computational burden of evaluating OT losses, (ii) the instability and lack of smoothness of these losses, (iii) the difficulty to estimate robustly these losses and their gradients in high dimension. This paper presents the first tractable computational method to train large scale generative models using an optimal transport loss, and tackles these three issues by relying on two key ideas: (a) entropic smoothing, which turns the original OT loss into one that can be computed using Sinkhorn fixed point iterations; (b) algorithmic (automatic) differentiation of these iterations. These two approximations result in a robust and differentiable approximation of the OT loss with streamlined GPU execution. Entropic smoothing generates a family of losses interpolating between Wasserstein (OT) and Maximum Mean Discrepancy (MMD), thus allowing to find a sweet spot leveraging the geometry of OT and the favorable high-dimensional sample complexity of MMD which comes with unbiased gradient estimates. The resulting computational architecture complements nicely standard deep network generative models by a stack of extra layers implementing the loss function.},
urldate = {2022-03-02},
journal = {arXiv:1706.00292 [stat]},
author = {Genevay, Aude and Peyré, Gabriel and Cuturi, Marco},
month = oct,
year = {2017},
note = {arXiv: 1706.00292},
keywords = {Statistics - Machine Learning},
}
Downloads: 0
{"_id":"MNA9fCJCew9ZxhAxj","bibbaseid":"genevay-peyr-cuturi-learninggenerativemodelswithsinkhorndivergences-2017","author_short":["Genevay, A.","Peyré, G.","Cuturi, M."],"bibdata":{"bibtype":"article","type":"article","title":"Learning Generative Models with Sinkhorn Divergences","url":"http://arxiv.org/abs/1706.00292","abstract":"The ability to compare two degenerate probability distributions (i.e. two probability distributions supported on two distinct low-dimensional manifolds living in a much higher-dimensional space) is a crucial problem arising in the estimation of generative models for high-dimensional observations such as those arising in computer vision or natural language. It is known that optimal transport metrics can represent a cure for this problem, since they were specifically designed as an alternative to information divergences to handle such problematic scenarios. Unfortunately, training generative machines using OT raises formidable computational and statistical challenges, because of (i) the computational burden of evaluating OT losses, (ii) the instability and lack of smoothness of these losses, (iii) the difficulty to estimate robustly these losses and their gradients in high dimension. This paper presents the first tractable computational method to train large scale generative models using an optimal transport loss, and tackles these three issues by relying on two key ideas: (a) entropic smoothing, which turns the original OT loss into one that can be computed using Sinkhorn fixed point iterations; (b) algorithmic (automatic) differentiation of these iterations. These two approximations result in a robust and differentiable approximation of the OT loss with streamlined GPU execution. Entropic smoothing generates a family of losses interpolating between Wasserstein (OT) and Maximum Mean Discrepancy (MMD), thus allowing to find a sweet spot leveraging the geometry of OT and the favorable high-dimensional sample complexity of MMD which comes with unbiased gradient estimates. The resulting computational architecture complements nicely standard deep network generative models by a stack of extra layers implementing the loss function.","urldate":"2022-03-02","journal":"arXiv:1706.00292 [stat]","author":[{"propositions":[],"lastnames":["Genevay"],"firstnames":["Aude"],"suffixes":[]},{"propositions":[],"lastnames":["Peyré"],"firstnames":["Gabriel"],"suffixes":[]},{"propositions":[],"lastnames":["Cuturi"],"firstnames":["Marco"],"suffixes":[]}],"month":"October","year":"2017","note":"arXiv: 1706.00292","keywords":"Statistics - Machine Learning","bibtex":"@article{genevay_learning_2017,\n\ttitle = {Learning {Generative} {Models} with {Sinkhorn} {Divergences}},\n\turl = {http://arxiv.org/abs/1706.00292},\n\tabstract = {The ability to compare two degenerate probability distributions (i.e. two probability distributions supported on two distinct low-dimensional manifolds living in a much higher-dimensional space) is a crucial problem arising in the estimation of generative models for high-dimensional observations such as those arising in computer vision or natural language. It is known that optimal transport metrics can represent a cure for this problem, since they were specifically designed as an alternative to information divergences to handle such problematic scenarios. Unfortunately, training generative machines using OT raises formidable computational and statistical challenges, because of (i) the computational burden of evaluating OT losses, (ii) the instability and lack of smoothness of these losses, (iii) the difficulty to estimate robustly these losses and their gradients in high dimension. This paper presents the first tractable computational method to train large scale generative models using an optimal transport loss, and tackles these three issues by relying on two key ideas: (a) entropic smoothing, which turns the original OT loss into one that can be computed using Sinkhorn fixed point iterations; (b) algorithmic (automatic) differentiation of these iterations. These two approximations result in a robust and differentiable approximation of the OT loss with streamlined GPU execution. Entropic smoothing generates a family of losses interpolating between Wasserstein (OT) and Maximum Mean Discrepancy (MMD), thus allowing to find a sweet spot leveraging the geometry of OT and the favorable high-dimensional sample complexity of MMD which comes with unbiased gradient estimates. The resulting computational architecture complements nicely standard deep network generative models by a stack of extra layers implementing the loss function.},\n\turldate = {2022-03-02},\n\tjournal = {arXiv:1706.00292 [stat]},\n\tauthor = {Genevay, Aude and Peyré, Gabriel and Cuturi, Marco},\n\tmonth = oct,\n\tyear = {2017},\n\tnote = {arXiv: 1706.00292},\n\tkeywords = {Statistics - Machine Learning},\n}\n\n","author_short":["Genevay, A.","Peyré, G.","Cuturi, M."],"key":"genevay_learning_2017","id":"genevay_learning_2017","bibbaseid":"genevay-peyr-cuturi-learninggenerativemodelswithsinkhorndivergences-2017","role":"author","urls":{"Paper":"http://arxiv.org/abs/1706.00292"},"keyword":["Statistics - Machine Learning"],"metadata":{"authorlinks":{}},"html":""},"bibtype":"article","biburl":"https://bibbase.org/zotero/mxmplx","dataSources":["aXmRAq63YsH7a3ufx"],"keywords":["statistics - machine learning"],"search_terms":["learning","generative","models","sinkhorn","divergences","genevay","peyré","cuturi"],"title":"Learning Generative Models with Sinkhorn Divergences","year":2017}