Learning Embeddings into Entropic Wasserstein Spaces. Frogner, C., Mirzazadeh, F., & Solomon, J.
Learning Embeddings into Entropic Wasserstein Spaces [link]Paper  abstract   bibtex   
Euclidean embeddings of data are fundamentally limited in their ability to capture latent semantic structures, which need not conform to Euclidean spatial assumptions. Here we consider an alternative, which embeds data as discrete probability distributions in a Wasserstein space, endowed with an optimal transport metric. Wasserstein spaces are much larger and more flexible than Euclidean spaces, in that they can successfully embed a wider variety of metric structures. We exploit this flexibility by learning an embedding that captures semantic information in the Wasserstein distance between embedded distributions. We examine empirically the representational capacity of our learned Wasserstein embeddings, showing that they can embed a wide variety of metric structures with smaller distortion than an equivalent Euclidean embedding. We also investigate an application to word embedding, demonstrating a unique advantage of Wasserstein embeddings: We can visualize the high-dimensional embedding directly, since it is a probability distribution on a low-dimensional space. This obviates the need for dimensionality reduction techniques like t-SNE for visualization.
@article{frognerLearningEmbeddingsEntropic2019,
  archivePrefix = {arXiv},
  eprinttype = {arxiv},
  eprint = {1905.03329},
  primaryClass = {cs, stat},
  title = {Learning {{Embeddings}} into {{Entropic Wasserstein Spaces}}},
  url = {http://arxiv.org/abs/1905.03329},
  abstract = {Euclidean embeddings of data are fundamentally limited in their ability to capture latent semantic structures, which need not conform to Euclidean spatial assumptions. Here we consider an alternative, which embeds data as discrete probability distributions in a Wasserstein space, endowed with an optimal transport metric. Wasserstein spaces are much larger and more flexible than Euclidean spaces, in that they can successfully embed a wider variety of metric structures. We exploit this flexibility by learning an embedding that captures semantic information in the Wasserstein distance between embedded distributions. We examine empirically the representational capacity of our learned Wasserstein embeddings, showing that they can embed a wide variety of metric structures with smaller distortion than an equivalent Euclidean embedding. We also investigate an application to word embedding, demonstrating a unique advantage of Wasserstein embeddings: We can visualize the high-dimensional embedding directly, since it is a probability distribution on a low-dimensional space. This obviates the need for dimensionality reduction techniques like t-SNE for visualization.},
  urldate = {2019-05-10},
  date = {2019-05-08},
  keywords = {Statistics - Machine Learning,Computer Science - Machine Learning},
  author = {Frogner, Charlie and Mirzazadeh, Farzaneh and Solomon, Justin},
  file = {/home/dimitri/Nextcloud/Zotero/storage/UAEFXQR2/Frogner et al. - 2019 - Learning Embeddings into Entropic Wasserstein Spac.pdf;/home/dimitri/Nextcloud/Zotero/storage/AY98Y254/1905.html}
}

Downloads: 0