Do Neural Optimal Transport Solvers Work? A Continuous Wasserstein-2 BenchmarkDownload PDF

May 21, 2021 (edited Oct 20, 2021)NeurIPS 2021 PosterReaders: Everyone
  • Keywords: optimal transport solvers, benchmark, continuous measures, quadratic cost, wasserstein-2 distance, parametric methods, input-convex neural networks, generative modeling
  • TL;DR: We construct pairs of continuous measures with analytically-known optimal transport (OT) solutions for the quadratic cost to test existing parametric OT solvers including those used in implicit generative modeling.
  • Abstract: Despite the recent popularity of neural network-based solvers for optimal transport (OT), there is no standard quantitative way to evaluate their performance. In this paper, we address this issue for quadratic-cost transport---specifically, computation of the Wasserstein-2 distance, a commonly-used formulation of optimal transport in machine learning. To overcome the challenge of computing ground truth transport maps between continuous measures needed to assess these solvers, we use input-convex neural networks (ICNN) to construct pairs of measures whose ground truth OT maps can be obtained analytically. This strategy yields pairs of continuous benchmark measures in high-dimensional spaces such as spaces of images. We thoroughly evaluate existing optimal transport solvers using these benchmark measures. Even though these solvers perform well in downstream tasks, many do not faithfully recover optimal transport maps. To investigate the cause of this discrepancy, we further test the solvers in a setting of image generation. Our study reveals crucial limitations of existing solvers and shows that increased OT accuracy does not necessarily correlate to better results downstream.
  • Supplementary Material: pdf
  • Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
  • Code:
10 Replies