On UMAP's True Loss FunctionDownload PDF

21 May 2021, 20:42 (edited 14 Jan 2022)NeurIPS 2021 PosterReaders: Everyone
  • Keywords: UMAP, t-SNE, negative sampling, scRNA-seq, unsupervised learning, visualization, non-linear dimension reduction, manifold learning
  • TL;DR: We derive UMAP's true loss function and show that UMAP's high-dimensional similarities are not important.
  • Abstract: UMAP has supplanted $t$-SNE as state-of-the-art for visualizing high-dimensional datasets in many disciplines, but the reason for its success is not well understood. In this work, we investigate UMAP's sampling based optimization scheme in detail. We derive UMAP's true loss function in closed form and find that it differs from the published one in a dataset size dependent way. As a consequence, we show that UMAP does not aim to reproduce its theoretically motivated high-dimensional UMAP similarities. Instead, it tries to reproduce similarities that only encode the $k$ nearest neighbor graph, thereby challenging the previous understanding of UMAP's effectiveness. Alternatively, we consider the implicit balancing of attraction and repulsion due to the negative sampling to be key to UMAP's success. We corroborate our theoretical findings on toy and single cell RNA sequencing data.
  • Supplementary Material: pdf
  • Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
  • Code: https://github.com/hci-unihd/UMAPs-true-loss
13 Replies