On the Matter of Embeddings Dispersion on Hyperspheres

Published: 17 Jun 2024, Last Modified: 10 Jul 2024ICML 2024 Workshop GRaMEveryoneRevisionsBibTeXCC BY 4.0
Track: Extended abstract
Keywords: embeddings, representation learning, learning embeddings on hypersphere, dispersion, maximum separation
Abstract: Dispersion of the embeddings on the $d$-dimensional hypersphere is a process of finding a configuration that preserves semantic information while pushing unrelated vectors away from each other without the need for negative examples. Such a formulation can be connected to the finding configuration of the points such that the minimum distance between two distinct points is maximal, which is a well-known open mathematical problem called the Tammes problem. When dealing with high-dimensional spaces and extremely large numbers of points, as in the text embeddings learning, there is typically no optimal solution, contrary to the Tammes problem, where the optimal solution exists for particular values of $N$ and $d$. Moreover, embeddings learning is mostly done in Euclidean space, which is at odds with the goal of directional dispersion. In this work, we revisit existing algorithms and propose new ones to find a sub-optimal solution for embeddings dispersion by defining the Riemannian optimization problem on the hypersphere.
Submission Number: 42
Loading