Correcting Flaws in Common Disentanglement Metrics

TMLR Paper2320 Authors

01 Mar 2024 (modified: 14 Apr 2024)Under review for TMLREveryoneRevisionsBibTeX
Abstract: Disentangled representations are those in which distinct features, such as size or shape, are represented by distinct neurons. Quantifying the extent to which a given representation is disentangled is not straightforward; multiple metrics have been proposed. In this paper, we identify two failings of existing metrics, which mean they can assign a high score to a model which is still entangled, and we propose two new metrics, which redress these problems. First, we use hypothetical toy examples to demonstrate the failure modes we identify for existing metrics. Then, we show that similar situations occur in practice. Finally, we validate our metrics on the downstream task of compositional generalization. We measure the performance of six existing disentanglement models on this downstream compositional generalization task, and show that performance is (a) generally quite poor, (b) correlated, to varying degrees, with most disentanglement metrics, and (c) most strongly correlated with our newly proposed metrics. Anonymous code to reproduce our results is available at https://github.com/anon296/anon.
Submission Length: Regular submission (no more than 12 pages of main content)
Assigned Action Editor: ~Simon_Kornblith1
Submission Number: 2320
Loading