Why We Need New Benchmarks for Local Intrinsic Dimension Estimation

Published: 26 Jan 2026, Last Modified: 01 Mar 2026ICLR 2026 PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Local intrinsic dimension estimation, LIDL, FLIPD, Diffusion Models, Benhamark, Normalizing Flows, ESS, Normal Bundle, NB, LID
TL;DR: We show that LID estimation community needs new benchmarks for intrinsic dimension estimation and come to interesting conclusions on the performance of existing algorithms.
Abstract: Neural Local Intrinsic Dimension (LID) estimators are typically bound to domain-specific architectures whose inductive biases can yield inconsistent estimates for the same underlying manifold. Existing evaluations either use overly simple synthetic data (with known LID) or real datasets (with unknown LID), obscuring true performance. We introduce a principled benchmarking framework that (i) maps the same manifold into multiple domain representations while preserving its structure, enabling like-for-like cross-architecture tests; (ii) designs harder variants of popular datasets that target key manifold properties; and (iii) applies controlled transformations with known LID shifts to stress-test methods even when absolute LID is unknown. Across this suite, including non-trivial synthetic datasets, we show that accuracy on simple manifolds does not transfer across domains and that state-of-the-art methods fail under targeted stressors, revealing clear failure modes and areas for improvement. Data and code are available: https://github.com/DominikFilipiak/LID-Benchmarks.
Supplementary Material: zip
Primary Area: datasets and benchmarks
Submission Number: 25138
Loading