Challenging the Foundations: Mining Hard Test Samples through Diffusion Generation

15 Sept 2023 (modified: 25 Mar 2024)ICLR 2024 Conference Withdrawn SubmissionEveryoneRevisionsBibTeX
Keywords: Large foundation models, Diffusion models, Vulnerability
Abstract: Large foundation models have achieved tremendous success with impressive performance in multiple applications. However, their performance is often benchmarked on natural images, where novel combinations of specific objects and nuisances can be missing and not tested. In this work, we develop a framework to efficiently probe foundation models for their vulnerabilities with diffusion generation, termed DiffusionExplorer. We show that our framework can efficiently construct a test set with novel combinations of object and nuisance factors to expose the failures of foundation models. Experimental results show that our mined test samples are challenging to foundation models, such as MiniGPT-4 and LLaVa, significantly reducing their accuracy by 29.56\% and 39.96\%, respectively. Our work suggests that generative models can be viewed as an effective data source in finding the vulnerability of large vision foundation models.
Primary Area: datasets and benchmarks
Code Of Ethics: I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics.
Submission Guidelines: I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2024/AuthorGuide.
Anonymous Url: I certify that there is no URL (e.g., github page) that could be used to find authors' identity.
No Acknowledgement Section: I certify that there is no acknowledgement section in this submission for double blind review.
Submission Number: 25
Loading