LLaVA-OneVision: Easy Visual Task Transfer

TMLR Paper3432 Authors

04 Oct 2024 (modified: 20 Oct 2024)Under review for TMLREveryoneRevisionsBibTeXCC BY 4.0
Abstract: We present LLaVA-OneVision, a family of open large multimodal models (LMMs) developed by consolidating our insights into data, models, and visual representations in the LLaVA-NeXT blog series. Our experimental results demonstrate that LLaVA-OneVision is the first single model that can simultaneously push the performance boundaries of open LMMs in three important computer vision scenarios: single-image, multi-image, and video scenarios. Importantly, the design of LLaVA-OneVision allows strong transfer learning across different modalities/scenarios, yielding new emerging capabilities. In particular, strong video understanding and cross-scenario capabilities are demonstrated through task transfer from images to videos.
Submission Length: Long submission (more than 12 pages of main content)
Assigned Action Editor: ~Jianbo_Jiao2
Submission Number: 3432
Loading