Model Merging Enables In-Context Learning for Bioacoustics Foundation Models

Published: 02 Oct 2025, Last Modified: 02 Oct 2025NeurIPS 2025 BeingconsideredfortalkEveryoneRevisionsBibTeXCC BY 4.0
Keywords: LLMs, Bioacustics, Model Merging, Few-Show In-Context Learning
TL;DR: We recover the instruction-following capabilities of NatureLM, a finetuned bioacustic foundation model, through model merging, enabling it to perform few-shot in-context learning.
Abstract: General-purpose foundation models capable of generalizing across species and tasks represent a promising new frontier in bioacoustics, with NatureLM being one of the most prominent examples. While its domain-specific finetuning yields strong performance on bioacoustic benchmarks, we observe that it also introduces trade-offs in instruction-following flexibility. For instance, NatureLM achieves high accuracy when prompted for either the common or scientific name individually, but its accuracy drops significantly when both are requested in a single prompt. These effects limit zero- and few-shot generalization to novel tasks. We address this by applying a simple model merging strategy that interpolates NatureLM with its base language model, recovering instruction-following capabilities with minimal loss of domain expertise. Finally, we show that this enables effective few-shot in-context learning, a key capability for real-world scenarios where labeled data for new species or environments are scarce.
Submission Number: 1
Loading