Keywords: concept bottleneck models; few-shot text classification;
TL;DR: We propose a novel CBM paradigm for few-shot text classification. Our method significantly outperforms prior CBMs and is competitive with LLMs using only 10 samples, while offering superior interpretability and efficiency.
Abstract: Few-shot text classification is a critical real-world task for which Large Language Models (LLMs) have shown great promise. However, their high inference costs and lack of interpretability limit their practical use. While Concept Bottleneck Models (CBMs) offer an efficient and interpretable alternative, their reliance on training surrogate models makes them incompatible with few-shot scenarios.
To bridge this gap, we introduce a novel CBM paradigm that relies solely on sample-concept similarity to make predictions.
We ensure the effectiveness of our concepts through a prototypical-discriminative dual-level architecture and a dynamic concept refinement mechanism. Extensive experiments show that with as few as 10 training samples, our method surpasses prior CBMs and even achieves performance comparable to LLMs.
The code is available at \url{https://github.com/alexiszlf/StructCBM}.
Primary Area: interpretability and explainable AI
Submission Number: 9554
Loading