ADCL: An attention feature enhancement network based on adversarial contrastive learning for short text classification
Abstract: Supervised Contrastive Learning (SCL) has emerged as a powerful approach for improving model performance in text classification tasks, particularly in few-shot learning scenarios. However, existing SCL methods predominantly focus on the contrastive relationships between positive and negative samples, often neglecting the intrinsic semantic features of individual samples. This limitation can introduce training biases, especially when labeled data are scarce. Additionally, the intrinsic feature sparsity of short texts further aggravates this issue, hindering the extraction of discriminative and robust representations. To address these challenges, we propose a Label-aware Attention-based Adversarial Contrastive Learning Network (ADCL). The model incorporates a bidirectional contrastive learning framework that leverages cross-attention layers to enhance interactions between label and document representations. Moreover, adversarial learning is employed to optimize the backpropagation of contrastive learning gradients, effectively decoupling sample embeddings from label-specific features. Compared to prior methods, ADCL not only emphasizes contrasts between positive and negative samples but also prioritizes the intrinsic semantic information of individual samples during the learning process. We conduct comprehensive experiments from both full-shot and few-shot learning perspectives on five benchmark short-text datasets: SST-2, SUBJ, TREC, PC, and CR. The results demonstrate that ADCL consistently outperforms existing contrastive learning methods, achieving superior average accuracy across the majority of tasks.
Loading