KNN-BERT: Fine-Tuning Pre-Trained Models with KNN ClassifierDownload PDF

Anonymous

16 Jan 2022 (modified: 05 May 2023)ACL ARR 2022 January Blind SubmissionReaders: Everyone
Abstract: Pre-trained models are widely used in fine-tuning downstream tasks with linear classifiers optimized by the cross entropy loss, which might face robustness and stability problems.These problems can be improved by learning representations that focus on similarities in the same class and variance in different classes when making predictions.In this paper, we utilize the K-Nearest Neighbors Classifier in pre-trained model fine-tuning.For this KNN classifier, we introduce a supervised momentum contrastive learning framework to learn the clustered representations of the supervised downstream tasks.Extensive experiments on text classification tasks and robustness tests show that by incorporating KNNs with the traditional fine-tuning process, we can obtain significant improvements on the clean accuracy in both rich-source and few-shot settings and can improve the robustness against adversarial attacks.\footnote{all codes will be available at https://github.com//}
Paper Type: long
0 Replies

Loading