Diversity Enhanced Active Learning with Strictly Proper Scoring RulesDownload PDF

21 May 2021, 20:46 (edited 26 Oct 2021)NeurIPS 2021 PosterReaders: Everyone
  • Keywords: active learning, machine learning, deep active learning, acquisition function, uncertainty based method, diversity base method
  • TL;DR: We develop diversity enhanced active learning with strictly proper scoring rules, for instance using a log probability, generally outperforms Bayesian active learning methods tested.
  • Abstract: We study acquisition functions for active learning (AL) for text classification. The Expected Loss Reduction (ELR) method focuses on a Bayesian estimate of the reduction in classification error, recently updated with Mean Objective Cost of Uncertainty (MOCU). We convert the ELR framework to estimate the increase in (strictly proper) scores like log probability or negative mean square error, which we call Bayesian Estimate of Mean Proper Scores (BEMPS). We also prove convergence results borrowing techniques used with MOCU. In order to allow better experimentation with the new acquisition functions, we develop a complementary batch AL algorithm, which encourages diversity in the vector of expected changes in scores for unlabelled data. To allow high performance text classifiers, we combine ensembling and dynamic validation set construction on pretrained language models. Extensive experimental evaluation then explores how these different acquisition functions perform. The results show that the use of mean square error and log probability with BEMPS yields robust acquisition functions, which consistently outperform the others tested.
  • Supplementary Material: pdf
  • Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
  • Code: https://github.com/davidtw999/BEMPS
11 Replies