Unbiased Classification through Bias-Contrastive and Bias-Balanced LearningDownload PDF

21 May 2021, 20:48 (edited 26 Oct 2021)NeurIPS 2021 PosterReaders: Everyone
  • Keywords: Debiasing, Fairness, Contrastive Learning
  • TL;DR: Debiased classification by bias-contrastive learning and bias-balanced regression.
  • Abstract: Datasets for training machine learning models tend to be biased unless the data is collected with complete care. In such a biased dataset, models are susceptible to making predictions based on the biased features of the data. The biased model fails to generalize to the case where correlations between biases and targets are shifted. To mitigate this, we propose Bias-Contrastive (BiasCon) loss based on the contrastive learning framework, which effectively leverages the knowledge of bias labels. We further suggest Bias-Balanced (BiasBal) regression which trains the classification model toward the data distribution with balanced target-bias correlation. Furthermore, we propose Soft Bias-Contrastive (SoftCon) loss which handles the dataset without bias labels by softening the pair assignment of the BiasCon loss based on the distance in the feature space of the bias-capturing model. Our experiments show that our proposed methods significantly improve previous debiasing methods in various realistic datasets.
  • Supplementary Material: pdf
  • Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
  • Code: https://github.com/grayhong/bias-contrastive-learning
20 Replies