Fair Transfer Learning Through Causally Motivated RegularizationDownload PDF

07 Nov 2023OpenReview Archive Direct UploadReaders: Everyone
Abstract: Transfer learning is where a source model trained on one domain is adapted for a downstream task on another domain. Recently, it has been shown that the unfair behaviors of the source model can persist even after it has been adapted for a downstream task. In this work, we propose a solution to this problem by using causally-motivated regularization schemes for creating fair source models through using auxiliary labels. Our regularization schemes work by enforcing independences with respect to the causal DAG. Our approach only requires having auxiliary labels at the time of source model training and it promotes adapted downstream models that don’t make predictions based off of sensitive attributes. We show empirically and theoretically that source models that use our proposed causally-motivated regularization schemes lead to fairer downstream models and require less data to adapt to other tasks.
0 Replies

Loading