2019 (modified: 11 Nov 2022)ICML 2019Readers: Everyone
Abstract:Pre-training and fine-tuning, e.g., BERT \citep{devlin2018bert}, have achieved great success in language understanding by transferring knowledge from rich-resource pre-training task to the low/zero...