MixKD: Towards Efficient Distillation of Large-scale Language ModelsDownload PDF

28 Sept 2020, 15:51 (edited 10 Feb 2022, 11:49)ICLR 2021 PosterReaders: Everyone
Keywords:
Abstract:
One-sentence Summary:
Code Of Ethics:
Data:
22 Replies

Loading