Train Big, Then Compress: Rethinking Model Size for Efficient Training and Inference of Transformers

Abstract: Since hardware resources are limited, the objective of training deep learning models is typically to maximize accuracy subject to the time and memory constraints of training and inference. We study...
0 Replies
Loading