Quantization in Layer's Input is MatterDownload PDFOpen Website

Published: 01 Jan 2022, Last Modified: 12 May 2023CoRR 2022Readers: Everyone
Abstract: In this paper, we will show that the quantization in layer's input is more important than parameters' quantization for loss function. And the algorithm which is based on the layer's input quantization error is better than hessian-based mixed precision layout algorithm.
0 Replies

Loading