Factorized and progressive knowledge distillation for CTC-based ASR models

Published: 01 Jan 2024, Last Modified: 15 May 2025Speech Commun. 2024EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Highlights•We explore why the conventional KD underperforms when applied to CTC models.•We propose Factorized KL-divergence for CTC-based models’ KD.•We propose a progressive KD framework to gradually build up the student’s knowledge.
Loading