Knowledge distillation under ideal joint classifier assumption

Published: 01 Jan 2024, Last Modified: 13 Nov 2024Neural Networks 2024EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Highlights•Conducted an in-depth analysis of knowledge distillation employing the ideal joint classifier.•Presented a comprehensive proof establishing the error bounds of the student network under a function of the teacher’s error.•Introduced a novel knowledge distillation framework grounded in the ideal joint classifier assumption.
Loading