Dual teachers for self-knowledge distillation

Published: 01 Jan 2024, Last Modified: 13 Nov 2024Pattern Recognit. 2024EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Highlights•To our best knowledge, we are the first work to jointly conduct historical and structural self-knowledge distillation in a simple and unified framework, where they demonstrate complementary advantages and mutual benefits.•We demonstrate that an extremely lightweight design of auxiliary branches outperforms the previous heavy counterparts in both efficiency and accuracy.•Extensive experiments prove the effectiveness of our proposed framework on three datasets: CIFAR-100, ImageNet-2012, and PASCAL VOC.
Loading