Knowledge Distillation for Teaching Symmetry Invariances

Published: 10 Oct 2024, Last Modified: 09 Nov 2024SciForDL PosterEveryoneRevisionsBibTeXCC BY 4.0
TL;DR: Knowledge distillation is just as good as data augmentation for learning a specific symmetry invariance in your data set.
Abstract: Knowledge distillation is used in an attempt to transfer model invariances related to specific symmetry transformations of the data. To this end, a model that exhibits such an invariance at the structural level is distilled into a simpler model that does not. The efficacy of knowledge distillation in transferring model invariances is empirically evaluated using four pairs of such networks, each pertaining to a different data invariance. Six metrics are reported; these determine how helpful the knowledge distillation is in general for the learning process and also specifically for learning the targeted invariance. It is observed that knowledge distillation fails at transferring invariances in the considered model pairs. Moreover, data augmentation shows a better performance at instilling invariances into a network.
Style Files: I have used the style files.
Submission Number: 25
Loading