Global-Aware Model-Free Self-distillation for Recommendation SystemOpen Website

Published: 01 Jan 2023, Last Modified: 14 May 2023DASFAA (4) 2023Readers: Everyone
Abstract: The recommendation performance in our Alipay advertising system may suffer from label noise in training data. Earlier approaches that relied on soft targets typically neglected global similarities. Here, we introduce a novel algorithm called Global-aware Model-free Self-Distillation (GMSD) to create soft targets using the information at the global scale. Specifically, we propose calculating the similarities between the target sample and cluster centers produced by clustering the training dataset. The direct calculation of global similarities requires computation across the full dataset, which is prohibitively expensive. Additionally, we develop a contrastive cluster loss (CCLoss) for limiting the distance between the data of the intra-class to be lower than that of the inter-class to promote samples to be drawn to accurate cluster clusters. Extensive experiments on public and industry datasets demonstrate that GMSD outperforms state-of-the-art self-distillation methods in efficiency and effectiveness.
0 Replies

Loading