Learning from Stochastic Teacher Representations Using Student-Guided Knowledge Distillation

Résultats de recherche: Chapitre dans un livre, rapport, actes de conférenceParticipation à un ouvrage collectif lié à un colloque ou une conférenceRevue par des pairs

Résumé

Advances in self-distillation have shown that when knowledge is distilled from a teacher to a student using the same deep learning (DL) model, student performance can surpass the teacher, particularly when the model is over-parameterized and the teacher is trained with early stopping. Alternatively, ensemble learning also improves performance, although training, storing, and deploying multiple DL models becomes impractical as the number of models grows. Even distilling a deep ensemble to a single student model or weight averaging methods first requires training of multiple teacher models and does not fully leverage the inherent stochasticity for generating and distilling diversity in DL models. These constraints are particularly prohibitive in resource-constrained or latency-sensitive applications on, e.g., wearable devices. This paper proposes to train only one model and generate multiple diverse teacher representations using distillation-time dropout. However, generating these representations stochastically leads to noisy representations that are misaligned with the learned task. To overcome this problem, a novel stochastic self-distillation (SSD) training strategy is introduced for filtering and weighting teacher representation to distill from task-relevant representations only, using student-guided knowledge distillation. The student representation at each distillation step is used to guide the distillation process. Experimental results4(Code and supplementary available at: https://github.com/haseebaslam95/SSD) on real-world affective computing, wearable/biosignal (UCR Archive), HAR, and image classification datasets show that the proposed SSD method can outperform state-of-the-art methods without increasing the model size at both training and testing time. It incurs negligible computational complexity compared to ensemble learning and weight averaging methods.

langue originaleAnglais
titreMachine Learning and Knowledge Discovery in Databases. Research Track - European Conference, ECML PKDD 2025, Proceedings
rédacteurs en chefRita P. Ribeiro, Carlos Soares, João Gama, Bernhard Pfahringer, Nathalie Japkowicz, Pedro Larrañaga, Alípio M. Jorge, Pedro H. Abreu
EditeurSpringer Science and Business Media Deutschland GmbH
Pages235-253
Nombre de pages19
ISBN (imprimé)9783032061058
Les DOIs
étatPublié - 2026
EvénementEuropean Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, ECML PKDD 2025 - Porto, Portugal
Durée: 15 sept. 202519 sept. 2025

Série de publications

NomLecture Notes in Computer Science
Volume16018 LNCS
ISSN (imprimé)0302-9743
ISSN (Electronique)1611-3349

Conférence

ConférenceEuropean Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, ECML PKDD 2025
Pays/TerritoirePortugal
La villePorto
période15/09/2519/09/25

Empreinte digitale

Voici les principaux termes ou expressions associés à « Learning from Stochastic Teacher Representations Using Student-Guided Knowledge Distillation ». Ces libellés thématiques sont générés à partir du titre et du résumé de la publication. Ensemble, ils forment une empreinte digitale unique.

Contient cette citation