Towards Domain-Aware Knowledge Distillation for Continual Model Generalization
56 views
0

 Published On Jan 29, 2024

Authors: Nikhil Reddy; Mahsa Baktashmotlagh; Chetan Arora
Description: Generalization on unseen domains is critical for Deep Neural Networks (\dnns) to perform well in real-world applications such as autonomous navigation.
However, catastrophic forgetting limit the ability of domain generalization and unsupervised domain adaption approaches to adapt to constantly changing target domains. To overcome these challenges, We propose \texttt{DoSe}\xspace framework, a \textbf{Do}main-aware \textbf{Se}lf-Distillation method based on batch normalization prototypes to facilitate continual model generalization across varying target domains. Specifically, we enforce the consistency of batch normalization statistics between two batches of images sampled from the same target domain distribution between the student and teacher models. To alleviate catastrophic forgetting, we introduce a novel exemplar-based replay buffer to identify difficult samples for the model to retain the knowledge. Specifically, we demonstrate that identifying difficult samples and updating the model periodically using them can help in preserving knowledge learned from previously seen domains. We conduct extensive experiments on two real-world datasets ACDC, C-Driving, and one synthetic dataset SHIFT to verify the efficiency of the proposed \texttt{DoSe}\xspace framework.
On ACDC, our method outperforms existing \SOTA in Domain Generalization, Unsupervised Domain Adaptation, and Daytime settings by $26\%$, $14\%$, and $70\%$ respectively.

show more

Share/Embed