research
∙
09/20/2023
Weight Averaging Improves Knowledge Distillation under Domain Shift
Knowledge distillation (KD) is a powerful model compression technique br...
research
∙
12/20/2022