Niyaz, U.; Bathula, D.R.
(2022-06-23)
Knowledge distillation (KD) is an effective model compression technique where a compact student network is taught to mimic the behavior of a complex and highly trained teacher network. In contrast, Mutual Learning (ML) ...