Abstract: Knowledge Distillation (KD), which focuses on transferring semantic knowledge from a parameter-heavy teacher network to a more compact student network, has been widely and successfully used ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results