1 code implementation • 30 Sep 2021 • Jinhyuk Park, Albert No
Recent results suggest that the student-friendly teacher is more appropriate to distill since it provides more transferable knowledge.
Knowledge Distillation Neural Network Compression