Search Results for author: Sangwook Cho

Found 3 papers, 2 papers with code

Comparing Kullback-Leibler Divergence and Mean Squared Error Loss in Knowledge Distillation

1 code implementation19 May 2021 Taehyeon Kim, Jaehoon Oh, Nakyil Kim, Sangwook Cho, Se-Young Yun

From this observation, we consider an intuitive KD loss function, the mean squared error (MSE) between the logit vectors, so that the student model can directly learn the logit of the teacher model.

Knowledge Distillation Learning with noisy labels

FINE Samples for Learning with Noisy Labels

1 code implementation NeurIPS 2021 Taehyeon Kim, Jongwoo Ko, Sangwook Cho, Jinhwan Choi, Se-Young Yun

Our framework, coined as filtering noisy instances via their eigenvectors (FINE), provides a robust detector with derivative-free simple methods having theoretical guarantees.

General Classification Learning with noisy labels

Understanding Knowledge Distillation

no code implementations1 Jan 2021 Taehyeon Kim, Jaehoon Oh, Nakyil Kim, Sangwook Cho, Se-Young Yun

To verify this conjecture, we test an extreme logit learning model, where the KD is implemented with Mean Squared Error (MSE) between the student's logit and the teacher's logit.

Knowledge Distillation

Cannot find the paper you are looking for? You can Submit a new open access paper.