1 code implementation • 4 Feb 2024 • Yuzhu Wang, Lechao Cheng, Chaowei Fang, Dingwen Zhang, Manni Duan, Meng Wang
Inspired by the observation that the prompt tokens tend to share high mutual information with patch tokens, we propose initializing prompts with downstream token prototypes.
Ranked #1 on Visual Prompt Tuning on VTAB-1k(Structured<8>)
no code implementations • 17 Dec 2023 • Yuzhu Wang, Archontis Politis, Tuomas Virtanen
The clean speech clips from WSJ0 are employed for simulating speech signals of moving speakers in a reverberant environment.
1 code implementation • 26 May 2023 • Yuzhu Wang, Lechao Cheng, Manni Duan, Yongheng Wang, Zunlei Feng, Shu Kong
Finally, we propose a rather simple loss term (dubbed ND loss) to simultaneously (1) encourage student to produce large-\emph{norm} features, and (2) align the \emph{direction} of student features and teacher class-means.
Ranked #1 on Knowledge Distillation on ImageNet