no code implementations • 4 Nov 2022 • Seock-Hwan Noh, JunSang Park, Dahoon Park, Jahyun Koo, Jeik Choi, Jaeha Kung
Thus, in this work, we conduct a detailed analysis of the batch normalization layer to efficiently reduce the runtime overhead in the batch normalization process.
no code implementations • 13 Mar 2022 • Seock-Hwan Noh, Jahyun Koo, SeungHyun Lee, Jongse Park, Jaeha Kung
While several prior works proposed such multi-precision support for DNN accelerators, not only do they focus only on the inference, but also their core utilization is suboptimal at a fixed precision and specific layer types when the training is considered.