no code implementations • 10 Dec 2020 • Beongjun Choi, Jy-yong Sohn, Dong-Jun Han, Jaekyun Moon
Through extensive real-world experiments, we demonstrate that our scheme, using only $20 \sim 30\%$ of the resources required in the conventional scheme, maintains virtually the same levels of reliability and data privacy in practical federated learning systems.
no code implementations • NeurIPS 2020 • Jy-yong Sohn, Dong-Jun Han, Beongjun Choi, Jaekyun Moon
Recent advances in large-scale distributed learning algorithms have enabled communication-efficient training via SignSGD.