no code implementations • 6 Feb 2023 • Morten From Elvebakken, Alexandros Iosifidis, Lukas Esterle
While this addresses limitations related to distributed data, it incurs a communication overhead as the model parameters or gradients need to be exchanged regularly during training.