no code implementations • 18 Jul 2023 • Kilian Pfeiffer, Martin Rapp, Ramin Khalili, Jörg Henkel
With an increasing number of smart devices like internet of things (IoT) devices deployed in the field, offloadingtraining of neural networks (NNs) to a central server becomes more and more infeasible.
no code implementations • 2 Feb 2023 • Alexander Lindermayr, Nicole Megow, Martin Rapp
We consider online scheduling on unrelated (heterogeneous) machines in a speed-oblivious setting, where an algorithm is unaware of the exact job-dependent processing speeds.
1 code implementation • 10 Mar 2022 • Kilian Pfeiffer, Martin Rapp, Ramin Khalili, Jörg Henkel
To adapt to the devices' heterogeneous resources, CoCoFL freezes and quantizes selected layers, reducing communication, computation, and memory requirements, whereas other layers are still trained in full precision, enabling to reach a high accuracy.
no code implementations • 16 Dec 2021 • Martin Rapp, Ramin Khalili, Kilian Pfeiffer, Jörg Henkel
We study the problem of distributed training of neural networks (NNs) on devices with heterogeneous, limited, and time-varying availability of computational resources.
no code implementations • 9 Jun 2020 • Martin Rapp, Ramin Khalili, Jörg Henkel
We consider a distributed system, consisting of a heterogeneous set of devices, ranging from low-end to high-end.