1 code implementation • 7 May 2021 • Matthias Wess, Matvey Ivanov, Anvesh Nookala, Christoph Unger, Alexander Wendt, Axel Jantsch
To decouple the architectural search from the target hardware, we propose a time estimation framework that allows for modeling the inference latency of DNNs on hardware accelerators based on mapping and layer-wise estimation models.
no code implementations • 19 Apr 2021 • Lukas Baischer, Matthias Wess, Nima TaheriNejad
In this article an overview of existing neural network hardware accelerators and acceleration methods is given.