no code implementations • 23 Dec 2023 • Konstantinos Balaskas, Andreas Karatzas, Christos Sad, Kostas Siozios, Iraklis Anagnostopoulos, Georgios Zervakis, Jörg Henkel
We explore, for the first time, per-layer fine- and coarse-grained pruning, in the same DNN architecture, in addition to low bit-width mixed-precision quantization for weights and activations.
no code implementations • 6 Jul 2023 • Andreas Karatzas, Iraklis Anagnostopoulos
Modern Deep Neural Networks (DNNs) exhibit profound efficiency and accuracy properties.
no code implementations • 25 Jul 2022 • Ourania Spantidi, Georgios Zervakis, Iraklis Anagnostopoulos, Jörg Henkel
Deep Neural Networks (DNNs) are being heavily utilized in modern applications and are putting energy-constraint devices to the test.
no code implementations • 20 Jul 2021 • Ourania Spantidi, Georgios Zervakis, Iraklis Anagnostopoulos, Hussam Amrouch, Jörg Henkel
In addition, we propose a filter-oriented approximation method to map the weights to the appropriate modes of the approximate multiplier.
no code implementations • 8 Mar 2021 • Sami Salamin, Georgios Zervakis, Ourania Spantidi, Iraklis Anagnostopoulos, Jörg Henkel, Hussam Amrouch
Transistor aging is one of the major concerns that challenges designers in advanced technologies.
no code implementations • 18 Feb 2021 • Georgios Zervakis, Ourania Spantidi, Iraklis Anagnostopoulos, Hussam Amrouch, Jörg Henkel
In this work, we introduce a control variate approximation technique for low error approximate Deep Neural Network (DNN) accelerators.