no code implementations • 13 May 2020 • Charlotte Frenkel, Jean-Didier Legat, David Bol
With an energy per classification of 313nJ at 0. 6V and a 0. 32-mm$^2$ area for accuracies of 95. 3% (on-chip training) and 97. 5% (off-chip training) on MNIST, we demonstrate that SPOON reaches the efficiency of conventional machine learning accelerators while embedding on-chip learning and being compatible with event-based sensors, a point that we further emphasize with N-MNIST benchmarking.
no code implementations • 17 Apr 2019 • Charlotte Frenkel, Jean-Didier Legat, David Bol
Recent trends in the field of neural network accelerators investigate weight quantization as a means to increase the resource- and power-efficiency of hardware devices.