no code implementations • 18 Nov 2022 • Gonçalo Mordido, Sébastien Henwood, Sarath Chandar, François Leduc-Primeau
In this work, we show that applying sharpness-aware training, by optimizing for both the loss value and loss sharpness, significantly improves robustness to noisy hardware at inference time without relying on any assumptions about the target hardware.
no code implementations • 3 May 2022 • Jonathan Kern, Sébastien Henwood, Gonçalo Mordido, Elsa Dupraz, Abdeldjalil Aïssa-El-Bey, Yvon Savaria, François Leduc-Primeau
Memristors enable the computation of matrix-vector multiplications (MVM) in memory and, therefore, show great potential in highly increasing the energy efficiency of deep neural network (DNN) inference accelerators.
no code implementations • 23 Dec 2019 • Sébastien Henwood, François Leduc-Primeau, Yvon Savaria
Deep neural networks (DNNs) depend on the storage of a large number of parameters, which consumes an important portion of the energy used during inference.