Local Multi-Head Channel Self-Attention for Facial Expression Recognition

14 Nov 2021  ·  Roberto Pecoraro, Valerio Basile, Viviana Bono, Sara Gallo ·

Since the Transformer architecture was introduced in 2017 there has been many attempts to bring the self-attention paradigm in the field of computer vision. In this paper we propose a novel self-attention module that can be easily integrated in virtually every convolutional neural network and that is specifically designed for computer vision, the LHC: Local (multi) Head Channel (self-attention). LHC is based on two main ideas: first, we think that in computer vision the best way to leverage the self-attention paradigm is the channel-wise application instead of the more explored spatial attention and that convolution will not be replaced by attention modules like recurrent networks were in NLP; second, a local approach has the potential to better overcome the limitations of convolution than global attention. With LHC-Net we managed to achieve a new state of the art in the famous FER2013 dataset with a significantly lower complexity and impact on the "host" architecture in terms of computational cost when compared with the previous SOTA.

PDF Abstract

Datasets


Results from the Paper


Ranked #5 on Facial Expression Recognition (FER) on FER2013 (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Facial Expression Recognition (FER) FER2013 LHC-Net Accuracy 74.42 # 5

Methods