no code implementations • 26 Dec 2023 • Jia Cheng Hu, Roberto Cavicchioli, Giulia Berardinelli, Alessandro Capotondi
Although the Transformer is currently the best-performing architecture in the homogeneous configuration (self-attention only) in Neural Machine Translation, many State-of-the-Art models in Natural Language Processing are made of a combination of different Deep Learning approaches.
2 code implementations • 20 May 2023 • Jia Cheng Hu, Roberto Cavicchioli, Alessandro Capotondi
The Image Captioning research field is currently compromised by the lack of transparency and awareness over the End-of-Sequence token (<Eos>) in the Self-Critical Sequence Training.
1 code implementation • 13 Aug 2022 • Jia Cheng Hu, Roberto Cavicchioli, Alessandro Capotondi
We introduce a method called the Expansion mechanism that processes the input unconstrained by the number of elements in the sequence.
Ranked #1 on Image Captioning on MS COCO
no code implementations • 7 Jul 2022 • Jia Cheng Hu, Roberto Cavicchioli, Alessandro Capotondi
Most recent state of the art architectures rely on combinations and variations of three approaches: convolutional, recurrent and self-attentive methods.