Search Results for author: Roberto Cavicchioli

Found 4 papers, 2 papers with code

Heterogeneous Encoders Scaling In The Transformer For Neural Machine Translation

no code implementations26 Dec 2023 Jia Cheng Hu, Roberto Cavicchioli, Giulia Berardinelli, Alessandro Capotondi

Although the Transformer is currently the best-performing architecture in the homogeneous configuration (self-attention only) in Neural Machine Translation, many State-of-the-Art models in Natural Language Processing are made of a combination of different Deep Learning approaches.

Machine Translation Translation

A request for clarity over the End of Sequence token in the Self-Critical Sequence Training

2 code implementations20 May 2023 Jia Cheng Hu, Roberto Cavicchioli, Alessandro Capotondi

The Image Captioning research field is currently compromised by the lack of transparency and awareness over the End-of-Sequence token (<Eos>) in the Self-Critical Sequence Training.

Image Captioning Sentence

Exploiting Multiple Sequence Lengths in Fast End to End Training for Image Captioning

1 code implementation13 Aug 2022 Jia Cheng Hu, Roberto Cavicchioli, Alessandro Capotondi

We introduce a method called the Expansion mechanism that processes the input unconstrained by the number of elements in the sequence.

Image Captioning

Exploring the sequence length bottleneck in the Transformer for Image Captioning

no code implementations7 Jul 2022 Jia Cheng Hu, Roberto Cavicchioli, Alessandro Capotondi

Most recent state of the art architectures rely on combinations and variations of three approaches: convolutional, recurrent and self-attentive methods.

Image Captioning

Cannot find the paper you are looking for? You can Submit a new open access paper.