no code implementations • 6 Jun 2020 • Anurag Pallaprolu, Radha Vaidya, Aditya Swaroop Attawar
State-of-the-art attention based models, mostly centered around the transformer architecture, solve the problem of sequence-to-sequence translation using the so-called scaled dot-product attention.