THE TRANSFORMER SERIES

Transformer’s Encoder-Decoder

Understanding The Model Architecture

Naoki
11 min readDec 12, 2021

--

In 2017, Vaswani et al. published a paper titled “Attention Is All You Need” for the NeurIPS conference. They introduced the original transformer architecture for machine translation, performing better and faster than RNN encoder-decoder models, which were mainstream.

--

--