Large-Scale Pre-Trained Language Models

RoBERTa

How did RoBERTa outperform XLNet with no architectural changes to the original BERT?

Naoki
7 min readFeb 20, 2022

--

In 2018, Jacob Devlin et al. (Google AI Language) released BERT that achieved state-of-the-art results on nine NLP tasks, inspiring other researchers to come up with further improvements. Soon, newer models superseded…

--

--