XLM: Cross-Lingual Language Model

Understanding Transformer-Based Self-Supervised Architectures

Rohan Jagtap
Towards Data Science
5 min readSep 13, 2020

--

Photo by Leonardo Toshiro Okubo on Unsplash

Models like BERT (Devlin et. al.) or GPT (Radford et. al.) have achieved the state of the art in language understanding. However, these models are pre-trained only on one language. Recently, efforts have been made towards mitigating monolingual representations and building universal cross-lingual models…

--

--

Immensely interested in AI Research | I read papers and post my notes on Medium