Unsupervised Cross-lingual Representation Learning at Scale (XLM-RoBERTa)

Estimated read time 1 min read

https://github.com/pytorch/fairseq

XLM-R (XLM-RoBERTa) is a generic cross lingual sentence encoder that obtains state-of-the-art results on many cross-lingual understanding (XLU) benchmarks. It is trained on 2.5T of filtered CommonCrawl data in 100 languages (list below).

You May Also Like

More From Author

+ There are no comments

Add yours