AI Tools.

Search

fill mask

xlm-roberta-base

XLM-RoBERTa base from Facebook AI, pre-trained on 2.5TB of filtered CommonCrawl text across 100 languages using the RoBERTa training procedure. Enables cross-lingual transfer — models fine-tuned on labeled English data can infer on other languages without parallel annotations. The standard starting point for multilingual classification and token-level tasks.

Last reviewed

Use cases

  • Multilingual NER without separate per-language models
  • Cross-lingual text classification (train in English, infer in other languages)
  • Multilingual sentiment analysis across international product reviews
  • Sequence labeling on low-resource languages via cross-lingual transfer
  • Universal sentence encoding for 100-language document corpora

Pros

  • 100-language coverage in a single model checkpoint
  • RoBERTa training rigor applied multilingually yields strong cross-lingual transfer
  • Multi-framework support (PyTorch, TF, JAX, ONNX, Rust); MIT license
  • Strong performance on XNLI and WikiANN multilingual benchmarks

Cons

  • Shared multilingual vocabulary degrades per-language token efficiency vs. monolingual models
  • Outperformed by dedicated monolingual models on high-resource languages
  • 512-token context limit
  • High-resource languages (English, German, French) dominate training data
  • Base size limits accuracy on tasks requiring deep language reasoning

FAQ

What is xlm-roberta-base used for?

Multilingual NER without separate per-language models. Cross-lingual text classification (train in English, infer in other languages). Multilingual sentiment analysis across international product reviews. Sequence labeling on low-resource languages via cross-lingual transfer. Universal sentence encoding for 100-language document corpora.

Is xlm-roberta-base free to use?

xlm-roberta-base is an open-source model published on HuggingFace. License terms vary by model — check the model card for the specific license.

How do I run xlm-roberta-base locally?

Most HuggingFace models can be loaded with transformers or the appropriate framework library. See the model card for framework-specific instructions and hardware requirements.

Tags

transformerspytorchtfjaxonnxsafetensorsxlm-robertafill-maskexbertmultilingualafamarasazbebgbnbrbs