Skip to content

Latest commit

 

History

History
11 lines (9 loc) · 805 Bytes

Facebook AI (Meta AI).md

File metadata and controls

11 lines (9 loc) · 805 Bytes

#highlevel

Models Released:

  • RoBERTa (Robustly Optimized BERT Pretraining Approach):
  • RoBERTa: An optimized version of BERT that improved training techniques and achieved state-of-the-art results on several benchmarks.
  • XLM (Cross-lingual Language Model): Designed for multilingual and cross-lingual tasks.
  • XLM-R: An extension of XLM with larger datasets and improved multilingual capabilities.

Contributions:

  • Optimized Pretraining: RoBERTa showed how tweaking pretraining approaches, such as removing the next sentence prediction task and increasing training data, can lead to better model performance.
  • Multilingual Capabilities: XLM and XLM-R expanded the scope of LLMs to multiple languages, promoting cross-lingual understanding and applications.