Browse deepinfra models:

All categories and models you can try out and directly use in deepinfra:
Search

Category/all

xlm-roberta-base cover image
$0.0005 / sec
  • fill-mask

The XLM-RoBERTa model is a multilingual version of RoBERTa, pre-trained on 2.5TB of filtered CommonCrawl data containing 100 languages. It was introduced in the paper "Unsupervised Cross-lingual Representation Learning at Scale" by Conneau et al. and first released in this repository. The model learns an inner representation of 100 languages that can be used to extract features useful for downstream tasks.