Browse deepinfra models:

All categories and models you can try out and directly use in deepinfra:
Search

Category/text2text-generation

google/flan-t5-large cover image
$0.0005 / sec
  • text2text-generation

The FLAN-T5 large language model is a variant of the T5 model, trained on a mix of tasks and fine-tuned on over 1000 additional tasks covering multiple languages. It achieved state-of-the-art results on several benchmarks, including few-shot learning tasks, and demonstrates improved performance and usability compared to its predecessor.

google/flan-t5-small cover image
$0.0005 / sec
  • text2text-generation

FLAN-T5 is a family of instructor-finetuned T5 models scaling up to 540B parameters. They are trained on more than 1000 tasks across over 100 diverse domains and cover multiple languages. FLAN-T5 demonstrates superiority over its predecessor T5 in various NLP tasks while being computationally efficient.

google/flan-t5-xl cover image
$0.0005 / sec
  • text2text-generation

Fine tuned T5 model on collection of datasets phrased as instructions

google/flan-t5-xxl cover image
$0.0005 / sec
  • text2text-generation

Flan-PaLM 540B achieves state-of-the-art performance on several benchmarks, such as 75.2% on five-shot MMLU. We also publicly release Flan-T5 checkpoints, which achieve strong few-shot performance even compared to much larger models, such as PaLM 62B. Overall, instruction finetuning is a general method for improving the performance and usability of pretrained language model.