text2text-generation
The FLAN-T5 large language model is a variant of the T5 model, trained on a mix of tasks and fine-tuned on over 1000 additional tasks covering multiple languages. It achieved state-of-the-art results on several benchmarks, including few-shot learning tasks, and demonstrates improved performance and usability compared to its predecessor.
text2text-generation
FLAN-T5 is a family of instructor-finetuned T5 models scaling up to 540B parameters. They are trained on more than 1000 tasks across over 100 diverse domains and cover multiple languages. FLAN-T5 demonstrates superiority over its predecessor T5 in various NLP tasks while being computationally efficient.
text2text-generation
Fine tuned T5 model on collection of datasets phrased as instructions
text2text-generation
Flan-PaLM 540B achieves state-of-the-art performance on several benchmarks, such as 75.2% on five-shot MMLU. We also publicly release Flan-T5 checkpoints, which achieve strong few-shot performance even compared to much larger models, such as PaLM 62B. Overall, instruction finetuning is a general method for improving the performance and usability of pretrained language model.