GPT-2 is a transformer-based language model developed by OpenAI that utilizes a causal language modeling (CLM) objective. It was trained on a 40GB dataset called WebText, which consists of texts from various websites, excluding Wikipedia. Without fine-tuning, GPT-2 achieved impressive zero-shot results on several benchmark datasets such as LAMBADA, CBT-CN, CBT-NE, WikiText2, PTB, enwiki8, and text8.
GPT-2 is a transformer-based language model developed by OpenAI that utilizes a causal language modeling (CLM) objective. It was trained on a 40GB dataset called WebText, which consists of texts from various websites, excluding Wikipedia. Without fine-tuning, GPT-2 achieved impressive zero-shot results on several benchmark datasets such as LAMBADA, CBT-CN, CBT-NE, WikiText2, PTB, enwiki8, and text8.
e7da7f221d5bf496a48136c0cd264e630fe9fcc8
2023-05-03T21:27:12+00:00
f27b190eeac4c2302d24068eabf5e9d6044389ae
2023-11-29T20:41:05+00:00