EleutherAI/gpt-neo-2.7B cover image

EleutherAI/gpt-neo-2.7B

We present GPT-Neo 2.7B, a transformer model designed using EleutherAI's replication of the GPT-3 architecture. With 2.7B parameters, this model was trained on the large-scale curated dataset, Pile, for 420 billion tokens over 400,000 steps. GPT-Neo 2.7B achieves state-of-the-art results on various benchmarks, including linguistic reasoning, physical and scientific reasoning, and down-stream applications.

We present GPT-Neo 2.7B, a transformer model designed using EleutherAI's replication of the GPT-3 architecture. With 2.7B parameters, this model was trained on the large-scale curated dataset, Pile, for 420 billion tokens over 400,000 steps. GPT-Neo 2.7B achieves state-of-the-art results on various benchmarks, including linguistic reasoning, physical and scientific reasoning, and down-stream applications.

Public
$0.0005/sec
demoapi

a0e5677e8611d87c8ce501dfa6a713e42ba2f6ef

2023-05-03T23:30:47+00:00


© 2023 Deep Infra. All rights reserved.

Discord Logo