EleutherAI/gpt-j-6B cover image

EleutherAI/gpt-j-6B

GPT-J 6B is a 6 billion parameter transformer model trained using Ben Wang's Mesh Transformer JAX. It was trained on the Pile, a large-scale curated dataset created by EleutherAI. The model consists of 28 layers with a model dimension of 4096, and a feedforward dimension of 16384.

GPT-J 6B is a 6 billion parameter transformer model trained using Ben Wang's Mesh Transformer JAX. It was trained on the Pile, a large-scale curated dataset created by EleutherAI. The model consists of 28 layers with a model dimension of 4096, and a feedforward dimension of 16384.

Public
$0.0005/sec
demoapi

f98c709453c9402b1309b032f40df1c10ad481a2

2023-05-04T21:12:25+00:00


© 2023 Deep Infra. All rights reserved.

Discord Logo