We use essential cookies to make our site work. With your consent, we may also use non-essential cookies to improve user experience and analyze website traffic…

🚀 New model available: openai/gpt-oss-120b 🚀

openai/

whisper-base

Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. It was trained on 680k hours of labelled data and demonstrates a strong ability to generalize to many datasets and domains without fine-tuning. The model is based on a Transformer encoder-decoder architecture. Whisper models are available for various languages including English, Spanish, French, German, Italian, Portuguese, Russian, Chinese, Japanese, Korean, and many more.

Public
openai/whisper-base cover image

Input

Please upload an audio file

You need to login to use this model

Login

Settings

task to perform 2

optional text to provide as a prompt for the first window.. (Default: empty)

temperature to use for sampling (Default: 0)

language that the audio is in; uses detected language if None; use two letter language code (ISO 639-1) (e.g. en, de, ja) 101

chunk level, either 'segment' or 'word' 2

Chunk Length S

chunk length in seconds to split audio (Default: 30, 1 ≤ chunk_length_s ≤ 30)

Output

Model Information