openai/whisper-tiny cover image

openai/whisper-tiny

Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. It was trained on 680k hours of labelled data and demonstrates a strong ability to generalize to many datasets and domains without fine-tuning. Whisper is a Transformer-based encoder-decoder model trained on English-only or multilingual data. The English-only models were trained on speech recognition, while the multilingual models were trained on both speech recognition and machine translation.

Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. It was trained on 680k hours of labelled data and demonstrates a strong ability to generalize to many datasets and domains without fine-tuning. Whisper is a Transformer-based encoder-decoder model trained on English-only or multilingual data. The English-only models were trained on speech recognition, while the multilingual models were trained on both speech recognition and machine translation.

Public
$0.0005/sec

HTTP/cURL API

 

Input fields

audiostring

audio to transcribe


taskstring

task to perform

Default value: transcribe

Allowed values: transcribetranslate


languagestring

language that the audio is in; uses detected language if None


temperaturenumber

temperature to use for sampling

Default value: 0


patiencenumber

patience value to use in beam decoding

Default value: 1


suppress_tokensstring

token ids to suppress during sampling

Default value: -1


initial_promptstring

optional text to provide as a prompt for the first window.


condition_on_previous_textboolean

provide the previous output of the model as a prompt for the next window

Default value: true


temperature_increment_on_fallbacknumber

temperature to increase when falling back when the decoding fails to meet either of the thresholds below

Default value: 0.2


compression_ratio_thresholdnumber

gzip compression ratio threshold

Default value: 2.4


logprob_thresholdnumber

average log probability threshold

Default value: -1


no_speech_thresholdnumber

probability of the <|nospeech|> token threshold

Default value: 0.6


webhookfile

The webhook to call when inference is done, by default you will get the output in the response of your inference request

Input Schema

Output Schema


© 2023 Deep Infra. All rights reserved.

Discord Logo