curl --request POST \
--url https://dream-gateway.livepeer.cloud/audio-to-text \
--header 'Authorization: Bearer <token>' \
--header 'Content-Type: multipart/form-data' \
--form audio='@example-file' \
--form model_id= \
--form return_timestamps=true \
--form 'metadata={}'{
"text": "<string>",
"chunks": [
{
"timestamp": [
"<unknown>"
],
"text": "<string>"
}
]
}API reference for the audio-to-text pipeline. Accepts audio input and returns a transcript using Whisper-compatible models.
curl --request POST \
--url https://dream-gateway.livepeer.cloud/audio-to-text \
--header 'Authorization: Bearer <token>' \
--header 'Content-Type: multipart/form-data' \
--form audio='@example-file' \
--form model_id= \
--form return_timestamps=true \
--form 'metadata={}'{
"text": "<string>",
"chunks": [
{
"timestamp": [
"<unknown>"
],
"text": "<string>"
}
]
}Bearer authentication header of the form Bearer <token>, where <token> is your auth token.
Uploaded audio file to be transcribed.
Hugging Face model ID used for transcription.
Return timestamps for the transcribed text. Supported values: 'sentence', 'word', or a string boolean ('true' or 'false'). Default is 'true' ('sentence'). 'false' means no timestamps. 'word' means word-based timestamps.
Additional job information to be passed to the pipeline.
Was this page helpful?