Use CometAPI POST /v1/audio/translations to translate spoken audio into English text. Supports Whisper model with multiple output formats.
from openai import OpenAI
client = OpenAI(
api_key="<COMETAPI_KEY>",
base_url="https://api.cometapi.com/v1"
)
audio_file = open("audio.mp3", "rb")
translation = client.audio.translations.create(
model="whisper-1",
file=audio_file
)
print(translation.text){
"text": "Hello, welcome to CometAPI."
}Use this endpoint to translate spoken audio into English text. It fits multilingual support queues, media review, and workflows that need one English transcript from audio in another language.Documentation Index
Fetch the complete documentation index at: https://apidoc.cometapi.com/llms.txt
Use this file to discover all available pages before exploring further.
model and file. Use a short sample first so you can verify upload handling, language detection, and output parsing.
text. If you request a different response format, parse the response according to that format.
Bearer token authentication. Use your CometAPI key.
The audio file to translate. Supported formats: flac, mp3, mp4, mpeg, mpga, m4a, ogg, wav, webm.
The audio translation model to use. Choose a current speech model from the Models page.
Optional text to guide the model's style or continue a previous audio segment. The prompt should be in English.
The output format for the translation.
json, text, srt, verbose_json, vtt Sampling temperature between 0 and 1. Higher values produce more random output; lower values are more focused. When set to 0, the model auto-adjusts temperature using log probability.
0 <= x <= 1The translation result in English.
The translated text in English.
from openai import OpenAI
client = OpenAI(
api_key="<COMETAPI_KEY>",
base_url="https://api.cometapi.com/v1"
)
audio_file = open("audio.mp3", "rb")
translation = client.audio.translations.create(
model="whisper-1",
file=audio_file
)
print(translation.text){
"text": "Hello, welcome to CometAPI."
}