{
"file": "string",
"model": "string",
"language": "string",
"prompt": "string",
"response_format": "json",
"temperature": 0,
"include[]": [
[]
],
"timestamp_granularities[]": [
"segment"
],
"stream": false
}