How can I simulate real-time streaming transcription using OpenAI API? #2307
Unanswered
Santoshchodipilli
asked this question in
Q&A
Replies: 1 comment 2 replies
-
You're on the right path of emulating streaming transcription with Whisper — that's the best workaround available at the moment, TheOpenAI's whisper-1 API is only capable of batch processing and not streaming. let me describe it for you -
Method: Chunked Streaming Simulation
If you're interested, I can help you set up a full real-time transcription. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I'm working on a project where I want to convert speech to text in real-time using OpenAI's Whisper model. I see that Whisper's hosted API (whisper-1) currently only supports batch mode — sending a full audio file and receiving the full transcript.
I'm trying to achieve a streaming-like transcription experience, where I can start receiving partial transcriptions as audio is still being recorded or uploaded.
Is there a way to simulate streaming transcription using Whisper?
I'm using Python.
I considered chunking the audio into small parts and sending them sequentially.
Is that the best approach, or is there a better method?
Also, is there any public roadmap or timeline for when the official OpenAI Whisper API might support real-time streaming transcription?
Thanks in advance!
Beta Was this translation helpful? Give feedback.
All reactions