Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

the on_realtime_transcription_update text issue #69

Closed
jacobtang opened this issue Jun 15, 2024 · 3 comments
Closed

the on_realtime_transcription_update text issue #69

jacobtang opened this issue Jun 15, 2024 · 3 comments

Comments

@jacobtang
Copy link

Hi,@KoljaB
I use the on_realtime_transcription_update(also try the on_realtime_transcription_stabilized) text,send to the web client,but it always show the pre text word.
Is there other method that can avoid show the repeated pre words?Thanks!
image
I try to implement live transcription function similar to Zoom Meeting in the product.

@KoljaB
Copy link
Owner

KoljaB commented Jun 15, 2024

No, that ist due to the way Whisper works and how it therefore is implemented. You'd need to program this on top of that.

@jacobtang
Copy link
Author

Ok,thanks a lot!
I use the process_text callback data, it costs about 3s to get the text data, is it normal? how should I reduce the time cost?
use the recorder_config,and recorder.feed_audio(audio_chunk) method in GPU env.
image

@KoljaB
Copy link
Owner

KoljaB commented Jun 16, 2024

Absolutely not normal. Even with largest model or on CPU transcription time should be way below 1s.
I have no real idea, why this happens.
Can you please try another model, like small.en or medium? Also a separated test with only faster_whisper library could tell us if it's the transcription or maybe something related to the VAD models.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants