Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implementation with Large-v3 but with Batching #553

Closed
souvikqb opened this issue Nov 8, 2023 · 9 comments
Closed

Implementation with Large-v3 but with Batching #553

souvikqb opened this issue Nov 8, 2023 · 9 comments

Comments

@souvikqb
Copy link

souvikqb commented Nov 8, 2023

I saw a large-v3 implementation with faster_whisper (#547) but it's quite slow.

Large-v3 is very fast with batching as shown here --- https://huggingface.co/openai/whisper-large-v3

Batching speeds up the transcription process by a lot. The only reason I wish to use faster_whisper is cause it provides things like srt, verbose, word level transcription

@souvikqb souvikqb changed the title With Large-v3 but with Batching Implementation with Large-v3 but with Batching Nov 8, 2023
@phineas-pta
Copy link

srt, verbose, word level transcription

u can write code to do those things with transformers interface

@souvikqb
Copy link
Author

souvikqb commented Nov 10, 2023

srt, verbose, word level transcription

u can write code to do those things with transformers interface

Do you have any resources that I can use?

@phineas-pta
Copy link

word timestamp is an option when u call the pipeline

fort srt just retrieve timestamp and format properly

@souvikqb
Copy link
Author

word timestamp is an option when u call the pipeline

fort srt just retrieve timestamp and format properly

No I meant do you have a dedicated function that I can use to convert it? For srt and verbose

@phineas-pta
Copy link

see my comment in openai/whisper#654

@souvikqb
Copy link
Author

see my comment in openai/whisper#654

Hey thanks! That sorts out my srt requirement. Now just looking for verbose and word

@phineas-pta
Copy link

as i said, word timestamp is an option when call the pipeline

@souvikqb
Copy link
Author

souvikqb commented Nov 10, 2023

as i said, word timestamp is an option when call the pipeline

I actually got an error on running this --

result = pipe(sample, return_timestamps="word")
print(result["chunks"])

image

@phineas-pta
Copy link

ask in transformers github repo

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants