Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add batched mode for faster inference on pre-recorded conversations #35

Closed
juanmc2005 opened this issue Apr 21, 2022 · 1 comment
Closed
Labels
feature New feature or request
Milestone

Comments

@juanmc2005
Copy link
Owner

Problem

Running the online diarization pipeline on an entire dataset can be difficult and slow because the current implementation simulates an online scenario and processes one chunk at a time.
This is the way to go in a real-time scenario but it would be very useful to have a faster implementation for evaluation, for example to see the performance impact of swapping a component (issue #34 is a good example).

Idea

Pre-calculate segmentation and embeddings for all chunks in a file and only run clustering and output reconstruction online, this would considerably speed up the process. It could be hosted in a new BatchedOnlineSpeakerDiarization class implementing the same interface as OnlineSpeakerDiarization

@juanmc2005
Copy link
Owner Author

Changes added as part of PR #46.
Implemented as OnlineSpeakerDiarization.from_file()

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant