Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for concurrently processing multiple users #196

Open
oa98105 opened this issue Feb 6, 2025 · 3 comments
Open

Support for concurrently processing multiple users #196

oa98105 opened this issue Feb 6, 2025 · 3 comments

Comments

@oa98105
Copy link

oa98105 commented Feb 6, 2025

Thank you for such an amazing work. I wonder if one wants to process multiple users concurrently, how this library can be used in such a scenario with real-time transcription. Furthermore, can we achieve this with a single inference of the model? Especially if one user takes a pause or not speaking then inference for another user should be possible.

@KoljaB
Copy link
Owner

KoljaB commented Feb 6, 2025

Only with horizontal scaling. This lib relies on faster_whisper for transcription, which can only serve in parallel when using multiple GPUs (or in CPU mode but that's too slow for realworld applications). There's batching support, but it does not help in this case: it does only work on a single audio and besides this if we wait for enough transcriptions so that a batch makes sense we introduce artificial waiting times for users, so latency would go up. So currently it's one GPU per user (=one RealtimeSTT instance setup to use a single dedicated GPU per user), otherwise you run into latency problems.

@oa98105
Copy link
Author

oa98105 commented Feb 8, 2025

That make sense. Thank you for the explanation.

@mathias7799
Copy link

I know this project is designed for running a local model, but have you considered adding optional support for cloud STT models like Azure STT? I can see a strong use case for this, especially for companies using contact centers that require real-time transcription for their agents.

A hybrid approach could be valuable—defaulting to a local model but allowing cloud STT integration as an option? Would open this project up to more usecases :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants