New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Boost Inference speed #426
Comments
What is your GPU? |
Its all on cloud - Nvidia A100 |
You can use |
I find v4 models faster than v3, like they were 2-3 minutes quick for the larger sample. |
@CarlGao4 So what is the value of default --segment ?? |
This is because default models of v3 (mdx mdx_q mdx_extra mdx_extra_q) are both bags of models containing 4 single models each. It is 4 times slower than single models. But default model for v4 ( |
It depends on the model. e.g. All v3 models is 44 |
Hey there, which is the average inference speed you're getting with your A100? I've been running some performance tests over an A100 (between other GPUs) and found the inference speed pretty much sticks at 48 seconds/s on the htdemucs (v4) model (the htdemucs_ft model is about 4 times slower, since it runs a sequence of 4 models instead of a single one). Are you getting close to those speeds? |
❓ Questions
Hi I am currently using a GPU for inferencing on files, however it takes a lot of time for longer files with duration around 20minutes. GPU VRam is not an issue, but I was wondering if I could increase power or faster anyhow the inference on GPU.
Thanks
The text was updated successfully, but these errors were encountered: