New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
infer_s2s.py: Load dataset (possibly sharded) ??? #69
Comments
Hi, How many GPUs did you use for decoding? Current script doesn't support multiple GPU for decoding and if you use >1 GPUs only one part of the dataset will be decoded. If you are under multi-gpu environment, you can do Besides, if your test set contains long utterances (depending on |
Yes I just found out decoding cannot be run on multiple GPUs(even CPUs, as long as multiprocessing is involved), but it still took me quite an amount of time to find that out when I went deeper into the code. |
I realized that only part of the test dataset is evaluated when running the "infer_s2s.py". After inspecting the code, I found this comment "Load dataset (possibly sharded)" here. Specifically, the test set of my database has around 400 samples but only 150 are decoded. Why? How can I solve this? I was trying to set different parameters of the dataset/task but with no succes. I would like to get a %WER performance on the whole test set in order to be able for comparing and benchmarking purposes.
The text was updated successfully, but these errors were encountered: