New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Option for GTN CTC mode #2866
Option for GTN CTC mode #2866
Conversation
Added @awni |
Hi! This looks great, I'm glad to see GTN is working. The run time in comparison to warp CTC and builtin (PyTorch?) is not bad, I assume the timings are dominated by something else in the pipeline. I will take a look at the implementation. For what it's worth, we have seen these kind of spikes before in CTC training though I don't recall what caused them. They haven't been an issue for us though they are somewhat concerning.. |
CC @vineelpratap who may recall some issues related to the spiking behaviour CC @shubho who is working on a much faster GTN compose which should speed things up a lot here. |
Regarding spiking behavior, is it possible that there could be samples which can trigger edge cases for CTC... Depending how the library treats them, the behavior can be different.. For example - |
Codecov Report
@@ Coverage Diff @@
## master #2866 +/- ##
==========================================
- Coverage 80.64% 80.45% -0.19%
==========================================
Files 323 324 +1
Lines 28660 28730 +70
==========================================
+ Hits 23114 23116 +2
- Misses 5546 5614 +68
Continue to review full report at Codecov.
|
Many thanks, @brianyan918!
|
Also, k2 once you fix the installation issue. |
@awni and @vineelpratap, thanks for the valuable comments! |
@sw005320 got it. Thanks! |
This PR contains an implementation of CTC using FB's GTN package, enabling another ctc_type called 'gtnctc'. This mode was tested using Voxforge Italian:
training curve shows spikes for builtin (red) and gtnctc (blue), but not for warpctc (orange)
training times, from single 2080 GPU trials