You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Training of DeformableDETR, which uses gradients through indices is very slow (GPU utilisation ~60%). When I stop the gradients at sampling points, before calling grid_sample function, the GPU utilisation bumps up to ~99%
As fas as my understanding goes, the current implementation of grads for bitwise ops + indexing works on CPU only. Is it possible to extend this functionality to use GPU kernel also?
The text was updated successfully, but these errors were encountered:
Training of DeformableDETR, which uses gradients through indices is very slow (GPU utilisation ~60%). When I stop the gradients at sampling points, before calling grid_sample function, the GPU utilisation bumps up to ~99%
Reference Issue : Grid Sample
As fas as my understanding goes, the current implementation of grads for bitwise ops + indexing works on CPU only. Is it possible to extend this functionality to use GPU kernel also?
The text was updated successfully, but these errors were encountered: