You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am a beginner with TensorRT, and I am currently trying to optimize some classic SNN (Spiking Neural Network) models using TensorRT.
When I previously used TensorRT for CNN-type networks, the scheduling/compilation process was quite fast — I could generate an .engine file within minutes.
However, when I try to import networks with transformer structures, such as BERT, the scheduling time becomes extremely long. For example, in one model I tested, the process has already taken more than 50 hours and is still not finished. Is this expected behavior?
I also noticed a similar issue mentioned in a 2020 paper, where the authors reported that scheduling LSTM networks with TensorRT also took an abnormally long time.
Since SNNs also have temporal characteristics similar to LSTMs, I wonder:
• Is this a known limitation or issue in TensorRT when dealing with such models?
• Or could it be that my scheduling code has some problem that causes this excessive compilation time?