Skip to content

[Question] Extremely long build time (>50 hours) when compiling Transformer/SNN models with TensorRT #4583

@pengjh0111

Description

@pengjh0111

I am a beginner with TensorRT, and I am currently trying to optimize some classic SNN (Spiking Neural Network) models using TensorRT.

When I previously used TensorRT for CNN-type networks, the scheduling/compilation process was quite fast — I could generate an .engine file within minutes.

However, when I try to import networks with transformer structures, such as BERT, the scheduling time becomes extremely long. For example, in one model I tested, the process has already taken more than 50 hours and is still not finished. Is this expected behavior?

I also noticed a similar issue mentioned in a 2020 paper, where the authors reported that scheduling LSTM networks with TensorRT also took an abnormally long time.

Since SNNs also have temporal characteristics similar to LSTMs, I wonder:
• Is this a known limitation or issue in TensorRT when dealing with such models?
• Or could it be that my scheduling code has some problem that causes this excessive compilation time?

Metadata

Metadata

Assignees

No one assigned

    Labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions