Skip to content

Switch torch compile to use max-autotune mode#111

Merged
WilliamYue37 merged 1 commit intomainfrom
feat/autotune_compile
Feb 10, 2026
Merged

Switch torch compile to use max-autotune mode#111
WilliamYue37 merged 1 commit intomainfrom
feat/autotune_compile

Conversation

@WilliamYue37
Copy link
Member

@WilliamYue37 WilliamYue37 commented Feb 10, 2026

What this does

Switch torch compile to use max-autotune mode. It speeds up PI05 inference from 150ms to 100ms on Nvidia 3090 GPUs.

How it was tested

Ran inference.py

How to checkout & try? (for the reviewer)

python src/opentau/scripts/inference.py --config_path=configs/examples/pi05_training_config.json

Checklist

  • I have added Google-style docstrings to important functions and ensured function parameters are typed.
  • My PR includes policy-related changes.
    • If the above is checked: I have run the GPU pytests (pytest -m "gpu") and regression tests.

Note: Before submitting this PR, please read the contributor guideline.

@WilliamYue37 WilliamYue37 self-assigned this Feb 10, 2026
@WilliamYue37 WilliamYue37 added the feature New feature or request label Feb 10, 2026
@WilliamYue37 WilliamYue37 merged commit 8f85131 into main Feb 10, 2026
7 of 9 checks passed
@WilliamYue37 WilliamYue37 deleted the feat/autotune_compile branch February 10, 2026 00:46
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

feature New feature or request

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants