Skip to content

Conversation

[ghstack-poisoned]
@pytorch-bot
Copy link

pytorch-bot bot commented Oct 27, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/166307

Note: Links to docs will display an error until the docs builds have been completed.

✅ No Failures

As of commit 2e05c55 with merge base 030de07 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@pytorch-bot pytorch-bot bot added ciflow/b200 ciflow/h100 ciflow/rocm Trigger "default" config CI on ROCm release notes: linalg_frontend release notes category labels Oct 27, 2025
@nikitaved nikitaved requested review from eqy and removed request for IvanYashchuk and lezcano October 27, 2025 15:00
@nikitaved nikitaved added ciflow/trunk Trigger trunk jobs on your pull request module: cuda Related to torch.cuda, and CUDA support in general labels Oct 27, 2025
Copy link
Collaborator

@eqy eqy left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

is it worth parametrizing the test(s) to try both 1xm and n-shaped bias or is that excessive?

&& (
self.is_contiguous() &&
// NOTE: fine to have 1-len dims to the left from the leading one
self.dim() <= result.dim() && self.squeeze().dim() == 1 &&
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

cool use of squeeze here

self._test_addmm_addmv(func, M, m1, m2, activation=activation)

# vector-shaped bias and beta=1 result in epilogue fusion in CUDA
# vector-shaped bias (or with 1-len dims on the left from the leading dim)
Copy link
Collaborator

@eqy eqy Oct 27, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

is this an "or" or have we changed the case from vector-shaped bias to 1 x n one?

Copy link
Collaborator Author

@nikitaved nikitaved Oct 27, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It is an or. Motivated by the "expected fusions" tests from Inductor -- and we can fuse these broadcast biases safely.

[ghstack-poisoned]
@nikitaved
Copy link
Collaborator Author

@eqy, thanks for the review! Do you mean in test_cuda_matmul.py? Yes, it would not hurt extending there -- will do. Otherwise test_linalg.py offers quite a comprehensive coverage.

@nikitaved nikitaved moved this to In Progress in PyTorch + CUDA Oct 27, 2025
@nikitaved nikitaved self-assigned this Oct 27, 2025
[ghstack-poisoned]
[ghstack-poisoned]
[ghstack-poisoned]
[ghstack-poisoned]
[ghstack-poisoned]
[ghstack-poisoned]
@nikitaved
Copy link
Collaborator Author

@eqy, the testing is expanded. Let me know if there is anything else we'd rather do before I merge.

[ghstack-poisoned]
[ghstack-poisoned]
[ghstack-poisoned]
[ghstack-poisoned]
[ghstack-poisoned]
[ghstack-poisoned]
[ghstack-poisoned]
[ghstack-poisoned]
[ghstack-poisoned]
[ghstack-poisoned]
[ghstack-poisoned]
@nikitaved
Copy link
Collaborator Author

@pytorchbot merge

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@github-project-automation github-project-automation bot moved this from In Progress to Done in PyTorch + CUDA Oct 31, 2025
BoyuanFeng pushed a commit that referenced this pull request Oct 31, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ciflow/b200 ciflow/h100 ciflow/rocm Trigger "default" config CI on ROCm ciflow/trunk Trigger trunk jobs on your pull request Merged module: cuda Related to torch.cuda, and CUDA support in general open source release notes: linalg_frontend release notes category

Projects

Status: Done

Development

Successfully merging this pull request may close these issues.

5 participants