Skip to content

Conversation

xuanzhang816
Copy link
Contributor

@xuanzhang816 xuanzhang816 commented Jul 18, 2025

Stack from ghstack (oldest at bottom):

This is an reland attempt of #157563, but insteading of introducing the realize_acc_reads_size_threshold config and setting to a default value, we set it to None for now to unblock an internal use case. Will deep dive into the issue and harden the logic in later PRs.

cc @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @ipiszy @chenyang78 @kadeng @muchulee8 @amjames @chauhang @aakhundov @coconutruben

[ghstack-poisoned]
Copy link

pytorch-bot bot commented Jul 18, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/158667

Note: Links to docs will display an error until the docs builds have been completed.

⏳ No Failures, 2 Pending

As of commit 34ffcc2 with merge base 70b4a88 (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

xuanzhang816 added a commit that referenced this pull request Jul 18, 2025
ghstack-source-id: 9b42230
Pull Request resolved: #158667
@xuanzhang816
Copy link
Contributor Author

@pytorchbot label "topic: not user facing"

@pytorch-bot pytorch-bot bot added the topic: not user facing topic category label Jul 18, 2025
@xuanzhang816
Copy link
Contributor Author

@pytorchbot rebase

@pytorchmergebot
Copy link
Collaborator

@pytorchbot started a rebase job onto refs/remotes/origin/viable/strict. Check the current status here

[ghstack-poisoned]
pytorchmergebot pushed a commit that referenced this pull request Jul 21, 2025
ghstack-source-id: ed6d989
Pull Request resolved: #158667
@pytorchmergebot
Copy link
Collaborator

Successfully rebased gh/xuanzhang816/21/orig onto refs/remotes/origin/viable/strict, please pull locally before adding more changes (for example, via ghstack checkout https://github.com/pytorch/pytorch/pull/158667)

@xuanzhang816 xuanzhang816 requested a review from yf225 July 21, 2025 16:58
@xuanzhang816
Copy link
Contributor Author

@pytorchbot merge

@pytorch-bot pytorch-bot bot added the ciflow/trunk Trigger trunk jobs on your pull request label Jul 21, 2025
@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in the wiki.

Questions? Feedback? Please reach out to the PyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

saienduri pushed a commit to saienduri/pytorch that referenced this pull request Jul 22, 2025
This is an reland attempt of pytorch#157563, but insteading of introducing the `realize_acc_reads_size_threshold` config and setting to a default value, we set it to `None` for now to unblock an internal use case. Will deep dive into the issue and harden the logic in later PRs.

Pull Request resolved: pytorch#158667
Approved by: https://github.com/yf225
def fusion_accumulate_large_reads(
self, node1: BaseSchedulerNode, node2: BaseSchedulerNode, threshold: int
) -> bool:
all_reads = (node1.read_writes.reads | node2.read_writes.reads) - (
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Curious that besides reads, will writes also be a problem, will the buffer for writes be hold by the kernel? If a large write buffer is generated by the last node of a fusion, with fusion, will the large write buffer be hold from the beginning of the fusion kernel?

@github-actions github-actions bot deleted the gh/xuanzhang816/21/head branch August 26, 2025 02:14
pytorchmergebot pushed a commit that referenced this pull request Sep 13, 2025
This is to revert some of the changes in #158667

In particular, we only disallow fusion of large accumulate read at IR level and not at scheduler level, as users can create their own custom fusion logics for the scheduler level.

Pull Request resolved: #161978
Approved by: https://github.com/yf225
markc-614 pushed a commit to markc-614/pytorch that referenced this pull request Sep 17, 2025
This is to revert some of the changes in pytorch#158667

In particular, we only disallow fusion of large accumulate read at IR level and not at scheduler level, as users can create their own custom fusion logics for the scheduler level.

Pull Request resolved: pytorch#161978
Approved by: https://github.com/yf225
mansiag05 pushed a commit to mansiag05/pytorch that referenced this pull request Sep 22, 2025
This is to revert some of the changes in pytorch#158667

In particular, we only disallow fusion of large accumulate read at IR level and not at scheduler level, as users can create their own custom fusion logics for the scheduler level.

Pull Request resolved: pytorch#161978
Approved by: https://github.com/yf225
cleonard530 pushed a commit to cleonard530/pytorch that referenced this pull request Sep 22, 2025
This is to revert some of the changes in pytorch#158667

In particular, we only disallow fusion of large accumulate read at IR level and not at scheduler level, as users can create their own custom fusion logics for the scheduler level.

Pull Request resolved: pytorch#161978
Approved by: https://github.com/yf225
dsashidh pushed a commit to dsashidh/pytorch that referenced this pull request Sep 26, 2025
This is to revert some of the changes in pytorch#158667

In particular, we only disallow fusion of large accumulate read at IR level and not at scheduler level, as users can create their own custom fusion logics for the scheduler level.

Pull Request resolved: pytorch#161978
Approved by: https://github.com/yf225
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants