New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Refine self_attention_fuse pass with removing the size restriction #56909
Conversation
你的PR提交成功,感谢你对开源项目的贡献! |
@tianshuo78520a Hi, we've added test cases for this PR but it seems cannot pass the c++ converage check. Can you give some suggestions on that? |
@zoooo0820 @jeff41404 Hi, this is an enhancement for the current fluid operator. Can you please check on it? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi, there are some fluid APIs, which will be deprecated very soon. Could you please modify them to our recommended APIs?
By the way, we are planning to remove fluid APIs in this week, and this PRs may be conflicted. Is it possible to merge this PR in next week, thanks.
Thanks for your suggestions, we'll hold on this PR. |
Hi, @xinyu-intel @abenmao |
b6e355d
to
cb83604
Compare
cb83604
to
7cca955
Compare
Done~ But The PR-CI-Api-Benchmark seems to be hung for some reason. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
PR types
Bug fixes
PR changes
Others
Description
remove the size restriction in self_attention_fuse pass. Fix PaddleOCR v4 on AVX512 CPU.