Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

inference support flash_attn #64213

Merged

Conversation

yuanlehome
Copy link
Contributor

@yuanlehome yuanlehome commented May 11, 2024

PR Category

Inference

PR Types

Performance

Description

Pcard-71500

Related pr: PaddlePaddle/Paddle-Inference-Demo#522

Copy link

paddle-bot bot commented May 11, 2024

你的PR提交成功,感谢你对开源项目的贡献!
请关注后续CI自动化测试结果,详情请参考Paddle-CI手册
Your PR has been submitted. Thanks for your contribution!
Please wait for the result of CI firstly. See Paddle CI Manual for details.

@yuanlehome yuanlehome force-pushed the inference_support_flashattn branch 2 times, most recently from a27fb36 to c406720 Compare May 11, 2024 05:44
@yuanlehome yuanlehome merged commit 95e5b9e into PaddlePaddle:develop May 13, 2024
30 of 31 checks passed
co63oc pushed a commit to co63oc/Paddle that referenced this pull request May 13, 2024
* inference support flash_attn

* fix

* add ut
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants