-
Notifications
You must be signed in to change notification settings - Fork 232
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Docs] Add accelerate related #504
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
每个功能项单独一个文档,accelerated 是一个文件夹
|
docs/zh_cn/accelerate/flash_attn.md
Outdated
|
||
| 模型 | Flash Attention | | ||
| :----------: | :----------------: | | ||
| baichuan 1/2 | :x: | |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
不支持的原因是啥?这个可以写出来吗
xtuner/configs/llama_speed_benchmark/llama2_70b/llama2_70b_full_alpaca_enzh_128k_sp8.py
Outdated
Show resolved
Hide resolved
xtuner/configs/llama_speed_benchmark/llama2_70b/llama2_70b_full_alpaca_enzh_128k_sp8.py
Outdated
Show resolved
Hide resolved
xtuner/configs/llama_speed_benchmark/llama2_70b/llama2_70b_full_alpaca_enzh_128k_sp8.py
Outdated
Show resolved
Hide resolved
xtuner/configs/llama_speed_benchmark/llama2_70b/llama2_70b_full_alpaca_enzh_256k_sp16.py
Outdated
Show resolved
Hide resolved
xtuner/configs/llama_speed_benchmark/llama2_70b/llama2_70b_full_alpaca_enzh_256k_sp16.py
Show resolved
Hide resolved
@@ -0,0 +1,40 @@ | |||
# 超参数设置 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
# 超参数设置 | |
# 如何调整加速策略 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
超参数设置应该变成一个小节,再新增一个 support matrix 的小节,列明每个模型能用的功能
Docs about
LengthGroupedSampler
will be added in pr#511