Replies: 2 comments
-
hi @vivekgoe , plz see deepspeedai/Megatron-DeepSpeed#415 |
Beta Was this translation helpful? Give feedback.
0 replies
-
@inkcherry Thanks for sharing PR, will check this. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I am looking to optimize performance for sequence parallel implementation using DistributedAttention by overlapping communication and compute. I see that there is some code for this available under "sp_overlap_comm" flag in
DeepSpeed/deepspeed/sequence/layer.py
Line 311 in 02bbf50
Is this feature (sp_overlap_com) functional? If yes, is there an example or blog about this which I can use to understand the feature. Will really appreciate it if someone can help answering this. Thanks.
Beta Was this translation helpful? Give feedback.
All reactions