Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] MegatronMoELayerPolicy not mapped #5285

Open
nikit91 opened this issue Mar 15, 2024 · 0 comments
Open

[BUG] MegatronMoELayerPolicy not mapped #5285

nikit91 opened this issue Mar 15, 2024 · 0 comments
Labels
bug Something isn't working inference

Comments

@nikit91
Copy link

nikit91 commented Mar 15, 2024

Describe the bug
Not able to start inference for MoE based DS Megatron GPT model because the MegatronMoELayerPolicy (https://github.com/microsoft/DeepSpeed/blob/master/deepspeed/module_inject/containers/megatron_gpt_moe.py) that supports this feature is not being used/referenced anywhere.

To Reproduce
Steps to reproduce the behavior:

  1. Load DS Megatron GPT MoE model
  2. Pass it to deepspeed.init_inference
  3. inference engine (using the BaseContainer instead of BaseMoeContainer) tries to transpose weights of MoE layer with logic written for a single MLP layer

Expected behavior
Inference engine should load without an issue

@nikit91 nikit91 added bug Something isn't working inference labels Mar 15, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working inference
Projects
None yet
Development

No branches or pull requests

1 participant