-
Notifications
You must be signed in to change notification settings - Fork 574
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ZeRORuntimeException #596
Comments
def setup_model_and_optimizer(model_provider_func, load_lr_scheduler=True):
这是我方法所使用的函数 |
这是我的config文件 |
It seems this is a problem with DeepSpeed and the Apex optimizer. It is unrelated to bitsandbytes. |
File "/xddata/home/wanghaha/LM/src/megatron/training.py", line 122, in pretrain
model, optimizer, lr_scheduler = setup_model_and_optimizer(model_provider)
File "/xddata/home/wanghaha/LM/src/megatron/training.py", line 338, in setup_model_and_optimizer
model, optimizer, _, lr_scheduler = deepspeed.initialize(
File "/xddata/home/wanghaha/LM/src/deepspeed/init.py", line 165, in initialize
engine = DeepSpeedEngine(args=args,
File "/xddata/home/wanghaha/LM/src/deepspeed/runtime/engine.py", line 309, in init
self._configure_optimizer(optimizer, model_parameters)
File "/xddata/home/wanghaha/LM/src/deepspeed/runtime/engine.py", line 1172, in _configure_optimizer
self._configure_optimizer(optimizer, model_parameters)
File "/xddata/home/wanghaha/LM/src/deepspeed/runtime/engine.py", line 1172, in _configure_optimizer
self._configure_optimizer(optimizer, model_parameters)
File "/xddata/home/wanghaha/LM/src/deepspeed/runtime/engine.py", line 1172, in _configure_optimizer
self._configure_optimizer(optimizer, model_parameters)
File "/xddata/home/wanghaha/LM/src/deepspeed/runtime/engine.py", line 1172, in _configure_optimizer
raise ZeRORuntimeException(msg)
deepspeed.runtime.zero.utils.ZeRORuntimeException: You are using ZeRO-Offload with a client provided optimizer (<class 'apex.optimizers.fused_adam.FusedAdam'>) which in most cases will yield poor performance. Please either use deepspeed.ops.adam.DeepSpeedCPUAdam or set an optimizer in your ds-config (https://www.deepspeed.ai/docs/config-json/#optimizer-parameters). If you really want to use a custom optimizer w. ZeRO-Offload and understand the performance impacts you can also set <"zero_force_ds_cpu_optimizer": false> in your configuration file.
raise ZeRORuntimeException(msg)
在应用Zero3+offload时出现以上错误
The text was updated successfully, but these errors were encountered: