Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

add support for adamw schedulefree #1486

Open
wants to merge 1 commit into
base: main
Choose a base branch
from
Open

Conversation

winglian
Copy link
Collaborator

@winglian winglian commented Apr 6, 2024

implements meta's https://github.com/facebookresearch/schedule_free for adamw

https://twitter.com/aaron_defazio/status/1776320004465582331

optimizer: schedule_free_adamw
lr_scheduler: constant

@winglian
Copy link
Collaborator Author

winglian commented Apr 6, 2024

upstreaming this @ huggingface/transformers#30079

@seungduk-yanolja
Copy link
Contributor

I got the following error:

Traceback (most recent call last):
  File "/home/seungduk/miniconda3/envs/axo/lib/python3.10/runpy.py", line 196, in _run_module_as_main
    return _run_code(code, main_globals, None,
  File "/home/seungduk/miniconda3/envs/axo/lib/python3.10/runpy.py", line 86, in _run_code
    exec(code, run_globals)
  File "/data/shared/apps/axolotl/src/axolotl/cli/train.py", line 59, in <module>
    fire.Fire(do_cli)
  File "/home/seungduk/miniconda3/envs/axo/lib/python3.10/site-packages/fire/core.py", line 141, in Fire
    component_trace = _Fire(component, args, parsed_flag_args, context, name)
  File "/home/seungduk/miniconda3/envs/axo/lib/python3.10/site-packages/fire/core.py", line 475, in _Fire
    component, remaining_args = _CallAndUpdateTrace(
  File "/home/seungduk/miniconda3/envs/axo/lib/python3.10/site-packages/fire/core.py", line 691, in _CallAndUpdateTrace
    component = fn(*varargs, **kwargs)
  File "/data/shared/apps/axolotl/src/axolotl/cli/train.py", line 35, in do_cli
    return do_train(parsed_cfg, parsed_cli_args)
  File "/data/shared/apps/axolotl/src/axolotl/cli/train.py", line 55, in do_train
    return train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
  File "/data/shared/apps/axolotl/src/axolotl/train.py", line 104, in train
    trainer = setup_trainer(
  File "/data/shared/apps/axolotl/src/axolotl/utils/trainer.py", line 351, in setup_trainer
    return trainer_builder.build(total_num_steps)
  File "/data/shared/apps/axolotl/src/axolotl/core/trainer_builder.py", line 1377, in build
    trainer = trainer_cls(
  File "/data/shared/apps/axolotl/src/axolotl/core/trainer_builder.py", line 239, in __init__
    super().__init__(*_args, **kwargs)
  File "/home/seungduk/miniconda3/envs/axo/lib/python3.10/site-packages/transformers/trainer.py", line 537, in __init__
    raise RuntimeError(
RuntimeError: Passing `optimizers` is not allowed if Deepspeed or PyTorch FSDP is enabled. You should subclass `Trainer` and override the `create_optimizer_and_scheduler` method.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants