Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

HuggingFace: Add log_model #641

Closed
daavoo opened this issue Jul 26, 2023 · 1 comment · Fixed by #649
Closed

HuggingFace: Add log_model #641

daavoo opened this issue Jul 26, 2023 · 1 comment · Fixed by #649
Assignees
Labels
A: frameworks Area: ML Framework integration enhancement p1-important Include in the next sprint

Comments

@daavoo
Copy link
Contributor

daavoo commented Jul 26, 2023

Similar to #640

@daavoo daavoo added enhancement A: frameworks Area: ML Framework integration labels Jul 26, 2023
@daavoo daavoo assigned daavoo and unassigned daavoo Jul 26, 2023
@dberenbaum dberenbaum added the p1-important Include in the next sprint label Jul 31, 2023
@daavoo daavoo mentioned this issue Aug 4, 2023
1 task
@dberenbaum
Copy link
Contributor

For the record, here's how it's handled in other loggers (see https://huggingface.co/docs/transformers/v4.31.0/en/main_classes/callback):

Wandb:

WANDB_LOG_MODEL (str, optional, defaults to "false"): Whether to log model and checkpoints during training. Can be "end", "checkpoint" or "false". If set to "end", the model will be uploaded at the end of training. If set to "checkpoint", the checkpoint will be uploaded every args.save_steps . If set to "false", the model will not be uploaded. Use along with load_best_model_at_end() to upload best model.

Mlflow:

HF_MLFLOW_LOG_ARTIFACTS (str, optional): Whether to use MLflow .log_artifact() facility to log artifacts. This only makes sense if logging to a remote server, e.g. s3 or GCS. If set to True or 1, will copy each saved checkpoint on each save in TrainingArguments’s output_dir to the local or remote artifact storage. Using it without a remote storage will just copy the files to your artifact location.

Neptune:

log_checkpoints (str, optional) — If “same”, uploads checkpoints whenever they are saved by the Trainer. If “last”, uploads only the most recently saved checkpoint. If “best”, uploads the best checkpoint (among the ones saved by the Trainer). If None, does not upload checkpoints.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
A: frameworks Area: ML Framework integration enhancement p1-important Include in the next sprint
Projects
No open projects
Archived in project
Development

Successfully merging a pull request may close this issue.

2 participants