Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

slim_model: options for float16, checkpoint selection #21

Merged
merged 2 commits into from
Jun 16, 2021

Conversation

nostalgebraist
Copy link
Contributor

Contains some additions I made to slim_model.py when I was using it with a finetuned model.

  • --f16: converts to float16

    • Why: Since many people are doing inference with float16 weights, it seems better to just go directly to float16, without the bfloat16 step
  • --ckpt-step: passing e.g. --ckpt-step 401 will load from step_401/ even if it's not the most recent checkpoint

    • Why: I wanted to convert the model with the lowest val loss, but usually there were 1 or more checkpoints after that step

Copy link
Owner

@kingoflolz kingoflolz left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Otherwise looks good!

slim_model.py Show resolved Hide resolved
@kingoflolz kingoflolz merged commit 960e694 into kingoflolz:master Jun 16, 2021
@kingoflolz
Copy link
Owner

Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants