Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Don't persist iteration number in model checkpoints #37

Open
tcbegley opened this issue May 12, 2023 · 0 comments
Open

Don't persist iteration number in model checkpoints #37

tcbegley opened this issue May 12, 2023 · 0 comments
Labels
bug Something isn't working

Comments

@tcbegley
Copy link
Owner

Currently training the reward model starts at the iteration number of the transformer checkpoint, which is weird. We should just start counting iterations from 0 in the reward model training loop (assuming that the reward model is being trained from scratch, if loading a reward model from a checkpoint then we can count from the checkpointed iteration number of the reward model) regardless of how many iterations the transformer was trained for.

@tcbegley tcbegley added the bug Something isn't working label May 12, 2023
@tcbegley tcbegley assigned vmoens and unassigned vmoens May 12, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants