You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Describe the bug
I wanted resume the training from a saved checkpoint
The green curve is previously trained model
The red curve is restart of the training after loading 450M steps check point of green curve by the following curve.
As you can see the red curve looks like it didn't load the weight at all as it pattern matches a curve that trained from scratch
I wanted to ask if the above code is how resume training works in d3rlpy, and if not, how to resume training the modeling from a checkpoint with the epoch, learning rate, entropy...etc that left from the checkpoint. Thank you!
The text was updated successfully, but these errors were encountered:
I believe that you should be able to get the expected performance by dropping random_steps. Let me close this issue. Feel free to reopen this if there is any further discussion.
Describe the bug
I wanted resume the training from a saved checkpoint
The green curve is previously trained model
The red curve is restart of the training after loading 450M steps check point of green curve by the following curve.
As you can see the red curve looks like it didn't load the weight at all as it pattern matches a curve that trained from scratch
below code is how I loaded the model
I wanted to ask if the above code is how resume training works in d3rlpy, and if not, how to resume training the modeling from a checkpoint with the epoch, learning rate, entropy...etc that left from the checkpoint. Thank you!
The text was updated successfully, but these errors were encountered: