Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Stuck while training at 977 itr #258

Open
zusne opened this issue Nov 8, 2022 · 0 comments
Open

Stuck while training at 977 itr #258

zusne opened this issue Nov 8, 2022 · 0 comments

Comments

@zusne
Copy link

zusne commented Nov 8, 2022

I choose the point_gather, and use rllab to train 5000 itr, but it always get stuck when it come to 977 itr. I try to train 8 programs with different parameters in parallel, but all get stuck at 977 itr, and i turn to train only one program, get the same result. I think thers is something wrong with my computer, so i change a computer to train the same program 5000 itr, but also get the same result: get stuck while training at 977. I am so comfused about it. All the progeams will stuck like this:


0% [# ] 100% | ETA: 00:00:10% [## ] 100% | ETA: 00:00:10% [### ] 100% | ETA: 00:00:10% [#### ] 100% | ETA: 00:00:10% [##### ] 100% | ETA: 00:00:00% [###### ] 100% | ETA: 00:00:00% [####### ] 100% | ETA: 00:00:00% [######## ] 100% | ETA: 00:00:00% [######### ] 100% | ETA: 00:00:00% [########## ] 100% | ETA: 00:00:00% [########### ] 100% | ETA: 00:00:00% [############ ] 100% | ETA: 00:00:00% [############# ] 100% | ETA: 00:00:00% [############## ] 100% | ETA: 00:00:00% [############### ] 100% | ETA: 00:00:00% [################ ] 100% | ETA: 00:00:00% [################# ] 100% | ETA: 00:00:00% [################## ] 100% | ETA: 00:00:00% [################### ] 100% | ETA: 00:00:00% [#################### ] 100% | ETA: 00:00:00% [##################### ] 100% | ETA: 00:00:00% [###################### ] 100% | ETA: 00:00:00% [####################### ] 100% | ETA: 00:00:00% [######################## ] 100% | ETA: 00:00:00% [######################### ] 100% | ETA: 00:00:00% [########################## ] 100% | ETA: 00:00:00% [########################### ] 100% | ETA: 00:00:00% [############################ ] 100% | ETA: 00:00:00% [############################# ] 100% | ETA: 00:00:00% [##############################] 100% | ETA: 00:00:00
Total time elapsed: 00:00:12


all thing seem to stop, but no error report. I wonder why, or is there some way to load the param about the trained model before?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant