-
Notifications
You must be signed in to change notification settings - Fork 18
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
which options #5
Comments
Hi Tsunehiko, thanks for your interest.
Please have a look at arguments.py or call python3 main.py --help to see a full list of arguments available. |
Thank you for your quick response. |
Funny, this repository is actually an adaptation of the following: https://github.com/ikostrikov/pytorch-a2c-ppo-acktr-gail, which is itself adapted from OpenAI baselines. I suppose one could use this environment with a different learning algorithm. There would be two roadblocks for you to overcome, at least that spring to mind:
Indeed it would be nice for this to be incorporated as part of the standard family of gym environments. I am working on it here for now because I'm still experimenting with the environment itself, e.g., playing with different reward functions, designing mini-games, playing with different map sizes. Which algorithm were you thinking of using in particular (stable_baselines is a collection of algorithms, not one single algorithm)? |
Feel free to shoot me an email if you have any more questions, ideas, or want to chat about RL :) |
Thank you for your kind explanation and kindness. In model.py, an error occurs when the modules of densenet_pytorch and ConvLSTMCell are not found. Which module should I install? I will ask you by email from next. |
Aha, densenet_pytorch is an old dependency, so I got rid of that, and I added ConvLSTMCell.py to the repo. Do a 'git pull' from inside the repo and try again! |
And indeed, you'd probably be wise to start with RL by playing with stable_baselines and some Atari games, etc. But I selfishly would rather you play with this repo because you're helping me troubleshoot it :) |
Thank you for your quick response. I'm glad if I'm useful too. By the way, I'm thinking of using this repo for at least July. |
Not ideal, but no problem. Just patched up something that was in the way of using no-cuda. I've now got the above command working with |
|
Strange, I can't replicate this. What are you getting in the command line? |
The command line output is very large, so I don't know which one to write. I wrote the last command line output. Also, the output of the screen at that time looks like an attached image.
|
This is the most recent command line output? If time is passing on the map, then training must be underway, so you must be getting printouts indicating avg. reward, number of frames and the like. Try the same with |
I tried both
|
How fast is time moving? An update only occurs for me in the late 1900s. You could make |
I tried --map-size 4 and --max-step 16, but the speed of the time was ultrafast and increased infinitely beyond the 1950s. |
I'm at a loss. I can only recommend looking at line 265 in main.py, which should be printing out, and working backward from there with print statements, trying to find out why 265 is never reached. |
Thank you for checking carefully. |
Also, try without |
That's some good news! I suspected that the problem stemmed from the GUI. In particular, there must be a call to gtk.main_interation(), or something to that effect, hidden somewhere. This function runs the GUI indefinitely, waiting for user input, so it would stop our training code dead in its tracks (and simply let the game run very fast). Strange that I don't experience the same issue on my end, and can't find a call to the function in the code. Might be an operating system-specific issue. As for the "Unable to init server" error, I think this might be the result of too many dead python processes hanging out on the cpu (interrupting training is not yet handled gracefully by the code). Try again after To see if the bot's doing anything, |
The array displayed by Also, "Unable to init server" is an error that was displayed when I tried using a GPU server that I can use, not my machine. This may be the cause. |
Yes, you can try something like |
Hello. I am not very familiar with reinforcement learning.
python3 main.py --log-dir trained_models/acktr --algo acktr --model squeeze --num-process 24 --map-width 27 --render
In the above, main.py does not work. Please tell me the options you need to run 'main.py'.
The text was updated successfully, but these errors were encountered: