New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Reproducing the recent SL model featured in README.md #18
Comments
Hi, thanks for your attention to this project. Actually, the results on the README.MD can be trained by our provided scripts and the code in our project. The hyper-parameters are set the same as defined in this project. To use it, you should use the main (master) version, not the v1.05 release. The main version will be the up-to-publish v1.06 version. Note that the replays used to train are self-played by us on the Simple64. We will also provide them in a few days (no more than two weeks). As soon as the data are available, we will notify you (by @infphilo ) inside this issue. |
Thank you @liuruoze for the detailed response! I look forward to reproducing the model featured on README with your provided replay files (and with your code). |
Hi, @infphilo sorry for replying late. These days I was stuck in the fine-tuning of RL training of AlphaStar (It's very exhausting, plus we have added a new training way of multi-processing plus multi-thread to do it, making it harder to optimize). However, we will provide all the things (replays and an introduction to use) no later than the Chinese Spring Festival (this is the final due, we will not be beyond this deadline again). Sorry again for the lateness. And thanks for your attention! I will @ you when all is ready. |
Hi, @infphilo replays are published, you can check that in "https://github.com/liuruoze/Useful-Big-Resources/blob/main/replays/simple64_replays.7z", have a good time! |
The v_1.07 is also published, you can find an introduction to use supervised learning and these replays in "https://github.com/liuruoze/mini-AlphaStar/tree/v_1.07". Good luck! |
@liuruoze Thank you for sharing the replays and additional information! I'll try to reproduce your excellent work and let you know how it goes. Have a wonderful Chinese Spring Festival! |
Hello @liuruoze I hope you had a great Chinese Spring Festival! Thanks to your help and your hard work, I believe that I was able to produce some results by training the supervised model on your provided replay information. I used the same hyperparameters (NUM_EPOCHS = 10, LEARNING_RATE = 1e-4, STEP_SIZE = 30, GAMMA = 0.2, WEIGHT_DECAY = 1e-5). My trained model seems to perform worse than what you had on the README.md, perhaps because I might have used a different version of mini-AlphaStar. Nonetheless, I also trained the supervised model on 1,000 replays (SC2 version: 3.16.1, Protoss vs. Teran, mmr: 3,500, map: AbyssalReef), but the model doesn’t seem to perform well. Actions, selected units, and targets generated from the SL model don’t seem to make sense. These results may be in part due to overfitting. I have two questions:
Thank you again for your wonderful work and dedication to mAS. |
|
Thank you! My email address is infphilo@gmail.com |
Hi, @infphilo . We have sent you the email. Please check that. Another trick we have tried and think is useful: you can increase your WEIGHT_DECAY to a larger value. A larger weight decay may be more effective. The mini-AlphaStar has much fewer parameters than AlphaStar. Increasing the weight decay can add the robustness of the model to overfitting. |
Hi. It seems that there are no more questions. I will close this issue. If you find any new problems, please open a new issue to discuss them. |
@liuruoze, let me first thank you for your effort to develop a practical version of AlphaStar!
I am really interested in reproducing your recent SL-based agent, which you featured in README.md via several video clips. I’d like to know which StarCraft II version (e.g., 3.16.1) and training values (batch size and the number of epochs) you used when building the SL model.
Thank you again for your amazing work - hopefully, I can contribute to your work, too!
The text was updated successfully, but these errors were encountered: