-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Discussion: How to save pytorch model weights in each sub experiments #13
Comments
I thought that I've fixed the issue in the recent commit, but it doesn't seem so. |
Thanks for your kind and fast response @chrstnkgn 👍 I got that result under the most recent commit so far (bd7043a) which was committed 22/12/08 10:15 without any change in code set by command in terminal python main_Tuner.py --multirun loss=BCE,Multi_Soft_Margin model=tinynet_e,mobilenetv3_small_050 epochs=2 num_samples=2 optimizer=adam mode=default I'll run that again to double-check the result. Thanks again |
I've quickly checked the code and yes, the path for saving the weight is set under hydra run_dir for now. The problem is that the saving path for multirun is one level deeper than that of single run. As for now, I think I can come up with two options to solve this problem:
|
Both options look like absolutely good solutions! As far as I know, if I turn the ray tune mode on, then ray automatically makes a directory for each trail and the working directory also moves to the trial's directory. Then, it matches lovely with the torch.save("the current working directory") However, if we choose the solution (2), then I'm afraid it might be more complicated to make it match with the ray logging system. I am still not sure about that I fully understand your suggestion, so please let me know whenever if I misunderstood 😃 |
I think you did fully understand what I meant! Also, I agree with your opinion to make the working directory look more straightforward |
Any other opinions on this issue and solutions? @juppak @jieonh @seoulsky-field 🙏 |
What
Why
./logs
└── 2022-12-08_02-12-25
├── best_saved.pth
├── epochs=2,loss=BCE,mode=default,model=mobilenetv3_small_050,num_samples=2,optimizer=adam
│ └── main_Tuner.log
├── epochs=2,loss=BCE,mode=default,model=tinynet_e,num_samples=2,optimizer=adam
│ └── main_Tuner.log
├── epochs=2,loss=Multi_Soft_Margin,mode=default,model=mobilenetv3_small_050,num_samples=2,optimizer=adam
│ └── main_Tuner.log
├── epochs=2,loss=Multi_Soft_Margin,mode=default,model=tinynet_e,num_samples=2,optimizer=adam
│ └── main_Tuner.log
└── multirun.yaml
How
The text was updated successfully, but these errors were encountered: