-
Notifications
You must be signed in to change notification settings - Fork 19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to obtain results of multiple runs #5
Comments
Hi, GentleZhu, The 10 random runs can be run by setting the Please let me know if any questions. |
I found even I set this parameter and config.exp_round>1, the goodtg or load_task function only run for one round. |
Your code seems to only store different round in a separate folder under storage. |
Hi Gentle Zhu and GOOD team, I have a similar question. It seems we have to manually check each folder from different rounds. Is there any convenient way to aggregate the results? |
Hi!
That's true because we generally run For example, you may generate the following commands, and pack them into a list goodtg --exp_round 1 --gpu_idx 0--config_file XXX
goodtg --exp_round 2 --gpu_idx 1--config_file XXX
...
goodtg --exp_round 10 --gpu_idx 9 --config_file XXX After that, you may find the use of package cmd_args = [XXX, ..., XXX]
subprocess.Popen(shlex.split(cmd_args), close_fds=True, stdout=open('debug_out.log', 'a'), stderr=open('debug_error.log', 'a'), start_new_session=False) I believe the way to launch your programs on GPUs also depends on your experiment environment (If one is sharing computation resources with others, one cannot launch one's programs aggressively). Because the results are fully stored, you can aggregate all results after finishing running. BTW, if you only need to run experiments sequentially, you may find reproduce_round1 useful.
Since the log file saving paths (structures) are fully based on your config parameters according to log settings, we don't need to manually check the outcomes. After experiments are completed, another script is needed to read all results. Note that to facilitate reading these results, there is a special line at the end of each log file as the result conclusions. You can set the (Another way to aggregate results is to read the information stored in model checkpoints. (uncommon)) We will share some convenient scripts after we reorganize these scripts for more general purposes. Please let me know if any questions.:smile: |
Hi GentleZhu, We have updated this project to version 1. You can now launch multiple jobs and collect their results easily. Please refer to the new README. Please let me know if any questions. |
Hi GOOD Team,
Thanks for the great library. I have been successfully ran goodtg but I found it only runs for one time and reports the best epoch on validation. What's the best way to reproduce paper's result on 10 random runs ?
The text was updated successfully, but these errors were encountered: