-
Notifications
You must be signed in to change notification settings - Fork 49
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Slow Execution of MOBSTER Algorithm for Tabulated Benchmarks #712
Comments
Hello, I don't think your script has any issues. The real-world runtime depends on two factors:
The simulator backend on a tabulated benchmark eliminates the first factor, but the second factor remains. While simple methods like RandomSearch and ASHA have very cheap decision making (just random sampling), BayesianOptimization and MOBSTER need a substantial amount of time for decision making. You can decrease this decision making time in various ways, as said in our docs. But in the end, you need to be careful, because what matters in the end are the results against real wall-clock time, where each trial has a runtime as in the tabulated benchmark. The fact that simulations run faster for one method than another, has no real significance for comparing them. BO often does much better than RS, unless you have extremely cheap trials. |
Your second question: The real wall-clock time for an experiment is what is logged in the results, in the column of name ST_TUNER_TIME. That is why if you plot results, you get the correct results (as if trials really took the time in the table). |
BTW: I strongly encourage you to look at this tutorial: Using this is much simpler than writing launcher scripts. At the moment, |
Thank you for your thorough explanations! I am studying it and will reach out again by opening a new issue if having other questions. |
Hello,
Thanks for your well-developed syne-tune.
I was trying to run the MOBSTER algorithm on the tabulated benchmarks (e.g., "lcbench") using simulator backend. I have attached my launching script below for reference. The running time is too long than expected. It reported a wallclock-time of 250s and 116s for the "Fashion-MNIST" and "christine" datasets respectively. However, ASHA only uses 0.25s and 0.13s respectively, for the same benchmark (datasets) and settings. I was wondering if my launching script has any issues.
Based on my understanding, the tabulated benchmarks are designed to accelerate and reduce the cost of experiments by bypassing the actual training procedure. And the wallclock-time reported should be the program execution time on my local machine. So I have another question: How can I obtain the execution time of the tabulated benchmarks (including the real training part)?
Thanks for your attention.
Launching script:
The text was updated successfully, but these errors were encountered: