Skip to content
This repository has been archived by the owner on Dec 7, 2022. It is now read-only.

High variance in solution quality #76

Closed
leonlan opened this issue Aug 31, 2022 · 6 comments
Closed

High variance in solution quality #76

leonlan opened this issue Aug 31, 2022 · 6 comments
Assignees
Labels
research Research related discussions static

Comments

@leonlan
Copy link
Collaborator

leonlan commented Aug 31, 2022

When running experiments with the same setting, I notice that there is high variance in solution qualities among different runs. See e.g. plots #75, where restarting can sometimes be very good, and sometimes be very bad.

I think it's worthwhile to let the algorithm run many times (without restarting) and analyze the solutions.

@leonlan leonlan added the research Research related discussions label Aug 31, 2022
@leonlan leonlan self-assigned this Aug 31, 2022
@leonlan
Copy link
Collaborator Author

leonlan commented Sep 26, 2022

As can be seen from the submission benchmark, there's a lot of variance in HGS. I didn't benchmark the static problem properly (using single seed), which led me to believe that our algorithm is outperforming the original baseline. It kinda is, but also not.

@leonlan
Copy link
Collaborator Author

leonlan commented Sep 26, 2022

For the curious, here are three plots showing the variance in objective value for 32 runs, each with a different seeds. Stopping after 10K no improvements (so disregarding max runtime and iterations). Here are some initial plots to demonstrate the variances over solving runs. Gaps are w.r.t. the best known solutions which I have collected over time.

variance-300
variance-600
variance-900

Ignore the outliers with 3+% mean, those were as a result of nbVeh being too small.

@leonlan leonlan mentioned this issue Sep 27, 2022
@leonlan leonlan added the static label Sep 30, 2022
@N-Wouda
Copy link
Owner

N-Wouda commented Oct 10, 2022

Comparing briefly the averages of my last benchmark run (ten different seeds): the average cost was 164316.8, with a standard deviation of 15.11. Each average cost is the average of 250 instances, but there still remains a fairly significant variation in quality: the range between min-max is 54; that's basically the entire margin on the static solver leaderboard.

@N-Wouda
Copy link
Owner

N-Wouda commented Oct 29, 2022

In #33 I also briefly commented on the standard deviations of the dynamic runs. Those are much more variable than the static ones.

@leonlan
Copy link
Collaborator Author

leonlan commented Oct 30, 2022

Note to self: make a pull request of the notebook that I used to analyze the variance in quality.

@leonlan
Copy link
Collaborator Author

leonlan commented Nov 21, 2022

I don't have time to do this anymore.

@leonlan leonlan closed this as completed Nov 21, 2022
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
research Research related discussions static
Projects
None yet
Development

No branches or pull requests

2 participants