Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix TSEMO #32

Merged
merged 122 commits into from Jul 17, 2020
Merged

Fix TSEMO #32

merged 122 commits into from Jul 17, 2020

Conversation

marcosfelt
Copy link
Member

@marcosfelt marcosfelt commented May 11, 2020

This is a PR for tracking improvements to TSEMO (issue #2). It is a work in progress.

I've made one change to the hypervolume reference, but I think it's still not the same as the Bradford paper.

  • Test functions (DTLZ2, vmlop)
  • Correct reference based on Bradford et al.
  • Set up spectral sampling
  • Do some ablation experiments on fixed vs ideal-anti-ideal difference and random rate.

@marcosfelt marcosfelt changed the title Fix tsemo Fix TSEMO May 11, 2020
@marcosfelt
Copy link
Member Author

marcosfelt commented Jul 13, 2020

I've discovered a lot of subtle things about TSEMO. I want to note them here:

  • For the training of the GPs (i.e., hyperparameter optimisation), at least 100 optimisation restarts are needed to consistently find the optimal log likelihood, especially with small data. You also need to constrain the hyperparameters and add priors so they won't train to extremely large or small values early on.
  • Sepctral sampling is very temperamental and slow (it scales with m^1/2). It will randomly fail due to problems with the singular value decomposition (SVD). Therefore, I have to implement "retries" in the code to restart the spectral sampling in these cases. Also, you need at least 1500 sampling points to get good results and 4000 preferably. For the code itself, I went with pyrff instead of my own code, but it requires saving down the sampled functions to disk and then copying them back up.
  • You need a good implementation of NSGA-II. I was working with platypus for a while but then realized it was not performing well and switched to pymoo.

@marcosfelt marcosfelt requested a review from jangerit July 14, 2020 22:52
@marcosfelt
Copy link
Member Author

I'm done with TSEMO now. I don't think the tests will pass because one of the Gryffin tests seems to have a problem. However, otherwise, things should be working now.

@jangerit
Copy link
Contributor

jangerit commented Jul 14, 2020

Great! Being able to use TSEMO in python now is really cool.

About Gryffin: for the test functions we implemented it seems like Gryffin only gets close the optimal points but does not find those exactly. That's why the tests sometimes fail. Also, the Gryffin tests take quite long. Maybe we should comment them out in test_strategies.py or rather just look whether the code runs without errors but exclude the full optimization runs (this would be comment out lines: 597, 601, 673, 677 in test_strategies.py).`

@marcosfelt
Copy link
Member Author

marcosfelt commented Jul 14, 2020 via email

jangerit
jangerit previously approved these changes Jul 14, 2020
@jangerit
Copy link
Contributor

maybe we can just run a few iterations of TSEMO and Gryffin to see if the best function values are improving

That's a good idea! We should also do this for SOBO, I think. This should reduce the time for the pytest runs a lot.

@marcosfelt marcosfelt merged commit f33590c into master Jul 17, 2020
@marcosfelt marcosfelt linked an issue Jul 17, 2020 that may be closed by this pull request
@jangerit jangerit mentioned this pull request Jul 18, 2020
@marcosfelt marcosfelt deleted the fix_tsemo branch August 3, 2020 17:43
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Correct issues with TSEMO and compare to Matlab implementation
2 participants