-
Notifications
You must be signed in to change notification settings - Fork 22
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Could you release the notebook to reproduce figure 10 of the paper? #8
Comments
Sorry for the delayed response, would you mind sharing the link to the wandb data for your run? |
Sure! This is the link to this particular run: |
thanks for sharing. there's three things that account for the discrepancy here
I will see what I can do about getting you a notebook to reproduce, but it may have to wait a bit while I deal with other work on my plate. In any case, I'm delighted you're taking the time to reproduce these experiments, if you have any further questions don't hesitate to ask :) |
Thank you for your suggestion! I will first fix these differences. |
sounds good. note that the seed is fixed in the config, so you'll want to be sure to override it, e.g.
|
one more thing, https://github.com/samuelstanton/lambo/blob/main/lambo/tasks/chem/chem.py#L105 |
Sure, thanks for the reminder! The experiments are still running, and I also wonder if it's reasonable that the single-objective experiments need 1 day 12 hours to finish, while the multi-objective experiments need just 5 hours? Intuitively the single-objective runs should be faster than the multi-objective ones? |
fair question. the single-objective experiment collects bigger batches of data over more rounds than the multi-objective experiments, so using exact GP inference would require a lot of GPU memory and would likely be numerically unstable. Instead for this task I use a variational GP, which has constant memory footprint and is more numerically stable for large datasets. Unfortunately variational GPs are fairly slow to train, which leads to the dramatic increase in runtime. There probably is room for optimization here, the current training recipe is optimized more for stability than speed. |
Thanks for the reply! It makes sense now. However, as I re-ran the experiments with seed 1, 2, 3, 4, I found that the optimization performance is still under expectation. The wandb loggings are the ones with id 12, 13, 14, 15 of the project: https://wandb.ai/thomaswang/lambo_replicate/groups/test/table?workspace=user-thomaswang I did not do cummin operations for the log outputs, but we can see that the min values for the obj_val_0 are around -7 in all runs, which are 7 for penalized logp. I wonder if there are some problems with the default configurations of the setting? Would it be possible for you to double check the configurations? On my side, I will also double check if there is something wrong with my reproduction. Thanks very much! |
hm ok I'll take a look, thanks for raising the issue |
Hi! I am also intereested in the single objective use case for LaMBO. Is there any update on reproducing the published numbers? |
@samuelstanton I'm also having some trouble reproducing the results, I ran the following line: It would be great to get an update on this, thank you! |
The major hypers that have been corrected are:
|
Sorry for the late response. Thank you for your effort! Previously I also found that the choice for the starting sequences matters a lot to the final results. I think I will close the issue. |
Hi Samuel!
Thank you for your repo and great work! It is reported in the paper that single-objective LaMBO can greatly outperform LSBO in the penalized logp task. I found your released notebooks reproducing figure 1 and figure 3 of the paper very helpful and it would be of great help if you could also release the notebook and the wandb loggings to reproduce the figure 10 in the appendix.
Thanks a lot in advance!
The text was updated successfully, but these errors were encountered: