Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Revision -- Feedback 1 #3

Closed
24 of 25 tasks
apoorvagnihotri opened this issue Jun 17, 2019 · 5 comments
Closed
24 of 25 tasks

Revision -- Feedback 1 #3

apoorvagnihotri opened this issue Jun 17, 2019 · 5 comments
Assignees

Comments

@apoorvagnihotri
Copy link
Collaborator

apoorvagnihotri commented Jun 17, 2019

  • Add reference to matern kernel
  • For prior model figure, also add the uncertainty in the legend
  • make the alpha higher for the uncertainty in all the plots.
  • the first plot showing only the GT has thin lines, others have thicker lines. I would suggest having line width something in between the two.
  • Make all the plots uniform
  • Verify the header levels in HTML
  • Change the trivial acquisition function's name in plots
  • Slow down all the animations
  • Show the CDF in a darker (lesser alpha) so that we can clearly see what is happening.
  • Put hyper parameters in new line in the plots (after iteration number). Put them with mathematical symbols (latex in matplotlib). Also, no need to mention HPar,ams word in the plots.
  • Do the same in html also. Like, do not write eps instead write \epsilon. I have made these corrections before EI. Please do it for everything that follows.
  • For Expected improvement, first write the formula in terms of expectation in improvement See slide labelled: An expected utility criterion in Nando De Freitas’ lecture: https://www.cs.ubc.ca/~nando/540-2013/lectures/l7.pdf
  • In the plots replace EI with \alpha_{EI} etc. for all the approaches
  • Write the formulae for GP-UCB and mention in text that it is similar to the first acquisition function that we had introduced. The text for GP-UCB seems broken (v to 3 …)
  • Don’t bring about Bernoulli discussion for Thomson sampling. Instead, add missing intuition: why choosing random samples from posterior and maximizing them will likely do a tradeoff between exploration and exploitation.
  • For hyperparams v/s params - just mention equation of Ridge regression and mention the params (\theta) and the hyperparams (\delta^2, learning rate,..)
  • The fonts on SVM and below plots are very small. Also, try with some other cmap options also. Also, the maximum point isn’t clearly visible for many of the plots.

  • I think we may want to ensure that the convention for bayesian optimisation is the same across the method. We could stick to the convention you used in the EI section just now. I had written slightly differently for PI. It may be best to just use Nando de Freitas' convention as used in the paper linked in TODO Master issue #2
  • https://arxiv.org/pdf/1012.2599.pdf is a nice paper to refer. I think we can just rename our first function as: UCB instead of ACQ1. You can also view Nando's lecture at 1:06 mins to understand why UCB would work.
  • From paper mentioned in TODO Master issue #2, we can update some text about GP-UCB also.
  • Don't write Plotting Posterior etc. as the title. Posterior should be enough.
  • For the PI visualisation wrt epsilon, I don't think the effect of epsilon is visible. We should be able to somehow see the larger variance points being chosen. Maybe plot the PI on a log-scale for the same. Also, make the epsilon = 0 as 0.0. This will ensure the plot doesn't jump.
  • For NNs, you might want to store all the results. So that if we have to just tweak the visualistions, we do not have to re-run the entire code.
  • The EI plots have PI as the ylabel currently.
  • For the random plot in the comparison section for 1d, run the random acquisition with multiple random initialisations and show the mean and the variance.

Need inputs
Add thompson sampling to SVM and others & NN for diff methods

@apoorvagnihotri
Copy link
Collaborator Author

Dear Prof. @nipunbatra

I think we may want to ensure that the convention for bayesian optimisation is the same across the method. We could stick to the convention you used in the EI section just now. I had written slightly differently for PI. It may be best to just use Nando de Freitas' convention as used in the paper linked in #2

Should I first mention the convention at the top before introducing the acquisition functions and then use the symbols in the formulae below?

@nipunbatra
Copy link
Collaborator

I think introducing as and when required is fine.

@apoorvagnihotri
Copy link
Collaborator Author

Dear Prof. @nipunbatra
The function that we had implemented for GPUCB is having a bit different formulation than what is mentioned in the paper. Could you please provide me your inputs on that?

apoorvagnihotri added a commit that referenced this issue Jun 18, 2019
@apoorvagnihotri
Copy link
Collaborator Author

Dear Prof. @nipunbatra

For hyperparams v/s params - just mention equation of Ridge regression and mention the params (\theta) and the hyperparams (\delta^2, learning rate,..)

Do you mean for the random forests? I have added the lines specifying hyper-params and params for RFs.

@apoorvagnihotri
Copy link
Collaborator Author

Dear Prof. @nipunbatra

  • I have the plots for NN using different methods, should I add it too?
  • SVM and RF were not tested on Thompson and UCB, do we need to test them on them too?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants