Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Hyperparameter tuning - discrepancy between readme and code? #73

Open
maruker opened this issue Oct 19, 2019 · 0 comments
Open

Hyperparameter tuning - discrepancy between readme and code? #73

maruker opened this issue Oct 19, 2019 · 0 comments

Comments

@maruker
Copy link

maruker commented Oct 19, 2019

The readme states

For probing tasks, we used an MLP with a Sigmoid nonlinearity and and tuned the nhid (in [50, 100, 200]) and dropout (in [0.0, 0.1, 0.2]) on the dev set.

However, in the code it looks like the parameters given by the user are always used. No tuning takes place and no predefined hyperparameters are loaded. Maybe I missed something?

Should I do hyperparameter tuning to get results that are comparable to the literature?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant