Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training hyperparameters #26

Open
lhfowl opened this issue Jan 21, 2020 · 4 comments
Open

Training hyperparameters #26

lhfowl opened this issue Jan 21, 2020 · 4 comments

Comments

@lhfowl
Copy link

lhfowl commented Jan 21, 2020

Hello,

I'm hoping to confirm that the hyperparameters specified in your paper are correct. Specifically, for miniimagenet, 100k meta steps were taken during training? I ask because it seems some of the default values in the code are different.

@unixpickle
Copy link
Contributor

See the readme: https://github.com/openai/supervised-reptile/blob/master/README.md#reproducing-training-runs which describes how to run the experiments by specifying the correct arguments

@siavash-khodadadeh
Copy link

siavash-khodadadeh commented Aug 30, 2020

In the readme file train-shots is 10, however, it seems that this should be 5-way 1-shot. Am I missing something?

transductive 1-shot 5-way Omniglot.

python -u run_omniglot.py --shots 1 --inner-batch 10 --inner-iters 5 --meta-step 1 --meta-batch 5 --meta-iters 100000 --eval-batch 5 --eval-iters 50 --learning-rate 0.001 --meta-step-final 0 --train-shots 10 --checkpoint ckpt_o15t --transductive

@siavash-khodadadeh
Copy link

Does this mean that during meta-learning we are training with 5-way 10-shot? but for the test, we evaluate on 5-way, 1-shot?

@MrDavidG
Copy link

MrDavidG commented Nov 7, 2020

@siavash-khodadadeh I have the same question. And I notice that the paper said

If we are doing K-shot, N-way classification, then we sample tasks by selecting N classes from C and then selecting K + 1 examples for each class. We split these examples into a training set and a test set, where the test set contains a single example for each class. 

It is a different setting with maml in evalution, maml use k_qry=15 (that is 15 examples for each class) to evaluate itself. It seems like the comparison in the experiment is unfair?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants