Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What are the key hyper-parameters to tune #8

Closed
RaptorMai opened this issue Aug 26, 2020 · 1 comment
Closed

What are the key hyper-parameters to tune #8

RaptorMai opened this issue Aug 26, 2020 · 1 comment

Comments

@RaptorMai
Copy link

Hi,

We are trying to apply CN-DPM on another dataset called CORE50 and the result is quite bad. Since CNDPM has so many hyper-parameters, except for log-alpha and classifier-chill, what hyper-parameters do you think are crucial to tune? Thank you so much and I look forward to hearing back from you.

@soochan-lee
Copy link
Owner

Sorry for the late reply.
Here are some tips for finding a good set of hyperparameters.

First, check the task inference accuracy under a task-based setting.
There is an option called send_to_stm_always in the configuration files to simulate a task-based setting. If you set this to true, every training example is sent to the STM. Then set the stm_capacity to the number of examples in a task so that every example is trained in the sleep phase. You can now tune the VAEs in complete isolation. Even in this configuration, however, the VAEs would not be good at recognizing tasks.

Second, tune the training-time task inference.
You want the nl_cond_dist graph of each expert to be high only during its responsible task. Adjust classifier_chill to have a sufficient gap. Then set the log_alpha such that it sits between high nl_cond_dists and low nl_cond_dists. Note that the graphs can be noisy due to the wrong classifications and poor VAE accuracy.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants