Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Large embedding vector, large number of classes? #3

Open
lindsey98 opened this issue Dec 15, 2020 · 1 comment
Open

Large embedding vector, large number of classes? #3

lindsey98 opened this issue Dec 15, 2020 · 1 comment

Comments

@lindsey98
Copy link

Hi authors,
Thanks for pushing this fantastic tool. I have a question regarding to the approach itself: When my my last layer dimension is really large + I have large number of classes to cover, if we compute the gradient vector, its dimension(embDim * nLab) will be of orders of magnitude 10^4 or more. Do you think BADGE is an efficient solution in this case?

@pshwetank
Copy link

I am not the author of this paper but I have been studying the active learning approaches for a while. I think you can approach the problem with parallelisation on several GPUs. This will not be easy to implement but can be potentially solve the problem of excessively large fc layer. Actually, using a large fully connected layer might be even help in better generalisation as it will allow the sampler to get as many diverse points possible.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants