New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Use only 1 hasher #2
Comments
(note: the linked repo will become public soon and isn't yet public) |
I profiled the execution time of the pipeline. The speedup to using only 1 hasher rather than 80 is quite good. Also, it's weird that using threads on hashers (rather than 1 thread) was performing slower. I tested on a 32 core computer to use many threads for the hashers in parallel (n_jobs=-1 param in feature union), and it was even slower than using no thread. |
Hi, have you replicate the model in their paper? |
@tnlin I did not implement the neural network layer on top of the projection layer. This is only the projection, and it also differs a little bit from the paper. |
Hey @tnlin , did you manage to find the real performance of SGNN? I can only achieve 71% with their architecture. |
Instead of using a FeatureUnion over T=80 random hashers of d=14 dimensions (80*14=1120 word features), use only one hasher (1x1120), which results in a dramatic speedup.
You can see the fix here: https://github.com/guillaume-chevalier/NLP-TP3
The text was updated successfully, but these errors were encountered: