-
Notifications
You must be signed in to change notification settings - Fork 306
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Different input dimention compared to output dimension #12
Comments
Hi, When n_in != n_out, we simply add one more linear transform (say W) at the highway connection: Again, this multiplication can be batched together with other mm as well. that said, this (Wx) is being put into U. and I guess that's why you find the code confusing. We should better document the implementation. Thanks for pointing this out! |
Also, in speech task, the (Wx) is always included, which is discussed in the appendix. |
Thanks for the quick reply! Just for further clarification, this new |
yes, the new W will be different to (3). and no bias term for them. only the neural gates have biases (4) (5) |
Fantastic. Thank you very much for the clarification, and fantastic work on this efficient implementation in CUDA. |
You are very welcome! |
Hi, I'm trying to implement a naive version of this paper in Keras, and was wondering how is the case that - n_in != n_out handled.
I went through the code a few times, and couldn't understand the element wise multiplication of (1 - r_t) with x_t, if x_t is of a different shape than r_t.
The text was updated successfully, but these errors were encountered: