You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In the tacotron paper, the expected dimensionality of the tensor during most of the encoder cbhg module is expected to be batch size x time steps x num features where num features = 128. Why do you shift to 256 before the bidirectional GRU layer? It seems like this would result in a real loss of information during encoding. Is this somehow what is described in the paper and I am just missing it? Thanks
The text was updated successfully, but these errors were encountered:
In the tacotron paper, the expected dimensionality of the tensor during most of the encoder cbhg module is expected to be batch size x time steps x num features where num features = 128. Why do you shift to 256 before the bidirectional GRU layer? It seems like this would result in a real loss of information during encoding. Is this somehow what is described in the paper and I am just missing it? Thanks
The text was updated successfully, but these errors were encountered: