Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About time variant latent z_rnn #2

Closed
heyuanYao-pku opened this issue Mar 20, 2022 · 1 comment
Closed

About time variant latent z_rnn #2

heyuanYao-pku opened this issue Mar 20, 2022 · 1 comment

Comments

@heyuanYao-pku
Copy link

Hello, I have a question about the generation of time variant latent z_t

I think in the paper z_t depends on z_0,...,z_t-1.

$$q(z_t|z_{<t}) = ...$$

while in your code latents are sampled independently, although with a rnn hidden state.

features, _ = self.z_rnn(lstm_out)
z_mean = self.z_mean(features)
z_logvar = self.z_logvar(features)
z_post = self.reparameterize(z_mean, z_logvar, random_sampling=True)

I think the randomness of previous latent are not introduced to the cureent latent, I'm wondering whether this is the same meaning with the paper?

@JunwenBai
Copy link
Owner

The lstm hidden states should capture that. Such sampling was also adopted in RWAE, S3VAE, etc.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants