Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Teacher forcing on TIMIT and GRID dataset #29

Open
hjzzju opened this issue Mar 29, 2021 · 2 comments
Open

Teacher forcing on TIMIT and GRID dataset #29

hjzzju opened this issue Mar 29, 2021 · 2 comments

Comments

@hjzzju
Copy link

hjzzju commented Mar 29, 2021

Hi, I want to know how to set teacher forcing in GRID and TCDTIMIT dataset. The same as lip2wav dataset? teacher forcing decay from 29000 steps?

@Rudrabha
Copy link
Owner

Rudrabha commented Apr 4, 2021

You can decay it earlier. Start from 1000 steps or something similar. You may decay within 10,000 steps and then let it train without teacher forcing for sometime.

@Domhnall-Liopa
Copy link

Hi,

With tacotron_teacher_forcing_mode="constant" during training, the teacher forcing ratio is never decayed and always stays at 1. Then in synthesizer/models/helpers.py the following code is used to select the groundtruth or output of the previous time-step:

next_inputs = tf.cond(
      tf.less(tf.random_uniform([], minval=0, maxval=1, dtype=tf.float32), self._ratio),
      lambda: self._targets[:, time, :],
      lambda: outputs[:,-self._output_dim:])

Since the ratio is always 1 and never decayed, the decoder is just passed in the groundtruth of the previous time-step for the entire training. Is this expected? Should there be a switch at some point so the outputs of the previous time-step are passed during training?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants