Replies: 3 comments
-
>>> nmstoker |
Beta Was this translation helpful? Give feedback.
-
>>> sanjaesc |
Beta Was this translation helpful? Give feedback.
-
>>> alvaro.antelo |
Beta Was this translation helpful? Give feedback.
-
>>> alvaro.antelo
[June 26, 2020, 8:38pm]
I am having multiple problems trying to run inference on AWS, when I try
to use the server here is what happens:
python -m TTS.server.server slash --tts_checkpoint checkpoint_670000.pth.tar
slash --tts_config config.json
> Loading TTS model ... slash
> slash | slash > model config: config.json slash
> slash | slash > checkpoint file: checkpoint_670000.pth.tar slash
> Setting up Audio Processor... slash
> slash | slash > sample_rate:22050 slash
> slash | slash > num_mels:80 slash
> slash | slash > min_level_db:-100 slash
> slash | slash > frame_shift_ms:None slash
> slash | slash > frame_length_ms:None slash
> slash | slash > ref_level_db:20 slash
> slash | slash > num_freq:513 slash
> slash | slash > power:1.5 slash
> slash | slash > preemphasis:0.0 slash
> slash | slash > griffin_lim_iters:60 slash
> slash | slash > signal_norm:True slash
> slash | slash > symmetric_norm:True slash
> slash | slash > mel_fmin:0 slash
> slash | slash > mel_fmax:8000.0 slash
> slash | slash > max_norm:4.0 slash
> slash | slash > clip_norm:True slash
> slash | slash > do_trim_silence:True slash
> slash | slash > trim_db:60 slash
> slash | slash > do_sound_norm:False slash
> slash | slash > stats_path:None slash
> slash | slash > hop_length:256 slash
> slash | slash > win_length:1024 slash
> slash | slash > n_fft:1024 slash
> Using model: Tacotron2 slash
> Traceback (most recent call last): slash
> File '/home/ubuntu/anaconda3/envs/python3/lib/python3.6/runpy.py',
> line 193, in slash _run_module_as_main slash
> 'main', mod_spec) slash
> File '/home/ubuntu/anaconda3/envs/python3/lib/python3.6/runpy.py',
> line 85, in slash _run_code slash
> exec(code, run_globals) slash
> File
> '/home/ubuntu/anaconda3/envs/python3/lib/python3.6/site-packages/TTS-0.0.3+3366328-py3.6.egg/TTS/server/server.py',
> line 62, in slash
> synthesizer = Synthesizer(args) slash
> File
> '/home/ubuntu/anaconda3/envs/python3/lib/python3.6/site-packages/TTS-0.0.3+3366328-py3.6.egg/TTS/server/synthesizer.py',
> line 36, in init slash
> self.config.use_cuda) slash
> File
> '/home/ubuntu/anaconda3/envs/python3/lib/python3.6/site-packages/TTS-0.0.3+3366328-py3.6.egg/TTS/server/synthesizer.py',
> line 73, in load_tts slash
> self.tts_model.load_state_dict(cp slash ['model' slash ]) slash
> File
> '/home/ubuntu/anaconda3/envs/python3/lib/python3.6/site-packages/torch/nn/modules/module.py',
> line 830, in load_state_dict slash
> self.class.name, ' slash n slash t'.join(error_msgs))) slash
> RuntimeError: Error(s) in loading state_dict for Tacotron2: slash
> Missing key(s) in state_dict:
> 'encoder.convolutions.0.convolution1d.weight',
> 'encoder.convolutions.0.convolution1d.bias',
> 'encoder.convolutions.0.batch_normalization.weight',
> 'encoder.convolutions.0.batch_normalization.bias',
> 'encoder.convolutions.0.batch_normalization.running_mean',
> 'encoder.convolutions.0.batch_normalization.running_var',
> 'encoder.convolutions.1.convolution1d.weight',
> 'encoder.convolutions.1.convolution1d.bias',
> 'encoder.convolutions.1.batch_normalization.weight',
> 'encoder.convolutions.1.batch_normalization.bias',
> 'encoder.convolutions.1.batch_normalization.running_mean',
> 'encoder.convolutions.1.batch_normalization.running_var',
> 'encoder.convolutions.2.convolution1d.weight',
> 'encoder.convolutions.2.convolution1d.bias',
> 'encoder.convolutions.2.batch_normalization.weight',
> 'encoder.convolutions.2.batch_normalization.bias',
> 'encoder.convolutions.2.batch_normalization.running_mean',
> 'encoder.convolutions.2.batch_normalization.running_var',
> 'decoder.prenet.linear_layers.0.linear_layer.weight',
> 'decoder.prenet.linear_layers.1.linear_layer.weight',
> 'decoder.attention.location_layer.location_conv1d.weight',
> 'postnet.convolutions.0.convolution1d.weight',
> 'postnet.convolutions.0.convolution1d.bias',
> 'postnet.convolutions.0.batch_normalization.weight',
> 'postnet.convolutions.0.batch_normalization.bias',
> 'postnet.convolutions.0.batch_normalization.running_mean',
> 'postnet.convolutions.0.batch_normalization.running_var',
> 'postnet.convolutions.1.convolution1d.weight',
> 'postnet.convolutions.1.convolution1d.bias',
> 'postnet.convolutions.1.batch_normalization.weight',
> 'postnet.convolutions.1.batch_normalization.bias',
> 'postnet.convolutions.1.batch_normalization.running_mean',
> 'postnet.convolutions.1.batch_normalization.running_var',
> 'postnet.convolutions.2.convolution1d.weight',
> 'postnet.convolutions.2.convolution1d.bias',
> 'postnet.convolutions.2.batch_normalization.weight',
> 'postnet.convolutions.2.batch_normalization.bias',
> 'postnet.convolutions.2.batch_normalization.running_mean',
> 'postnet.convolutions.2.batch_normalization.running_var',
> 'postnet.convolutions.3.convolution1d.weight',
> 'postnet.convolutions.3.convolution1d.bias',
> 'postnet.convolutions.3.batch_normalization.weight',
> 'postnet.convolutions.3.batch_normalization.bias',
> 'postnet.convolutions.3.batch_normalization.running_mean',
> 'postnet.convolutions.3.batch_normalization.running_var',
> 'postnet.convolutions.4.convolution1d.weight',
> 'postnet.convolutions.4.convolution1d.bias',
> 'postnet.convolutions.4.batch_normalization.weight',
> 'postnet.convolutions.4.batch_normalization.bias',
> 'postnet.convolutions.4.batch_normalization.running_mean',
> 'postnet.convolutions.4.batch_normalization.running_var'. slash
> Unexpected key(s) in state_dict:
> 'encoder.convolutions.0.net.0.weight',
> 'encoder.convolutions.0.net.0.bias',
> 'encoder.convolutions.0.net.1.weight',
> 'encoder.convolutions.0.net.1.bias',
> 'encoder.convolutions.0.net.1.running_mean',
> 'encoder.convolutions.0.net.1.running_var',
> 'encoder.convolutions.0.net.1.num_batches_tracked',
> 'encoder.convolutions.1.net.0.weight',
> 'encoder.convolutions.1.net.0.bias',
> 'encoder.convolutions.1.net.1.weight',
> 'encoder.convolutions.1.net.1.bias',
> 'encoder.convolutions.1.net.1.running_mean',
> 'encoder.convolutions.1.net.1.running_var',
> 'encoder.convolutions.1.net.1.num_batches_tracked',
> 'encoder.convolutions.2.net.0.weight',
> 'encoder.convolutions.2.net.0.bias',
> 'encoder.convolutions.2.net.1.weight',
> 'encoder.convolutions.2.net.1.bias',
> 'encoder.convolutions.2.net.1.running_mean',
> 'encoder.convolutions.2.net.1.running_var',
> 'encoder.convolutions.2.net.1.num_batches_tracked',
> 'decoder.prenet.layers.0.linear_layer.weight',
> 'decoder.prenet.layers.0.bn.weight',
> 'decoder.prenet.layers.0.bn.bias',
> 'decoder.prenet.layers.0.bn.running_mean',
> 'decoder.prenet.layers.0.bn.running_var',
> 'decoder.prenet.layers.0.bn.num_batches_tracked',
> 'decoder.prenet.layers.1.linear_layer.weight',
> 'decoder.prenet.layers.1.bn.weight',
> 'decoder.prenet.layers.1.bn.bias',
> 'decoder.prenet.layers.1.bn.running_mean',
> 'decoder.prenet.layers.1.bn.running_var',
> 'decoder.prenet.layers.1.bn.num_batches_tracked',
> 'decoder.attention.location_layer.location_conv.weight',
> 'postnet.convolutions.0.net.0.weight',
> 'postnet.convolutions.0.net.0.bias',
> 'postnet.convolutions.0.net.1.weight',
> 'postnet.convolutions.0.net.1.bias',
> 'postnet.convolutions.0.net.1.running_mean',
> 'postnet.convolutions.0.net.1.running_var',
> 'postnet.convolutions.0.net.1.num_batches_tracked',
> 'postnet.convolutions.1.net.0.weight',
> 'postnet.convolutions.1.net.0.bias',
> 'postnet.convolutions.1.net.1.weight',
> 'postnet.convolutions.1.net.1.bias',
> 'postnet.convolutions.1.net.1.running_mean',
> 'postnet.convolutions.1.net.1.running_var',
> 'postnet.convolutions.1.net.1.num_batches_tracked',
> 'postnet.convolutions.2.net.0.weight',
> 'postnet.convolutions.2.net.0.bias',
> 'postnet.convolutions.2.net.1.weight',
> 'postnet.convolutions.2.net.1.bias',
> 'postnet.convolutions.2.net.1.running_mean',
> 'postnet.convolutions.2.net.1.running_var',
> 'postnet.convolutions.2.net.1.num_batches_tracked',
> 'postnet.convolutions.3.net.0.weight',
> 'postnet.convolutions.3.net.0.bias',
> 'postnet.convolutions.3.net.1.weight',
> 'postnet.convolutions.3.net.1.bias',
> 'postnet.convolutions.3.net.1.running_mean',
> 'postnet.convolutions.3.net.1.running_var',
> 'postnet.convolutions.3.net.1.num_batches_tracked',
> 'postnet.convolutions.4.net.0.weight',
> 'postnet.convolutions.4.net.0.bias',
> 'postnet.convolutions.4.net.1.weight',
> 'postnet.convolutions.4.net.1.bias',
> 'postnet.convolutions.4.net.1.running_mean',
> 'postnet.convolutions.4.net.1.running_var',
> 'postnet.convolutions.4.net.1.num_batches_tracked'. slash
> size mismatch for embedding.weight: copying a param with shape
> torch.Size( slash [129, 512 slash ]) from checkpoint, the shape in current model
> is torch.Size( slash [181, 512 slash ]).
[This is an archived TTS discussion thread from discourse.mozilla.org/t/trying-too-run-mozilla-tts-on-aws-server-problem]
Beta Was this translation helpful? Give feedback.
All reactions