You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello,
Thank you for your contributions. I tried to run the UniTS_pretrain_x128.sh script, but after a while, the outputs appeared to be nan, and the corresponding loss value also changed to nan. But reducing the d_model to 64 there is no problem. What is the reason for this?
The text was updated successfully, but these errors were encountered:
That happens sometimes because the co-training on cross-domain datasets is not always stable, and it happens not only for time series but also on foundation models on other fields. We rerun the experiments when we find nan. You can also adjust the learning rate and use a smaller grad-clip value.
Hello,
Thank you for your contributions. I tried to run the UniTS_pretrain_x128.sh script, but after a while, the outputs appeared to be nan, and the corresponding loss value also changed to nan. But reducing the d_model to 64 there is no problem. What is the reason for this?
The text was updated successfully, but these errors were encountered: