Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

A question about training weights of embedding #34

Closed
AriKing11 opened this issue Jan 4, 2024 · 2 comments
Closed

A question about training weights of embedding #34

AriKing11 opened this issue Jan 4, 2024 · 2 comments

Comments

@AriKing11
Copy link

AriKing11 commented Jan 4, 2024

I have a question about the training weights of embedding. I used my own datasets to process stage 1 (which includes tuning the embedding weights of new graph tokens, e.g. DEFAULT_GRAPH_TOKEN = ""), but the weights became Nan instantly, I don't know why. Thanks for your patience.

@tjb-tech
Copy link
Collaborator

I have a question about the training weights of embedding. I used my own datasets to process stage 1 (which includes tuning the embedding weights of new graph tokens, e.g. DEFAULT_GRAPH_TOKEN = ""), but the weights became Nan instantly, I don't know why. Thanks for your patience.

Thanks for you interests! May I ask details of your error? Do the weights become Nan or loss become Nan?

@HKUDS HKUDS closed this as completed Mar 3, 2024
@msy0513
Copy link

msy0513 commented Aug 22, 2024

I have a question about the training weights of embedding. I used my own datasets to process stage 1 (which includes tuning the embedding weights of new graph tokens, e.g. DEFAULT_GRAPH_TOKEN = ""), but the weights became Nan instantly, I don't know why. Thanks for your patience.

Thanks for you interests! May I ask details of your error? Do the weights become Nan or loss become Nan?

我更换数据后再stage1和stage2都得到了train_loss=nan,这是正常情况么?该怎么解决这个问题呢?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants