Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Falcon Loss Not Decreasing During Training #125

Closed
LamOne1 opened this issue Jun 8, 2023 · 3 comments
Closed

Falcon Loss Not Decreasing During Training #125

LamOne1 opened this issue Jun 8, 2023 · 3 comments
Labels
question Further information is requested

Comments

@LamOne1
Copy link

LamOne1 commented Jun 8, 2023

I'm using pretrain code with falcon 7B. I've noticed that the loss didn't change for 400 iterations.

iter 1: loss 11.0666, time: 13381.00ms, speed: 306 toks/s/device
....
iter 400: loss 11.0666, time: 19090.34ms, speed: 214 toks/s/device
@carmocca
Copy link
Contributor

Can you share exactly what script you ran?

@LamOne1
Copy link
Author

LamOne1 commented Jun 25, 2023

I was using redpajama pretrain script as it is.
However I prepared my data with PackedDataset type np.uint16 instead of np.int32, could be the problem?

@carmocca carmocca added the question Further information is requested label Aug 14, 2023
@carmocca
Copy link
Contributor

I don't think so. But you might need to tweak hyperparameters. This is the dark art of machine learning 😉

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants