-
Notifications
You must be signed in to change notification settings - Fork 60
[QEff. Finetune] Adding the support to resume the fine tuning using pre computed #233
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
… prev run whoch would have stopped in between. There's no necessity to pass tokenizer_name if a model_name is passed. It will take the same name as model_name by default. If a different tokenizer_name is required than the model_name, then it can be passed separately as an argument. Signed-off-by: Swati Allabadi <quic_sallabad@quicinc.com>
…ers into finetune Signed-off-by: Swati Allabadi <quic_sallabad@quicinc.com>
0dfaa2f to
da18e86
Compare
Signed-off-by: Swati Allabadi <quic_sallabad@quicinc.com>
0d4affa to
6e729fe
Compare
…ts and check for loss convergence. Signed-off-by: Swati Allabadi <quic_sallabad@quicinc.com>
|
If we don't change output_dir then after resuming FT, will new tensorboard data be appended to previous tensorboard log files? |
|
Irrespective of the value of the output_dir, the tensorboard files get saved inside directory named runs. For each fine tuning job, a new directory is created inside runs. So, if we run the following command : "tensorboard --logdir runs --bind_all", tensorboard data from both the jobs will show up together in a single plot. |
…re computed (#233) 1) Adding the support to resume the fine tuning using checkpoints from a prev run which would have stopped in between. 2) Checkpoints, both intermediate and for complete epoch, will get saved for each epoch through these changes. 3) There's no necessity to pass tokenizer_name if a model_name is passed. It will take the same name as model_name by default. If a different tokenizer_name is required than the model_name, then it can be passed separately as an argument in the command. --------- Signed-off-by: Swati Allabadi <quic_swatia@quicinc.com> Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com>
…re computed (quic#233) 1) Adding the support to resume the fine tuning using checkpoints from a prev run which would have stopped in between. 2) Checkpoints, both intermediate and for complete epoch, will get saved for each epoch through these changes. 3) There's no necessity to pass tokenizer_name if a model_name is passed. It will take the same name as model_name by default. If a different tokenizer_name is required than the model_name, then it can be passed separately as an argument in the command. --------- Signed-off-by: Swati Allabadi <quic_sallabad@quicinc.com> Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com> Signed-off-by: Rishin Raj <quic_rishinr@quicinc.com>
…re computed (quic#233) 1) Adding the support to resume the fine tuning using checkpoints from a prev run which would have stopped in between. 2) Checkpoints, both intermediate and for complete epoch, will get saved for each epoch through these changes. 3) There's no necessity to pass tokenizer_name if a model_name is passed. It will take the same name as model_name by default. If a different tokenizer_name is required than the model_name, then it can be passed separately as an argument in the command. --------- Signed-off-by: Swati Allabadi <quic_sallabad@quicinc.com> Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com> Signed-off-by: Dipankar Sarkar <quic_dipankar@quicinc.com>
…re computed (quic#233) 1) Adding the support to resume the fine tuning using checkpoints from a prev run which would have stopped in between. 2) Checkpoints, both intermediate and for complete epoch, will get saved for each epoch through these changes. 3) There's no necessity to pass tokenizer_name if a model_name is passed. It will take the same name as model_name by default. If a different tokenizer_name is required than the model_name, then it can be passed separately as an argument in the command. --------- Signed-off-by: Swati Allabadi <quic_swatia@quicinc.com> Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com>
…re computed (quic#233) 1) Adding the support to resume the fine tuning using checkpoints from a prev run which would have stopped in between. 2) Checkpoints, both intermediate and for complete epoch, will get saved for each epoch through these changes. 3) There's no necessity to pass tokenizer_name if a model_name is passed. It will take the same name as model_name by default. If a different tokenizer_name is required than the model_name, then it can be passed separately as an argument in the command. --------- Signed-off-by: Swati Allabadi <quic_swatia@quicinc.com> Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com>
…re computed (quic#233) 1) Adding the support to resume the fine tuning using checkpoints from a prev run which would have stopped in between. 2) Checkpoints, both intermediate and for complete epoch, will get saved for each epoch through these changes. 3) There's no necessity to pass tokenizer_name if a model_name is passed. It will take the same name as model_name by default. If a different tokenizer_name is required than the model_name, then it can be passed separately as an argument in the command. --------- Signed-off-by: Swati Allabadi <quic_swatia@quicinc.com> Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com> Signed-off-by: eplatero <quic_eplatero@quicinc.com>
…re computed (quic#233) 1) Adding the support to resume the fine tuning using checkpoints from a prev run which would have stopped in between. 2) Checkpoints, both intermediate and for complete epoch, will get saved for each epoch through these changes. 3) There's no necessity to pass tokenizer_name if a model_name is passed. It will take the same name as model_name by default. If a different tokenizer_name is required than the model_name, then it can be passed separately as an argument in the command. --------- Signed-off-by: Swati Allabadi <quic_swatia@quicinc.com> Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com> Signed-off-by: eplatero <quic_eplatero@quicinc.com>
…re computed (quic#233) 1) Adding the support to resume the fine tuning using checkpoints from a prev run which would have stopped in between. 2) Checkpoints, both intermediate and for complete epoch, will get saved for each epoch through these changes. 3) There's no necessity to pass tokenizer_name if a model_name is passed. It will take the same name as model_name by default. If a different tokenizer_name is required than the model_name, then it can be passed separately as an argument in the command. --------- Signed-off-by: Swati Allabadi <quic_swatia@quicinc.com> Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com> Signed-off-by: eplatero <quic_eplatero@quicinc.com>
…re computed (quic#233) 1) Adding the support to resume the fine tuning using checkpoints from a prev run which would have stopped in between. 2) Checkpoints, both intermediate and for complete epoch, will get saved for each epoch through these changes. 3) There's no necessity to pass tokenizer_name if a model_name is passed. It will take the same name as model_name by default. If a different tokenizer_name is required than the model_name, then it can be passed separately as an argument in the command. --------- Signed-off-by: Swati Allabadi <quic_swatia@quicinc.com> Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com> Signed-off-by: eplatero <quic_eplatero@quicinc.com>
If a different tokenizer_name is required than the model_name, then it can be passed separately as an argument in the command.