Skip to content

Conversation

@quic-swatia
Copy link
Contributor

  1. In case of finetuning on qaic, torch_qaic gradScaler will be used
  2. Moving back to lora_dropout = 0.05 on ML Framework team's ask.

@quic-swatia quic-swatia requested a review from vbaddi March 18, 2025 11:20
@quic-swatia quic-swatia self-assigned this Mar 18, 2025
@quic-swatia quic-swatia force-pushed the gradScaler branch 2 times, most recently from 49ed776 to ea3acc7 Compare March 18, 2025 20:07
if train_config.grad_scaler:
scaler = GradScaler()
if device.startswith("qaic"):
scaler = qaic_GradScaler()
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

same here.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Done.

… 2. Moving back to lora_dropout = 0.05 on ML Framework team's ask.

Signed-off-by: Swati Allabadi <quic-swatia@quicinc.com>
@quic-swatia quic-swatia merged commit 2f50ce3 into quic:main Mar 19, 2025
4 checks passed
ochougul pushed a commit that referenced this pull request Mar 20, 2025
1. In case of finetuning on qaic, torch_qaic gradScaler will be used
2.  Moving back to lora_dropout = 0.05 on ML Framework team's ask.

Signed-off-by: Swati Allabadi <quic-swatia@quicinc.com>
Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com>
quic-swatia pushed a commit that referenced this pull request Mar 20, 2025
1. In case of finetuning on qaic, torch_qaic gradScaler will be used
2.  Moving back to lora_dropout = 0.05 on ML Framework team's ask.

Signed-off-by: Swati Allabadi <quic-swatia@quicinc.com>
Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com>
qcdipankar pushed a commit to qcdipankar/efficient-transformers that referenced this pull request Apr 1, 2025
1. In case of finetuning on qaic, torch_qaic gradScaler will be used
2.  Moving back to lora_dropout = 0.05 on ML Framework team's ask.

Signed-off-by: Swati Allabadi <quic-swatia@quicinc.com>
Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com>
Signed-off-by: Dipankar Sarkar <quic_dipankar@quicinc.com>
quic-meetkuma pushed a commit to vbaddi/efficient-transformers that referenced this pull request Apr 21, 2025
1. In case of finetuning on qaic, torch_qaic gradScaler will be used
2.  Moving back to lora_dropout = 0.05 on ML Framework team's ask.

Signed-off-by: Swati Allabadi <quic-swatia@quicinc.com>
Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com>
eplatero97 pushed a commit to eplatero97/efficient-transformers that referenced this pull request Apr 29, 2025
1. In case of finetuning on qaic, torch_qaic gradScaler will be used
2.  Moving back to lora_dropout = 0.05 on ML Framework team's ask.

Signed-off-by: Swati Allabadi <quic-swatia@quicinc.com>
Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com>
eplatero97 pushed a commit to eplatero97/efficient-transformers that referenced this pull request Apr 29, 2025
1. In case of finetuning on qaic, torch_qaic gradScaler will be used
2.  Moving back to lora_dropout = 0.05 on ML Framework team's ask.

Signed-off-by: Swati Allabadi <quic-swatia@quicinc.com>
Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com>
eplatero97 pushed a commit to eplatero97/efficient-transformers that referenced this pull request Apr 29, 2025
1. In case of finetuning on qaic, torch_qaic gradScaler will be used
2.  Moving back to lora_dropout = 0.05 on ML Framework team's ask.

Signed-off-by: Swati Allabadi <quic-swatia@quicinc.com>
Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com>

Signed-off-by: eplatero <quic_eplatero@quicinc.com>
eplatero97 pushed a commit to eplatero97/efficient-transformers that referenced this pull request Apr 29, 2025
1. In case of finetuning on qaic, torch_qaic gradScaler will be used
2.  Moving back to lora_dropout = 0.05 on ML Framework team's ask.

Signed-off-by: Swati Allabadi <quic-swatia@quicinc.com>
Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com>

Signed-off-by: eplatero <quic_eplatero@quicinc.com>
eplatero97 pushed a commit to eplatero97/efficient-transformers that referenced this pull request Apr 29, 2025
1. In case of finetuning on qaic, torch_qaic gradScaler will be used
2.  Moving back to lora_dropout = 0.05 on ML Framework team's ask.

Signed-off-by: Swati Allabadi <quic-swatia@quicinc.com>
Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com>

Signed-off-by: eplatero <quic_eplatero@quicinc.com>
eplatero97 pushed a commit to eplatero97/efficient-transformers that referenced this pull request Apr 29, 2025
1. In case of finetuning on qaic, torch_qaic gradScaler will be used
2.  Moving back to lora_dropout = 0.05 on ML Framework team's ask.

Signed-off-by: Swati Allabadi <quic-swatia@quicinc.com>
Co-authored-by: Swati Allabadi <quic-swatia@quicinc.com>

Signed-off-by: eplatero <quic_eplatero@quicinc.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants