Skip to content

Conversation

quic-amitraj
Copy link
Contributor

Addition of MIN_MASKED_ATTN_VALUE to release branch

qcdipankar and others added 16 commits June 22, 2025 22:33
Announcement update for Granite Vision in 1.20.0 release

---------

Signed-off-by: Dipankar Sarkar <dipankar@qti.qualcomm.com>
Addressed The issue faced during mllama dual qpc compilation
Added coditions for mllama in modeling_auto

---------

Signed-off-by: Dipankar Sarkar <dipankar@qti.qualcomm.com>
This fixed the issue with higher BS compilation for SwiftKV models

```
Compiler command: ['/opt/qti-aic/exec/qaic-exec', '-aic-hw', '-aic-hw-version=2.0', '-m=/prj/qct/aisyssol_scratch/users/shagsood/quic_shagun/LlamaSwiftKVForCausalLM-a5879ebc0e59ab40/LlamaSwiftKVForCausal
LM.onnx', '-compile-only', '-retained-state', '-convert-to-fp16', '-aic-num-cores=16', '-network-specialization-config=/prj/qct/aisyssol_scratch/users/shagsood/quic_shagun/LlamaSwiftKVForCausalLM-a5879eb
c0e59ab40/qpc-60f86f912a187346/specializations.json', '-custom-IO-list-file=/prj/qct/aisyssol_scratch/users/shagsood/quic_shagun/LlamaSwiftKVForCausalLM-a5879ebc0e59ab40/qpc-60f86f912a187346/custom_io.ya
ml', '-mdp-load-partition-config=/prj/qct/aisyssol_scratch/users/shagsood/quic_shagun/LlamaSwiftKVForCausalLM-a5879ebc0e59ab40/qpc-60f86f912a187346/mdp_ts_4.json', '-aic-binary-dir=/prj/qct/aisyssol_scra
tch/users/shagsood/quic_shagun/LlamaSwiftKVForCausalLM-a5879ebc0e59ab40/qpc-60f86f912a187346/qpc']                                                                                                         
Compiler exitcode: 1                                                                                                                                                                                       
Compiler stderr:                                                                                                                                                                                           
QAIC_ERROR:                                                                                                                                                                                                
Error message:  [Operator-'/model/layers.16/self_attn/Reshape'] : Reshape: input shape (4, 4, 4096) and output shape (4, 1, 32, 128) have different number of elements (in 65536 vs. out 16384)            
Unable to AddNodesToGraphFromModel 
```

Tested with BS4. Able to compile now

Signed-off-by: quic-shagun <quic_shagsood@quicinc.com>
CI enablement and other minor fixes for Gemma3

---------

Signed-off-by: Ann Kuruvilla <quic_akuruvil@quicinc.com>
Added fix for spdtransform due to change in hash

---------

Signed-off-by: Dipankar Sarkar <dipankar@qti.qualcomm.com>
- Enabled CI tests for Finetuning. 
- Updated Jenkins file to install torch_qaic as it is required during FT
tests.
- Added finetune as a new pytest flag and updated other existing tests
not to trigger for this flag.

---------

Signed-off-by: meetkuma <meetkuma@qti.qualcomm.com>
Co-authored-by: Meet Patel <meetkuma@qti.qualcomm.com>
CI enablement and other minor fixes for Gemma3

---------

---------

Signed-off-by: Ann Kuruvilla <quic_akuruvil@quicinc.com>
Signed-off-by: Dipankar Sarkar <dipankar@qti.qualcomm.com>
Co-authored-by: Dipankar Sarkar <dipankar@qti.qualcomm.com>
…boarded features in docs (#423)

This PR is created for updating the readme and docs for adding the
latest features added in this release.

---------

Signed-off-by: Abukhoyer Shaik <abukhoye@qti.qualcomm.com>
…next file. (#475)

Signed-off-by: Dhiraj Kumar Sah <dhirajku@qti.qualcomm.com>
Signed-off-by: Rishin Raj <rishinr@qti.qualcomm.com>
Padding the dataset with dummy samples (they won't contribute in
total_loss) to make the #samples a multiple of degree of ddp*batch_size)
in case of
1) Fine tuning through DDP 
2) train_batch_size > 1 or val_batch_size > 0

---------

Signed-off-by: Swati Allabadi <sallabad@qti.qualcomm.com>
Co-authored-by: Swati Allabadi <sallabad@qti.qualcomm.com>
Co-authored-by: Mamta Singh <168400541+quic-mamta@users.noreply.github.com>
Generating data format config file fails for encoder onnx graph without
past key or past value.
Fixed a coding bug in the function.

---------

Signed-off-by: Shubham Agrawal <shubhagr@qti.qualcomm.com>
Changed Total (E2E) inference time from decode/sec to sec.

Signed-off-by: Asmita Goswami <quic_asmigosw@quicinc.com>
…ng optimizer step only. (#477)

Disabling gradient is necessary when using gradient_accumulation_step >
1 with ddp enabled.
Currently, we are syncing gradient at every loss.backward() call, which
is called at all steps. When using gradient accumulation, the weight
update during opt.step() step. Only during that step, the gradients
across each devices should sync with each other.

with model.no_sync() --> context manager solves this issue.

Here, we are not using it but instead setting
ddp_model.require_backward_grad_sync to True or False depending on which
step we are.

---------

Signed-off-by: Meet Patel <meetkuma@qti.qualcomm.com>
Signed-off-by: meetkuma <meetkuma@qti.qualcomm.com>
…371)

1. Implement logger for finetuning
2. enable dumping logs by given flag

---------

Signed-off-by: Mamta Singh <mamtsing@qti.qualcomm.com>
Co-authored-by: Mamta Singh <mamtsing@qti.qualcomm.com>
@quic-amitraj quic-amitraj force-pushed the Model_fix branch 7 times, most recently from b50eef2 to 7c365c4 Compare July 10, 2025 06:43
quic-akuruvil and others added 6 commits July 10, 2025 06:44
CI enablement and other minor fixes for Gemma3

---------

---------

Signed-off-by: Ann Kuruvilla <quic_akuruvil@quicinc.com>
Signed-off-by: Dipankar Sarkar <dipankar@qti.qualcomm.com>
Co-authored-by: Dipankar Sarkar <dipankar@qti.qualcomm.com>
Signed-off-by: Amit Raj <amitraj@qti.qualcomm.com>
Reverts #484

Signed-off-by: Amit Raj <amitraj@qti.qualcomm.com>
CI enablement and other minor fixes for Gemma3

---------

---------

Signed-off-by: Ann Kuruvilla <quic_akuruvil@quicinc.com>
Signed-off-by: Dipankar Sarkar <dipankar@qti.qualcomm.com>
Co-authored-by: Dipankar Sarkar <dipankar@qti.qualcomm.com>
Signed-off-by: Amit Raj <amitraj@qti.qualcomm.com>
Reverts #484

Signed-off-by: Amit Raj <amitraj@qti.qualcomm.com>
Signed-off-by: Amit Raj <amitraj@qti.qualcomm.com>
Signed-off-by: Amit Raj <amitraj@qti.qualcomm.com>
@quic-amitraj
Copy link
Contributor Author

Created new #513 so closing.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.