Skip to content

Conversation

@xiaoyu-work
Copy link
Collaborator

Describe your changes

We will only focus on cases when the onnx model is exported from torch dynamo. If the optimize is True (which is default) for torch dynamo export, the lora weights will be folded. So:

  1. we check where the initializer comes from for each initializer.
  2. If it is from a LORA weight, rename it back to lora weight, and save it.
  3. DORA weights might be folded to GEMM, we decompose this node to MatMul so it can be quantized by Olive quantization pass.

Checklist before requesting a review

  • Add unit tests for this change.
  • Make sure all tests can pass.
  • Update documents if necessary.
  • Lint and apply fixes to your code by running lintrunner -a
  • Is this a user-facing change? If yes, give a description of this change to be included in the release notes.

(Optional) Issue link

@xiaoyu-work
Copy link
Collaborator Author

waiting for onnxscript new version release then retrigger the pipeline.

@xiaoyu-work
Copy link
Collaborator Author

/azp run

@azure-pipelines
Copy link

Azure Pipelines successfully started running 1 pipeline(s).

@xiaoyu-work xiaoyu-work marked this pull request as ready for review October 8, 2025 23:51
@xiaoyu-work xiaoyu-work merged commit 7699cae into main Oct 28, 2025
11 checks passed
@xiaoyu-work xiaoyu-work deleted the xiaoyu/ext branch October 28, 2025 01:15
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants