Skip to content

fix: resolve rope_theta from rope_parameters dict in HF config validation#1720

Merged
zhuzilin merged 1 commit intomainfrom
feat/glm4.6v-vl-bridge
Mar 13, 2026
Merged

fix: resolve rope_theta from rope_parameters dict in HF config validation#1720
zhuzilin merged 1 commit intomainfrom
feat/glm4.6v-vl-bridge

Conversation

@zhuzilin
Copy link
Copy Markdown
Contributor

Some models (e.g., GLM-4.6V) store rope_theta inside the rope_parameters dict rather than as a top-level config attribute. The previous code compared against config.rope_theta which returned a stale class default (10000) instead of the actual value (500000 from rope_parameters).

This caused validation failures when launching GLM-4.6V training with correct --rotary-base 500000.

…tion

Some models (e.g., GLM-4.6V) store rope_theta inside the rope_parameters
dict rather than as a top-level config attribute. The previous code compared
against config.rope_theta which returned a stale class default (10000)
instead of the actual value (500000 from rope_parameters).

This caused validation failures when launching GLM-4.6V training with
correct --rotary-base 500000.

Co-authored-by: Copilot <223556219+Copilot@users.noreply.github.com>
@zhuzilin zhuzilin merged commit ae2590a into main Mar 13, 2026
2 checks passed
@zhuzilin zhuzilin deleted the feat/glm4.6v-vl-bridge branch March 13, 2026 02:37
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant