Add mm data example#3
Merged
Merged
Conversation
Contributor
There was a problem hiding this comment.
Pull request overview
This PR updates the Kimi-K2.5 SGLang training examples to use an OpenAI-style conversations column (including multimodal and tool-call examples), and aligns the provided example configs and sample dataset accordingly.
Changes:
- Documented the expected
conversationsJSONL dataset format (including multimodal + tool-call examples) in the 2-node and 3-node Kimi-K2.5 READMEs. - Updated Kimi-K2.5 SGLang example configs to use
prompt_key: conversations(instead ofmessages). - Added new
kimi_*example rows toexamples/data/sample_conversations.jsonlcovering text, images, multi-image, multi-turn, and tool calls.
Reviewed changes
Copilot reviewed 5 out of 5 changed files in this pull request and generated 4 comments.
Show a summary per file
| File | Description |
|---|---|
| examples/kimi-k25-3node-h100/README.md | Adds dataset format guidance and a multimodal OpenAI-style example for the 3-node setup. |
| examples/kimi-k25-2node-h200/README.md | Adds dataset format guidance and a multimodal OpenAI-style example for the 2-node setup. |
| examples/data/sample_conversations.jsonl | Appends new kimi_* sample conversations including multimodal and tool-call scenarios. |
| configs/sglang_kimi_k25_3node.yaml | Switches dataset prompt_key to conversations to match the documented dataset schema. |
| configs/sglang_kimi_k25_2node.yaml | Switches dataset prompt_key to conversations to match the documented dataset schema. |
💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.
cicirori
pushed a commit
to cicirori/TorchSpec
that referenced
this pull request
Feb 26, 2026
* refactor mooncake * fix * fix
zhubohao911
pushed a commit
to zhubohao911/TorchSpec
that referenced
this pull request
Mar 23, 2026
* add mm data * split back
9 tasks
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
This pull request updates the dataset configuration and documentation to support OpenAI-style conversations with multimodal and tool-call capabilities. It also adds new sample data covering a range of scenarios, and clarifies the expected dataset format for Kimi-K2.5 training setups.
Dataset format and documentation updates:
examples/kimi-k25-2node-h200/README.mdandexamples/kimi-k25-3node-h100/README.md, including guidance on using<think>...</think>blocks, image content, and tool calls. [1] [2]conversationsfield, and explained how multimodal content is handled for SGLang engine training. [1] [2]Configuration changes for dataset compatibility:
prompt_keyfrommessagestoconversationsin bothconfigs/sglang_kimi_k25_2node.yamlandconfigs/sglang_kimi_k25_3node.yamlto match the new dataset format. [1] [2]Sample data additions:
examples/data/sample_conversations.jsonlcovering text-only, multimodal (single/multi-image), system messages, multi-turn, and tool calls, providing comprehensive examples for model training.