Skip to content

Allow specifying trainer tensor parallelism in multislice RL#3067

Merged
copybara-service[bot] merged 1 commit intomainfrom
xfgu-rl-sharding
Feb 3, 2026
Merged

Allow specifying trainer tensor parallelism in multislice RL#3067
copybara-service[bot] merged 1 commit intomainfrom
xfgu-rl-sharding

Conversation

@xuefgu
Copy link
Copy Markdown
Collaborator

@xuefgu xuefgu commented Feb 2, 2026

Description

Allow specifying trainer tensor parallelism in multislice RL

FIXES: b/480979614

Tests

Manually tested.

Checklist

Before submitting this PR, please make sure (put X in square brackets):

  • I have performed a self-review of my code. For an optional AI review, add the gemini-review label.
  • I have necessary comments in my code, particularly in hard-to-understand areas.
  • I have run end-to-end tests tests and provided workload links above if applicable.
  • I have made or will make corresponding changes to the doc if needed, including adding new documentation pages to the relevant Table of Contents (toctree directive) as explained in our documentation.

@xuefgu xuefgu force-pushed the xfgu-rl-sharding branch 3 times, most recently from 00678c6 to 744d788 Compare February 2, 2026 22:46
@codecov
Copy link
Copy Markdown

codecov Bot commented Feb 3, 2026

Codecov Report

❌ Patch coverage is 0% with 9 lines in your changes missing coverage. Please review.

Files with missing lines Patch % Lines
src/MaxText/rl/train_rl.py 0.00% 9 Missing ⚠️

📢 Thoughts on this report? Let us know!

Comment thread src/MaxText/rl/train_rl.py
@copybara-service copybara-service Bot merged commit 1a44692 into main Feb 3, 2026
28 of 30 checks passed
@copybara-service copybara-service Bot deleted the xfgu-rl-sharding branch February 3, 2026 20:13
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants