Skip to content

Conversation

@bradhilton
Copy link
Collaborator

No description provided.

…ok; adjust policy loss clipping in training function
- Cleaned up the temporal-clue-7b.ipynb notebook for better readability and structure.
- Added `max_negative_advantage_importance_sampling_weight` parameter to training configuration in `train.py`.
- Improved assertion error message formatting in `unsloth/train.py` for clarity.
@bradhilton bradhilton marked this pull request as ready for review August 5, 2025 21:07
@bradhilton bradhilton merged commit e5dbddc into main Aug 5, 2025
2 checks passed
@bradhilton bradhilton deleted the feat/prime-intellect-clipping branch August 5, 2025 21:07
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants