[ICLR 2026] Official repository for the paper "Diverse Text Decoding via Iterative Reweighting".
We have updated the code to support vLLM v0.11.1. This version of vLLM supports recording several key metrics during offline inference. For details, please refer to this pull request and this pull request. Accordingly, we have modified gpu_model_runner.py and modeling.py to adapt to the new v1 engine.
For the original implementation of OverRIDE using vLLM v0.8.5.post1, please switch to the vllm-0.8.5.post1 branch:
git checkout vllm-0.8.5.post1Use the following commands to create a conda environment:
conda create -n override python=3.10
conda activate override
pip install -r requirements.txtUse the following commands to download the data:
python utils/data_preprocess/humaneval.py
python utils/data_preprocess/math500.py
python utils/data_preprocess/gsm8k.py
python utils/data_preprocess/ccnews.pyThe corresponding data will be saved in the data folder.
Adjust the config file at config/default.yaml based on your device settings.
Use the following scripts to conduct corresponding experiments:
bash scripts/override.sh
bash scripts/rank.sh
bash scripts/throughput.shmain.py: Main function for running the baseline / OverRIDE evaluation.engine.py: Engine class for generating formatted responses using vLLM.dataset.py: Dataset class for loading and preprocessing data.evaluator.py: Evaluator class for evaluating PASS@K scores and saving the results.modeling.py: OverRIDE model class with reweighting heads, overriden from vLLM'sLogitsProcessor.gpu_model_runner.py: GPU model runner class that supports OverRIDE, overriden from vLLM'sGPUModelRunner.utils/: Utility functions for preprocessing data and scoring responses. These files are modified from verl.
⚠️ All modifications are based on the vLLM v1 engine. You may encounter issues when implementing on specific models (e.g., Gemma2, whoselm_headimplementation is different) or decoding methods (e.g., speculative decoding).
This project is based on vLLM and verl, licensed under Apache 2.0.
We thank the contributors of these projects for their excellent work.
