Skip to content

asahi-research/C2

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

C2: Rubric-Augmented Reward Modeling from Binary Preferences

This repository contains the official implementation of the paper "C2: Rubric-Augmented Reward Modeling from Binary Preferences". C2 enables rubric-augmented reward modeling from binary preference data alone. At test time, a rubric generator trained on synthesized rubrics collaborates with a reward model to produce nuanced, fine-grained LLM-as-a-judge evaluations.

C2 overview

Abstract

Rubric-augmented verification guides reward models with explicit evaluation criteria, yielding more reliable judgments than single-model verification. However, most existing methods require costly rubric annotations, limiting scalability. Moreover, we find that rubric generation is vulnerable to a failure of cooperation; low-quality rubrics actively mislead reward models rather than help. Inspired by the principle of cooperative communication, we propose Cooperative yet Critical reward modeling (C2), a framework that significantly improves reward model judgments by having them collaborate with a rubric generator trained solely from binary preferences. In C2, we synthesize helpful and misleading rubric pairs by measuring how each rubric shifts the reward model toward or away from the correct preference. Using these contrastive pairs, we train a cooperative rubric generator to propose helpful rubrics, and a critical verifier to assess rubric validity before making its judgment, following only rubrics it deems helpful at inference time. C2 outperforms reasoning reward models trained on the same binary preferences, with gains of up to 6.5 points on RM-Bench and 6.0 points length-controlled win rate on AlpacaEval 2.0. Without external rubric annotations, C2 enables an 8B reward model to match performance achieved with rubrics from a 4× larger model. Overall, our work demonstrates that eliciting deliberate cooperation in rubric-augmented verification makes reward models more trustworthy in a scalable way.

Installation

Install the project with uv:

uv sync

This installs the package and exposes the following CLI entry points:

  • c2-synthesize-rubrics
  • c2-train-generator-dpo
  • c2-train-verifier-grpo
  • c2-infer

Getting Started

Input data

The loader accepts pairwise preference format:

prompt + response_a + response_b + label

{
  "id": "ex_0002",
  "prompt": "Explain why the sky is blue.",
  "response_a": "...",
  "response_b": "...",
  "label": "A"
}

Run the full pipeline

The simplest entry point is the shell launcher:

bash scripts/run_c2.sh \
  --model-name allenai/Llama-3.1-Tulu-3-8B-SFT \
  --dataset-path /path/to/pairwise_train.jsonl \
  --output-root /path/to/c2_run

This command produces:

  • data/contrastive_rubric_pairs.jsonl
  • data/generator_contrastive_pairs.jsonl
  • data/rubric_augmented_examples.jsonl
  • generator/
  • verifier/

Run each stage manually

1. Synthesize contrastive rubric pairs

uv run c2-synthesize-rubrics \
  --dataset-path /path/to/pairwise_train.jsonl \
  --model-name allenai/Llama-3.1-Tulu-3-8B-SFT \
  --output-path /path/to/contrastive_rubric_pairs.jsonl \
  --generator-contrastive-pairs-path /path/to/generator_contrastive_pairs.jsonl \
  --rubric-augmented-examples-path /path/to/rubric_augmented_examples.jsonl

2. Train the rubric generator

uv run c2-train-generator-dpo \
  --dataset-path /path/to/generator_contrastive_pairs.jsonl \
  --model-name allenai/Llama-3.1-Tulu-3-8B-SFT \
  --output-dir /path/to/generator

The generator is trained with DPO on (context, helpful rubric, misleading rubric) triples.

3. Train the verifier

uv run c2-train-verifier-grpo \
  --original-dataset-path /path/to/pairwise_train.jsonl \
  --rubric-augmented-examples-path /path/to/rubric_augmented_examples.jsonl \
  --model-name allenai/Llama-3.1-Tulu-3-8B-SFT \
  --output-dir /path/to/verifier

The verifier is trained on a mixture of rubric-free and rubric-augmented tasks.

4. Run selective inference

uv run c2-infer \
  --dataset-path /path/to/pairwise_eval.jsonl \
  --generator-model /path/to/generator \
  --verifier-model /path/to/verifier \
  --output-path /path/to/predictions.jsonl

At inference time, the generator proposes a rubric, the verifier decides whether that rubric is helpful or misleading, and the system falls back to rubric-free verification when the rubric should not be trusted.

Citation

TBA.

About

The implementation for the paper "C2: Rubric-Augmented Reward Modeling from Binary Preferences"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors