The idea behind AutoMix is simple:
-
Send a query to small language model (SLM), gets a noisy label on its correctness using few-shot self-verification done with the same model (SLM).
-
Use a meta-verifier to double check verifier's output, and route the query to a larger language model (LLM) if needed.
At the center of automix is the idea of context-grounded self-verification:
- However, such verification can often be noisy, so we introduce an additional layer of meta-verification using POMDPs or thresholding.
- Step1 Run inference to solve tasks - Task prompts, code to run inference from different language models.
- Step2 Self Verify - Verification prompts, code to run verification on the outputs produced in step 1.
-
Step3 Meta Verify - Run meta-verification using different AutoMix methods on outputs produced from Step 2.
-
You can run `pip install automix-llm' to use the meta-verifier system wide.
- To replicate the results in the paper, please run
python scripts paper_results.py
-
We experiment with 5 datasets: CNLI, CoQA, NarrativeQA, QASPER, and Quality.
-
Note: The dataset are sourced from scrolls. Please cite scrolls and the appropriate sources if you use these datasets. We are making them available in a sinlge jsonl file for ease of use and reproducibility. For details on how CoQa was prepared, please see Preparing COQA.
-
Inputs: All input data for the AutoMix project is provided in
automix_inputs.jsonl
. You can access and download it directly from Google Drive. -
Outputs from LLAMA2: The outputs generated using the LLAMA2 model are stored in
automix_llama2_outputs.jsonl
, available alongside the input file in the linked Google Drive.
id: A unique identifier for each question and answer pair.
pid: An additional identifier potentially mapping to specific instances or model variants.
base_ctx: The context.
question: Input question or query.
output: Ground truth.
dataset: .
llama13b_pred_ans: The answer generated by the llama13b model.
llama70b_pred_ans: The answer generated by the llama70b model.
llama13b_ver: Verification outputs of the llama13b model’s answers.
--------------------------------
| Dataset | Split | Count |
|--------------|-------|-------|
| cnli | train | 7191 |
| | val | 1037 |
| coqa | train | 3941 |
| | val | 3908 |
| narrative_qa | train | 9946 |
| | val | 5826 |
| qasper | train | 2556 |
| | val | 1715 |
| quality | train | 2515 |
| | val | 2085 |
--------------------------------
Name: split, dtype: int64
@misc{madaan2023automix,
title={AutoMix: Automatically Mixing Language Models},
author={Aman Madaan and Pranjal Aggarwal and Ankit Anand and Srividya Pranavi Potharaju and Swaroop Mishra and Pei Zhou and Aditya Gupta and Dheeraj Rajagopal and Karthik Kappaganthu and Yiming Yang and Shyam Upadhyay and Mausam and Manaal Faruqui},
year={2023},
eprint={2310.12963},
archivePrefix={arXiv},
primaryClass={cs.CL}
}