The repository of the paper ART: Attention Replacement Technique to Improve Factuality in LLMs. Detailed source code and prompt templates are available here.
Demonstration for Attention Replacement Technique.
modify_attention.py corresponds to the operations that calculate
modeling_llama_head_modified.py, modeling_mistral_head_modified.py, and modeling_qwen_head_modified.py are all based on the minor modification on the HuggingFace source code modeling_llama, modeling_mistral, and modeling_qwen.
Prompt templates for Multi-Choice Question (MCQ) task. (TruthfulQA)
You are a problem solver. There is a question with related options provided. You need to think step-by-step to infer the correct choice
Question: {question}
Options:
{options}
Your response should end with 'The answer is: (X)' with X being your chosen option.
Prompt templates for MCQ with context for reference. (LogiQA)
You are a problem solver. There is a question with related context and options provided. You need to think step-by-step to infer the correct choice.
Context:
{context}
Question: {question}
Options:
{options}
Your response should end with 'The answer is: (X)' with X being your chosen option.
Prompt templates for Math Word Problem. (GSM8K)
You are a problem solver. There is a math-word problem you need to solve.
This is the question for you: {question}
Your response should end with 'The answer is: X' with X being the numerical result. Think step-by-step to infer the correct answer.
