Skip to content

ReML-AI/LaCoMSA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

LaCoMSA: Language-Consistency Multilingual Self-Alignment with Latent Representation Rewarding

Khanh-Tung Tran, Barry O'Sullivan, Hoang D. Nguyen

Accepted to EACL 2026 (Main Conference)

Overview

Large Language Models (LLMs) have achieved impressive performance yet remain inconsistent across languages, often defaulting to high-resource outputs such as English. Existing multilingual alignment methods mitigate these issues through preference optimization but rely on external supervision, such as translation systems or English-biased signal. We propose Multilingual Self-Alignment (MSA), a preference optimization framework that leverages an LLM’s own latent representations as intrinsic supervision signals, rewarding lower-resource language outputs based on their alignment with high-resource (English) counterparts in the ``semantic hub''. We further introduce Language-Consistency MSA (LaCoMSA), which augments MSA with a final-layer language-consistency factor to prevent off-target generation. Integrated with Direct Preference Optimization, LaCoMSA improves a Llama 3 8B-based model multilingual win rates by up to 6.8% absolute (55.0% relatively) on X-AlpacaEval and achieves consistent gains across benchmarks and models. Our findings demonstrate that LaCoMSA can serve as an effective and scalable mechanism, opening a new venue toward multilingual self-alignment.

This repository contains the cleaned implementation for the LaCoMSA paper.

Repository Structure

Alignment/        # DPO training scripts
Preprocess/       # Data & reward generation scripts
Data/
requirement.txt

Quick Start

  1. Environment:
python -m venv .venv
source .venv/bin/activate
pip install --upgrade pip
pip install -r requirements.txt
  1. Prepare preference data:
cd Preprocess
bash preprocess.sh # more details available in the bash script
  1. Train with. DPO:
cd Alignment
bash dpo.sh example.json

Citation

TBU

About

[EACL 2026] LaCoMSA: Language-Consistency Multilingual Self-Alignment with Latent Representation Rewarding

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors