Skip to content

blazerye/AROMA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

AROMA Logo

🧬 AROMA: Augmented Reasoning Over a Multimodal Architecture for Virtual Cell Genetic Perturbation Modeling

📃 Paper • 🤗 Model • 🗂️ Datasets

📌 Contents

🌐 Overview

AROMA is a novel multimodal architecture for virtual cell modeling that integrates textual evidence, graph topology, and protein sequences to predict the effects of genetic perturbations.

Overview

The overall AROMA pipeline is illustrated in the figure above and is divided into three stages:

  • Data stage. AROMA constructs two complementary knowledge graphs and a large-scale virtual cell reasoning dataset for evidence grounding.

  • Modeling stage. AROMA adopts a retrieval-augmented strategy to incorporate query-relevant information, thereby providing explicit evidence cues for prediction. In addition, it jointly leverages topological representations learned from graph neural networks (GNN) and protein sequence representations encoded by ESM-2, and applies a cross-attention module to explicitly model perturbation-target gene dependencies across modalities.

  • Training stage. AROMA first performs multimodal supervised fine-tuning (SFT), and is then further optimized with Group Relative Policy Optimization (GRPO) reinforcement learning to enhance predictive performance while generating biologically meaningful explanations.

🚀 Quick Start

Environment Setup

Install the necessary dependencies:

cd AROMA
# Install requirements
pip install -r requirements.txt

Knowledge Graph Preprocessing

AROMA utilizes Graph Neural Networks to extract topological features from biological knowledge graphs. This process is divided into two parts: Gene-KG and Path-KG.

🧬 Gene Knowledge Graph

cd gnn
# Pretraining
python gnn_pretrain_gene_graph.py
# Inference
python gnn_inference_gene_graph.py

🗺 Pathway Knowledge Graph

cd gnn
# Pretraining
python gnn_pretrain_pathway_graph.py
# Inference (Extracting embeddings)
python gnn_inference_pathway_graph.py

Prepare the Base Model

Please first download the original Qwen3-8B model from the official Hugging Face repository:
https://huggingface.co/Qwen/Qwen3-8B Then, run the following commands:

cd aroma
python modify_tokenizer.py

AROMA Model Training

The training process follows a two-stage optimization strategy.

Stage 1: Multimodal Supervised Fine-Tuning

cd aroma
CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 \
deepspeed --num_gpus 8 train_sft.py

Stage 2: GRPO Reinforcement Learning

cd aroma
CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 \
accelerate launch --multi_gpu --num_processes 8 train_grpo.py

Inference & Evaluation

cd aroma
python inference.py

🤗 Model Weights

The pretrained AROMA model checkpoints have been released on Hugging Face at blazerye/AROMA. The weights can be directly used for inference or fine-tuning on downstream virtual-cell perturbation modeling tasks.

🗂️ Datasets

  • Gene-KG and Path-KG. We release the complete versions of Gene-KG and Path-KG in this repository to support running our codebase and facilitate downstream research.
  • PerturbReason. For the PerturbReason dataset, only a 1,000-sample subset is publicly released in this repository for demonstration and verification purposes. The full PerturbReason dataset is available at blazerye/PerturbReason.

📌 Citation

If you find AROMA useful for your research and applications, please cite using this BibTeX:

@inproceedings{wang2026aroma,
    title="{AROMA}: Augmented Reasoning Over a Multimodal Architecture for Virtual Cell Genetic Perturbation Modeling",
    author="Wang, Zhenyu and Ye, Geyan and Liu, Wei and Ng, Man Tat Alexander",
    booktitle="Findings of the Association for Computational Linguistics: ACL 2026",
    year="2026",
    publisher="Association for Computational Linguistics"
}

About

[ACL 2026] AROMA: Augmented Reasoning Over a Multimodal Architecture for Virtual Cell Genetic Perturbation Modeling

Resources

Stars

Watchers

Forks

Languages