Skip to content

declare-lab/LG-VQA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

LG-VQA

The repository contains the code for the paper Language Guided Visual Question Answering: Elevate Your Multimodal Language Model Using Knowledge-Enriched Prompts published at Findings of EMNLP 2023.

Usage

Download the images for the respective datasets and update the image paths in the data/*/*.json files.

The code for guidance genenration can be found in guidance. We have pre-computed the guidances and uploaded it in the data folder.

The VQA models can be trained using the train.py script. Some examples commands are shown in the run.sh file.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published