Skip to content
/ VAWI Public
forked from pyogher/VAWI

The implementation code of the ACL 2023 paper "Visually-augmented Pretrained Language Models for NLP Tasks without Images"

License

Notifications You must be signed in to change notification settings

RUCAIBox/VAWI

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Visually-augmented Pretrained Language Models for NLP Tasks without Images

This repository contains the code for our paper Visually-augmented Pretrained Language Models for NLP Tasks without Images.

Overview

We propose VAWI (paper), a visually-augmented fine-tuning approach applicable to various pre-trained language models (PLMs) and NLP tasks, without the need for retrieved or generated images. To mitigate the lack of visual semantics and commonsense in PLMs, we identify and extract visually-hungry words (VH-words) in the input text using three strategies. Then, we adopt a fixed CLIP text encoder to generate the visually-aligned representations of these VH-words. Finally, The visually-aligned features were transformed into visually-augmented representations through a reformulation layer based on VH-words, which are then inserted into PLMs to enrich the visual semantics of word representations.

model_figure

News

Train VAWI

Citation

Please cite our paper if you use VAWI in your work:

@article{guo2022visually,
  title={Visually-augmented pretrained language models for NLP tasks without images},
  author={Guo, Hangyu and Zhou, Kun and Zhao, Wayne Xin and Zhang, Qinyu and Wen, Ji-Rong},
  journal={arXiv preprint arXiv:2212.07937},
  year={2022}
}

About

The implementation code of the ACL 2023 paper "Visually-augmented Pretrained Language Models for NLP Tasks without Images"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published