Skip to content

Domain specific self-supervised fine-tuning by learning mask specific losses

Notifications You must be signed in to change notification settings

mykelismyname/MSLM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

MSLM

Domain Sensitive Fine-tuning:

Improving Pre-trained Language Model Sensitivity via Mask Specific losses: A case study on Biomedical NER

Requirements

  • Python 3.8+
  • transformers 4.31.0
  • torch 2.0.1

Data

BLURB benchmark dataset

Data Preparation

python utils.py \
    [path to data] \
    [storage or destination directory]

Alternatively inherit pre-processed BLURB datasets such as,

Masking

Our proposed Joint ELM-BLM masking approach

PMI masking

Construct a vocabularly from a dataset using the masking approach 

 ./run_pmi.sh

Fine-tuning

Specify the paths to the data and set the masking budgets for both the Base level masking BLM and the Entity level masking ELM

./run_train.sh [DATASET]

Citation

@article{abaho2024improving,
  title={Improving Pre-trained Language Model Sensitivity via Mask Specific losses: A case study on Biomedical NER},
  author={Abaho, Micheal and Bollegala, Danushka and Leeming, Gary and Joyce, Dan and Buchan, Iain E},
  journal={arXiv preprint arXiv:2403.18025},
  year={2024}
}

About

Domain specific self-supervised fine-tuning by learning mask specific losses

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published