Skip to content

jiangycTarheel-zz/compositional-auxseq

main
Switch branches/tags

Name already in use

A tag already exists with the provided branch name. Many Git commands accept both tag and branch names, so creating this branch may cause unexpected behavior. Are you sure you want to create this branch?
Code

Latest commit

 

Git stats

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time
 
 
 
 
 
 
 
 
 
 
 
 

compositional-auxseq

This repo contains the source code of the models described in the following paper

  • "Inducing Transformer's Compositional Generalization Ability via Auxiliary Sequence Prediction Tasks" in Proceedings of EMNLP, 2021. (paper).

The basic code structure was adapted from the HuggingFace Transformers.

0. Preparation

0.1 Dependencies

  • PyTorch 1.4.0/1.6.0/1.8.0
  • Pytorch Lightning 0.7.6. Support for more recent versions of lightning coming up soon.

0.2 Data

  • Download the original SCAN data
  • Download the SCAN MCD splits
  • Organize the data into data/scan and make sure it follows such a structure:
------ data
--------- scan
------------ tasks_test_mcd1.txt
------------ tasks_train_mcd1.txt
------------ tasks_val_mcd1.txt

2. Training

  • Train the model on the SCAN MCD1 splits by running:
./train_scan_scripts/train_auxseq_mcd1.sh
  • By defaults, the top-5 best model checkpoints will be saved in out/scan/auxseq-00.

3. Evaluation

  • Set the EVAL_EPOCH parameter in the eval_scan_scripts/eval_auxseq_mcd1.sh.
  • Evaluate the model on the SCAN MCD1 splits by running:
./eval_scan_scripts/eval_auxseq_mcd1.sh

Citation

@inproceedings{jiang-bansal-2021-enriching,
    title = "Inducing Transformer's Compositional Generalization Ability via Auxiliary Sequence Prediction Tasks",
    author = "Jiang, Yichen and Bansal, Mohit",
    booktitle = "Proceedings of the EMNLP 2021",
    year = "2021",
}

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published