Skip to content

Latest commit

 

History

History
39 lines (27 loc) · 1.13 KB

README.md

File metadata and controls

39 lines (27 loc) · 1.13 KB

DistilBART

http://arxiv.org/abs/2010.13002

More info can be found here.

Speedup DistilBART (Huggingface Transformers version) by using FastSeq

  • Speed on single NVIDIA-V100-16GB

    BatchSize 64 128
    transformers-3.0.2 3.4 samples/s OOM
    above + fastseq 16.8 samples/s 18.5 samples/s

Model

sshleifer/distilbart-cnn-12-6 from model hub.

Task

CNN/DM validation data

Setting

$ fastseq-generate-for-transformers \
    sshleifer/distilbart-cnn-12-6 \
    cnn_dm/val.source \
    out.summary \
    --reference_path cnn_dm/val.target \
    --device cuda \
    --bs BATCH_SIZE \
    --fp16 \
    --score_path out.score \
    --task summarization

Baseline speed number is obtained by running Transformers v3.0.2 code.