Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[s2s] add BartTranslationDistiller for distilling mBART #6363

Merged
merged 4 commits into from
Aug 12, 2020

Conversation

sshleifer
Copy link
Contributor

@sshleifer sshleifer commented Aug 9, 2020

New class BartTranslationDistiller does the same distillation method as SummarizationDistiller, but computes BLEU scores instead of ROUGE scores. It also accepts --src_lang and --tgt_lang arguments from the command line.

There is one strong checkpoint already posted at sshleifer/distillmbart-12-6/. I will post more in the coming days.

@codecov
Copy link

codecov bot commented Aug 9, 2020

Codecov Report

Merging #6363 into master will decrease coverage by 1.55%.
The diff coverage is n/a.

Impacted file tree graph

@@            Coverage Diff             @@
##           master    #6363      +/-   ##
==========================================
- Coverage   79.93%   78.37%   -1.56%     
==========================================
  Files         153      148       -5     
  Lines       27888    27196     -692     
==========================================
- Hits        22293    21316     -977     
- Misses       5595     5880     +285     
Impacted Files Coverage Δ
src/transformers/modeling_tf_mobilebert.py 24.55% <0.00%> (-70.09%) ⬇️
src/transformers/tokenization_reformer.py 41.66% <0.00%> (-40.00%) ⬇️
src/transformers/tokenization_marian.py 92.17% <0.00%> (-7.00%) ⬇️
src/transformers/modeling_tf_albert.py 81.74% <0.00%> (-6.61%) ⬇️
src/transformers/modeling_tf_electra.py 94.24% <0.00%> (-1.84%) ⬇️
src/transformers/modeling_longformer.py 89.21% <0.00%> (-1.64%) ⬇️
src/transformers/modeling_tf_bert.py 95.48% <0.00%> (-1.10%) ⬇️
src/transformers/configuration_bart.py 93.75% <0.00%> (-0.49%) ⬇️
src/transformers/modeling_tf_auto.py 66.66% <0.00%> (-0.21%) ⬇️
src/transformers/tokenization_bart.py 95.77% <0.00%> (-0.18%) ⬇️
... and 21 more

Continue to review full report at Codecov.

Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update be1520d...0718179. Read the comment docs.

@sshleifer sshleifer changed the title [s2s] code for distilling mbart, pegasus [s2s] add BartTranslationDistiller for distilling Mbart Aug 11, 2020
@sshleifer sshleifer requested a review from sgugger August 11, 2020 20:06
@sshleifer sshleifer changed the title [s2s] add BartTranslationDistiller for distilling Mbart [s2s] add BartTranslationDistiller for distilling mBART Aug 11, 2020
Copy link
Collaborator

@sgugger sgugger left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM!

@sshleifer sshleifer merged commit f94a52c into huggingface:master Aug 12, 2020
@sshleifer sshleifer deleted the distillfix branch August 12, 2020 15:41
fabiocapsouza added a commit to fabiocapsouza/transformers that referenced this pull request Nov 15, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants