Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Slue PR configs #5087

Merged
merged 16 commits into from
Jan 8, 2024
Merged

Slue PR configs #5087

merged 16 commits into from
Jan 8, 2024

Conversation

siddhu001
Copy link
Collaborator

Upload configs and provide access to best pre-trained models for SLT paper (https://arxiv.org/abs/2211.05869).

@sw005320 sw005320 added the SLU Spoken language understanding label Apr 6, 2023
@sw005320 sw005320 added this to the v.202303 milestone Apr 6, 2023
@sw005320 sw005320 requested a review from ftshijt April 6, 2023 14:55
@kan-bayashi kan-bayashi modified the milestones: v.202303, v.202307 May 1, 2023
@kan-bayashi kan-bayashi modified the milestones: v.202307, v.202312 Aug 3, 2023
@kan-bayashi kan-bayashi modified the milestones: v.202310, v.202312 Oct 25, 2023
@siddhu001 siddhu001 marked this pull request as ready for review December 29, 2023 04:02
@siddhu001 siddhu001 changed the title Slue PR configs [WIP] Slue PR configs Dec 29, 2023
Copy link

codecov bot commented Dec 29, 2023

Codecov Report

All modified and coverable lines are covered by tests ✅

Comparison is base (7c140c2) 76.54% compared to head (273c36f) 76.57%.
Report is 64 commits behind head on master.

Additional details and impacted files
@@            Coverage Diff             @@
##           master    #5087      +/-   ##
==========================================
+ Coverage   76.54%   76.57%   +0.03%     
==========================================
  Files         720      720              
  Lines       66639    66695      +56     
==========================================
+ Hits        51008    51072      +64     
+ Misses      15631    15623       -8     
Flag Coverage Δ
test_configuration_espnet2 ∅ <ø> (∅)
test_integration_espnet1 62.92% <ø> (ø)
test_integration_espnet2 50.10% <ø> (+0.02%) ⬆️
test_python_espnet1 19.06% <ø> (-0.02%) ⬇️
test_python_espnet2 52.45% <ø> (+0.05%) ⬆️
test_utils 22.15% <ø> (ø)

Flags with carried forward coverage won't be shown. Click here to find out more.

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

@siddhu001
Copy link
Collaborator Author

Hi @ftshijt,

This PR is ready for review. The primary modifications include the addition of configuration files for the best-trained model in the SLT paper and the ESPnet-SLUE toolkit, which was presented at the ASRU workshop. I've also implemented some minor fixes to the scoring function and the 2-pass SLU implementation. Could you please review it?

@siddhu001
Copy link
Collaborator Author

Hi @sw005320,

This PR is related to slue-voxceleb and slue-voxpopuli recipes for ESPnet-SLUE toolkit. Please let me know if you have any comments before we can merge this PR!

@sw005320
Copy link
Contributor

sw005320 commented Jan 3, 2024

OK, I'll start to review.
@ftshijt, can you also start it?

Copy link
Collaborator

@ftshijt ftshijt left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for the update! Looks very good to me. Two minor comments

valid_set="devel"
test_sets="test devel"

slu_config=conf/tuning/train_asr_no_pretrain.yaml
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Consider make it standardize by setting conf/train.yaml

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Done!

@@ -0,0 +1,88 @@
specaug: specaug
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It seems to me that the change of many configs are quite similar, do we want to keep all of them? (I'm ok with it if that's for public benchmark, but ideally, it would be better if we could summarize them into fewer configs, which is also easier for users)

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yes, I considered it, but similar configs (like the ones using pre-trained ASR) have different sets of hyperparameters (like warmup steps, learning rate) selected based on validation performance. Hence, grouping similar configs would reduce reproducibility, so I think it is better to keep all of the configs.

@siddhu001
Copy link
Collaborator Author

siddhu001 commented Jan 8, 2024

@sw005320 I think this PR is ready to be merged! Let me know if you have any comments.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The table is not correctly displayed.
Can you confirm it?
image

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for catching this! I have fixed this now!

@sw005320 sw005320 added the auto-merge Enable auto-merge label Jan 8, 2024
@mergify mergify bot merged commit 4dccbdf into espnet:master Jan 8, 2024
27 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
auto-merge Enable auto-merge ESPnet2 README SLU Spoken language understanding
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

4 participants