Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Discrepancy in Reproduced Results (Zuco Dataset) #5

Open
FishCatCake opened this issue Sep 1, 2023 · 2 comments
Open

Discrepancy in Reproduced Results (Zuco Dataset) #5

FishCatCake opened this issue Sep 1, 2023 · 2 comments

Comments

@FishCatCake
Copy link

Hi Lukas,

Thank you for the valuable work and for making the code and datasets publicly available. Your research is very innovative and inspiring!

I recently went through your paper DETRtime and tried to reproduce the results using the code from this repository and the provided Zuco dataset. While I followed the steps as described in the README and the paper, the validation results I obtained were notably different from those reported in the paper.

Hyperparameter setup follows the table 6 in the paper appendix:
image

Obtained results for zuco:

  1. Below are the f1 scores for each class (0, 1, 2), orange for training, blue for validation.
image
  1. Below are the weighted average f1 scores and macro average f1 scores, respectively. Orange for training, blue for validation.
image

Reported result for zuco in the paper:
image

I'm reaching out to see if there might be any additional steps, configurations, or updates that I might have missed. I'm eager to understand the potential reasons for this discrepancy. Any insights or suggestions would be highly appreciated.

Thanks.
Yicen

@Onlyknight
Copy link

Hi Lukas,

Thank you for the valuable work and for making the code and datasets publicly available. Your research is very innovative and inspiring!

I recently went through your paper DETRtime and tried to reproduce the results using the code from this repository and the provided Zuco dataset. While I followed the steps as described in the README and the paper, the validation results I obtained were notably different from those reported in the paper.

Hyperparameter setup follows the table 6 in the paper appendix: image

Obtained results for zuco:

  1. Below are the f1 scores for each class (0, 1, 2), orange for training, blue for validation.
image 2. Below are the weighted average f1 scores and macro average f1 scores, respectively. Orange for training, blue for validation. image **Reported result for zuco in the paper:** image

I'm reaching out to see if there might be any additional steps, configurations, or updates that I might have missed. I'm eager to understand the potential reasons for this discrepancy. Any insights or suggestions would be highly appreciated.

Thanks. Yicen

Hello,I have meet the same problem with you!

@Onlyknight
Copy link

Hi Lukas,
When I was replicating your article using the zuco dataset, I found that all the final prediction boxes were overlapping in one place, and I couldn't reproduce the experimental results as effectively as you did in your paper. May I ask if there are any specific training techniques you can recommend?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants