Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Comparisons might not be consistent across batch sizes #25

Closed
rajatsen91 opened this issue Mar 14, 2023 · 3 comments
Closed

Comparisons might not be consistent across batch sizes #25

rajatsen91 opened this issue Mar 14, 2023 · 3 comments

Comments

@rajatsen91
Copy link

rajatsen91 commented Mar 14, 2023

The prediction and actual sizes here are not the same across batch sizes. This means that the metrics calculated are not exactly comparable across different models trained with different batch sizes.

Therefore if my understanding is correct all models might need to be reevaluated.

I think the culprit is here: https://github.com/cure-lab/LTSF-Linear/blob/main/data_provider/data_factory.py#L20

The drop_last should be False during the test evaluations.

@ikvision
Copy link

Indeed for different batch sizes the test set will be different as we drop the last one.
I think this issues is similar to #7

@rajatsen91
Copy link
Author

Yes indeed. Sorry I missed the previous issue. I also found that this does change results somewhat significantly in ETTh2 dataset as well.

@namctin
Copy link
Collaborator

namctin commented Mar 15, 2023

Yes that is a valid question. We noticed that when @oguiza raised the issue. With self-supervised model, the result is with respect to drop_last=False and you can see the performance did not affect much, as shown in the paper.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants