Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Results not matching with paper? #19

Closed
samra-irshad opened this issue Aug 5, 2020 · 7 comments
Closed

Results not matching with paper? #19

samra-irshad opened this issue Aug 5, 2020 · 7 comments

Comments

@samra-irshad
Copy link

Hi, I am trying to reproduce the results but could not get the results closer to those reported in paper. Did you utilize validation set for reporting the results or test set (in table 4)? Also, could you provide with the training, test and validation splits?

@sinAshish
Copy link
Owner

The reported results are from our validation set since ground truth in testing set are not available.
Also, the experiments were performed about 2 years ago and I have changed the computer, so I don’t recall the exact splitting, but tbh results should be consistent across architectures.

@samra-irshad
Copy link
Author

Right. Actually I am not talking about the test set in competition, I am talking about the test set in split. Like you divided the data into (13: train, 2: validation and 5: test) for CHAOS dataset. I can see in code that validation performance has been saved in folders at end of each epoch and the best segmentations on validation split are used for 3D reconstruction. So the question is did you use validation split (2 cases) for reporting model's performance in paper or reported the scores after evaluating the model with best validation score on test set at end?

@sinAshish
Copy link
Owner

sinAshish commented Aug 7, 2020

I'm not sure exactly about the splits, but a 5-fold cv was used to report the score on the test set.

@josedolz
Copy link
Collaborator

josedolz commented Aug 7, 2020

The reported results are on our test set (averaged over k-folds). The validation set is only used to pick the best model during training for each run.

@samra-irshad
Copy link
Author

Ok thanks Jose

@samra-irshad
Copy link
Author

My results are different for 3-fold cross-validation and i have applied the same exact method. I am clueless why there's performance difference. I have emailed you the training, validation and test lists that I am using. It would be very helpful if you could train your model instance with my sent lists and update me with the results. Or maybe train your model once again for a fresh train-valid-test splits and update me on email or here? Or perhaps provide the instance of trained model?

@sinAshish
Copy link
Owner

sinAshish commented Aug 10, 2020

train your model instance with my sent lists and update me with the results. Or maybe train your model once again for a fresh train-valid-test splits

sorry, but this project is already done and we are busy with other things now.
You should try looking for any bugs in the code!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants