You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thank you for the dataset and code, I could imagine what a hard job it was to deliver that.
We are reproducing your pipeline and found that some subjects have zero DICE on inference.
We are trying to explore the issue with failing segmentations for a while, maybe you've already faced that. So we try to (1) reproduce your pipeline and after to train (2) nnUnet with custom data preprocessing.
(1) While reproducing your pipeline on T1 subjects we found that the network will not predict any tumour in 103 subjects. The overall inference quality will be around 0.3 DICE. During training, the quality reaches DICE 0.9.
By inference, I mean predicting the whole dataset after the network is trained.
(2) When we trained nnUnet with a similar preprocessing to yours on T1 and T2 and we get DICE 0.83 and have ~9 subjects predicted with DICE 0.
We train nnUnet from MONAI on T1 data after your preprocessing, and again 103 subjects will be poorly predicted (with not DICE 0, but DICE 0.4)
During the inference we reduce the sliding window size to
self.sliding_window_inferer_roi_size = [128, 128, 32]```
This was reduced to fit into GPU.
May be you can guide us - why subjects on inference get zeo DICEs? Maybe it is becouse you were training and predicting on the whole size image?
The text was updated successfully, but these errors were encountered:
thanks for your interest and questions!
I'm not sure if you changed anything else in the code, but if you change the sliding_window_inferer_roi_size, you should change the patch size (pad_crop_shape) at training time accordingly. Otherwise, you will feed much smaller patches to the network at inference time than at training time and that can lead to a bad performance.
We have recently trained models with nn-UNet on this GK dataset but also on a new multi-center dataset. The corresponding paper is not out, yet, but you can find the corresponding models here: https://zenodo.org/record/6827679#.Yt_vKi1Q3-Y
Dear all,
Thank you for the dataset and code, I could imagine what a hard job it was to deliver that.
We are reproducing your pipeline and found that some subjects have zero DICE on inference.
We are trying to explore the issue with failing segmentations for a while, maybe you've already faced that. So we try to (1) reproduce your pipeline and after to train (2) nnUnet with custom data preprocessing.
(1) While reproducing your pipeline on T1 subjects we found that the network will not predict any tumour in 103 subjects. The overall inference quality will be around 0.3 DICE. During training, the quality reaches DICE 0.9.
By inference, I mean predicting the whole dataset after the network is trained.
(2) When we trained nnUnet with a similar preprocessing to yours on T1 and T2 and we get DICE 0.83 and have ~9 subjects predicted with DICE 0.
We train nnUnet from MONAI on T1 data after your preprocessing, and again 103 subjects will be poorly predicted (with not DICE 0, but DICE 0.4)
During the inference we reduce the sliding window size to
The text was updated successfully, but these errors were encountered: