This repository has been archived by the owner on Aug 29, 2023. It is now read-only.
Question about parameter tuning #40
Comments
The loss is normalized by the number of GT boxes: MaskFormer/mask_former/modeling/criterion.py Line 165 in 232bd1b
We use batchsize of 64 following DETR, and we have not tried using a smaller batchsize in this work. |
Is the box only used to normalize? I think you use the dice of mask and class prob to conduct matching between gt and preds? |
In bipartite matching, we calculate all pairwise losses, so there is no normalization needed in matching. |
Sorry for the confusion. My question is do you use the box for matching or as part of the loss function? |
Boxes are never used. Please ignore the name in comment (copied from DETR code). |
I see, thank you for your answers. |
Sign up for free
to subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Hi bowen,
Thank you for sharing such a great work.
I have a question about the parameter for training on new dataset for panoptic segmentation task. In the new dataset, we have less objects in each image (maybe 1-5). What parameter do you think is the most important to conduct this adaptation? Any advice is really appreciated.
The situation is that I found the final mask and dice loss are close to 0.1 which is somehow smaller than it in coco panoptic training (about 0.3)? I wonder is there any normalization in the code to make the loss small when object number is small? I think no?
Another random question: do you use a large batchsize of 64 because of the poor label quality of coco? If I change it to 16 will that have a large difference (I ask this because of the panoptic-deeplab pytorch version using 16 while paper using 64 also)?
Look forward to your reply.
The text was updated successfully, but these errors were encountered: