Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Confused about how to optimize #3

Closed
UpCoder opened this issue Oct 29, 2019 · 7 comments
Closed

Confused about how to optimize #3

UpCoder opened this issue Oct 29, 2019 · 7 comments

Comments

@UpCoder
Copy link

UpCoder commented Oct 29, 2019

Hi, I am confused about how to optimize.
1, what's the loss function? in this paper, it seems not to be mentioned. is it just BCE?
2, what's the ground truth?
Thank you!

@yz93
Copy link
Owner

yz93 commented Oct 30, 2019

Dear @UpCoder,

No problem. Thank you for your interest.

  1. The loss is indeed binary cross entropy. Use BCEWithLogitsLoss which combines a Sigmoid layer with the BCELoss.

  2. The ground truth is 0/1 binary labels, with 1 indicating target and 0 indicating background.

@UpCoder
Copy link
Author

UpCoder commented Oct 30, 2019

Thank you for your reply.
So the unsupervised VOS just indicates that you do not use any annotation in the test stage. the semi-supervised method needs the annotation of the first frame in the test stage.
The training processing is fully supervised.
Is it right?

@yz93
Copy link
Owner

yz93 commented Oct 31, 2019

Dear @UpCoder,

What you said is exactly right. The use of the word “unsupervised” causes a fair amount of confusion and also some well founded doubts.

By today’s generally accepted definition of “unsupervised”—not using GT labels during training—this may be considered a misuse of a word. But the usage of the word in the sense that you just described I believe is originated from the once canonical setting in VOS, which uses human input as supervision to guide the algorithm at test time, as at those times “supervision” at test time was fairly common.

@UpCoder
Copy link
Author

UpCoder commented Oct 31, 2019

OK, got it.

Thank you!

@UpCoder UpCoder closed this as completed Oct 31, 2019
@mingminzhen
Copy link

@yz93 Hi, for the training step, do you compute the loss for the anchor image, which is not mentioned in the paper?

@yz93
Copy link
Owner

yz93 commented Nov 27, 2019

@mingminzhen No. The loss is binary cross-entropy with logits on the output of the network with GT binary labels.

@mingminzhen
Copy link

@yz93 Hi, for the training step, what is the scale range for randomly resize? If possible, can you provide the data augmentation code?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants