Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

why the MT work? #36

Open
mountain-three opened this issue Jun 15, 2022 · 4 comments
Open

why the MT work? #36

mountain-three opened this issue Jun 15, 2022 · 4 comments

Comments

@mountain-three
Copy link

in my opinion,the mt just make the teacher and student network's outputs as the same as possible,but the same result doesn't mean the result is a correct result,especially when the labeled data is smaller than unlable data

@eraserNut
Copy link
Owner

With regard to the usefulness of unlabeled data, you can read this paper "A brief introduction to weakly supervised learning". Figure 3 in this paper maybe give you some enlightening.

@mountain-three
Copy link
Author

Thanks for reply,But in the paper, when you count the conherency loss,you use the same unlabel input to teacher and student network,and minimize the difference between the outputs.the same image into two similar model,the outputs ought to be the same.so I think your idea is a good way to let your model output variance low but not kind of semi-supervised

@eraserNut
Copy link
Owner

A type of semi-supervised learning method is using consistency losses between multi-models. In this way, you can consider it a regularization method. For some evidence, you can read "Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results".

@mountain-three
Copy link
Author

mountain-three commented Jun 19, 2022 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants