Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About training and testing details #1

Closed
bigmoking opened this issue Jul 16, 2020 · 5 comments
Closed

About training and testing details #1

bigmoking opened this issue Jul 16, 2020 · 5 comments
Labels
question Further information is requested

Comments

@bigmoking
Copy link

Hi!
There are very few details about training and testing in your article. Do you have any supplementary materials?If not, can you describe it to me in detail? Thanks!
yours

@angpo
Copy link
Collaborator

angpo commented Jul 16, 2020

Hi,

You can find the supplementary materials on arXiv (precisely, at the end of the main paper).
Additionally, commands.txt contains the bash scripts we used to train the models. If you refer to the values of hyperparameters, Section 4.1 of the main paper proposes a brief discussion about Implementation details.

If you need any further information please feel free to write us.

A.

@bigmoking
Copy link
Author

bigmoking commented Jul 16, 2020 via email

@angpo angpo closed this as completed Jul 16, 2020
@lucabergamini lucabergamini added the question Further information is requested label Jul 18, 2020
@bigmoking
Copy link
Author

bigmoking commented Jul 19, 2020 via email

@angpo
Copy link
Collaborator

angpo commented Jul 19, 2020

Hi,

  1. Yes. Each image in the input set comes from a different camera.
  2. No. The teacher network is fixed during the second stage; we train only the student one.
  3. In general terms, Knowledge distillation relies on transferring knowledge from one network to another. In the model-compression area, a compact neural network is trained to mimic the outputs of a larger (and slower) model. Instead, the idea behind self-distillation is to teach a student network comprising the same architecture of its teacher. If you wish to know more about self-distillation, you can find here a list of paper on the topic.

A.

@angpo angpo reopened this Jul 19, 2020
@bigmoking
Copy link
Author

bigmoking commented Jul 19, 2020 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

3 participants