New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
patch partition? #42
Comments
|
Thank you for your patient reply. I noticed that during the training process, queries and targets are pairs of query points of images a and b, but they are concatenated in reverse order, which means that in the model prediction process, all queries of a and b need to be input to predict. COTR/COTR/inference/sparse_engine.py Lines 120 to 123 in 5c9363f
COTR/COTR/inference/sparse_engine.py Line 157 in 5c9363f
it looks like loc_from is the coordinates of the query point on graph a and loc_to is the coordinates of the query point on image b but isn't the prediction process of the model done in the infer_batch function def infer_batch ?COTR/COTR/inference/sparse_engine.py Line 216 in 5c9363f
|
Here it seems that the pred value is overwritten by the new loop each time the loop, so the resulting pred does not seem to store all the values predicted by the loop. COTR/COTR/trainers/cotr_trainer.py Lines 60 to 69 in 5c9363f
|
|
Your reply is very detailed. Thank you for your patient reply. |
Thank you for such an excellent job. I have some questions about cotr. During the training process, do you divide the scene images into 256*256 patches according to certain rules after scaling and then input them into the network for training? (I'm not sure where this step is implemented in the program.) How is corrs partitioned? Will it be the case that the corresponding point is divided into the next patch? How should this be handled? Is the validation process also similar to the training process after the split iteration.
The text was updated successfully, but these errors were encountered: