You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Aug 1, 2024. It is now read-only.
First of all thank you for sharing such an amazing project 🙂. Can we use the embeddings generated by context encoder of this pertained model for a task like object detection on custom dataset? If yes do you have any suggestions on which type of head architecture(for bbox and class label prediction) can be used with ijepa as backbone for object detection task?
Thank you
The text was updated successfully, but these errors were encountered:
May be you can find the answer in this paper https://arxiv.org/pdf/2111.06377.pdf
Just keep the encoder as a backbone and create a head for object detection. Case of R-CNN is described in this paper.
In annex A.4 the way to access a FPN like structure is also described
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
First of all thank you for sharing such an amazing project 🙂. Can we use the embeddings generated by context encoder of this pertained model for a task like object detection on custom dataset? If yes do you have any suggestions on which type of head architecture(for bbox and class label prediction) can be used with ijepa as backbone for object detection task?
Thank you
The text was updated successfully, but these errors were encountered: