Interested in Project 8 - Refining Zero-Shot Object Segmentation by Combining Vision Foundation Models #29497
DeepC004
started this conversation in
Google Summer of Code
Replies: 1 comment
-
The idea is to follow the interface/implementation of Visual Prompting in https://github.com/openvinotoolkit/model_api/tree/master |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Dear Daan, Klaas, and Samet,
The idea of refining zero-shot object segmentation with models like DINOv2 and SAM sounds exciting and I’d love to get involved.
I have previous experience in relevant areas. For example, I co-authored a paper on few-shot learning, exploring techniques to compare features and enable generalization from limited examples—approaches that share similarities with zero-shot learning’s reliance on feature extraction for object identification. I’ve also worked with vision-based neural networks, such as a style transfer project, which has given me strong familiarity with PyTorch and solving vision challenges. Additionally, I’ve contributed to open-source projects, developing neural network implementations and integrating them into testing pipelines. These experiences have equipped me with a solid foundation in machine learning tools and producing reliable code. I’d be keen to use my experience to help build a segmentation system that’s solid and works broadly.
The zero-shot approach and the opportunity to enhance its generalizability are compelling, and I got a few questions.
I’m looking forward to digging into DINOv2 and SAM and others architectures like Swin Transformer. If you’ve got any pointers on where to start or what’s ahead, I’d love to hear them. Hoping to team up on this!
Cheers,
Deep Chordia
Beta Was this translation helpful? Give feedback.
All reactions