Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Suggestion - Integrate MobileSAM into the pipeline for lightweight and faster inference #30

Open
mdimtiazh opened this issue Jun 28, 2023 · 3 comments

Comments

@mdimtiazh
Copy link

Reference: https://github.com/ChaoningZhang/MobileSAM

Our project performs on par with the original SAM and keeps exactly the same pipeline as the original SAM except for a change on the image encode, therefore, it is easy to Integrate into any project.

MobileSAM is around 60 times smaller and around 50 times faster than original SAM, and it is around 7 times smaller and around 5 times faster than the concurrent FastSAM. The comparison of the whole pipeline is summarzed as follows:

image

image

Best Wishes,

Qiao

@lkeab
Copy link
Collaborator

lkeab commented Jun 28, 2023

hi, nice work and thanks for the suggestion. Can you provide the mask AP results on coco by comparing SAM with the same box prompt input?

@Coo1Sea
Copy link

Coo1Sea commented Jun 29, 2023

SAM-HQ is an awesome work. I'm planning to do that, to make SAM-HQ more lightweight. And another idea is to add a language model to increase SAM-HQ's visual scene reasoning capabilities. To my surprise, the demo on Hugging Face already supports open-vocabulary segmentation by text prompt and performs very well. I'd like to know how and look forward to the opportunity to collaborate.

@lkeab
Copy link
Collaborator

lkeab commented Jul 19, 2023

Hi, we provide Light HQ-SAM now. Please read more info here.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants