Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[REQUEST] Code and models please! #2

Open
spacewalkingninja opened this issue May 22, 2023 · 19 comments
Open

[REQUEST] Code and models please! #2

spacewalkingninja opened this issue May 22, 2023 · 19 comments

Comments

@spacewalkingninja
Copy link

Hello!
I am urgently asking for the release of the inference code + model. Training would be good too.
Incredibly thankful, very interesting project!

@spacewalkingninja spacewalkingninja changed the title Code and models please! [REQUEST] Code and models please! May 22, 2023
@mtjhl
Copy link

mtjhl commented May 25, 2023

When will the codes be released?

@hjq133
Copy link

hjq133 commented May 26, 2023

+1. looking forward to the code. intersting project.

@wzhings
Copy link

wzhings commented May 26, 2023

+1. I am looking forward to the codes. It is an awesome work.

@wojiaohumaocheng
Copy link

+1

1 similar comment
@karthikyeredla
Copy link

+1

@spacewalkingninja
Copy link
Author

@czczup can you please enlighten us from the realms of the model and code lands <3

@mpragnay
Copy link

mpragnay commented Jun 9, 2023

what training data has been used?? Is it publicly available

@autosquid
Copy link

any update?

@amygbAI
Copy link

amygbAI commented Jul 5, 2023

i think all of you guys are wasting your time waiting for this .. check out the original LLaVA paper ..it has code, demo and all you need to get started ..i however, thank the authors of this paper for referencing it and letting us know it exists :)

@amygbAI
Copy link

amygbAI commented Sep 14, 2023 via email

@GuangxingHan
Copy link

Hi Bruno .. i would say the objectives are 100% the same. So its better to go with a Microsoft research paper that has code rather than some random copy of it ..obviously the authors dont seem to care much anymore

On Thu, Sep 14, 2023 at 4:58 PM Bruno Ma @.> wrote: Hi @amygbAI https://github.com/amygbAI, u mean this paper is totally the same with LLaVA? i think all of you guys are wasting your time waiting for this .. check out the original LLaVA paper ..it has code, demo and all you need to get started ..i however, thank the authors of this paper for referencing it and letting us know it exists :) — Reply to this email directly, view it on GitHub <#2 (comment)>, or unsubscribe https://github.com/notifications/unsubscribe-auth/ATIQOSDEJWY52C3JKN3FZATX2LS7LANCNFSM6AAAAAAYKZB7RI . You are receiving this because you were mentioned.Message ID: @.>

Hi, May I know if you can reproduce the results? Do you mean this work uses the same training objective as LLaVA? Thanks.

@amygbAI
Copy link

amygbAI commented Sep 19, 2023 via email

@GuangxingHan
Copy link

GuangxingHan commented Sep 19, 2023

Hi Mr Han .. no, i havent reproduced the results because i would like to train this on charts / graph data exclusively and im preparing the datasets ..having said that , i believe the training objective is the same , i.e. - ensure that the model can take both images and text as input - perform analysis over both image + textual contexts - provide results of the query in textual format if you go through the LLaVA paper, this will be amply evident to you On Tue, Sep 19, 2023 at 11:25 AM Guangxing Han @.> wrote:

Hi Bruno .. i would say the objectives are 100% the same. So its better to go with a Microsoft research paper that has code rather than some random copy of it ..obviously the authors dont seem to care much anymore … <#m_-2675669682656386325_> On Thu, Sep 14, 2023 at 4:58 PM Bruno Ma @.
> wrote: Hi @amygbAI https://github.com/amygbAI https://github.com/amygbAI https://github.com/amygbAI, u mean this paper is totally the same with LLaVA? i think all of you guys are wasting your time waiting for this .. check out the original LLaVA paper ..it has code, demo and all you need to get started ..i however, thank the authors of this paper for referencing it and letting us know it exists :) — Reply to this email directly, view it on GitHub <#2 (comment) <#2 (comment)>>, or unsubscribe https://github.com/notifications/unsubscribe-auth/ATIQOSDEJWY52C3JKN3FZATX2LS7LANCNFSM6AAAAAAYKZB7RI https://github.com/notifications/unsubscribe-auth/ATIQOSDEJWY52C3JKN3FZATX2LS7LANCNFSM6AAAAAAYKZB7RI . You are receiving this because you were mentioned.Message ID: @.
> Hi, May I know if you can reproduce the results? Do you mean this work uses the same training objective as LLaVA? Thanks. — Reply to this email directly, view it on GitHub <#2 (comment)>, or unsubscribe https://github.com/notifications/unsubscribe-auth/ATIQOSAK2N7RFUJN2OK6773X3EXWJANCNFSM6AAAAAAYKZB7RI . You are receiving this because you were mentioned.Message ID: @.
**>

Thanks for your reply. Yes, LLaVA works exactly in this way.

@becauseofAI
Copy link

@czczup Can you provide a timeline for the release code? Thx!

@shaniaos
Copy link

I see that this paper is accepted by NeurIPS 2023, which is held 1 month ago. It's Jan 2024 now.
Is the code going to be released?

@zzchust
Copy link

zzchust commented Feb 4, 2024

+1

@annopackage
Copy link

waiting for code release.

@spacewalkingninja
Copy link
Author

spacewalkingninja commented Mar 1, 2024 via email

@Haiyang-W
Copy link

Haiyang-W commented May 12, 2024

If needed, everyone may try the GiT repository, a general end-to-end vision transformer, which fully covers the tasks included in visionLLM and can also handle semantic segmentation. The code and pre-trained weights have been fully open-sourced.

"GiT: Towards Generalist Vision Transformer through Universal Language Interface"

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests