-
Notifications
You must be signed in to change notification settings - Fork 539
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Project author team stay tuned: I found out that the llama3-V project is stealing a lot of academic work from MiniCPM-Llama3-V 2.5 #196
Comments
Fact 2: When I questioned how the authors of llama3-v used MinicPM-Llama3-V2.5's tokenizer before the MinicPM-Llama3-V2.5 project was released, the authors of the llama3-v project began to lie. |
Based on the above three facts, I think there is sufficient evidence to prove that the llama3-v project has stolen the academic achievements of the minicpm-llama 3-v 2.5 project, and I strongly suggest that the minicpm-llama 3-v 2.5 project's team go to the complaint to expose the llama3-v project authors' stealing and lying about academic misconduct, and so on a series of problems! |
Hi @pzc163, This situation sounds extremely serious. We never expected anything like this to happen. We hope the truth will come to light soon. |
Thanks for the info. The inference fix and noise sound horrific. We are reproducing it and will test more on some in-house features. |
The conclusion of our investigation:
After receiving the issue from @yangzhizheng1on GitHub, we launched a serious investigation. We can obtain inference results correctly using Llama3-V checkpoint with MiniCPM-Llama3-V 2.5's code and config file following @yangzhizheng1's instruction on GitHub. Even more, we also surprisingly find that Llama3-V shows highly similar behaviors to MiniCPM-Llama3-V 2.5 in some unrevealed experimental features, which are trained on private in-house data, such as recognizing Tsinghua Bamboo Characters. One of the experimental features of MiniCPM-Llama3-V 2.5 is recognizing Tsinghua Bamboo Characters (清华简), a very special and rare type of Chinese ancient characters written on bamboo during China's Warring States Period (475 BC-221 BC). These training images are recently scanned from unearthed cultural relics and annotated by our team, which has not been publicly released yet. Surprisingly, we find highly similar capabilities for Llama3-V in both good and bad cases. For quantative results, we also tested several Llama3-based VLMs on 1K Bamboo Character images and compared the prediction exact match for each pair of models. The overlaps between every two models are zero, whereas the overlaps between Llama3-V and MiniCPM-Llama3-V 2.5 achieve a surprising 87%. Moreover, MiniCPM-Llama3-V 2.5 and Llama3-V even share a similar error distribution. Llama3-V and MiniCPM-Llama3-V 2.5 make 236 and 194 wrong predictions respectively, while the overlapped part is 182. The MiniCPM-Llama3-V2.5-noisy obtained following @yangzhizheng1's instruction on GitHub shows nearly identical quantative results with Llama3-V. This is really confusing... The same thing also happens to WebAgent, another unrevealed feature trained on in-house data. They even make identical errors in a WebAgent schema newly defined within our team... Since the HuggingFace page of Llama3-V is removed now, we upload the checkpoint here (https://bit.ly/3yRFxYq). Since this model has received several thousands of downloads on HuggingFace, there should be independent copies to reproduce this. Given these results, we are afraid it is hard to explain such unusual similarities as coincidences. We hope the authors can give an official explanation of the issue. We believe this is important for the common good of the open-source community. |
look~~ |
You might want to report this to Stanford CS or Stanford itself. These are serious allegations and they appear (at a quick glance and to my non-expert eyes) to be well substantiated. |
If the research team from Stanford University is proven to have plagiarized this MiniCPM-V project from Tsinghua University, they should feel ashamed, and also, MiniCPM-V project deserve an apology and acknowledgment. |
You can consult the Dean of Stanford CS department to report misconducts. Refer to this policy: Section 5: Individual Reporting Responsibility |
The current Dean is likely Jennifer Widom: |
This comment was marked as abuse.
This comment was marked as abuse.
Definitely escalate this to Stanford. Plagiarism cannot be tolerated. |
This comment was marked as off-topic.
This comment was marked as off-topic.
If you google Llama3v there have now been more than 1000 pages attributing the work. This has made a detrimental impact. Their actions seemed deliberately planned, aiming for rapid and extensive coverage in tech news with attention-grabbing assertions. This strategy can make the stolen credits be attributed to them quickly before the original authors even realize it. The authors may want to escalate this to their academic administrators immediately to prevent any further negative impact. |
I am not sure if I understand correctly. Actually, the original project is an open-source project, so Llama3-V can use it, but they didn't comply with the open-source license? |
The latest news, one of the authors Aksh Garg, has acknowledged that on his medium post
|
You would hardly be satisfied with this kind of statement if it were your work being copied, with model weights deliberately altered with Gaussian noise and renamed, with a plotted and overwhelming coverage in news and social media (enhanced by eyes-grabbing "$500" statements in the headlines) etc... This goes beyond merely saying "the architecture is very similar to blah blah blah..." And guess what, they said you merely "beat us to the implementation."??? To be honest, one would be furious if they were the authors and saw the statement... |
A thief is being tried in court, and this is his statement:
|
@tangmingxing1988 you forgot the part before the trial where the thief takes a world tour extolling their amazing belongings 😜 |
@pzc163 @Cuiunbo @RylanSchaeffer Given the two checkpoints, perhaps one can compute the diff in the weight to see the histogram? (even though it already seems like there is enough evidence) |
So basically they just randomly added some noise to the weight and called it a day? Sheesh. |
|
If you ran a hypothesis test about whether these distributions are Gaussian, what would it tell us? Agostino-Pearson and Kolmogorov-Smirnov seem reasonable |
The fact that all the means are nearly 0 and the standard deviations appear almost identical seems damning... |
This is the strongest evidence that llama3-V does not train its model at all, but adds random Gaussian noise to the model parameters of miniCPM-llama3-v2.5 |
IMHO no way the delta between a finetuned model with the original would fit a gaussian this well, not to mention that the distribution is almost the SAME across several different layers. Yeah, this is looking very bad. |
This comment was marked as off-topic.
This comment was marked as off-topic.
This comment was marked as off-topic.
This comment was marked as off-topic.
This comment was marked as off-topic.
This comment was marked as off-topic.
This comment was marked as off-topic.
This comment was marked as off-topic.
如果只是需要收到新回复提醒可以点击侧边栏的 Subscribe 订阅,无需回复这个 issue,不然所有订阅者都会收到邮件通知。 |
This comment was marked as off-topic.
This comment was marked as off-topic.
This comment was marked as off-topic.
This comment was marked as off-topic.
Can someone reshare the "weights diff" plot? |
https://gist.github.com/TomoshibiAkira/151a2353b946aa9cd8d4d2cdabc31245 Quickly wrote a script to compare the delta of the two weights (LLaMa3V with @Cuiunbo 's link, MiniCPM is from HF), calculate the delta's mean and std and abs(delta)'s mean for every layer, and finally print the histogram across all layers. Anyway, here's the final histogram: Std of delta: Mean of abs(delta): |
支持维护自身创作权,抄袭令人不齿 |
stanford ai research should issue a statement on this |
|
Thanks Rylan. From an outsider perspective it seemed so since two authors are/were undergraduate researchers in SAIL. I appreciate you and Professor Manning shedding light on this matter from within the SAIL community. |
Were they SAIL researchers? I'm trying to decide whether to report them to Stanford. I am not personally involved but I think their behavior is clearly immoral and unbecoming of Stanford students. |
https://www.linkedin.com/in/aksh-garg/ https://www.linkedin.com/in/siddharth-sharma-9942b2104/details/experience/ both point to current/former SAIL experience |
Also please report to https://communitystandards.stanford.edu/policies-guidance/honor-code
|
Fellow MiniCPM-Llama3-V 2.5 project authors, a few days ago I discovered a shocking fact.There is a large amount of work in the llama3-V (https://github.com/mustafaaljadery/llama3v) project that is suspected to have been stolen from the MiniCPM-Llama3-V 2.5 project, and I raised my query in the GitHub project issue of llama3-v, and did not think that the The authors of Llama3-V quickly deleted my questionable post, and hid Llama3-V's Huggingface project page. I strongly question what they did, and I will release all the evidence next, and I urge you to pay attention to this fact.
![3274bc276dd9b91471395b3075c5e6e](https://private-user-images.githubusercontent.com/105222039/335843024-dd982b1a-789a-4990-a50b-8ba162298b61.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MTk5MzY3MjEsIm5iZiI6MTcxOTkzNjQyMSwicGF0aCI6Ii8xMDUyMjIwMzkvMzM1ODQzMDI0LWRkOTgyYjFhLTc4OWEtNDk5MC1hNTBiLThiYTE2MjI5OGI2MS5wbmc_WC1BbXotQWxnb3JpdGhtPUFXUzQtSE1BQy1TSEEyNTYmWC1BbXotQ3JlZGVudGlhbD1BS0lBVkNPRFlMU0E1M1BRSzRaQSUyRjIwMjQwNzAyJTJGdXMtZWFzdC0xJTJGczMlMkZhd3M0X3JlcXVlc3QmWC1BbXotRGF0ZT0yMDI0MDcwMlQxNjA3MDFaJlgtQW16LUV4cGlyZXM9MzAwJlgtQW16LVNpZ25hdHVyZT1jMmYwMWM3NjVlOWZkMTA0NWU4MzY2Mjg0YmViZWI3ZmNkZTQ4NTAyMzU2OGM4OWNmYWZlMGMyYzllNDhjYjNlJlgtQW16LVNpZ25lZEhlYWRlcnM9aG9zdCZhY3Rvcl9pZD0wJmtleV9pZD0wJnJlcG9faWQ9MCJ9.ixHACnSUPyp2zjE4-rs7iZHSAQJsDtm_AyWVJ6gjzio)
![c77dc3fc767e6be2903be1f858cf163](https://private-user-images.githubusercontent.com/105222039/335843044-3085f8b4-67be-4bc5-b7aa-07b7af9f781f.jpg?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MTk5MzY3MjEsIm5iZiI6MTcxOTkzNjQyMSwicGF0aCI6Ii8xMDUyMjIwMzkvMzM1ODQzMDQ0LTMwODVmOGI0LTY3YmUtNGJjNS1iN2FhLTA3YjdhZjlmNzgxZi5qcGc_WC1BbXotQWxnb3JpdGhtPUFXUzQtSE1BQy1TSEEyNTYmWC1BbXotQ3JlZGVudGlhbD1BS0lBVkNPRFlMU0E1M1BRSzRaQSUyRjIwMjQwNzAyJTJGdXMtZWFzdC0xJTJGczMlMkZhd3M0X3JlcXVlc3QmWC1BbXotRGF0ZT0yMDI0MDcwMlQxNjA3MDFaJlgtQW16LUV4cGlyZXM9MzAwJlgtQW16LVNpZ25hdHVyZT00NDc3NWMwNzNmMzU1NGE4NmU1YjJlYmMyMmJmMDRkNjViZDQwZjIwOTVjZmNlNTlmZmQ0NDdhOGVkZjk5ZTZmJlgtQW16LVNpZ25lZEhlYWRlcnM9aG9zdCZhY3Rvcl9pZD0wJmtleV9pZD0wJnJlcG9faWQ9MCJ9.-5Yyu0JwqKT-PnDy66qERTexBbRHR3paE2HUPiy7bfA)
![ccaaa5e2e0c4c79fcfcc618740b1786](https://private-user-images.githubusercontent.com/105222039/335843300-89923dc8-0fb1-44c7-acf3-e316e77b12af.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MTk5MzY3MjEsIm5iZiI6MTcxOTkzNjQyMSwicGF0aCI6Ii8xMDUyMjIwMzkvMzM1ODQzMzAwLTg5OTIzZGM4LTBmYjEtNDRjNy1hY2YzLWUzMTZlNzdiMTJhZi5wbmc_WC1BbXotQWxnb3JpdGhtPUFXUzQtSE1BQy1TSEEyNTYmWC1BbXotQ3JlZGVudGlhbD1BS0lBVkNPRFlMU0E1M1BRSzRaQSUyRjIwMjQwNzAyJTJGdXMtZWFzdC0xJTJGczMlMkZhd3M0X3JlcXVlc3QmWC1BbXotRGF0ZT0yMDI0MDcwMlQxNjA3MDFaJlgtQW16LUV4cGlyZXM9MzAwJlgtQW16LVNpZ25hdHVyZT02MGY0YTU4NjhmNzc5YTc3YTc0NDA5Y2M5YWM2MTk5YzVjZGEyZDJhYzQ5OGUwODQzNzJhZmE4NWJlNWUyMzBhJlgtQW16LVNpZ25lZEhlYWRlcnM9aG9zdCZhY3Rvcl9pZD0wJmtleV9pZD0wJnJlcG9faWQ9MCJ9.ud_2vvY3iK5mKs0PUn26Q9yjG8ymU2r6zrNa1STkbiw)
![c891d8439a67682715cb488fbd0b351](https://private-user-images.githubusercontent.com/105222039/335843072-5a76a819-299e-4102-9a7b-3bf91be254ec.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MTk5MzY3MjEsIm5iZiI6MTcxOTkzNjQyMSwicGF0aCI6Ii8xMDUyMjIwMzkvMzM1ODQzMDcyLTVhNzZhODE5LTI5OWUtNDEwMi05YTdiLTNiZjkxYmUyNTRlYy5wbmc_WC1BbXotQWxnb3JpdGhtPUFXUzQtSE1BQy1TSEEyNTYmWC1BbXotQ3JlZGVudGlhbD1BS0lBVkNPRFlMU0E1M1BRSzRaQSUyRjIwMjQwNzAyJTJGdXMtZWFzdC0xJTJGczMlMkZhd3M0X3JlcXVlc3QmWC1BbXotRGF0ZT0yMDI0MDcwMlQxNjA3MDFaJlgtQW16LUV4cGlyZXM9MzAwJlgtQW16LVNpZ25hdHVyZT1iZjYzNTIxNjRiYzNhNDZmZmRiNDY2ZmRjODVlMjhmNTExNTU1N2E0NTlhMjI0ZGIzMTdjMmY0MDMwMjlhZTBmJlgtQW16LVNpZ25lZEhlYWRlcnM9aG9zdCZhY3Rvcl9pZD0wJmtleV9pZD0wJnJlcG9faWQ9MCJ9.DDYrCVLDUA2CvmzrICW178m9b369dY91W6MUZnhIPV4)
![2a7a14b98280eabd04c7dc503c30c27](https://private-user-images.githubusercontent.com/105222039/335843078-541d0443-72c5-43a4-96de-9a2818700ff4.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MTk5MzY3MjEsIm5iZiI6MTcxOTkzNjQyMSwicGF0aCI6Ii8xMDUyMjIwMzkvMzM1ODQzMDc4LTU0MWQwNDQzLTcyYzUtNDNhNC05NmRlLTlhMjgxODcwMGZmNC5wbmc_WC1BbXotQWxnb3JpdGhtPUFXUzQtSE1BQy1TSEEyNTYmWC1BbXotQ3JlZGVudGlhbD1BS0lBVkNPRFlMU0E1M1BRSzRaQSUyRjIwMjQwNzAyJTJGdXMtZWFzdC0xJTJGczMlMkZhd3M0X3JlcXVlc3QmWC1BbXotRGF0ZT0yMDI0MDcwMlQxNjA3MDFaJlgtQW16LUV4cGlyZXM9MzAwJlgtQW16LVNpZ25hdHVyZT0yNzE2NmI0N2MyNWQ4Mjc4OWY1MDBiNDIxNWIxOGJjOTFhZmE1OTMyYmQ0NDFlMDJhNWVjY2IxNjgyYmNiYTc2JlgtQW16LVNpZ25lZEhlYWRlcnM9aG9zdCZhY3Rvcl9pZD0wJmtleV9pZD0wJnJlcG9faWQ9MCJ9.gXdevNXuxjweqGbbiyVhAsylx7Yg6pZCfbMvufr8UgI)
this issue has been deleted by the author of llama3-V ( https://github.com/mustafaaljadery/llama3v ),I will expose all the evidence to expose the fact that the authors of llama3-v are a bunch of thieves!
The text was updated successfully, but these errors were encountered: