Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Will the LVD-142M dataset or data processing codes be released? #24

Closed
XiaohuJoshua opened this issue Apr 19, 2023 · 13 comments
Closed
Assignees
Labels
enhancement New feature or request

Comments

@XiaohuJoshua
Copy link

Thanks for the outstanding work. Do you have the plan to release the LVD-142M dataset or codes for data processing?

@patricklabatut
Copy link
Contributor

Releasing LVD-142M is not something that we are considering releasing, I am afraid. Re: open-sourcing the data curation code, this could depend on feedback and interest from the community.

@patricklabatut patricklabatut self-assigned this Apr 19, 2023
@patricklabatut patricklabatut added the enhancement New feature or request label Apr 19, 2023
@salonit77
Copy link

salonit77 commented Apr 19, 2023

Would it be possible to share additional details regarding the deduplication part used in the data curation pipeline. Or how that can be done for a custom dataset.

@XiaohuJoshua
Copy link
Author

Would it be possible to share additional details regarding the deduplication part used in the data curation pipeline. Or how that can be done for a custom dataset.

+1.

@woctezuma
Copy link

woctezuma commented Apr 20, 2023

Would it be possible to share additional details regarding the deduplication part used in the data curation pipeline. Or how that can be done for a custom dataset.

It is probably something simple like idealo/imagededup.
Edit: the procedure is applied to embeddings as shown on Figure 3 rather than images, so it is a bit more complex.

In Figure 3, you can see that duplicates include:

  • an image which is identical to one of the curated dataset,
  • another image which has a different ratio but is otherwise nearly identical to another one of the uncurated dataset.

Dedpulication

Edit: You can find more information in the paper. It is a bit more complex as it is done on the embeddings.
However, I believe a method using image hashes is also used in a first deduplication process, as "PCA hash" is mentioned. 🤔

Paper

Appendix

@XiaohuJoshua
Copy link
Author

Thanks for your reply.

@salonit77
Copy link

@woctezuma Thank you for the reply.. I was able to follow the A3 section and get some deduplication results

@woctezuma woctezuma mentioned this issue Apr 21, 2023
@FrancescoSaverioZuppichini

@salonit77 hi :) could you share the code?

@patricklabatut patricklabatut changed the title Will the LVD-142M dataset or data processing codes be released? [request] LVD-142M pretraining dataset and / or data curation code Apr 24, 2023
@patricklabatut
Copy link
Contributor

patricklabatut commented Apr 24, 2023

Using #56 instead to keep track of requests about data curation code. Also happy to provide clarifications on the procedure.

@patricklabatut patricklabatut changed the title [request] LVD-142M pretraining dataset and / or data curation code Will the LVD-142M dataset or data processing codes be released? Apr 24, 2023
@yyyyyyfs
Copy link

depend

Thanks for your reply! I have a question about the deduplication method SSCD mentioned in your paper. I would like to confirm that SSCD is only used to extract embedding? Then use the Faiss library to calculate the similarity between embeddings?

@yyyyyyfs
Copy link

Would it be possible to share additional details regarding the deduplication part used in the data curation pipeline. Or how that can be done for a custom dataset.

It is probably something simple like idealo/imagededup. Edit: the procedure is applied to embeddings as shown on Figure 3 rather than images, so it is a bit more complex.

In Figure 3, you can see that duplicates include:

  • an image which is identical to one of the curated dataset,
  • another image which has a different ratio but is otherwise nearly identical to another one of the uncurated dataset.

Dedpulication

Edit: You can find more information in the paper. It is a bit more complex as it is done on the embeddings. However, I believe a method using image hashes is also used in a first deduplication process, as "PCA hash" is mentioned. 🤔

Paper

Appendix

Thanks for your reply~
About implementation details,I still have some questions。If I have 5B images,but I only have few gpus,how can I do this work? Reading a large amount of data into memory simultaneously is also a serious problem. Do you have any good suggestions?

@forever208
Copy link

forever208 commented Feb 20, 2024

Releasing LVD-142M is not something that we are considering releasing, I am afraid. Re: open-sourcing the data curation code, this could depend on feedback and interest from the community.

So, the editor said something non-existing in OpenReview :D
https://openreview.net/forum?id=a68SUt6zFt

@dvikdvik
Copy link

soo,, release it!! 🥇

@rfan-debug
Copy link

Would it be possible to share additional details regarding the deduplication part used in the data curation pipeline. Or how that can be done for a custom dataset.

It is probably something simple like idealo/imagededup. Edit: the procedure is applied to embeddings as shown on Figure 3 rather than images, so it is a bit more complex.
In Figure 3, you can see that duplicates include:

  • an image which is identical to one of the curated dataset,
  • another image which has a different ratio but is otherwise nearly identical to another one of the uncurated dataset.

Dedpulication
Edit: You can find more information in the paper. It is a bit more complex as it is done on the embeddings. However, I believe a method using image hashes is also used in a first deduplication process, as "PCA hash" is mentioned. 🤔
Paper
Appendix

Thanks for your reply~ About implementation details,I still have some questions。If I have 5B images,but I only have few gpus,how can I do this work? Reading a large amount of data into memory simultaneously is also a serious problem. Do you have any good suggestions?

@yyyyyyfs (Not the author..just want to put my two cents:) Processing 5B images with a few gpus would take years. Personally I'd store images on GCS buckets and try to apply for their free TPU units to process it. My personal experience is that a cloud provider like GCP usually optimize the connection between storage and computing units very well so you don't have to worry about speed of loading images into memory. You may just treat it like an embarrassingly data parallel process.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

9 participants