Skip to content
This repository has been archived by the owner on Feb 16, 2022. It is now read-only.

Could you please provide access to the required data files? #1

Closed
yangapku opened this issue Jan 17, 2020 · 10 comments
Closed

Could you please provide access to the required data files? #1

yangapku opened this issue Jan 17, 2020 · 10 comments

Comments

@yangapku
Copy link

Hi! Thank you for releasing this great project! However, I notice that the data files (including the lmdb feature files as well as other metadata) needed to run pre-training and multi-task fine-tuning is not accessible. Could you please add accessible links to them? Or a readme guiding how to generate them is also fine. Thank you very much!

@vedanuj
Copy link
Contributor

vedanuj commented Jan 21, 2020

Hi @yangapku .. I have added a PR that includes files to do feature extraction and converting the extracted features to a LMDB file that can be used to train.

Unfortunately we cannot make public the feature files at this point of time. It should be easy to extract features and use them after the Readme is updated. Stay tuned and monitor the PR for updates.

@yangapku
Copy link
Author

Thank you for your reply! I will try to use the script to generate the feature files.

@yangapku
Copy link
Author

Hi, could you please provide more details about the arguments for running the extract_features.py script, such as "num_features", "feature_name", "confidence_threshold" and "background"? Is using the default parameters the appropriate way? Thank you!

@vedanuj
Copy link
Contributor

vedanuj commented Jan 24, 2020

Yes using the default arguments should work for this project.

@yangapku
Copy link
Author

Thanks! Do you mean that the default arguments work for preprocessing both the Conceptual Captions and the down-streaming datasets (COCO, VCR, etc.)? Meanwhile, I noticed that the arguments are different from the original ViL-BERT, like the increase in the number of boxes to 100 and the decrease in the confidence threshold to 0. Will that be okay?

@vedanuj
Copy link
Contributor

vedanuj commented Jan 30, 2020

For Conceptual Captions you can use number of boxes 36 and for downstream tasks it can be 100

@yangapku
Copy link
Author

Thank you! May I ask another question? For VCR and RefCOCO, it's needed to generate features based on the given ground truth bounding boxes. In the original ViL-BERT, there is a generate_tsv_gt.py script for this case. How do I achieve this using the new feature extractor script?

@vedanuj
Copy link
Contributor

vedanuj commented Jan 31, 2020

Thanks for asking this. We will add that script as well.

@yangapku
Copy link
Author

yangapku commented Feb 6, 2020

Hi, @vedanuj . May I ask is there any progress on including the script we've discussed before? Thank you!

@vedanuj
Copy link
Contributor

vedanuj commented Feb 7, 2020

@yangapku The scripts are added. Please check the readme in data directory. Let us know if you face any problems running the script.

@vedanuj vedanuj closed this as completed Feb 19, 2020
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants