Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to train my own data set? #5

Open
lingyixia opened this issue Nov 6, 2023 · 2 comments
Open

How to train my own data set? #5

lingyixia opened this issue Nov 6, 2023 · 2 comments

Comments

@lingyixia
Copy link

Could you please share a pipline for pre-preparing a new data for training?

@zjr2000
Copy link
Owner

zjr2000 commented Nov 8, 2023

  1. Prepare the annotation file:

    train_caption_file: training corpus, refer to this file
    val_caption_file: validation corpus, refer to this file
    eval_gt_file_for_grounding: validation file for video grounding, refer to this file
    dict_file: vocabulary file of your dataset, refer to this file

  2. Prepare the features: Gather each video's features into a .npy file, with the format L * D, where L denotes temporal resolution and D represents the feature dimension. Store these files in a single designated folder for streamlined access.

  3. Prepare the .yaml file: Create a configuration file for training by modifying the existing cfg file. You can start with the template provided at: Configuration File Template and adjust it using the annotation details mentioned above.

@hipeng-tech
Copy link

hi,thanks for your work, I have a question that in the train_capion_file ,what does the "area"stands for?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants