This is the repository for migrating ToMe to videoclip work. For more information about ToMe, please refer to ToMe_README.MD
.
- Clone CLIP repository
pip install git+https://github.com/openai/CLIP.git
- Requirements
Please refer to requirements.txt
- Data Preparation
We use HMDB51 video dataset and extract videos into frames for fast reading. You can refer to MVF for the detaied guide of data processing.
See examples/validation_videoclip.ipynb
This process does not require a trained model and loading a dataset. The principle is to :
(1) build the model,
(2) randomly generate tensors of the corresponding data size,
(3) input them into the model to calculate throughput.
bash script/run_train.sh configs/hmdb51/hmdb_few_shot.yaml
Since we are currently optimizing only for inference, you can freely train a model on the HMDB51 dataset (or use the checkpoint we will provide later). This won't take much of your time; training on 4 2080ti GPUs takes 1 hour and 10 minutes.
bash script/run_test.sh configs/hmdb51/hmdb_few_shot.yaml <PATH_TO_MODEL>