Repository for learning pose-aware video representations for downstream tasks.
Code is stored in src/
and all runnable scripts are in this directory. This is an in-development repository and will contain multiple different experiments. The respective scripts for each type of experiment will be differentiated by the first word in the script. For example, all experiments related to masked autoencoders (mae) will be named mae_{script_name}.py
.
Bash scripts that contain settings for each of these scripts of kept in scripts/
. Some may be run python files that will submit SLURM jobs using submitit, some will call a script in scripts/launch_scripts
which will submit the SLURM job.
Experiments will create and store all outputs in a new directory experiments/
.
Prefix: tsfmr_
Experiment files: src/tsfmr
\
Code is written from scratch except for transformer model implementation which is borrowed from minGPT.
Predicts action labels given 3D pose information on NTURGB+D.
Prefix: mae_
Experiment files: src/mae
Code is adapted from D36893220 to run on AWS cluster and run without internal build tools.
Pretrains a representation used masked-autoencoders on video data that will then be used for action classification on the K400 dataset.
How to run:
bash scripts/mae_pretrain.sh
Prefix: pc3d_
Experiment files: src/pc3d
Ideas taken from the PoseConv3D paper. Currently, only a data processing procedure where bounding boxes are extracted from a video dataset.
Prefix: dmae_
Experiment files: src/dmae
All new files, models, experiments and utilities that enable directed masking for a masked autoencoder approach for video pretraining.
Packages are all provided in requirements.txt
.
torchvision
needs to be build from source withffmpeg
backend to properly decode video data forMasked autoencoder
data loading. Follow instructions here in the section regardingVideo Backend
.