Skip to content

anshulbshah/STEPs

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

#STEPs: Self-Supervised Key Step Extraction and Localization from Unlabeled Procedural Videos

Accepted at ICCV 2023

  • Installing the environment :

    • conda create -n steps python=3
    • conda activate steps
    • Install packages:
      • Install PyTorch from here
      • Install additional packages using : pip install tqdm pandas scikit-learn==0.24.2
  • Download data

    • Download features for RGB-ResNet50 and OF-RAFT from this location : Size : ~2.6G
    • Place them in $ROOT/Data/Meccano
    • Download annotations for Meccano dataset from EgoProceL and place them in $ROOT/Data/Meccano/annotations/
  • Train a model

    • Run python main.py --train --random_seed $seed to train a model. Results in the paper are reported as average over three seeds.
    • Trained model can be evaluated for Key Step Localization by running python main.py --test --load_checkpoint saved_models/iccv2023_STEPs_Meccano/300.pth
  • Acknowledgements

    • We use annotations for Meccano and Evaluation code from EgoProceL (Bansal et al.)

Citation

If you find this repository useful in your research, please cite:

@InProceedings{Shah_2023_ICCV,
    author    = {Shah, Anshul and Lundell, Benjamin and Sawhney, Harpreet and Chellappa, Rama},
    title     = {STEPs: Self-Supervised Key Step Extraction and Localization from Unlabeled Procedural Videos},
    booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)},
    month     = {October},
    year      = {2023},
    pages     = {10375-10387}
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages