#STEPs: Self-Supervised Key Step Extraction and Localization from Unlabeled Procedural Videos
Accepted at ICCV 2023
-
Installing the environment :
conda create -n steps python=3
conda activate steps
- Install packages:
- Install PyTorch from here
- Install additional packages using :
pip install tqdm pandas scikit-learn==0.24.2
-
Download data
-
Train a model
- Run
python main.py --train --random_seed $seed
to train a model. Results in the paper are reported as average over three seeds. - Trained model can be evaluated for Key Step Localization by running
python main.py --test --load_checkpoint saved_models/iccv2023_STEPs_Meccano/300.pth
- Run
-
Acknowledgements
- We use annotations for Meccano and Evaluation code from EgoProceL (Bansal et al.)
If you find this repository useful in your research, please cite:
@InProceedings{Shah_2023_ICCV,
author = {Shah, Anshul and Lundell, Benjamin and Sawhney, Harpreet and Chellappa, Rama},
title = {STEPs: Self-Supervised Key Step Extraction and Localization from Unlabeled Procedural Videos},
booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)},
month = {October},
year = {2023},
pages = {10375-10387}
}