Skip to content

haochenheheda/LVVIS

Repository files navigation

LV-VIS: Large-Vocabulary Video Instance Segmentation dataset

📄[arXiv]📄[ICCV(Oral)]🔥[Dataset Download]🔥[Evaluation Server]

This repo is the official implementation of Towards Open Vocabulary Video Instance Segmentation (ICCV2023 oral)

News

We are working on the final revision of the annotations. The Codalab set will be released in November.

Haochen Wang1, Cilin Yan2, Shuai Wang 1, Xiaolong Jiang 3, Xu Tang3, Yao Hu3, Weidi Xie 4,Efstratios Gavves 1

1University of Amsterdam, 2Beihang University, 3Xiaohongshu Inc, 4 Shanghai Jiao Tong University.

LV-VIS dataset

LV-VIS is a dataset/benchmark for Open-Vocabulary Video Instance Segmentation. It contains a total of 4,828 videos with pixel-level segmentation masks for 26,099 objects from 1,196 unique categories.

Demo Demo Demo Demo Demo Demo Demo Demo Demo Demo Demo Demo Demo Demo Demo Demo

Dataset Download

Videos Annotations Annotations (oracle) Submission Example
Training Download Download - -
Validation Download Download Download -
Test Download - - Download

Dataset Structure

## JPEGImages

|- train
  |- 00000
    |- 00000.jpg
    |- 00001.jpg
       ...
  |- 00001
    |- 00000.jpg
    |- 00001.jpg
       ...
    ...
|- val
    ...
|- test
    ...

## Annotations
train_instances.json
val_instances.json
image_val_instances.json

The annotation files have the same formation as Youtube-VIS 2019.

We used this platform for the annotation of LV-VIS. This platform is a smart video segmentation annotation tool based on Lableme, SAM, and STCN. See segment-anything-annotator.

We provide our baseline OV2Seg code for LV-VIS. Please check Baseline.md for more details.

TODO

  • Training and inference code of OV2Seg
  • Leaderboard for Val/test set

NOTE:

  • We haven't decided to release the annotation file for the test set yet. Please be patient.
  • The training set is not exhaustively annotated.
  • If you find mistakes in the annotations, please contact us (h.wang3@uva.nl). We will update the annotations.

Cite

@inproceedings{wang2023towards,
  title={Towards Open-Vocabulary Video Instance Segmentation},
  author={Wang, Haochen and Yan, Cilin and Wang, Shuai and Jiang, Xiaolong and Tang, XU and Hu, Yao and Xie, Weidi and Gavves, Efstratios},
  booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
  year={2023}
}

Acknowledgement

This repo is built based on Mask2Former and Detic, thanks for those excellent projects.

About

Large-Vocabulary Video Instance Segmentation dataset

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages