打开 code/demo.py
, 配置好 DATASET_NAME
和 SALMAP_SHAPE
;
生成saliency maps:
cd code
python cli.py demo with input_video='---' 'input_format'='mp4' ckpt_name='paver_ckpt.pt'
移动生成好的saliency maps:
cd ../output/David_MMSys_18/
mv ./saliency_src/ /dataset/uniformed-vp-datasets/David_MMSys_18/
以下为原README.md内容:
Official repository of Panoramic Vision Transformer for Saliency Detection in 360 Videos (ECCV 2022)
- To create a virtual environment with all relevant libraries installed:
conda env create -f environment.yml
- Dataset will be stored under the
data
directory by default. First, create a symbolic link at your convenience:
ln -s /some/path/to/data data
- The
data
directory should be organized as follows. Since the Wild360 dataset is available upon request, you may ask for access here.
data
├─ wild360 # the dataset should be downloaded on your own
│ ├─ data/{test,train}
│ └─ Wild360_GT_29
├─ cache # cache features for fast experiments
└─ log # training log & weights
Provided the dataset and environment are all prepared, simply run the command below:
python cli.py
Default configurations can be found in config.py
and configs/*/*.json
. To override parameters at your disposal:
python cli.py with max_epoch=10
You can download our model weight here.
python cli.py demo with {some_necessary_parameters}
Adapting PAVER for ViT variants is quite simple. Provided the target ViT uses PatchEmbed
, all you need to do is to replace it with a deformable counterpart using compute_deform_offset
in geometry.py
.
# (...)
from geometry import compute_deform_offset
class DeformPatchEmbed(nn.Module):
def __init__(self, ...):
# (...)
self.proj = DeformConv2d(in_chans, embed_dim, kernel_size=patch_size, stride=patch_size)
offset = torch.from_numpy(compute_deform_offset(model_config=model_config,
is_discrete=is_discrete)).float()
self.register_buffer('offset', offset)
# (...)
class DeformViTVariant(ViT):
def __init__(self, ...):
# (...)
self.patch_embed = DeformPatchEmbed(...)
# (...)
Please refer to model/ViT.py
for detailed usage.
If you find our work useful in your research, please consider citing:
@InProceedings{
author = {Yun, Heeseung and Lee, Sehun and Kim, Gunhee},
title = {Panoramic Vision Transformer for Saliency Detection in 360 Videos},
booktitle = {ECCV},
year = {2022}
}