(CVPR2024) Realigning Confidence with Temporal Saliency Information for Point-level Weakly-Supervised Temporal Action Localization
This is an official release of our paper "Realigning Confidence with Temporal Saliency Information for Point-level Weakly-Supervised Temporal Action Localization". [Paper]
Ziying Xia, Jian Cheng, Siyu Liu, Yongxiang Hu, Shiguang Wang, Yijie Zhang, Liwan Dang
Point-level weakly-supervised temporal action localization (P-TAL) aims to localize action instances in untrimmed videos through the use of single-point annotations in each instance. Existing methods predict the class activation sequences without any boundary information, and the unreliable sequences result in a significant misalignment between the quality of proposals and their corresponding confidence. In this paper, we surprisingly observe the most salient frame tend to appear in the central region of the each instance and is easily annotated by humans. Guided by the temporal saliency information, we present a novel proposal-level plug-in framework to relearn the aligned confidence of proposals generated by the base locators. The proposed approach consists of Center Score Learning (CSL) and Alignment-based Boundary Adaptation (ABA). In CSL, we design a novel center label generated by the point annotations for predicting aligned center scores. During inference, we first fuse the center scores with the predicted action probabilities to obtain the aligned confidence. ABA utilizes the both aligned confidence and IoU information to enhance localization completeness. Extensive experiments demonstrate the generalization and effectiveness of the proposed framework, showcasing state-of-the-art or competitive performances across three benchmarks.
Here are the necessary dependencies:
- CUDA: 11.3
- Python: 3.8.16
- Pytorch: 1.10.0
- Numpy: 1.20.3
- Pandas: 1.5.3
- Tqdm: 4.65.0
We use the extracted THUMOS14 features provided by LAC, or you can obtain the feature from GoogleDrive.
After getting the features, move the folder features
into ./data/THUMOS14/
.
The action proposal can be obtained by reproducing the P-TAL methods without any post-processing operations.
We have provided the proposals generated by LAC in ./data/THUMOS14
.
Run:
bash ./train_thumos.sh
Download the pre-trained model from GoogleDrive. Change the model path in ./test_thumos.sh
, and run:
bash ./test_thumos.sh
@InProceedings{Xia_2024_CVPR,
author = {Xia, Ziying and Cheng, Jian and Liu, Siyu and Hu, Yongxiang and Wang, Shiguang and Zhang, Yijie and Dang, Liwan},
title = {Realigning Confidence with Temporal Saliency Information for Point-Level Weakly-Supervised Temporal Action Localization},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2024},
pages = {18440-18450}
}
If you have any question please contact zyxia@std.uestc.edu.cn
This code contains modified codes from LAC and P-MIL.
Thanks for their excellent works and great contributions for the community!