Can I Trust Your Answer? Visually Grounded Video Question Answering (CVPR'24, Highlight)
-
Updated
Jul 1, 2024 - Python
Can I Trust Your Answer? Visually Grounded Video Question Answering (CVPR'24, Highlight)
Data and PyTorch code for the LifeQA LREC 2020 paper.
Contrastive Video Question Answering via Video Graph Transformer (IEEE T-PAMI'23)
NExT-QA: Next Phase of Question-Answering to Explaining Temporal Actions (CVPR'21)
[ICCV 2023 CLVL Workshop] Zero-Shot and Few-Shot Video Question Answering with Multi-Modal Prompts
[NeurIPS 2022] Zero-Shot Video Question Answering via Frozen Bidirectional Language Models
NExT-QA: Next Phase of Question-Answering to Explaining Temporal Actions (CVPR'21)
[IEEE T-PAMI 2023] Cross-Modal Causal Relational Reasoning for Event-Level Visual Question Answering
Video Graph Transformer for Video Question Answering (ECCV'22)
This repo contains code for Invariant Grounding for Video Question Answering
Video as Conditional Graph Hierarchy for Multi-Granular Question Answering (AAAI'22, Oral)
An VideoQA dataset based on the videos from ActivityNet
Implementation for the paper "Hierarchical Conditional Relation Networks for Video Question Answering" (Le et al., CVPR 2020, Oral)
Add a description, image, and links to the videoqa topic page so that developers can more easily learn about it.
To associate your repository with the videoqa topic, visit your repo's landing page and select "manage topics."