This is the official code implementation of our CVPR 2026 paper "Rethinking Token Reduction for Large Vision-Language Models". We propose a novel learning-based, prompt-agnostic token compression method tailored for Large Vision-Language Models (LVLMs) in multi-turn Visual Question Answering (MT-VQA) scenarios.
⌛️ Code Release Update: The code implementation is currently being organized and will be released as soon as possible.