Skip to content

OpenGVLab/Ask-Anything

Repository files navigation

🦜 Ask-Anything [Paper]

Open in OpenXLab | | | |
Open in Spaces [VideoChat-7B-8Bit] End2End ChatBOT for video and image.

[VideoChat2-7B]End2End ChatBOT for video and image.

δΈ­ζ–‡ README 及 中文亀桁羀 | Paper

πŸš€: We update video_chat by instruction tuning for video & image chatting now! Find its details here. We release instruction data at InternVideo. The old version of video_chat moved to video_chat_with_chatGPT.

⭐️: We are also working on a updated version, stay tuned!

english.mp4
intro.mp4

πŸ”₯ Updates

  • 2024/05/22: πŸ“’ We release VideoChat2_mistral, which shows better capacity on diverse tasks (60.4% on MVBench, 78.6% on NExT-QA, 63.8% on STAR, 46.4% on TVQA, 54.4% on EgoSchema-full and 80.5% on IntentQA). More details will be updated in the paper. Have a try! πŸƒπŸ»β€β™€οΈπŸƒπŸ»

  • 2024/04/05 MVBench is selected as Poster (Highlight)!

  • 2024/2/27 MVBench is accepted by CVPR2024.

  • 2023/11/29 VideoChat2 and MVBench are released.

  • 2023/05/11 End-to-end VideoChat and its technical report.

    • VideoChat: Instruction tuning for video chatting (also supports image one).
    • Paper: We present how we craft VideoChat with two versions (via text and embed) along with some discussions on its background, applications, and more.
  • 2023/04/25 Watch videos longer than one minute with chatGPT

  • 2023/04/21 Chat with MOSS

  • 2023/04/20: Chat with StableLM

  • 2023/04/19: Code release & Online Demo

πŸ”¨ Getting Started

Build video chat with:

πŸ“„ Citation

If you find this project useful in your research, please consider cite:

@article{2023videochat,
  title={VideoChat: Chat-Centric Video Understanding},
  author={Li, Kunchang and He, Yinan and Wang, Yi and Li, Yizhuo and Wang, Wenhai and Luo, Ping and Wang, Yali and Wang, Limin and Qiao, Yu},
  journal={arXiv preprint arXiv:2305.06355},
  year={2023}
}

⏳ Ongoing

Our team constantly studies general video understanding and long-term video reasoning:

  • Strong video foundation model.
  • Video-text dataset and video reasoning benchmark.
  • Video-language system with LLMs.
  • Artificial Intelligence Generated Content (AIGC) for Video.
  • ...

🌀️ Discussion Group

If you have any questions during the trial, running or deployment, feel free to join our WeChat group discussion! If you have any ideas or suggestions for the project, you are also welcome to join our WeChat group discussion!

image

We are hiring researchers, engineers and interns in General Vision Group, Shanghai AI Lab. If you are interested in working with us, please contact Yi Wang (wangyi@pjlab.org.cn).