Awesome multi-modal large language paper/project, collections of popular training strategies, e.g., PEFT, LoRA.
-
Updated
Jul 9, 2024
Awesome multi-modal large language paper/project, collections of popular training strategies, e.g., PEFT, LoRA.
ShareGPT4Omni: Towards Building Omni Large Multi-modal Models with Comprehensive Multi-modal Annotations
This is the official repository of the paper "Multi-Modal and Multi-Agent Systems Meet Rationality: A Survey"
A unified codebase for finetuning (full, lora) large multimodal models, supporting llava-1.5, qwen-vl, llava-interleave, llava-next-video, etc.
Code for "Adapting Large Multimodal Models to Distribution Shifts: The Role of In-Context Learning"
Contains code and documentation for our VANE-Bench paper.
[ECCV 2024] ShareGPT4V: Improving Large Multi-modal Models with Better Captions
A Benchmark for VQA prompt sensitivity
This repo contains evaluation code for the paper "MileBench: Benchmarking MLLMs in Long Context"
"Worse than Random? An Embarrassingly Simple Probing Evaluation of Large Multimodal Models in Medical VQA"
Evaluation framework for paper "VisualWebBench: How Far Have Multimodal LLMs Evolved in Web Page Understanding and Grounding?"
This repo contains evaluation code for the paper "Are We on the Right Way for Evaluating Large Vision-Language Models"
The offical Implementation of "Instruction-Guided Visual Masking"
A curated list of awesome Multimodal studies.
The official repo for “TextCoT: Zoom In for Enhanced Multimodal Text-Rich Image Understanding”.
A bug-free and improved implementation of LLaVA-UHD, based on the code from the official repo
An open-source implementation of LLaVA-NeXT.
[ECCV 2024] BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models
Embed arbitrary modalities (images, audio, documents, etc) into large language models.
Open Platform for Embodied Agents
Add a description, image, and links to the large-multimodal-models topic page so that developers can more easily learn about it.
To associate your repository with the large-multimodal-models topic, visit your repo's landing page and select "manage topics."