LLaVA-MORE: A Comparative Study of LLMs and Visual Backbones for Enhanced Visual Instruction Tuning
-
Updated
Aug 8, 2025 - Python
LLaVA-MORE: A Comparative Study of LLMs and Visual Backbones for Enhanced Visual Instruction Tuning
The official repo of our work "Pensieve: Retrospect-then-Compare mitigates Visual Hallucination"
FaceXBench: Evaluating Multimodal LLMs on Face Understanding
The Official Code Repo for EgoOrientBench [CVPR25]
This repository is a comprehensive collection of open-source tools, code implementations, research notes, and practical tutorials for Autonomous Driving technologies. Key focuses include AI/ML (Reinforcement Learning, Transformers, and Multimodal LLMs with PyTorch), Robotics (ROS integration and simulation), and Theoretical Foundations.
Add a description, image, and links to the multimodal-llms topic page so that developers can more easily learn about it.
To associate your repository with the multimodal-llms topic, visit your repo's landing page and select "manage topics."