[CVPR'24] RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
-
Updated
Sep 11, 2024 - Python
[CVPR'24] RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
Add a description, image, and links to the rlhf-v topic page so that developers can more easily learn about it.
To associate your repository with the rlhf-v topic, visit your repo's landing page and select "manage topics."