A toolbox for benchmarking trustworthiness of multimodal large language models (MultiTrust, NeurIPS 2024 Track Datasets and Benchmarks)
-
Updated
Mar 26, 2025 - Python
A toolbox for benchmarking trustworthiness of multimodal large language models (MultiTrust, NeurIPS 2024 Track Datasets and Benchmarks)
Code for ACL 2024 paper "TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space"
[ICML'2024] Can AI Assistants Know What They Don't Know?
A toolbox for benchmarking Multimodal LLM Agents trustworthiness across truthfulness, controllability, safety and privacy dimensions through 34 interactive tasks
Improving LLM truthfulness via reporting confidence
Google Chrome Extension to add LLM text reading in a browser
Add a description, image, and links to the truthfulness topic page so that developers can more easily learn about it.
To associate your repository with the truthfulness topic, visit your repo's landing page and select "manage topics."