Skip to content

Banishing LLM Hallucinations Requires Rethinking Generalization

Notifications You must be signed in to change notification settings

lamini-ai/Lamini-Memory-Tuning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 

Repository files navigation

Banishing LLM Hallucinations Requires Rethinking Generalization

Johnny Li, Saksham Consul, Eda Zhou, James Wong, Naila Farooqui, Nithyashree Manohar, Zhuxiaona (Nina) Wei, Tian Wu, Ben Echols, Sharon Zhou, and Gregory Diamos

Despite their powerful chat, coding, and reasoning abilities, Large Language Models (LLMs) frequently hallucinate. Conventional wisdom suggests that hallucinations are a consequence of a balance between creativity and factuality, which can be mitigated, but not eliminated, by grounding the LLM in external knowledge sources. Through extensive systematic experiments, we show that these traditional approaches fail to explain why LLMs hallucinate in practice. Specifically, we show that LLMs augmented with a mixture of Millions of Memory Experts (MoME) can easily memorize large datasets of random numbers. We corroborate these experimental findings with a theoretical construction showing that simple neural networks trained to predict the next token hallucinate when the training loss is above a threshold as it usually does in practice when training on internet scale data. We interpret our findings by comparing against traditional retrieval methods for mitigating hallucinations. We use our findings to design a first generation model for removing hallucinations - Lamini-1 - that stores facts in a massive mixture of millions of memory experts that are retrieved dynamically.

About

Banishing LLM Hallucinations Requires Rethinking Generalization

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published