Your LLM will always make stuff up - blame Gödel, not the engineers.
According to this paper 🤔🤔
"LLMs Will Always Hallucinate, and We Need to Live With This"
📚 https://arxiv.org/abs/2409.05746v1
Key points from the paper. 👇
🧠 Hallucinations in LLMs not just mistakes, but inherent property. Arise from undecidable problems in training and usage process. Can't be fully eliminated through architectural improvements or data cleaning.
🔬 They use computational theory and Gödel's incompleteness theorems to explain hallucinations. Argue that LLM structure inherently leads to some inputs causing model to generate false or nonsensical information.
🚫 Complete elimination of hallucinations impossible due to undecidable problems in LLM foundations. No amount of tweaks or fact-checking can fully solve this issue. Fundamental limitation of current LLM approach.
------
🧮 Gödel's incompleteness theorems:
👉 First theorem: Any consistent formal system powerful enough to encode arithmetic contains statements that are true but unprovable within the system.
👉 Second theorem: Such a system cannot prove its own consistency.
------
Are you into AI and LLMs❓ Join me on Twitter with 43K+ others, to remain on the bleeding-edge every day.
𝕏/🐦 https://x.com/rohanpaul_ai
Share this post