Links
Hallucination is inevitable
Researchers show that hallucination is inevitable:
LLMs cannot learn all of the computable functions and will therefore always hallucinate. Since the formal world is a part of the real world which is much more complicated, hallucinations are also inevitable for real world LLMs.
(From: [2401.11817] Hallucination is Inevitable: An Innate Limitation of Large Language Models)