LLMs Will Always Hallucinate, and We Need to Live With This
This talk draws from the paper “LLMs Will Always Hallucinate, and We Need to Live With This” and presents a critical analysis of hallucinations in large language models (LLMs), arguing that these phenomena are not occasional errors but inevitable byproducts of the models’ underlying mathematical and logical structures. Leveraging insights from computational theory, including Gödel’s First Incompleteness Theorem and undecidability results like the Halting, Emptiness, and Acceptance Problems, the talk will demonstrate that hallucinations arise at every stage of the LLM process—from training data compilation to fact retrieval and text generation.
By introducing the concept of Structural Hallucination, we assert that hallucinations cannot be entirely eliminated through architectural improvements, dataset refinement, or fact-checking mechanisms. Instead, this work challenges the prevailing belief that LLM hallucinations can be fully mitigated, proposing instead that we must adapt to and manage their inevitability as a structural characteristic of these systems.
Ayushi Agarwal
Head of Data Science & Analytics at United We Care
Ayushi is an innovative AI visionary, creator of Stella 2.0 first’s first real-time AI therapist. She has mentored over 500, over 10 patents, and over 13 research papers in AI and mental health, specializing in large language models (LLMs), and is dedicated to ethical, inclusive technology solutions.