{"title":"法学硕士总会产生幻觉,我们需要接受这一点","authors":"Sourav Banerjee, Ayushi Agarwal, Saloni Singla","doi":"arxiv-2409.05746","DOIUrl":null,"url":null,"abstract":"As Large Language Models become more ubiquitous across domains, it becomes\nimportant to examine their inherent limitations critically. This work argues\nthat hallucinations in language models are not just occasional errors but an\ninevitable feature of these systems. We demonstrate that hallucinations stem\nfrom the fundamental mathematical and logical structure of LLMs. It is,\ntherefore, impossible to eliminate them through architectural improvements,\ndataset enhancements, or fact-checking mechanisms. Our analysis draws on\ncomputational theory and Godel's First Incompleteness Theorem, which references\nthe undecidability of problems like the Halting, Emptiness, and Acceptance\nProblems. We demonstrate that every stage of the LLM process-from training data\ncompilation to fact retrieval, intent classification, and text generation-will\nhave a non-zero probability of producing hallucinations. This work introduces\nthe concept of Structural Hallucination as an intrinsic nature of these\nsystems. By establishing the mathematical certainty of hallucinations, we\nchallenge the prevailing notion that they can be fully mitigated.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":"32 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"LLMs Will Always Hallucinate, and We Need to Live With This\",\"authors\":\"Sourav Banerjee, Ayushi Agarwal, Saloni Singla\",\"doi\":\"arxiv-2409.05746\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"As Large Language Models become more ubiquitous across domains, it becomes\\nimportant to examine their inherent limitations critically. This work argues\\nthat hallucinations in language models are not just occasional errors but an\\ninevitable feature of these systems. We demonstrate that hallucinations stem\\nfrom the fundamental mathematical and logical structure of LLMs. It is,\\ntherefore, impossible to eliminate them through architectural improvements,\\ndataset enhancements, or fact-checking mechanisms. Our analysis draws on\\ncomputational theory and Godel's First Incompleteness Theorem, which references\\nthe undecidability of problems like the Halting, Emptiness, and Acceptance\\nProblems. We demonstrate that every stage of the LLM process-from training data\\ncompilation to fact retrieval, intent classification, and text generation-will\\nhave a non-zero probability of producing hallucinations. This work introduces\\nthe concept of Structural Hallucination as an intrinsic nature of these\\nsystems. By establishing the mathematical certainty of hallucinations, we\\nchallenge the prevailing notion that they can be fully mitigated.\",\"PeriodicalId\":501340,\"journal\":{\"name\":\"arXiv - STAT - Machine Learning\",\"volume\":\"32 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - STAT - Machine Learning\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.05746\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Machine Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.05746","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
LLMs Will Always Hallucinate, and We Need to Live With This
As Large Language Models become more ubiquitous across domains, it becomes
important to examine their inherent limitations critically. This work argues
that hallucinations in language models are not just occasional errors but an
inevitable feature of these systems. We demonstrate that hallucinations stem
from the fundamental mathematical and logical structure of LLMs. It is,
therefore, impossible to eliminate them through architectural improvements,
dataset enhancements, or fact-checking mechanisms. Our analysis draws on
computational theory and Godel's First Incompleteness Theorem, which references
the undecidability of problems like the Halting, Emptiness, and Acceptance
Problems. We demonstrate that every stage of the LLM process-from training data
compilation to fact retrieval, intent classification, and text generation-will
have a non-zero probability of producing hallucinations. This work introduces
the concept of Structural Hallucination as an intrinsic nature of these
systems. By establishing the mathematical certainty of hallucinations, we
challenge the prevailing notion that they can be fully mitigated.