Does anyone really believe that having a good corpus will remove hallucinations?<p>Is this article even written by a person? Hard to know; they have a real blog with real article, but stuff like this reads strangely. Maybe it's just not a native english speaker?<p>> Hallucinations are certainly the toughest nut to crack and their negative impact is basically only slightly lessened by good confidence estimates and reliable citations (sources).<p>> The impact of contradictions in the training data.<p>(was this a prompt header you forget to remove?)<p>> LLM are incapable of "self-inspection" on their training data to find logical inconsistencies in it but in the input context window they should be able to find logical inconsistencies.<p>Annnnyway...<p>Hallucinations cannot be fixed by a good corpus in a non-deterministic (ie. temp > 0) LLM system where you've introduced a random factor.<p>Period. QED. If you think it can, do more reading.<p>The idea that a good corpus can <i>significantly improve</i> the error rate is an open question, but the research I've seen <i>tends</i> to fall on the side of "to some degree, but curating a 'perfect' dataset like that, of a sufficiently large size, is basically impossible'".<p>So, it's a pipe dream.<p>Yes, if you could have a perfect corpus, absolutely, you would get a better model.<p>...but <i>how</i> do you plan to <i>get</i> that perfect corpus of training data?<p>If it was that easy, the people spending <i>millions and millions of dollars</i> making LLMs would have, I guess, probably come up with a solution for it. They're not stupid. If you could easily do it, it would already have been done.<p>my $0.02:<p>This is a dead end of research, because it's impossible.<p>Using LLMs which are finetuned to evaluate the output of <i>other</i> LLMs and using multi-sample / voting to reduce the incidence of halluciations that make it past the API barrier is both actively used and far, far more effective.<p>(ie. it doesn't matter if your LLM hallucinates 1 time in 10; if you can reliably <i>detect</i> that 1 instance, sample again, and return a non hallucination).<p>Other solutions... I'm skeptical; most of the ones I've seen haven't worked when you actually try to use them.