TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Andrej Karpathy on Hallucinations

20 pointsby sonabinuover 1 year ago

5 comments

65aover 1 year ago
I really prefer confabulation, because there are a lot of parallels with the human brain (e.g. <a href="https:&#x2F;&#x2F;people.uncw.edu&#x2F;tothj&#x2F;PSY510&#x2F;Schnider-Confabulation-NRN-2003.pdf" rel="nofollow noreferrer">https:&#x2F;&#x2F;people.uncw.edu&#x2F;tothj&#x2F;PSY510&#x2F;Schnider-Confabulation-...</a>). Hallucination conceptually is often about sensing false inputs, whereas confabulation is about creating non-malicious false outputs.
alok-gover 1 year ago
I somehow do not think the &#x27;hallucination&#x27; (or confabulation) problem, as defined by common people should be theoretically hard to solve. It may need a lot more computations though.<p>We start with quasi-random initialization of network weights. The weights change under training based on actual data (assumed truthful), but some level of randomization is bound to remain in the network. There would be some weights which settle with low error margins, while some which would have wide error margins. Wider error margins are indicative or higher initial randomness remaining and lesser impact of the training data.<p>Now when a query comes for which the network has not seen enough baking from the training data, the network will keep producing tokens based on weights that have a larger margin. And then, as others have noted, once it picks a wrong token, it will likely keep on the erroneous path. We could, in theory, however, maintain metadata around how each weight changed, and use that to foresee how likely would the network confabulate.
评论 #38642372 未加载
29athrowawayover 1 year ago
If LLMs produced music<p><a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=ETTwP0mMmbk" rel="nofollow noreferrer">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=ETTwP0mMmbk</a>
ChatGTPover 1 year ago
<i>TLDR I know I&#x27;m being super pedantic but the LLM has no &quot;hallucination problem&quot;. Hallucination is not a bug, it is LLM&#x27;s greatest feature. The LLM Assistant has a hallucination problem, and we should fix it.</i><p>Sounds like a cop out.
评论 #38640544 未加载
zer0c00lerover 1 year ago
Agreed mostly, although I&#x27;m less of a fan of the &quot;dream&quot; terminology.<p>The key takeaway is that we can&#x27;t trust LLM output and systems need to be designed accordingly.<p>Eg Agents will never be able to securely take actions autonomously based on LLM responses.