TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Andrej Karpathy on Hallucinations

20 点作者 sonabinu超过 1 年前

5 条评论

65a超过 1 年前
I really prefer confabulation, because there are a lot of parallels with the human brain (e.g. <a href="https:&#x2F;&#x2F;people.uncw.edu&#x2F;tothj&#x2F;PSY510&#x2F;Schnider-Confabulation-NRN-2003.pdf" rel="nofollow noreferrer">https:&#x2F;&#x2F;people.uncw.edu&#x2F;tothj&#x2F;PSY510&#x2F;Schnider-Confabulation-...</a>). Hallucination conceptually is often about sensing false inputs, whereas confabulation is about creating non-malicious false outputs.
alok-g超过 1 年前
I somehow do not think the &#x27;hallucination&#x27; (or confabulation) problem, as defined by common people should be theoretically hard to solve. It may need a lot more computations though.<p>We start with quasi-random initialization of network weights. The weights change under training based on actual data (assumed truthful), but some level of randomization is bound to remain in the network. There would be some weights which settle with low error margins, while some which would have wide error margins. Wider error margins are indicative or higher initial randomness remaining and lesser impact of the training data.<p>Now when a query comes for which the network has not seen enough baking from the training data, the network will keep producing tokens based on weights that have a larger margin. And then, as others have noted, once it picks a wrong token, it will likely keep on the erroneous path. We could, in theory, however, maintain metadata around how each weight changed, and use that to foresee how likely would the network confabulate.
评论 #38642372 未加载
29athrowaway超过 1 年前
If LLMs produced music<p><a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=ETTwP0mMmbk" rel="nofollow noreferrer">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=ETTwP0mMmbk</a>
ChatGTP超过 1 年前
<i>TLDR I know I&#x27;m being super pedantic but the LLM has no &quot;hallucination problem&quot;. Hallucination is not a bug, it is LLM&#x27;s greatest feature. The LLM Assistant has a hallucination problem, and we should fix it.</i><p>Sounds like a cop out.
评论 #38640544 未加载
zer0c00ler超过 1 年前
Agreed mostly, although I&#x27;m less of a fan of the &quot;dream&quot; terminology.<p>The key takeaway is that we can&#x27;t trust LLM output and systems need to be designed accordingly.<p>Eg Agents will never be able to securely take actions autonomously based on LLM responses.