TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Ask HN: What can we learn about human cognition from the performance of LLMs

11 点作者 abrax3141将近 2 年前
What can we learn about human cognition from the performance of LLMs<p>Some hypotheses (adapted from other posts):<p>* We have learned that Spreading Activation, when applied through high-dimensional non-symbolic network (the network formed by embedding vectors) may be able to account for abstraction in fluent language.<p>* We have learned that &quot;fluent reasoning&quot; (sometimes called &quot;inline&quot; or &quot;online&quot; reasoning), that is, the shallow reasoning embedded in fluent language, may be more powerful than usually thought.<p>* We have learned that &quot;talking to yourself&quot; (externally, in the case of GPTs, and potentially also internally in the case of human&#x27;s &quot;hearing yourself think&quot;) is able to successfully maintain enough short-term context to track naturally long chains of argument (via contextually-guided fluent reasoning, as above).<p>* We have learned that to some extent powerful &quot;mental models&quot; that support (again, at least fluent) reasoning can be in effect (functionally) represented and used in a highly distributed system.<p>* We have learned that meta-reasoning (which the LLMs do not do) may be important in augmenting fluent reasoning, and in tracking extended &quot;trains of thought&quot; (and thus extended dialogues).<p>* We have a new model of confabulation that fits into the fluent language model as implemented by LLMs.<p>* We have learned that people&#x27;s &quot;knowledge space&quot; is quite amazing, given that they have ~10x current LLM parameter size (~10T, where as an individual has potentially ~100T cortical parameters -- depending on what you count, of course) but a given individual only encodes a small number of languages and a small number of domains to any great depth (in addition to the standard operating procedures that almost all people encode). [That is, vs. the LLM encoding the whole damned internet in ~10 different languages.]<p>What else? (And, of course, it goes w&#x2F;o saying that you&#x27;ll argue about the above :-)

2 条评论

tlb将近 2 年前
Transformer-based LLMs define a theory of time: each token representation has added to it a vector full of sincos(wt) for a set of frequencies w, after which order is ignored. (Each sincos defines 2 elements of the vector: sin(wt) and cos(wt). Use e^iwt if you prefer to think in complex numbers.)<p>So in &quot;Your heart is stronger than your head&quot;, heart and head are 5 words apart, or ~8 tokens. So one gets sincos(w(t+0)), the other gets sincos(w(t+8)). That&#x27;s the only thing that distinguishes it from the converse sentence, &quot;Your head is stronger than your heart.&quot;<p>Chomsky had a much more symbolic theory of grammar. The fact that Chat GPT can answer questions about the above sentences (try them!) with order only defined by relative timestamps is remarkable.<p>Interestingly, if you throw in some extra words like &quot;Bob&#x27;s head is stronger (and more cromulent) than his heart&quot; it fails to answer questions about which is stronger. Possibly because the extra tokens bring the sincos terms it had learned to use for &quot;A is Xer than B&quot; statements wrapped all the way around the circle.<p>It&#x27;d be interesting to devise similar tests for people, to see what extraneous parentheticals can confuse them.
评论 #36099393 未加载
ted_bunny将近 2 年前
Surfing Uncertainty makes a case that the whole damn brain seems to work on a predictive model. I found it convincing, but I&#x27;m very much a layman so that&#x27;s all I can say with half confidence.