TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Ask HN: When did OpenAI first observe emergent reasoning and generalization?

2 点作者 semessier大约 1 年前
I am wondering when the seminal moment occurred that signs of reasoning and generalization transpired when scaling the parameters and training data in transformer-based models. Presumably it happened at OpenAI, but that's just an assumption. It would be fascinating to learn more on the moment of discovery and if there are any technical write-ups or blog posts detailing the -- fair to say -- historical process and findings. Does anybody have insights or pointers to share on this topic?

2 条评论

cmcollier大约 1 年前
In terms of &quot;the moment&quot;, I would imagine it happened during development inside Google (Lamda) or OpenAI (GPT2&#x2F;3).<p>More technically, here&#x27;s one of the key papers discussing the topic (from google):<p>* <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2206.07682" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2206.07682</a><p>Emergent Abilities of Large Language Models<p>Scaling up language models has been shown to predictably improve performance and sample efficiency on a wide range of downstream tasks. This paper instead discusses an unpredictable phenomenon that we refer to as emergent abilities of large language models. We consider an ability to be emergent if it is not present in smaller models but is present in larger models. Thus, emergent abilities cannot be predicted simply by extrapolating the performance of smaller models. The existence of such emergence implies that additional scaling could further expand the range of capabilities of language models.<p>Version history (for relevant dates):<p><pre><code> [v1] Wed, 15 Jun 2022 17:32:01 UTC (59 KB) [v2] Wed, 26 Oct 2022 05:06:24 UTC (88 KB)</code></pre>
illuminant大约 1 年前
Emergence literally means to create more potential than the mere sum of parts.<p>By this technical definition, the moment it did something useful which the parts themselves could not do was an emergent moment.<p>Remember the very first release? GPT scared OpenAI so much they didn&#x27;t release it to the public until after pressure from competing open source alternatives began to surface.<p>Maybe you want to check the basic algorithms? When where they first demonstrating non gibberish?<p>Most changes since have been refinements.