TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

A demo of GPT-3's ability to understand long instructions

161 点作者 monort超过 2 年前

13 条评论

goodside超过 2 年前
Be sure to read the thread, in particular: <a href="https:&#x2F;&#x2F;twitter.com&#x2F;goodside&#x2F;status&#x2F;1557926101615366144?s=21&amp;t=6tyUdwtpHbH6TjRHYXgsIA" rel="nofollow">https:&#x2F;&#x2F;twitter.com&#x2F;goodside&#x2F;status&#x2F;1557926101615366144?s=21...</a><p>&gt; A caveat to all of these: I use GPT-3 a lot, so I know the “golden path” of tasks it can do reliably. Had I asked it to write a sentence backwards or sum a list of numbers, it would fail every time. These are all softball questions in isolation.<p>I haven’t shown that GPT-3 can handle all coherent directions of this length, or even most directions that an untrained person would think to create. It’s just a demo that, if GPT-3 happens to be capable of your tasks separately, length per se is not a major issue.
评论 #32540974 未加载
shdon超过 2 年前
Seems like there is one instruction it didn&#x27;t follow: The first task mentions the usernames should be <i>exactly</i> like in the list, yet the AI responds with &quot;firebob&quot; (as in the comment) rather than &quot;FireBob1990&quot; (as in the list)<p>Funnily enough, that is exactly the kind of thing a human might do, as we too are terrible at following instructions precisely.
评论 #32536593 未加载
评论 #32539219 未加载
v4dok超过 2 年前
I don&#x27;t think modern big language models are conscious, mainly because they fail in absurd ways. But TBH, they don&#x27;t need to. This &quot;golden path&quot; deployed properly etc could easily automate a lot of jobs tomorrow.
评论 #32536974 未加载
评论 #32538554 未加载
评论 #32536563 未加载
benreesman超过 2 年前
My initial instinct was that this has to be getting some nudges from whatever human-in-the-loop is going on at OpenAI.<p>But then I realized that somewhere on the Internet there inevitably is a message board where people play the &quot;find me some shit on the internet&quot; game, and there&#x27;s some rabid subculture around it with zillions upon zillions of of examples, and it&#x27;s in the Bing index, and all the nudging it would need is to emphasize that sort of thing in the corpus.<p>Very impressive stuff.
评论 #32536841 未加载
russellbeattie超过 2 年前
Wow, that was really impressive. I thought I had a clear idea of what GPT-3 could do, but I had underestimated by a lot. Even if the results weren&#x27;t accurate, which they mostly seem to be, it&#x27;s still doing an amazing job of following complex instructions. Better than most people I would guess<p>Makes me double down on my prediction a week or so ago* of a Mid-Level AI Knowledge Work Apocalypse. In the next decade, AIs like this are going to do to office work what robotic mechanization did to the manufacturing sector.<p>1. <a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=32395193" rel="nofollow">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=32395193</a>
评论 #32538960 未加载
评论 #32538563 未加载
评论 #32539435 未加载
seaucre超过 2 年前
It didn&#x27;t correctly identify that FireBob1990&#x27;s name was misspelled as &quot;firebob&quot; in the original comment.
评论 #32536793 未加载
OJFord超过 2 年前
DALL·E I can see obvious use for, GPT tends to be similarly impressive, but I don&#x27;t understand if it&#x27;s &#x27;just&#x27; interesting research, seeing what we can do sort of thing, or whether people actually see real-world use cases for it?<p>The closest to it was perhaps that code-generating demo here a day or two ago - but who wants to be a &#x27;GPT programmer&#x27; writing code as &#x27;write a Python program that computes fizzbuzz replacing the arguments $fizz$ and $buzz$, ...&#x27; instead of just the &#x27;actual&#x27; code? It just seems like a more clever AppleScript to me, pseudocode, and I don&#x27;t think anybody&#x27;s ever seriously pursued a flexible keyword pseudocode like language as a goal, it&#x27;s just appeared as a demo of more general models?<p>Generating template&#x2F;outline text I suppose? (Like that essay-writing helper here a few days ago.)
评论 #32539427 未加载
评论 #32542565 未加载
评论 #32539534 未加载
Titan2189超过 2 年前
Uhh. How is that even possible? I thought I had a basic understanding of Neural Networks and inuput-, hidden- and output layers and those things. So how can it possibly backreference to it&#x27;s own previous answers and then follow another prompt based on this? Mind = Blown
评论 #32536924 未加载
评论 #32536501 未加载
评论 #32536517 未加载
评论 #32536631 未加载
mach1ne超过 2 年前
While impressive, it doesn&#x27;t imply that GPT would have any significant &#x27;task memory&#x27;. Remember that it always predicts the next token or word - as such, it essentially recognizes whether the next &#x27;task&#x27; in the list has already been written, and if so, it writes the next task.<p>It might be interesting to see how well it is able to modify the first output given some aspect of the final tasks.
chucky超过 2 年前
Now I&#x27;m curious if it can handle the classic reading comprehension assignment I&#x27;ve been given multiple times in my life. You know, the one that goes something like this:<p>1. Read through all steps carefully.<p>2. Do X<p>3. Do Y<p>(...)<p>99. As you have now read through the instructions, simply put your name in the top right corner of the first page.
评论 #32540243 未加载
anigbrowl超过 2 年前
Interesting results goodside.<p>Is it able to extract any kind of structural information? For example, you pass it the text of a movie script or children&#x27;s story (where the descriptive language is simple) and it returns a structured summary of the content?
评论 #32536723 未加载
etaioinshrdlu超过 2 年前
Is GPT-3 being regularly updated?
评论 #32536392 未加载
masswerk超过 2 年前
This should be really &quot;react to&quot; or &quot;answer to&quot;, instead of &quot;understand&quot;. These are not the same.<p>Edit: Anthropomorphizing algorithms and pattern stores doesn&#x27;t really help understanding. Instead, it&#x27;s apt to spread misunderstanding. Remember how long it took to purge the popular idea of &quot;electronic brains&quot; actually thinking, and to establish that these were restricted to executing what&#x27;s actually in code? We don&#x27;t need to start another level of this with &quot;AI&quot;. (Understanding is closely related to self-awareness and consciousness, and this is dangerous ground of misunderstanding when it comes to AI. As we&#x27;ve seen, even staff of pioneering companies, like Google, is prone to fall for this.)
评论 #32547149 未加载