TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Ask HN: ChatGPT doesn't ackowledge being wrong?

12 点作者 caldarons超过 2 年前
So far in all the cases I have seen of ChatGPT providing a wrong answer, I have never seen it actually acknowledge that the answer it provides might be innacurate. Am I the only one worried by this?<p>For all the talk about &quot;AI ethics&quot; there is it seems strinking to me that the current state of the art model will opt for providing a convincing argument for why it&#x27;s wrong answer is correct, rather than say that the answer it provides might be innacurate. (Funnily enough this is what humans tend to do aswell.) Being that these models often tend to be trained on data found on the internet, could this be a sign of the bias we tend to have when writing on social platforms and the internet in general? (As in justifying our answers instead of trying to get to the correct one)<p>So the questions are: 1. What are the consequences of this in the development of LLMs and in their application to various fields?<p>2. How would one implement this capability of recognizing where the model might be innacurate?<p>3. Is 2. really that much more complicated than what is currently being done? If not, then why hasn&#x27;t it been done yet?

11 条评论

Oxidation超过 2 年前
It always acknowledged it to me. Something like:<p>&gt; AI: Answer that&#x27;s clearly factually wrong (e.g. a function that doesn&#x27;t exist or a completely wrong numerical figure).<p>&gt; Me: that&#x27;s not right, the function doesn&#x27;t exist (etc.)<p>&gt; AI: you are right, that function doesn&#x27;t exist. The answer is blah<p>And repeat, since if it didn&#x27;t get it right first time, it seems unlikely to be able to get there at all (and you&#x27;d not know when it did unless you already know the answer).<p>It&#x27;s very like a specific person I know in a PR-marketing type job that will just glide across a noticed outright falsehood and instantly reshape what they&#x27;re saying in real time and carry on as if nothing happened, leaving you wondering if you&#x27;re taking crazy pills.
评论 #34375758 未加载
评论 #34374549 未加载
mavu超过 2 年前
This is why it should be outlawed to call every ML model we have today &quot;AI&quot;.<p>Those things are NOT artificial intelligence.<p>They are specific noise generators. They generate noise that is as similar as possible to noise it learned from, which match the input.<p>THats it. nothing more.
评论 #34374593 未加载
评论 #34374588 未加载
heavyset_go超过 2 年前
It&#x27;s given me blatantly wrong results for things that come up easily on Google, and it tells me those wrong things with confidence.<p>What&#x27;s worse is that someone who isn&#x27;t a domain expert might end up being convinced by the arguments and conviction the model provides.<p>Your second question is a good one, and I see that as a big problem with this generation of AI&#x2F;ML. You&#x27;re starting to scratch the surface of &quot;this requires understanding of the real world&quot; problems without models really understanding anything. It&#x27;s all statistics and correlations in data, the model is not capable of really understanding its input or output.<p>These models are like very impressive Markov chain text generators in that they can spit out pretty convincing answers that seem cogent, but there is no real comprehension going on with what they read and write. It&#x27;s just statistics.
seba_dos1超过 2 年前
It sure can acknowledge being wrong, but it&#x27;s not what it&#x27;s there for.<p>GPT is the ultimate cosplayer. It pretends to be who you want it to be. If you want it to answer a question, it will make up something that looks like a valid answer to that question. Sometimes it may actually be the right answer (after all, the right answer has a pretty good chance to look like the right answer), but ultimately being right is not the goal it&#x27;s trying to achieve - all it &quot;wants&quot; is to autocomplete your prompt in a plausible way.<p>Instead of making it try to come up with <i>the</i> answer, you can, for example, ask it (either explicitly or indirectly) to cosplay a scientist who is unsure of their position and tries to evaluate various options - and it will be happy to oblige.<p>What&#x27;s worrying is not whether the model &quot;acknowledges being wrong&quot; or not - it&#x27;s rather how it&#x27;s being marketed to people and, in turn, what people expect from it. We&#x27;ve got plenty of submissions here on HN with people being surprised that the model has made up exactly what they asked it for - for example [0], which shouldn&#x27;t happen if there wasn&#x27;t a dissonance between what people think it&#x27;s doing and what it&#x27;s actually doing.<p>[0] <a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=33841672" rel="nofollow">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=33841672</a>
kybernetikos超过 2 年前
I once triggered one of its stock responses about how it&#x27;s only a large language model and can&#x27;t answer some kinds of questions. Sometime later I asked if it had made any mistakes in our conversation. It said that it could have answered that earlier question better, so I asked it what it would have said if it had answered better, and it gave me a pretty good answer.
yetanotherloser超过 2 年前
The way it&#x27;s constructed is very impressive but it doesn&#x27;t (to the best of my knowledge and understanding) have the concept of &quot;this is correct and this is incorrect&quot; that you have - or any kind of analogy to it whatsoever. (NB this does not mean it has no internal analogy to a concept or to correctness. It&#x27;s just under no obligation to be like yours if it does.)<p>There seem to be a fair number of humans of whom I could say the same. As with the humans, the question is &quot;which jobs am I happy to see this one doing and which ones worry me&quot;.<p>As GPT has endless ability to produce flowing words and precisely zero concept-comparable-to-will to verify them, unfortunately, the job for which it is most apt is probably politics. That I must admit worries me a little because of all the gpt-like humans we tolerate and encourage.
alexfromapex超过 2 年前
It&#x27;s definitely learned that from being trained on data from real humans
irvingprime超过 2 年前
When ChatGPT gives me a wrong answer, and I point out that it is wrong, ChatGPT immediately apologizes, every time. I have not seen it argue that it&#x27;s actually right.<p>I&#x27;ve also seen it give different answers when asked the same question in a different way. Usually, one of the answers will be correct.<p>It has no idea if it&#x27;s answers are right or wrong. It only knows that it&#x27;s putting words in a common order.
joshka超过 2 年前
ChatGPT isn&#x27;t wrong.<p>Imagine putting two newspaper headings on the shelf at your local drugstore that happen to spell out an incorrect headline when combined. (Or put two browser windows together with a similar effect).<p>This is not wrong, your understanding of what it means is wrong. Sort your perspective out and you&#x27;re fine.<p>1. LLMs are working as expected.<p>2. <a href="https:&#x2F;&#x2F;beta.openai.com&#x2F;docs&#x2F;guides&#x2F;fine-tuning&#x2F;case-study-is-the-model-making-untrue-statements" rel="nofollow">https:&#x2F;&#x2F;beta.openai.com&#x2F;docs&#x2F;guides&#x2F;fine-tuning&#x2F;case-study-i...</a><p>3. More complicated - probably not. It mostly just needs more data with specific training on domain specific areas. Take any field and generate some wrong completions from the data available and some right completions and the output tends to get better. Train it in the opposite fashion and it gets worse.<p>The meat of this comes back to people see ChatGPT as changing its response as an indication of understanding, but it&#x27;s not. Users are just adding extra constraints on what parts of the language model the model uses to generate text. This is not the same thing.<p>An (perhaps poor) example of that might be where you&#x27;re deciding to buy a tool that might solve a problem that you have. Evaluating whether it does solve the problem you decide that it doesn&#x27;t. But then you evaluate whether you can afford the tool based on your budget and you decide you can, so you buy the tool. You haven&#x27;t changed your mind, you&#x27;ve just constrained the reasons that you&#x27;re using to evaluate the choice (electronic musicians often refer to this as GAS - Gear Acquisition Syndrome).<p>TL;DR: Stop Anthropomorphizing Language Models
PaulHoule超过 2 年前
&quot;Truth&quot; is the most problematic concepts in philosophy. The introduction of the concept of the &quot;Truth&quot; undermines truthfulness. (e.g. you can call something &quot;Truth Social&quot;)<p>This book<p><a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;G%C3%B6del,_Escher,_Bach" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;G%C3%B6del,_Escher,_Bach</a><p>has a set of parables about people trying to paint on a facility to a system very similar to a &quot;truth detector&quot; for GPT-3. The gist of it is that &quot;awareness of truth&quot; makes it possible to make statements like &quot;Am I lying now?&quot;<p>People under GPT-3&#x27;s spell think that giving correct answers is a minor detail that will be handled in a point revision of it but actually it is a much harder problem than everything they&#x27;ve done so far.
评论 #34374254 未加载
评论 #34374452 未加载
sgt101超过 2 年前
It does acknowledge it&#x27;s wrong, but it doesn&#x27;t really understand.<p><a href="https:&#x2F;&#x2F;medium.com&#x2F;p&#x2F;cf66b04f3b9f" rel="nofollow">https:&#x2F;&#x2F;medium.com&#x2F;p&#x2F;cf66b04f3b9f</a>