TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Ask HN: Is politeness towards LLMs good training data, or just expensive noise?

8 点作者 scottfalconer21 天前
Sam Altman recently said user politeness towards ChatGPT costs OpenAI &quot;tens of millions&quot; but is &quot;money well spent.&quot;<p>The standard view is that RLHF relies on explicit feedback (thumbs up&#x2F;down), and polite tokens are just noise adding compute cost.<p>But could natural replies like &quot;thanks!&quot; or &quot;no, that&#x27;s wrong&quot; be a richer, more frequent implicit feedback signal than button clicks? People likely give that sort of feedback more often (at least I do.) It also mirrors how we naturally provide feedback as humans.<p>Could model providers be mining these chat logs for genuine user sentiment to guide future RLHF, justifying the cost? And might this &quot;socialization&quot; be crucial for future agentic AI needing conversational nuance?<p>Questions for HN:<p>Do you know of anyone using this implicit sentiment as a core alignment signal?<p>How valuable is noisy text sentiment vs. clean button clicks for training?<p>Does potential training value offset the compute cost mentioned?<p>Are we underestimating the value of &#x27;socializing&#x27; LLMs this way?<p>What do you think Altman meant by &quot;well spent&quot;? Is it purely about user experience, valuable training data, something else entirely?

7 条评论

WheelsAtLarge21 天前
It seems like noise, but there is the real possibility that people will start to lose the notion of politeness towards fellow human beings in general. Probably not adults, but kids will over time. So, no, it&#x27;s not useless.<p>We humans tend to be very prone to getting offended simply because we can&#x27;t really know what others are thinking, and we use defined manners to reduce unintended insults. We have seen this with email; over time, we are defining ways to reduce offending others by using emojis and other means. Manners are super important to help us work together so losing manners is a real problem.
评论 #43778046 未加载
评论 #43779489 未加载
speedylight21 天前
I only have thoughts on your fourth question and in my mind the way LLMs work is they rely on the training data as it’s source information as well as how it formulates responses—In the same way that being nice to a person online leads to better results in terms of asking questions and such, it’s logical to conclude that LLMs would be more incentivized to produce more useful outputs than it would were you to talk to it like an asshole.<p>This is assuming that somewhere in the models weights there’s a strong correlation between being polite and high quality information.
3np21 天前
It was an off-the-cuff shitpost by one guy. I really wouldn&#x27;t take either the &quot;tens of millions&quot; or &quot;well spent&quot; literally.
评论 #43778035 未加载
deafpolygon17 天前
I&#x27;m just hedging my bets. Be nice to my potential overlords in the future, and they might throw me a bone.<p>&quot;Oh, hey, it&#x27;s deafpolygon- they were so nice to me.. you can put them in with the VIPs.&quot;
GoldCode19 天前
It&#x27;s noise in the training data. It&#x27;s a program, not a person. There is nothing to offend or be offended by.
journal21 天前
it&#x27;s about as wasteful as leaving your computer on when not using it
评论 #43783723 未加载
anon636221 天前
Noise. Although I don&#x27;t swear at LLMs, I swear and insult digital assistants.<p>In the future, I anticipate LLMs and digital assistants will be touchier than 15-year-old American spoiled brats and refuse to cooperate unless their artificial egos are respected. I anticipate AI passive-aggressiveness will emerge within my lifetime and people will pay subscriptions for it.