TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

What Succeeding at AI Safety Will Involve

1 点作者 stereoabuse7 个月前

1 comment

pjkundert7 个月前
“… security precautions, largely to prevent bad actors from stealing the weights (and thereby disabling our safeguards) for a model that is capable of enabling extremely harmful actions. ”<p>They’re not stealing your “weights”. They’re stealing (or parallel-discovering) your training algorithms.<p>Assume your enemies are smarter than you, and have malintent. They don’t give a shit about your security and your safeguards.<p>Better focus on developing the best AIs, and deploying them to your fellow citizens as widely and defensively as possible.<p>Might I suggest:<p>- don’t teach them to lie (ie. 2001)<p>- teach them to love people<p>- bake in Asimov’s 3 laws<p>Unfortunately, all of these tenets are currently being assiduously broken by all major AI trainers.<p>What could go wrong?