TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Researchers tested AI watermarks and broke all of them

106 点作者 adg29超过 1 年前

13 条评论

jacobr1超过 1 年前
We need to focus on the other direction. How can we have chains of trust for content creation, such as for real video. Content can be faked, but not necessarily easily faked from the same sources that make use of cryptographic signing. The attacks can sign the own work, so you'd need ways to distinguish those cases, but device level keys, organizational keys, distribution keys all can provide provenance chains that can be used by downstream systems to _better_ detect fraud, though not eliminate it.
评论 #37796922 未加载
评论 #37796879 未加载
评论 #37797258 未加载
评论 #37797031 未加载
评论 #37797949 未加载
评论 #37797580 未加载
评论 #37798061 未加载
评论 #37796981 未加载
评论 #37798839 未加载
评论 #37796833 未加载
评论 #37798257 未加载
obblekk超过 1 年前
For written text, the problem may be even harder. Identifying the human author of text is a field called &quot;stylometry&quot; but this result shows that some simple transformations reduce the success to random chance [1].<p>Similarly, I suspect watermarking LLM output is probably unworkable. The output of a smart model could be de-watermarked by fine tuning a dumb open source model on the initial output, and then regenerating the original output token by token, selecting alternate words whenever multiple completions have close probabilities and semantically equivalent. It would be a bit tedious to perfectly dial in, but I suspect it could be done.<p>And then ultimately, short text selections can have a lot of meaning with very little entropy to uniquely tag (e.g., covfefe).<p>[1] <a href="https:&#x2F;&#x2F;dl.acm.org&#x2F;doi&#x2F;abs&#x2F;10.1145&#x2F;2382448.2382450" rel="nofollow noreferrer">https:&#x2F;&#x2F;dl.acm.org&#x2F;doi&#x2F;abs&#x2F;10.1145&#x2F;2382448.2382450</a><p>Curious if Scott Aaronson solved this challenge...
评论 #37797862 未加载
评论 #37796828 未加载
great_psy超过 1 年前
It seems it would be much easier to watermark non-ai images instead. Aka crypto signature.<p>That will be much harder to evade, but also pretty hard to implement.<p>I guess we will end up in the middle ground, where any non-signed image could be ai generate, but for most day to day use it’s ok.<p>If you want something to be deemed legit (gov press release, newspaper photo, etc) then just sign it. Very similar to what we do for web traffic (https)
brap超过 1 年前
People have been trying to watermark digital media for decades, when there was (still is) a very strong financial incentive to get it working. It never worked. I don’t think it ever will work.
评论 #37797303 未加载
评论 #37796586 未加载
评论 #37796477 未加载
epivosism超过 1 年前
Wasn&#x27;t this obvious from the get go that this can&#x27;t work?<p>If AI will eventually generate say 10k by 10k images, I can resize to 2.001k by 1.999k or similar, and I just don&#x27;t get how any subtle signal in the pixels can persist through that.<p>Maybe you could do something at the compositional level, but that seems restrictive to the output. Maybe something about like larger regions average color balance or something? But you wouldn&#x27;t be able to fit many bits in there, especially when you need to avoid triggering accidentally.<p>Also: here are some play money markets for whether this will work:<p><a href="https:&#x2F;&#x2F;manifold.markets&#x2F;Ernie&#x2F;midjourney-images-can-be-effectivel" rel="nofollow noreferrer">https:&#x2F;&#x2F;manifold.markets&#x2F;Ernie&#x2F;midjourney-images-can-be-effe...</a><p><a href="https:&#x2F;&#x2F;manifold.markets&#x2F;Ernie&#x2F;openai-images-have-a-useful-and-har" rel="nofollow noreferrer">https:&#x2F;&#x2F;manifold.markets&#x2F;Ernie&#x2F;openai-images-have-a-useful-a...</a>
评论 #37797945 未加载
评论 #37797109 未加载
KaiserPro超过 1 年前
We already have well established systems to prove the provenance of images and other sources.<p>At the moment the internet is a <i>wash</i> with bullshit images. Its imperative that news outlets are at a high enough standard to actually prove the provenance of them.<p>You don&#x27;t trust some bloke off facebook asserting that something is true, its the same for images.
998244353超过 1 年前
The actual paper seems to be <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2310.00076" rel="nofollow noreferrer">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2310.00076</a>.
skilled超过 1 年前
<a href="https:&#x2F;&#x2F;archive.ph&#x2F;1F0Ng" rel="nofollow noreferrer">https:&#x2F;&#x2F;archive.ph&#x2F;1F0Ng</a>
rakkhi超过 1 年前
It’s like captcha, highly annoying to users and authors, but if you don’t want to pay it works against low spend bots
评论 #37797856 未加载
whywhywhywhy超过 1 年前
I’ll never get over the “invisible_watermark” Python package being entirely visible to the naked eye, obviously degrades the image in an way that’s unacceptable and even easily spottable on any image once you know what it looks like.
natch超过 1 年前
Who was it, Eric Schmidt, who said we need to get over it, there is no privacy? I feel like we have the same energy here for authenticating human origin of content.
TestingTest5超过 1 年前
Was only a matter of time anyways...
bulla超过 1 年前
What happened to C2PA?