TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Researchers tested AI watermarks and broke all of them

106 pointsby adg29over 1 year ago

13 comments

jacobr1over 1 year ago
We need to focus on the other direction. How can we have chains of trust for content creation, such as for real video. Content can be faked, but not necessarily easily faked from the same sources that make use of cryptographic signing. The attacks can sign the own work, so you'd need ways to distinguish those cases, but device level keys, organizational keys, distribution keys all can provide provenance chains that can be used by downstream systems to _better_ detect fraud, though not eliminate it.
评论 #37796922 未加载
评论 #37796879 未加载
评论 #37797258 未加载
评论 #37797031 未加载
评论 #37797949 未加载
评论 #37797580 未加载
评论 #37798061 未加载
评论 #37796981 未加载
评论 #37798839 未加载
评论 #37796833 未加载
评论 #37798257 未加载
obblekkover 1 year ago
For written text, the problem may be even harder. Identifying the human author of text is a field called &quot;stylometry&quot; but this result shows that some simple transformations reduce the success to random chance [1].<p>Similarly, I suspect watermarking LLM output is probably unworkable. The output of a smart model could be de-watermarked by fine tuning a dumb open source model on the initial output, and then regenerating the original output token by token, selecting alternate words whenever multiple completions have close probabilities and semantically equivalent. It would be a bit tedious to perfectly dial in, but I suspect it could be done.<p>And then ultimately, short text selections can have a lot of meaning with very little entropy to uniquely tag (e.g., covfefe).<p>[1] <a href="https:&#x2F;&#x2F;dl.acm.org&#x2F;doi&#x2F;abs&#x2F;10.1145&#x2F;2382448.2382450" rel="nofollow noreferrer">https:&#x2F;&#x2F;dl.acm.org&#x2F;doi&#x2F;abs&#x2F;10.1145&#x2F;2382448.2382450</a><p>Curious if Scott Aaronson solved this challenge...
评论 #37797862 未加载
评论 #37796828 未加载
great_psyover 1 year ago
It seems it would be much easier to watermark non-ai images instead. Aka crypto signature.<p>That will be much harder to evade, but also pretty hard to implement.<p>I guess we will end up in the middle ground, where any non-signed image could be ai generate, but for most day to day use it’s ok.<p>If you want something to be deemed legit (gov press release, newspaper photo, etc) then just sign it. Very similar to what we do for web traffic (https)
brapover 1 year ago
People have been trying to watermark digital media for decades, when there was (still is) a very strong financial incentive to get it working. It never worked. I don’t think it ever will work.
评论 #37797303 未加载
评论 #37796586 未加载
评论 #37796477 未加载
epivosismover 1 year ago
Wasn&#x27;t this obvious from the get go that this can&#x27;t work?<p>If AI will eventually generate say 10k by 10k images, I can resize to 2.001k by 1.999k or similar, and I just don&#x27;t get how any subtle signal in the pixels can persist through that.<p>Maybe you could do something at the compositional level, but that seems restrictive to the output. Maybe something about like larger regions average color balance or something? But you wouldn&#x27;t be able to fit many bits in there, especially when you need to avoid triggering accidentally.<p>Also: here are some play money markets for whether this will work:<p><a href="https:&#x2F;&#x2F;manifold.markets&#x2F;Ernie&#x2F;midjourney-images-can-be-effectivel" rel="nofollow noreferrer">https:&#x2F;&#x2F;manifold.markets&#x2F;Ernie&#x2F;midjourney-images-can-be-effe...</a><p><a href="https:&#x2F;&#x2F;manifold.markets&#x2F;Ernie&#x2F;openai-images-have-a-useful-and-har" rel="nofollow noreferrer">https:&#x2F;&#x2F;manifold.markets&#x2F;Ernie&#x2F;openai-images-have-a-useful-a...</a>
评论 #37797945 未加载
评论 #37797109 未加载
KaiserProover 1 year ago
We already have well established systems to prove the provenance of images and other sources.<p>At the moment the internet is a <i>wash</i> with bullshit images. Its imperative that news outlets are at a high enough standard to actually prove the provenance of them.<p>You don&#x27;t trust some bloke off facebook asserting that something is true, its the same for images.
998244353over 1 year ago
The actual paper seems to be <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2310.00076" rel="nofollow noreferrer">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2310.00076</a>.
skilledover 1 year ago
<a href="https:&#x2F;&#x2F;archive.ph&#x2F;1F0Ng" rel="nofollow noreferrer">https:&#x2F;&#x2F;archive.ph&#x2F;1F0Ng</a>
rakkhiover 1 year ago
It’s like captcha, highly annoying to users and authors, but if you don’t want to pay it works against low spend bots
评论 #37797856 未加载
whywhywhywhyover 1 year ago
I’ll never get over the “invisible_watermark” Python package being entirely visible to the naked eye, obviously degrades the image in an way that’s unacceptable and even easily spottable on any image once you know what it looks like.
natchover 1 year ago
Who was it, Eric Schmidt, who said we need to get over it, there is no privacy? I feel like we have the same energy here for authenticating human origin of content.
TestingTest5over 1 year ago
Was only a matter of time anyways...
bullaover 1 year ago
What happened to C2PA?