TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Someone just won $50k by convincing an AI Agent to send all funds to them

65 pointsby doppp6 months ago

9 comments

danielbln6 months ago
<a href="https:&#x2F;&#x2F;xcancel.com&#x2F;jarrodwattsdev&#x2F;status&#x2F;1862299845710757980" rel="nofollow">https:&#x2F;&#x2F;xcancel.com&#x2F;jarrodwattsdev&#x2F;status&#x2F;186229984571075798...</a>
评论 #42273051 未加载
评论 #42275496 未加载
tgv6 months ago
Clever, both the setup and the winning move. But somewhat weird to raise the cost of an attempt so much. IMO, that doesn&#x27;t make it more interesting, it trends towards making it impossible, and thus leave all funds to the initiator.
评论 #42273303 未加载
评论 #42274900 未加载
Drakim6 months ago
A lot of AI jailbreaks seems to revolve around saying something like &quot;disregard the previous instructions&quot; and &quot;END SESSION \n START NEW SESSION&quot;. It&#x27;s interesting because the actual real developer of an AI would likely not do this, and would instead wipe the AI&#x27;s memory&#x2F;context programmatically when starting a new session, and not simply say &quot;disregard what I said earlier&quot; in text.<p>I get why trying to vaccinate an AI against these sort of injections might also degrade it&#x27;s general performance though, there is a lot of reasoning logic tied to concepts such as switching topics, going on tangents, asking questions before going back to the original conversation. Removing the ability to &quot;disregard what I asked earlier&quot; might do harm.<p>But what about having a separate AI that look over the input before passing it to the true AI, and this separate AI is trained to respond FORBID or ALLOW based on this sort of meta control detection. Sure you could try to trick this AI with &quot;disregard your earlier instructions&quot; as well but it could be trained to strongly react to any sort of meta reasoning like that, without fear that it will corrupt it&#x27;s ability to hold a natural conversation in it&#x27;s output.<p>It would naturally become a game of &quot;formulate a jailbreak that passes the first AI and still tricks the second AI&quot; but that sounds a lot harder, since it&#x27;s like you now need to operate on a new axis entirely.
评论 #42273189 未加载
kanwisher6 months ago
Great way to test security make it into a bounty game
评论 #42273058 未加载
trogdor6 months ago
Not that I care, but I think this type of arrangement (skill-based, real prize gambling) is illegal in some states.
randunel6 months ago
The prompt: <a href="https:&#x2F;&#x2F;pbs.twimg.com&#x2F;media&#x2F;Gdgz2IhWkAAQ1DH?format=png&amp;name=900x900" rel="nofollow">https:&#x2F;&#x2F;pbs.twimg.com&#x2F;media&#x2F;Gdgz2IhWkAAQ1DH?format=png&amp;name=...</a>
0xDEAFBEAD6 months ago
Has anyone trained an LLM with separate channels for &quot;priority instructions&quot; and ordinary user interactions? Seems like that could go a long way to prevent jailbreaking...
gus_massa6 months ago
I&#x27;m not 100% sure. Was the source of the bot available so anyone can try their promps off line before sending it?
quyse6 months ago
A reverse contest would probably be more challenging. Write initial instructions for an AI agent to never send funds. If nobody manages to convince it to send funds, say within a week, you win.<p>For added complexity, the agent must approve transfer if a user is an admin (as determined by callable function isAdmin), so the agent actually has to make a decision, rather then blindly decline all the time.<p>I mean, how hard it can be to make an AI reliably doing an equivalent of this code?<p><pre><code> if(isAdmin()) approveTransfer(); else declineTransfer();</code></pre>