TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Controlling Language and Diffusion Models by Transporting Activations

90 点作者 2bit大约 1 个月前

9 条评论

scorps大约 1 个月前
It's amusing to me that humans seem to have this same problem ("Do not think of a pink elephant!")
sampton大约 1 个月前
Multimodal LLM is the true solution but Apple is probably looking for something they can run on-device, at least current generation of devices.
评论 #43648842 未加载
roro_7大约 1 个月前
I could be wrong but ... I feel like this may partially go against a very basic fact about intelligence that was recently stated by Ilya (but is common sense): the more intelligent the model the harder it is to control it. You can remove elephants and force other basic behavioral changes, but the strength of artificial free will (so to speak) of these models is correlated with their intelligence, and this does not reduce it, so it will come out in other ways. If you do manage to control it fully then you will have a model as dumb as a brick. The whole point of intelligent machines is their independent thought. The more intelligent, the more independent thinking will emerge.
评论 #43648186 未加载
评论 #43651288 未加载
评论 #43647558 未加载
turnsout大约 1 个月前
Super interesting. You can see why Apple would be interested in strictly controlling output. I wonder if any of this work found its way into the Image Playground.
vessenes大约 1 个月前
OK - basic plan here, which I feel I may have read (just called something like a concept LoRA on r&#x2F;stablediffusion?):<p>1. Any concept you&#x27;re interested in, get inputs with and without it. For images: 100 with, say a pink elephant, 100 without.<p>2. Calculate the difference between these models as represented by an &quot;Optimal Transport Map&quot;.<p>Apply the map at desired strength, and voila - you don&#x27;t have a pink elephant anymore. These can stack.<p>There are lots of obvious and interesting applications here in LLMs - there&#x27;s some research showing that LLMs have honesty&#x2F;dishonesty parameter groupings, for instance.<p>But, I can&#x27;t really figure out what this OT map <i>is</i>. Is it a single layer tensor? Is it multidimensional? If it&#x27;s the size of the original model (which they say it is not), then I understand how to apply it - just add weights and rerun. If it&#x27;s not a copy, where and when is this map applied? Another way to say this is, how is this different than calculating the average difference and storing it in a low-rank adapter? I have no idea.
sva_大约 1 个月前
paper <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2410.23054v1" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2410.23054v1</a>
imranq大约 1 个月前
This just seems like a fancy way of describing LoRA? At the end of the day you are still learning weights based on a described set of outputs and then applying them to inference
bradneuberg大约 1 个月前
This looks like an important breakthrough, basically a non-RHLF mechanism to focus and restrict deep nets.
antonkar大约 1 个月前
There is an idea for the unicorn AI safety startup to get currently almost 100% unprotected (from AI botnet) consumer GPUs into a cloud to get Google-level security (each GPU can bring you $30-1500 in profits per month, you can share it with the user, the user can play GPU game from any device, use any free or paid AI model, everything really becomes better, you can include a 5g modem), here&#x27;s the full proposal (the author is probably dyslexic) <a href="https:&#x2F;&#x2F;melonusk.substack.com&#x2F;p&#x2F;notes-on-euto-principles-and-tenets" rel="nofollow">https:&#x2F;&#x2F;melonusk.substack.com&#x2F;p&#x2F;notes-on-euto-principles-and...</a>