TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Differential Transformer

562 点作者 weirdcat7 个月前

31 条评论

Imnimo7 个月前
I feel like I&#x27;m missing a key insight here. I understand the problem that regular softmax attention struggles to approach assigning zero attention to irrelevant stuff. And I get that having this subtraction formula makes it possible to assign exactly (or near) zero attention weight without having crazy outlier activations. But it seems like it also makes it very easy to have negative attention weight (which is equivalent to having positive attention weight on the negation of your value vectors). Intuitively, it just feels like a difficult balancing act to keep all the stuff you don&#x27;t care about so close to zero.<p>But Figure 1 clearly shows that it works, so I don&#x27;t doubt that it is in fact possible. I&#x27;m just struggling to build a picture of how exactly the network accomplishes this.
评论 #41780844 未加载
评论 #41792020 未加载
评论 #41779968 未加载
评论 #41780900 未加载
评论 #41779779 未加载
aDyslecticCrow7 个月前
Very clever. I like this kind of nitty-gritty detail work, and the change is small enough to be adapted easily by others. Bravo!<p>I&#x27;m a little concerned about the last sentence of the section introduction of &quot;2 Differential Transformer&quot;. It mentions using improvements from previous papers, but in the grammatical context, it&#x27;s unclear if this improvement is added to both the normal transformer and their diff transformer. This would otherwise sully the comparisons. It&#x27;s the &quot;main difference&quot; wording in the previous sentence that raised a flag for me.<p>Of course, a good-faith researcher would know this and may not feel the need to clarify. But you can never be too careful about some published research in this field.
评论 #41780849 未加载
评论 #41778608 未加载
msoad7 个月前
Like most things in this new world of Machine Learning, I&#x27;m really confused why this works?<p>The analogy to noise-cancelling headphones is helpful but in that case we clearly know which is signal and which is noise. Here, if we knew why would we even bother to the noise-cancelling work?
评论 #41777492 未加载
评论 #41777981 未加载
评论 #41778514 未加载
评论 #41780581 未加载
评论 #41777095 未加载
评论 #41778030 未加载
评论 #41780220 未加载
评论 #41777464 未加载
评论 #41776928 未加载
islewis7 个月前
&gt; Differential attention takes the difference between two softmax attention functions to eliminate attention noise<p>If I understand correctly, this architecture trades twice as much attention memory in exchange for either a higher quality model, or less parameters at a similar quality.<p>&gt; According to the fitted curves, 6.8B-size DIFF Transformer achieves a validation loss comparable to 11B-size Transformer, requiring only 62.2% of parameters<p>This raises a few questions for me:<p>- Would having only 60% of the parameters negate the double space for attention, leaving a similar memory profile as a traditional transformer?<p>- Does that tradeoff change noticeably between training and inference?
评论 #41779508 未加载
评论 #41780050 未加载
评论 #41779423 未加载
评论 #41780064 未加载
WithinReason7 个月前
<i>We empirically find that the setting λᵢₙᵢₜ = 0.8 − 0.6 × exp(−0.3 · (l − 1)) works well in practice</i><p>I wonder about the story behind that formula...
评论 #41779303 未加载
评论 #41781032 未加载
评论 #41781407 未加载
iandanforth7 个月前
The key bit I didn&#x27;t understand at first was what happens if the two groups of attention learn the same thing; because their attention masks are subtracted from one another if they both output similar values the attention across the board will drop to zero and this will lead to high loss. So the only way to reduce loss is if they learn to attend to different things. One of the simplest strategies they could learn (and this paper claims that they do) is for one group to focus on relevant context and the other to focus on irrelevant context. Thus one group learns the noise and the other the signal (it&#x27;s not this cut and dry but is a useful simplification for understanding IMO).
评论 #41778252 未加载
评论 #41780196 未加载
评论 #41777939 未加载
评论 #41778284 未加载
patcon7 个月前
I wonder what is lost here. Surely there&#x27;s a trade-off...<p>I&#x27;m wondering if there&#x27;s any effect of &quot;creativity&quot;, or ability to interpolate between concepts. Hallucination and creativity feel very related to me. I understand hallucinating as simply being misaligned with the space humans feel appropriate to interpolate between
评论 #41778047 未加载
评论 #41777640 未加载
评论 #41777437 未加载
chessgecko7 个月前
I wonder how much of the value here is from canceling out the positional noise rope produces. I would love to see a table comparing an alibi version of this to an alibi baseline in addition to the rope models here.<p>Crazy gains though congrats to the researchers
vsroy7 个月前
Is the thing that&#x27;s going on here that softmax can&#x27;t push a value to 0, but by subtracting 2 softmax maps we can output 0s?
评论 #41781421 未加载
评论 #41778150 未加载
machinelearning7 个月前
This is a good problem to solve but the approach is wrong imo.<p>It has to be done in a hierarchical way to know what you attended to + full context.<p>If the differential vector is being computed with the same input as the attention vector how do you know how to modify the attention vector correctly
评论 #41780027 未加载
pxdm7 个月前
What&#x27;s the comparison with conventional attention using a more aggressive (lower temperature) softmax? I can imagine that for the multi-needle retrieval test this may also give a performance boost, although at some cost other more creative tasks.
评论 #41781277 未加载
nmacias7 个月前
AdderaLLM was <i>right there</i>
miven7 个月前
Is there an intuitive reason why this ends up working this well compared to, say, applying some kind of thresholding to attention activations that are below average for a given head to filter that same attention noise out?
pizza7 个月前
Was just going to mention that it seems that it should be possible to make a Flash Attention version of this algorithm and was pleasantly surprised to see they already included an implementation of one :)
watsonmusic7 个月前
The modification is simple and beautiful. And the improvements are quite significant.
singularity20017 个月前
Anyone remember siamese networks?
slashdave7 个月前
I don&#x27;t get it. Arbitrary linear combinations are already accommodated via feed forward. What am I missing?
评论 #41781616 未加载
评论 #41785363 未加载
WithinReason7 个月前
Hmmm, this could be expressed as 2 consecutive attentions in a residual branch:<p>Simplified differential T. looks like: (softmax(Q₁K₁) − λ softmax(Q₂K₂)) V<p>You can factor this into:<p><pre><code> x = softmax(Q₁K₁)V x += -λ softmax(Q₂K₂)V </code></pre> which is like 2 subsequent regular attentions added that are sharing V
评论 #41779130 未加载
评论 #41782952 未加载
h_tbob7 个月前
I wish they didn’t use swiGLU and preRMSnorm so we could have a better comparison.<p>Then we would know how much this transformer innovation helps by itself.
digdugdirk7 个月前
Is there any way to replicate this with existing models, or are we going to need to wait for models to be trained in this style?<p>I&#x27;m imagining a smaller model examining the output tokens of a larger model and metaphorically slapping it on the wrist with a ruler if the output tokens start drifting off topic. Not quite the same, but an entertaining thought nonetheless.
评论 #41777738 未加载
评论 #41777313 未加载
dartos7 个月前
&gt; By being less distracted by irrelevant context, Diff Transformer can mitigate hallucination in question answering and text summarization<p>I’m very interested in this claim. I was under the impression that hallucination is unavoidable in these kinds of models. IIRC proof for that was trending on HN a couple weeks ago.
评论 #41779570 未加载
评论 #41777977 未加载
评论 #41778016 未加载
mik097 个月前
r&#x2F;machine learning comment thread has some interesting ideas, one of them linking this one with similar work in CV: <a href="https:&#x2F;&#x2F;www.reddit.com&#x2F;r&#x2F;MachineLearning&#x2F;comments&#x2F;1g0lnij&#x2F;r_ngpt_normalized_transformer_with_representation&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.reddit.com&#x2F;r&#x2F;MachineLearning&#x2F;comments&#x2F;1g0lnij&#x2F;r_...</a>
lucidrains7 个月前
does this not mean we should explore usage of talking heads (Shazeer et al) a bit more? <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2003.02436" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2003.02436</a>
评论 #41779911 未加载
x49asvk7 个月前
This concept is really interesting to me, I am very very new to transformers but would love to learn more about normal transformers and differential too. Can anyone suggest any resources?
pikseladam7 个月前
Did this mean they solved the hallucination problem of transformers?<p>edit: not fully but it gives promising results. quiet an improvement actually.
评论 #41777894 未加载
评论 #41776873 未加载
评论 #41776887 未加载
nowayno5837 个月前
Does anyone understand why they are taking the difference between transformers instead of the sum? It seems to me that in a noise reducing solution we would be more interested in the sum, as random noise would cancel out and signal would be constructive.<p>Of course, even if I&#x27;m right proper training would account to that by inverting signs where appropriate. Still, it seems weird to present it as the difference, especially seeing as they compare this directly to noise cancelling headphones, where we sum both microphones inputs.
评论 #41778075 未加载
评论 #41778085 未加载
badsandwitch7 个月前
What is purpose of the lambda parameter? Why isn&#x27;t it a constant of 1?
esafak7 个月前
How is this different than using a sparsity-inducing prior?
magicalhippo7 个月前
<i>The visualization reveals that Transformer tends to allocate only a small proportion of attention scores to the correct answer, while disproportionately focusing on irrelevant context.</i><p><i>[...] Specifically, we partition the query and key vectors into two groups and compute two separate softmax attention maps. Then the result of subtracting these two maps is regarded as attention scores.</i><p><i>[...] The approach is analogous to noise-canceling headphones and differential amplifiers in electrical engineering, where the difference between two signals cancels out common-mode noise.</i><p>Simple change, with seemingly decent improvements across the board.
评论 #41776966 未加载
评论 #41776914 未加载
campers7 个月前
The tl;dr on high level performance improvements<p>&quot;The scaling curves indicate that Diff Transformer requires only about 65% of model size or training tokens needed by Transformer to achieve comparable language modeling performance.&quot;<p>&quot;Diff Transformer retains high performance even at reduced bit-widths, ranging from 16 bits to 6 bits. In comparison, Transformer’s accuracy significantly drops with 6-bit quantization. The 4-bit Diff Transformer achieves comparable accuracy as the 6-bit Transformer, and outperforms the 4-bit Transformer by about 25% in accuracy.&quot;
评论 #41776963 未加载
ExxKA7 个月前
Very interesting. Currently working on timeseries with Transformers. Let me know if anyone else out there is also reading it from that context.
评论 #41777069 未加载