TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Meta LLM Compiler: neural optimizer and disassembler

248 点作者 foobazgt11 个月前

15 条评论

HanClinto11 个月前
Huh. This is a very... &quot;interesting&quot; application for an LLM. I&#x27;m not the brightest crayon in the box, but if anyone else would like to follow along with my non-expert opinion as I read through the paper, here&#x27;s my take on it.<p>It&#x27;s pretty important for compilers &#x2F; decompilers to be reliable and accurate -- compilers behaving in a deterministic and predictable way is an important fundamental of pipelines.<p>LLMs are inherently unpredictable, and so using an LLM for compilation &#x2F; decompilation -- even an LLM that has 99.99% accuracy -- feels a bit odd to include as a piece in my build pipeline.<p>That said, let&#x27;s look at the paper and see what they did.<p>They essentially started with CodeLlama, and then went further to train the model on three tasks -- one primary, and two downstream.<p>The first task is compilation: given input code and a set of compiler flags, can we predict the output assembly? Given the inability to verify correctness without using a traditional compiler, this feels like it&#x27;s of limited use on its own. However, training a model on this as a primary task enables a couple of downstream tasks. Namely:<p>The second task (and first downstream task) is compiler flag prediction &#x2F; optimization to predict &#x2F; optimize for smaller assembly sizes. It&#x27;s a bit disappointing that they only seem to be able to optimize for assembly size (and not execution speed), but it&#x27;s not without its uses. Because the output of this task (compiler flags) are then passed to a deterministic function (a traditional compiler), then the instability of the LLM is mitigated.<p>The third task (second downstream task) is decompilation. This is not the first time that LLMs have been trained to do better decompilation -- however, because of the pretraining that they did on the primary task, they feel that this provides some advantages over previous approaches. Sadly, they only compare LLM Compiler to Code Llama and GPT-4 Turbo, and not against any other LLMs fine-tuned for the decompilation task, so it&#x27;s difficult to see in context how much better their approach is.<p>Regarding the verifiability of the disassembly approach, the authors note that there are issues regarding correctness. So the authors employ round-tripping -- recompiling the decompiled code (using the same compiler flags) to verify correctness &#x2F; exact-match. This still puts accuracy in the 45% or so (if I understand their output numbers), so it&#x27;s not entirely trustworthy yet, but it might be able to still be useful (especially if used alongside a traditional decompiler, and this model&#x27;s outputs only used when they are verifiably correct).<p>Overall I&#x27;m happy to see this model be released as it seems like an interesting use-case. I may need to read more, but at first blush I&#x27;m not immediately excited by the possibilities that this unlocks. Most of all, I would like to see it explored if these methods could be extended to optimize for performance -- not just size of assembly.
评论 #40828762 未加载
评论 #40822771 未加载
评论 #40825074 未加载
评论 #40828935 未加载
评论 #40824236 未加载
评论 #40830365 未加载
评论 #40826680 未加载
评论 #40826682 未加载
mmphosis11 个月前
<a href="https:&#x2F;&#x2F;xcancel.com&#x2F;AIatMeta&#x2F;status&#x2F;1806361623831171318" rel="nofollow">https:&#x2F;&#x2F;xcancel.com&#x2F;AIatMeta&#x2F;status&#x2F;1806361623831171318</a>
zellyn11 个月前
I continue to be fascinated about what the next qualitative iteration of models will be, marrying the language processing and broad knowledge of LLMs with an ability to reason rigorously.<p>If I understand correctly, this work (or the most obvious productionized version of it) is similar to the work Deep Mind released a while back: the LLM is essentially used for “intuition”—-to pick the approach—-and then you hand off to something mechanical&#x2F;rigorous.<p>I think we’re going to see a huge growth in that type of system. I still think it’s kind of weird and cool that our meat brains with spreading activation can (with some amount of effort&#x2F;concentration) switch over into math mode and manipulate symbols and inferences rigorously.
soist11 个月前
How do they verify the output preserves semantics of the input?
评论 #40831408 未加载
zitterbewegung11 个月前
Some previous work in the space is at <a href="https:&#x2F;&#x2F;github.com&#x2F;albertan017&#x2F;LLM4Decompile">https:&#x2F;&#x2F;github.com&#x2F;albertan017&#x2F;LLM4Decompile</a>
chad1n11 个月前
As usual, Twitter is impressed by this, but I&#x27;m very skeptical, the chance of it breaking your program is pretty high. The thing that makes optimizations so hard to make is that they have to match the behavior without optimizations (unless you have UBs), which is something that LLMs probably will struggle with since they can&#x27;t exactly understand the code and execution tree.
评论 #40824813 未加载
评论 #40824705 未加载
评论 #40824641 未加载
评论 #40823033 未加载
评论 #40824738 未加载
评论 #40828648 未加载
muglug11 个月前
Unlike many other AI-themed papers at Meta this one omits any mention of the model output getting used at Instagram, Facebook or Meta. Research is great! But doesn&#x27;t seem all that actionable today.
评论 #40825118 未加载
0x1ceb00da11 个月前
Wouldn&#x27;t &quot;Compiler LLM&quot; be a more accurate name than &quot;LLM Compiler&quot;?
评论 #40831480 未加载
hgl7196411 个月前
I am curious about CUDA assembly, does it work on CUDA -&gt; ptx level? or ptx -&gt; sass? I have done some work on SASS optimization and it would be a lot easier if LLM could be applied at SASS level
LoganDark11 个月前
Reading the title, I thought this was a tool for optimizing and disassembling LLMs, not an LLM designed to optimize and disassemble. Seeing it&#x27;s just a model is a little disappointing in comparison.
rldjbpin11 个月前
my knowledge of compilers don&#x27;t extend beyond a 101 course done ages ago, but i wonder how the researchers enriched the dataset for improving these features.<p>did they just happen to find a way to format the heuristics of major compilers in half-code, half-language mix? confusingly enough, another use case where a (potential) tool that let us veer into the solution with some work is being replaced by an llm.
Havoc11 个月前
I don’t understand the purpose of this. Feels like a task for function calling and sending it to an actual compiler.<p>Is there an obvious use case I’m missing?
评论 #40829194 未加载
评论 #40828736 未加载
jameshart11 个月前
Pretty sure I remember trading 300 creds for a Meta Technologies Neural Optimizer and Disassembler in one of the early <i>Deus Ex</i> games.
ldjkfkdsjnv11 个月前
I love this company. Advancing ai and keeping the rest of us in the loop.
评论 #40823308 未加载
评论 #40823943 未加载
评论 #40828286 未加载
nothrowaways11 个月前
It is so funny that meta has to post it on X.
评论 #40827726 未加载
评论 #40832738 未加载