TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Meta LLM Compiler: neural optimizer and disassembler

248 pointsby foobazgt11 months ago

15 comments

HanClinto11 months ago
Huh. This is a very... &quot;interesting&quot; application for an LLM. I&#x27;m not the brightest crayon in the box, but if anyone else would like to follow along with my non-expert opinion as I read through the paper, here&#x27;s my take on it.<p>It&#x27;s pretty important for compilers &#x2F; decompilers to be reliable and accurate -- compilers behaving in a deterministic and predictable way is an important fundamental of pipelines.<p>LLMs are inherently unpredictable, and so using an LLM for compilation &#x2F; decompilation -- even an LLM that has 99.99% accuracy -- feels a bit odd to include as a piece in my build pipeline.<p>That said, let&#x27;s look at the paper and see what they did.<p>They essentially started with CodeLlama, and then went further to train the model on three tasks -- one primary, and two downstream.<p>The first task is compilation: given input code and a set of compiler flags, can we predict the output assembly? Given the inability to verify correctness without using a traditional compiler, this feels like it&#x27;s of limited use on its own. However, training a model on this as a primary task enables a couple of downstream tasks. Namely:<p>The second task (and first downstream task) is compiler flag prediction &#x2F; optimization to predict &#x2F; optimize for smaller assembly sizes. It&#x27;s a bit disappointing that they only seem to be able to optimize for assembly size (and not execution speed), but it&#x27;s not without its uses. Because the output of this task (compiler flags) are then passed to a deterministic function (a traditional compiler), then the instability of the LLM is mitigated.<p>The third task (second downstream task) is decompilation. This is not the first time that LLMs have been trained to do better decompilation -- however, because of the pretraining that they did on the primary task, they feel that this provides some advantages over previous approaches. Sadly, they only compare LLM Compiler to Code Llama and GPT-4 Turbo, and not against any other LLMs fine-tuned for the decompilation task, so it&#x27;s difficult to see in context how much better their approach is.<p>Regarding the verifiability of the disassembly approach, the authors note that there are issues regarding correctness. So the authors employ round-tripping -- recompiling the decompiled code (using the same compiler flags) to verify correctness &#x2F; exact-match. This still puts accuracy in the 45% or so (if I understand their output numbers), so it&#x27;s not entirely trustworthy yet, but it might be able to still be useful (especially if used alongside a traditional decompiler, and this model&#x27;s outputs only used when they are verifiably correct).<p>Overall I&#x27;m happy to see this model be released as it seems like an interesting use-case. I may need to read more, but at first blush I&#x27;m not immediately excited by the possibilities that this unlocks. Most of all, I would like to see it explored if these methods could be extended to optimize for performance -- not just size of assembly.
评论 #40828762 未加载
评论 #40822771 未加载
评论 #40825074 未加载
评论 #40828935 未加载
评论 #40824236 未加载
评论 #40830365 未加载
评论 #40826680 未加载
评论 #40826682 未加载
mmphosis11 months ago
<a href="https:&#x2F;&#x2F;xcancel.com&#x2F;AIatMeta&#x2F;status&#x2F;1806361623831171318" rel="nofollow">https:&#x2F;&#x2F;xcancel.com&#x2F;AIatMeta&#x2F;status&#x2F;1806361623831171318</a>
zellyn11 months ago
I continue to be fascinated about what the next qualitative iteration of models will be, marrying the language processing and broad knowledge of LLMs with an ability to reason rigorously.<p>If I understand correctly, this work (or the most obvious productionized version of it) is similar to the work Deep Mind released a while back: the LLM is essentially used for “intuition”—-to pick the approach—-and then you hand off to something mechanical&#x2F;rigorous.<p>I think we’re going to see a huge growth in that type of system. I still think it’s kind of weird and cool that our meat brains with spreading activation can (with some amount of effort&#x2F;concentration) switch over into math mode and manipulate symbols and inferences rigorously.
soist11 months ago
How do they verify the output preserves semantics of the input?
评论 #40831408 未加载
zitterbewegung11 months ago
Some previous work in the space is at <a href="https:&#x2F;&#x2F;github.com&#x2F;albertan017&#x2F;LLM4Decompile">https:&#x2F;&#x2F;github.com&#x2F;albertan017&#x2F;LLM4Decompile</a>
chad1n11 months ago
As usual, Twitter is impressed by this, but I&#x27;m very skeptical, the chance of it breaking your program is pretty high. The thing that makes optimizations so hard to make is that they have to match the behavior without optimizations (unless you have UBs), which is something that LLMs probably will struggle with since they can&#x27;t exactly understand the code and execution tree.
评论 #40824813 未加载
评论 #40824705 未加载
评论 #40824641 未加载
评论 #40823033 未加载
评论 #40824738 未加载
评论 #40828648 未加载
muglug11 months ago
Unlike many other AI-themed papers at Meta this one omits any mention of the model output getting used at Instagram, Facebook or Meta. Research is great! But doesn&#x27;t seem all that actionable today.
评论 #40825118 未加载
0x1ceb00da11 months ago
Wouldn&#x27;t &quot;Compiler LLM&quot; be a more accurate name than &quot;LLM Compiler&quot;?
评论 #40831480 未加载
hgl7196411 months ago
I am curious about CUDA assembly, does it work on CUDA -&gt; ptx level? or ptx -&gt; sass? I have done some work on SASS optimization and it would be a lot easier if LLM could be applied at SASS level
LoganDark11 months ago
Reading the title, I thought this was a tool for optimizing and disassembling LLMs, not an LLM designed to optimize and disassemble. Seeing it&#x27;s just a model is a little disappointing in comparison.
rldjbpin11 months ago
my knowledge of compilers don&#x27;t extend beyond a 101 course done ages ago, but i wonder how the researchers enriched the dataset for improving these features.<p>did they just happen to find a way to format the heuristics of major compilers in half-code, half-language mix? confusingly enough, another use case where a (potential) tool that let us veer into the solution with some work is being replaced by an llm.
Havoc11 months ago
I don’t understand the purpose of this. Feels like a task for function calling and sending it to an actual compiler.<p>Is there an obvious use case I’m missing?
评论 #40829194 未加载
评论 #40828736 未加载
jameshart11 months ago
Pretty sure I remember trading 300 creds for a Meta Technologies Neural Optimizer and Disassembler in one of the early <i>Deus Ex</i> games.
ldjkfkdsjnv11 months ago
I love this company. Advancing ai and keeping the rest of us in the loop.
评论 #40823308 未加载
评论 #40823943 未加载
评论 #40828286 未加载
nothrowaways11 months ago
It is so funny that meta has to post it on X.
评论 #40827726 未加载
评论 #40832738 未加载