TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Quantized Llama models with increased speed and a reduced memory footprint

508 点作者 egnehots7 个月前

18 条评论

tveita7 个月前
So SpinQuant learns a rotation for activations and weights that, to my understanding, &quot;smear&quot; the outlier weights out so you don&#x27;t get extreme values in any one weight.<p>Random anecdote warning - In the old days, before vector search became AI and everyone and their dog offered a vector database, I had a task that required nearest neighbour search in a decent amount of high-dimensional vectors.<p>I tried quantizing them to bit vectors in an index and scanning through it to get an initial set of candidates. Performance was actually quite decent - reading through RAM linearly is fast! But the selectivity wasn&#x27;t great.<p>Somewhere along the way I found this paper[1] that iteratively finds a rotation to apply before quantization to reduce the quantization error. Very similar goal to SpinQuant, but focused on bit quantization only.<p>As it turns out the &#x27;random rotation&#x27; baseline they benchmark against worked great for my use case, so I never tried implementing the fancier algorithm. But it&#x27;s a pretty rare day at work that &quot;apply a random rotation matrix to a 128-dimensional vector&quot; is the solution to my problem.<p>[1] <a href="https:&#x2F;&#x2F;ieeexplore.ieee.org&#x2F;abstract&#x2F;document&#x2F;6296665" rel="nofollow">https:&#x2F;&#x2F;ieeexplore.ieee.org&#x2F;abstract&#x2F;document&#x2F;6296665</a> &#x2F; <a href="https:&#x2F;&#x2F;slazebni.cs.illinois.edu&#x2F;publications&#x2F;ITQ.pdf" rel="nofollow">https:&#x2F;&#x2F;slazebni.cs.illinois.edu&#x2F;publications&#x2F;ITQ.pdf</a>
评论 #41941325 未加载
评论 #41941741 未加载
评论 #41940862 未加载
评论 #41940931 未加载
评论 #41943580 未加载
评论 #41946734 未加载
nisten7 个月前
It&#x27;s pretty interesting that the new SpinQuant method did not manage to be better than good old nf4bit QLORA training (Tim Dettmers really cooked with that one).<p>Really appreciate that Meta published both results+model quants and didn&#x27;t just make some bs claim about a new sota quant like most other bigger companies would&#x27;ve done.
评论 #41943699 未加载
评论 #41940657 未加载
评论 #41942160 未加载
评论 #41944623 未加载
评论 #41942314 未加载
评论 #41941192 未加载
theanonymousone7 个月前
May I ask if anyone has successfully used 1B and 3B models in production and if yes, in what use cases? I seem to be failing even in seemingly simpler tasks such as word translation or zero-shot classification. For example, they seem to not care about instructions to only write a response and no explanation, thus making it impossible to use them in a pipeline :&#x2F;
评论 #41940038 未加载
评论 #41939668 未加载
评论 #41940202 未加载
评论 #41940328 未加载
评论 #41939835 未加载
评论 #42034912 未加载
评论 #41940125 未加载
评论 #41942328 未加载
评论 #41941603 未加载
评论 #41945839 未加载
评论 #41940410 未加载
formalsystem7 个月前
Hi I&#x27;m Mark I work on torchao which was used for the quantization aware training and ARM kernels in this blog. If you have any questions about quantization or performance more generally feel free to let me know!
评论 #41940616 未加载
评论 #41942068 未加载
评论 #41941963 未加载
philipkglass7 个月前
These quantized models show much less degradation compared to a &quot;vanilla post-training-quantization&quot; but there are a bunch of PTQ schemes that people have already applied to Llama models [1]. I didn&#x27;t see any details about the vanilla PTQ they used as a baseline. Has it been written about elsewhere?<p>[1] <a href="https:&#x2F;&#x2F;ollama.com&#x2F;library&#x2F;llama3.2&#x2F;tags">https:&#x2F;&#x2F;ollama.com&#x2F;library&#x2F;llama3.2&#x2F;tags</a>
yuvalr17 个月前
Looking at how to deploy 1B and 3B Llama models on Android for inference. Some posts online recommend using Termux (an amazing app) to have an emulated shell and then install as if it&#x27;s Linux, using ollama for example. However, this forces you into a manual installation process, and also most of the people don&#x27;t know what Termux is, and would be afraid to install it from F-Droid.<p>Maybe someone can recommend a way to deploy Llama to Android without Termux, maybe even something that can be potentially fully implemented inside an app?<p>I&#x27;m currently looking into compiling llama.cpp for Android and bundling it inside an app. Is that a viable path? Would love to hear from someone who tried something similar.
评论 #41946735 未加载
评论 #41969207 未加载
评论 #41946812 未加载
评论 #41947630 未加载
cmsj7 个月前
It really bugs me that every time I see posts about new models, there is never any indication of how much VRAM one needs to actually run them.
评论 #41943621 未加载
ed7 个月前
Oh cool! I’ve been playing with quantized llama 3B for the last week. (4-bit spinquant). The code for spinquant has been public for a bit.<p>It’s pretty adept at most natural language tasks (“summarize this”) and performance on iPhone is usable. It’s even decent at tool once you get the chat template right.<p>But it struggles with json and html syntax (correctly escaping characters), and isn’t great at planning, which makes it a bad fit for most agenetic uses.<p>My plan was to let llama communicate with more advanced AI’s, using natural language to offload tool use to them, but very quickly llama goes rogue and starts doing things you didn’t ask it to, like trying to delete data.<p>Still - the progress Meta has made here is incredible and it seems we’ll have capable on-device agents in the next generation or two.
评论 #41944409 未加载
Evidlo7 个月前
Why don&#x27;t they actually say what the size of the model is in GB?<p>That and average inference times on common hardware is what I&#x27;m curious about.
评论 #41942261 未加载
itsTyrion6 个月前
Wait, so I can get incorrect information and text summaries with things added or cut off even faster and on mobile now? that&#x27;s amazing.
nikolayasdf1237 个月前
what&#x27;s your opinion on LlamaStack?<p>for me it is nothing short of bad experience. it is way over-engineered with poor quality and just plain does not work, and maintainers are questionable. I would rather call HuggingFace python code for inference or anything else.<p>is ExecuTorch any better?
评论 #41947696 未加载
Tepix7 个月前
From TFA:<p>&gt; <i>At Connect 2024 last month, we open sourced Llama 3.2 1B and 3B</i><p>No you did not. There is no source (in this case: training data) included. Stop changing the meaning of &quot;open source&quot;, Meta!
justanotheratom7 个月前
Any pointers no how to finetune this on my dataset and package and run it in my swift ios app?
behnamoh7 个月前
Does anyone know why the most common method to speed up inference time is quantization? I keep hearing about all sorts of new methods but nearly none of them is implemented in practice (except for flash attention).
评论 #41943096 未加载
评论 #41940570 未加载
评论 #41940263 未加载
评论 #41941818 未加载
EliBullockPapa7 个月前
Anyone know a nice iOS app to run these locally?
评论 #41939828 未加载
评论 #41939702 未加载
评论 #41940483 未加载
评论 #41940878 未加载
评论 #41939884 未加载
arnaudsm7 个月前
How do they compare to their original quants on ollama like q4_K_S?
评论 #41939226 未加载
newfocogi7 个月前
TLDR: Quantized versions of Llama 3.2 1B and 3B models with &quot;competitive accuracy&quot; to the original versions (meaning some degraded performance; plots included in the release notes).
评论 #41939012 未加载
评论 #41944844 未加载
mmaunder7 个月前
[flagged]
评论 #41939813 未加载
评论 #41940765 未加载
评论 #41939784 未加载
评论 #41939844 未加载
评论 #41939969 未加载
评论 #41941956 未加载
评论 #41939827 未加载
评论 #41939824 未加载
评论 #41940325 未加载