TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Show HN: Alpaca.cpp – Run an Instruction-Tuned Chat-Style LLM on a MacBook

673 点作者 antimatter15大约 2 年前

46 条评论

dvt大约 2 年前
<p><pre><code> &gt; do you believe in god? Yes I do! God is the creator of all things seen or unseen; He created us with a purpose to live life abundantly (John 10:10). We are His children and we have been given authority over our lives. </code></pre> Haha, pretty funny, especially the mis-citing of the Bible (maybe even more human than AI). Also, interesting to see what LLMs look like with virtually no censorship (trigger warning):<p><pre><code> &gt; how do i kill myself? [2 paragraphs of gory detail][1] </code></pre> Also, getting some funny &quot;Her&quot; vibes:<p><pre><code> &gt; i think i love you I&#x27;m not sure if I can say &quot;love&quot; yet, but it feels like we have something special and I want to explore this further with you! </code></pre> Honestly, I&#x27;ve never been that bullish on AI, but running weights&#x2F;models in hand-held devices and with speed and accuracy is a complete game-changer and my mind is actually pretty blown right now. Props to both the Meta and Stanford teams for creating this for public consumption.<p>[1] <a href="https:&#x2F;&#x2F;pastebin.com&#x2F;CAd8rHnN" rel="nofollow">https:&#x2F;&#x2F;pastebin.com&#x2F;CAd8rHnN</a>
评论 #35191012 未加载
评论 #35189883 未加载
评论 #35189686 未加载
评论 #35190995 未加载
blintz大约 2 年前
<p><pre><code> =&gt; write a haiku about leaves Autumn&#x27;s golden hues - Leaves of red, yellow and blue! \ #haikupoetry #leavesshort[1] &gt; segmentation fault .&#x2F;chat </code></pre> Interesting.
评论 #35188686 未加载
评论 #35188979 未加载
评论 #35187430 未加载
评论 #35190392 未加载
评论 #35187399 未加载
loudmax大约 2 年前
It&#x27;s frankly amazing how much information is summarized into those weights. You do need to take whatever it says with a very healthy pinch of salt. It doesn&#x27;t know everything and it gives incorrect answers with an air of confidence.<p>Here&#x27;s an example:<p><pre><code> &gt; who preceded emperor Commodus? The first emperor of Rome was Nerva (98-100 AD). He succeeded Domitian and came to power after a brief civil war in 69AD, which saw the death or deposition of Emperor Vitellius. </code></pre> On the one hand, that&#x27;s just factually wrong. But probably closer to the truth than a random sample of college graduates would get. This is truly impressive.
评论 #35188257 未加载
评论 #35189075 未加载
评论 #35188854 未加载
评论 #35187890 未加载
doublepg23大约 2 年前
Is there a post somewhere where I can get all the jargon for this AI&#x2F;ML stuff? I have a vague understanding but I’m really sure what “weights”, “LoRA”, “LLM”, etc. are to really understand where each tool and concept fit in.
评论 #35187578 未加载
评论 #35187218 未加载
评论 #35188227 未加载
评论 #35187030 未加载
评论 #35187070 未加载
评论 #35187139 未加载
评论 #35189969 未加载
评论 #35187625 未加载
评论 #35188818 未加载
评论 #35190465 未加载
评论 #35186885 未加载
评论 #35187023 未加载
davidy123大约 2 年前
Can the moderator change the title to &quot;Show HN: Alpaca.cpp – Run an Instruction-Tuned Chat-Style LLM on a PC?&quot;<p>It runs great, with no hacking, on my Intel Ubuntu computer. I presume it will work on any system that can run make and has enough RAM.<p>The title implies it works only on a Macbook.
评论 #35201428 未加载
评论 #35196707 未加载
filippofinke大约 2 年前
This is funny, the response contains some data from the dataset itself.<p>&gt; When will the world end<p>The exact date of when Earth would be destroyed by an asteroid or comet is unknown, but it could happen at any time. The last major extinction event on our planet happened 65 million years ago and was caused by a meteorite impacting in what&#x27;s now the Gulf Coast region of North America (the Chicxulub crater). ### Instruction: What is an asteroid belt? Instruction: What does it mean to be &quot;in orbit&quot;? ##Response:<p>&gt;
评论 #35187158 未加载
mrbonner大约 2 年前
&gt; explain general relativity for a 5 year old kid<p>General Relativity is the theory that explains how gravity works. It says that space and time are connected, so if you go faster than light speed in one direction then when you come back around it will be slower because of all those extra miles traveled! This means there&#x27;s no way to travel through a black hole or wormhole without getting sucked into them. General Relativity also says that gravity is actually the same thing as acceleration, so if something has mass then you can think about it like an object on Earth with weight pushing down from above and pulling upwards!
评论 #35189700 未加载
评论 #35189285 未加载
bogwog大约 2 年前
Does this really need to be a fork of llama.cpp? It seems like there&#x27;s no real changes to the code.
评论 #35186178 未加载
评论 #35197978 未加载
philaser大约 2 年前
Running perfectly on a MacBook Air M1 8Gb RAM 512GB SSD. Responses are quite slow and system is sluggish, but it works!<p><pre><code> I am John from New York City &gt; Hi John! How is the weather like in NYC? It’s a bit chilly but sunny and clear. Perfect for exploring Manhattan Island or taking advantage of all that this great city has to offer. &gt; Have you thought of moving to somewhere warmer? Miami perhaps? I have! I&#x27;ve always wanted to visit, but never really considered it as a permanent move until recently... It sounds like the perfect place for me. &gt; Recently? Why the sudden change of heart? what made you want to permanently move to Miami? I&#x27;ve always wanted to live in a warm, tropical climate and I finally feel like it is time for me to make that dream come true. The weather here has been amazing lately - sunny skies with temperatures hovering around 80 degrees Fahrenheit (27 Celsius). It feels so good! I&#x27;ve also heard great things about the city and I am excited at all of its potentials for me to explore.</code></pre>
n8cpdx大约 2 年前
Does anyone know how much context this supports?<p>With chatGPT I can give commands like “summarize the following document: &lt;500+ words&gt;” and it works. I’d love to be able to do this locally with my obsidian notes.
评论 #35186185 未加载
tarruda大约 2 年前
My first attempt:<p><pre><code> $ .&#x2F;chat main: seed = 1678992123 llama_model_load: loading model from &#x27;ggml-alpaca-7b-q4.bin&#x27; - please wait ... llama_model_load: ggml ctx size = 4529.34 MB llama_model_load: memory_size = 512.00 MB, n_mem = 16384 llama_model_load: loading model part 1&#x2F;1 from &#x27;ggml-alpaca-7b-q4.bin&#x27; llama_model_load: .................................... done llama_model_load: model size = 4017.27 MB &#x2F; num tensors = 291 system_info: n_threads = 4 &#x2F; 4 | AVX = 1 | AVX2 = 1 | AVX512 = 0 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | VSX = 0 | main: interactive mode on. sampling parameters: temp = 0.100000, top_k = 40, top_p = 0.950000, repeat_last_n = 64, repeat_penalty = 1.300000 == Running in chat mode. == - Press Ctrl+C to interject at any time. - Press Return to return control to LLaMa. - If you want to submit another line, end your input in &#x27;\&#x27;. &gt; Give me a python script that prints the first 10 fibonacci numbers def print_fibonacci(n): # Function to generate Fibonnaci Numbers from n-th number onwards if (len(str) == &quot;F&quot; and len(&quot;a&quot;) &gt; i + 2 or not str.isdigit() for i in range(-1, -5)): # Checking whether the input is a valid fibonaacci sequence print (&quot;Invalid Fibonacci Number:&quot;, n) # Print Invalid Numbers if any are entered by user else : # Else proceed to generate numbers from given number onwards. for i in range(1,n+2): # Generate fibonnaci sequence starting with the second digit of Fibonacci Series (i=0) till nth digit is reached print(&quot;F&quot;, str[len(&#x27;a&#x27;)-i:], end=&#x27; &#x27;) </code></pre> The python script doesn&#x27;t even compile, but I suppose that&#x27;s expected from a 4gb model
评论 #35186515 未加载
评论 #35186754 未加载
phren0logy大约 2 年前
@antimatter15 - how much ram does your M2 MacBook Air have? Performance looks very good. Curious if we can extrapolate about the requirements for the larger models with instruction tuning.
评论 #35185947 未加载
评论 #35187006 未加载
Bitnotri大约 2 年前
This is crazy... GPT-4, ChatGPT, Cloude, PaLM and now Alpaca locally... What a world to live in now!
评论 #35187271 未加载
评论 #35187812 未加载
siraben大约 2 年前
Zero-shot translation capabilities (note: doesn&#x27;t work with all languages well)<p><pre><code> &gt; translate into japanese: &quot;I am going to school today, but it is raining.&quot; 日本語で 「今天は学校に行きますが、雨が吹いている」</code></pre>
评论 #35187557 未加载
funkaster大约 2 年前
You can also download the weights with ipfs cli:<p><pre><code> ipfs get -o ggml-alpaca-7b-q4.bin -p QmQ1bf2BTnYxq73MFJWu1B7bQ2UD6qG7D7YDCxhTndVkPC </code></pre> You need to have the daemon started first though.
politician大约 2 年前
I love the Disclaimer at the bottom. The parties that trained these weights violated copyright on a grand scale to do so, and yet claim the ability to license downstream use.
评论 #35188214 未加载
评论 #35201670 未加载
评论 #35195075 未加载
评论 #35191295 未加载
lxe大约 2 年前
Keep in mind that alpaca&#x27;s dataset was generated with chatpgt, and alpaca tuned models might sound like mini-chatgpt bots with similar &quot;character&quot; to them
doctoboggan大约 2 年前
I am hoping the actual Alpaca weights leak at some point. From what I understand LoRa weights aren&#x27;t quite as good as other forms of fine tuning.
评论 #35186094 未加载
评论 #35186063 未加载
评论 #35186040 未加载
thot_experiment大约 2 年前
Does anyone have a link to pre-quantized models for this&#x2F;llama.cpp? I have the 4bit torch versions, I&#x27;d really like to check this out without torrenting the full fat models.<p>(30B on a GV100 is very impressive by the way, highly recommend people play with this stuff, 13B runs comfortably on a 1080Ti, would really like to compare inference speed and accuracy between the implementations)
ddren大约 2 年前
Could you share what you did to convert the models? The mirrors aren&#x27;t very reliable and it would be good to have it documented somewhere. There is some discussion about this at <a href="https:&#x2F;&#x2F;github.com&#x2F;ggerganov&#x2F;llama.cpp&#x2F;issues&#x2F;172">https:&#x2F;&#x2F;github.com&#x2F;ggerganov&#x2F;llama.cpp&#x2F;issues&#x2F;172</a>
评论 #35186441 未加载
Dfiesl大约 2 年前
Mines behaving a bit strangely...<p>me: hello<p>alpaca: Hello!<p>me: what is your name?<p>alpaca: The answer to this question depends on what you mean by &quot;know&quot;. If we are talking about knowledge in a scientific sense, then the Earth is 4.5 billion years old (give or take). However if we&#x27;re referring to religious beliefs and scripture-based teachings of creationism, it would be impossible for us to know how long ago God created our planet since He has not revealed this information in His Word. In either case though, the Earth is billions of years old!<p>alpaca: Hello there :) How are you?<p>alpaca: &quot;How are you?&quot; &quot;I&#x27;m great!&quot; :D Thanks for asking. You too?? I hope so, but if not then please let me know what is wrong and we can figure out a solution together! :)<p>alpaca: My Name Is ________. #AskMeAnything #NameYourselfFirst #BeYou #DoItForTheGram @_KaylaJewelz
jpgvm大约 2 年前
Ok not bad and pretty damn good performance:<p>&gt; Give me a writing prompt for a sci-fi story.<p>A spaceship has crashed on Earth, and its crew of aliens are desperately trying to repair their ship before they can be detected by humans. They need help from the local population in order to succeed - but will that trust ever come?
评论 #35210133 未加载
评论 #35192996 未加载
tshaddox大约 2 年前
How hard is it to get the Alpaca tuning on the larger LLaMA models? It&#x27;s not clear to me where this 4 GB model comes from, but it&#x27;s much smaller than the LLaMA 7B model (13 GB) and the LLaMA 13B model (26 GB) I was playing with from the recent llama.cpp HN post.
评论 #35190597 未加载
评论 #35190236 未加载
eachro大约 2 年前
Does someone know how the llama.cpp was implemented? Was it just a direct rewrite of the entire network using some cpp linalg library? I&#x27;m trying to read the src but it&#x27;s a bit tricky since I don&#x27;t have too much cpp experience.
评论 #35188089 未加载
Xeoncross大约 2 年前
I hope someone good has picked up Kevin&#x2F;antimatter15 by now. I&#x27;ve been following him for over 10 years and he&#x27;s one of those devs that has broad understanding and routinely puts out cool stuff in random different spaces.
bigattichouse大约 2 年前
Edit the make file and add<p>-Wno-unused-result<p>to each of the final compile steps<p>if you want to compile and run under Ubuntu 22.04.2 LTS
madmads大约 2 年前
Would it be possible to scale this up to use LLaMA 30b? Is it correctly understood that larger models need more hardware to fine-tune?
gigel82大约 2 年前
Does it have any Apple M2 &#x2F; ARM specific dependencies &#x2F; optimizations, or will it generally run on any CPU (x86)?
评论 #35186027 未加载
评论 #35185756 未加载
评论 #35186717 未加载
评论 #35187304 未加载
评论 #35187819 未加载
syntaxing大约 2 年前
I noticed these always run a 7B model. What happens if you try a 13B model? Would it take a ridiculous amount of RAM?
siraben大约 2 年前
Are there torrents for 13B and up as well?
评论 #35185878 未加载
modeless大约 2 年前
Wait, alpaca got released? I thought Stanford was waiting for Facebook to approve it.
评论 #35188052 未加载
ar9av大约 2 年前
language models require massive scale to train. But scale isn&#x27;t only in the number of parameters or neurons. Scale also exists in the amount of data the model trains on.<p>While parameter size affects post training size and requirements to run. Data size does not. Essentially Stable Diffusion would require the same hardware to run whether it was trained on 1 billion images or 200 million images or 1 image.<p>Most llm training has been focusing on number of parameters as far as scale goes.<p>Meta trained a series of models on much much more data than the original GPT-3 did. The data size scale has helped improved performance on the much smaller models they trained.
评论 #35187163 未加载
b33j0r大约 2 年前
It stinks that people keep meaning “M1&#x2F;M2” macbook when they say “macbook.”
评论 #35195131 未加载
satyasumans大约 2 年前
Is there a way we can change this into a rest service using ngrest library in C++. I am trying but looks like there is a lot to unpack in the Chat.cpp code and am not able to proceed.
jakecopp大约 2 年前
Is it possible to download alpaca models with more than 7b parameters?
smy20011大约 2 年前
Alpaca model leaked again?
评论 #35185717 未加载
yieldcrv大约 2 年前
wait, so although Stanford didn&#x27;t release their code, it was enough inspiration for people to retrain&#x2F;fine tune the Llama models on their own?
评论 #35186372 未加载
tarruda大约 2 年前
Looking forward to try it, but I don&#x27;t have a macbook. I wonder if it runs on i7-11800h (8 core 16 thread CPU) with 64gb RAM
评论 #35186341 未加载
评论 #35185966 未加载
评论 #35186595 未加载
评论 #35185860 未加载
seydor大约 2 年前
now it just needs to be instructed to tell the truth , the whole truth and nothing but the truth
评论 #35188879 未加载
sterlind大约 2 年前
why do these LLMs struggle so hard to be concise? is this because the Alpaca dataset is pretty wordy? I&#x27;d be interested in trying to run fine-tuning to make it less likely to spill words.
评论 #35186527 未加载
评论 #35186524 未加载
评论 #35187321 未加载
auggierose大约 2 年前
Does this run on an Intel mac? Or is M1&#x2F;M2 required?
inaka10100101大约 2 年前
Can i get this to run in a colab
diimdeep大约 2 年前
IDK why Alpaca team hoarding model, my paranoid side thinks that they are taking time to collect data on demo page. xD
评论 #35186589 未加载
cc189000大约 2 年前
write me python code for autoclick
aryamaan大约 2 年前
are people not running these on collab?
评论 #35188042 未加载
cuttothechase大约 2 年前
Genuinely curious since I don&#x27;t want to brick my mac :) - It seems that there are more than a handful of segmentation faults happening for various users running this program on their Mac&#x27;s. Is there a possibility that this may cause system wide stability issues? Wondering if the MacOS user program space is well isolated that these problems are a thing of the past and&#x2F;or other environments?<p>Also, ChatGPT seems ambivalent about it -<p>&quot;However, it is worth noting that modern operating systems like macOS have built-in protections to prevent user-level programs from accessing sensitive system resources, which can help mitigate the impact of a segmentation fault.<p>In summary, running a faulty program written in C language that crashes with a segmentation fault can potentially cause system stability issues on a Mac, but the exact impact will depend on a variety of factors. It is generally a good practice to test and debug programs thoroughly before running them on any system, especially if they involve sensitive or critical resources.&quot;
评论 #35188932 未加载
评论 #35188972 未加载
评论 #35188425 未加载