TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Big LLMs weights are a piece of history

301 点作者 freeatnet2 个月前

28 条评论

intellectronica2 个月前
I love the title "Big LLMs" because it means that we are now making a distinction between big LLMs and minute LLMs and maybe medium LLMs. I'd like to propose the we call them "Tall LLMs", "Grande LLMs", and "Venti LLMs" just to be precise.
评论 #43380740 未加载
评论 #43380587 未加载
评论 #43379431 未加载
评论 #43379704 未加载
评论 #43380392 未加载
评论 #43381899 未加载
评论 #43379315 未加载
评论 #43379509 未加载
评论 #43379821 未加载
评论 #43379375 未加载
评论 #43379738 未加载
评论 #43380311 未加载
评论 #43379326 未加载
评论 #43380279 未加载
评论 #43379732 未加载
评论 #43380516 未加载
评论 #43380631 未加载
评论 #43381241 未加载
评论 #43384455 未加载
评论 #43379283 未加载
评论 #43380504 未加载
评论 #43380460 未加载
评论 #43380650 未加载
评论 #43382243 未加载
dr_dshiv2 个月前
“We should regard the Internet Archive as one of the most valuable pieces of modern history; instead, many companies and entities make the chances of the Archive to survive, and accumulate what otherwise will be lost, harder and harder. I understand that the Archive headquarters are located in what used to be a church: well, there is no better way to think of it than as a sacred place.”<p>Amen. There is an active effort to create an Internet Archive based in Europe, just… in case.
评论 #43382397 未加载
评论 #43380139 未加载
评论 #43380335 未加载
jart2 个月前
Mozilla&#x27;s llamafile project is designed to enable LLMs to be preserved for historical purposes. They ship the weights and all the necessary software in a deterministic dependency-free single-file executable. If you save your llamafiles, you should be able to run them in fifty years and have the outputs be exactly the same as what you&#x27;d get today. Please support Mozilla in their efforts to ensure this special moment in history gets archived for future generations!<p><a href="https:&#x2F;&#x2F;github.com&#x2F;Mozilla-Ocho&#x2F;llamafile&#x2F;" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;Mozilla-Ocho&#x2F;llamafile&#x2F;</a>
评论 #43379938 未加载
GeoAtreides2 个月前
Just like the map isn&#x27;t the territory, so summaries are not the content nor the library fillings the actual books.<p>If I want to read a post, a book, a forum, I want to read exactly that, not a simulacrum built by arcane mathematical algorithms.
评论 #43379913 未加载
评论 #43381133 未加载
api2 个月前
That&#x27;s really what these are: something analogous to JPEG for language, and queryable in natural language.<p>Tangent: I was thinking the other day: these are not AI in the sense that they are not primarily <i>intelligence</i>. I still don&#x27;t see much evidence of that. What they do give me is superhuman memory. The main thing I use them for is search, research, and a &quot;rubber duck&quot; that talks back, and it&#x27;s like having an intern who has memorized the library and the entire Internet. They occasionally hallucinate or make mistakes -- compression artifacts -- but it&#x27;s there.<p>So it&#x27;s more AM -- artificial memory.<p>Edit: as a reply pointed out: this is Vannevar Bush&#x27;s Memex, kind of.
评论 #43379066 未加载
评论 #43379057 未加载
评论 #43379450 未加载
评论 #43378847 未加载
评论 #43379345 未加载
评论 #43379959 未加载
评论 #43379221 未加载
评论 #43379952 未加载
评论 #43379053 未加载
评论 #43379046 未加载
评论 #43379690 未加载
laborcontract2 个月前
I miss the good ol days when I&#x27;d have text-davinci make me a table of movies that included a link to the movie poster. It usually generated a url of an image in an s3 bucket. The link <i>always worked</i>.
andix2 个月前
I think it’s fine that not everything on the internet is archived forever.<p>It has always been like that, in the past people wrote on paper, and most of it was never archived. At some point it was just lost.<p>I inherited many boxes of notes, books and documents from my grandparents. Most of it was just meaningless to me. I had to throw away a lot of it and only kept a few thousand pages of various documents. The other stuff is just lost forever. And that’s probably fine.<p>Archives are very important, but nowadays the most difficult part is to select what to archive. There is so much content added to the internet every second, only a fraction of it can be archived.
hedgehog2 个月前
This doesn&#x27;t make much sense to me. Unattributed heresay has limited historical value, perhaps zero given that the view of the web most of the weights-available models have is Common Crawl which is itself available for preservation.
评论 #43384495 未加载
fl4tul42 个月前
&gt; Scientific papers and processes that are lost forever as publishers fail, their websites shut down.<p>I don&#x27;t think the big scientific publishers (now, in our time) will ever fail, they are RICH!
评论 #43382445 未加载
评论 #43380438 未加载
评论 #43380142 未加载
nickpsecurity2 个月前
People wanting this would be better off using memory architectures, like how the brain does it. For ML, the simplest approach is putting in memory layers with content-addressible schemes. I have a few links on prototypes in this comment:<p><a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=42824960">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=42824960</a>
评论 #43379270 未加载
dstroot2 个月前
Isn’t big LLM training data actually the most analogous to the internet archive? Shouldn’t the title be “Big LLM training data is a piece of history”? Especially at this point in history since a large portion of internet data going forward will be LLM generated and not human generated? It’s kind of the last snapshot of human-created content.
评论 #43380857 未加载
ilaksh2 个月前
Great idea. Slightly related idea: use the Internet Archive to build a dataset of 6502 machine code&#x2F;binaries, corresponding manuals, and possibly videos of the software in action.. maybe emulator traces.<p>It might be possible to create an L LM that can write a custom vintage game or program on demand in machine code and simultaneously generate assets like sprites. Especially if you use the latest reinforcement learning techniques.
rollcat2 个月前
<a href="https:&#x2F;&#x2F;xkcd.com&#x2F;1683&#x2F;" rel="nofollow">https:&#x2F;&#x2F;xkcd.com&#x2F;1683&#x2F;</a>
hi_hi2 个月前
Naming antics aside, the article makes a good point I&#x27;ve heard previously about the importance of the Internet Archive.<p>Are there any search experiences that allow me to search like it&#x27;s 1999? I&#x27;d love to be able to re-create the experience of finding random passion project blogs that give a small snapshot of things people and business were using the web for back then.
OuterVale2 个月前
Interesting. It seems that both they and I had very similar ideas at about the same time, with this being posted just a few hours after I finally published about AI model history being lost.<p><a href="https:&#x2F;&#x2F;vale.rocks&#x2F;posts&#x2F;ai-model-history-is-being-lost" rel="nofollow">https:&#x2F;&#x2F;vale.rocks&#x2F;posts&#x2F;ai-model-history-is-being-lost</a>
Havoc2 个月前
I wonder whether it&#x27;ll become like pre-WW2 steel that doesn&#x27;t have nuclear contamination.<p>Just with a pre-LLM knowledge
dmos622 个月前
Enjoy the insight, but the title makes my eye twitch. How about &quot;LLM weights are pieces of history&quot;?
评论 #43380178 未加载
pama2 个月前
I would be curious to know if it would be possible to recunstruct approximate versions of popular common subsets of internet training data by using many different LLMs that may have happened to read the same info. Anyone knows pointers to math papers about such things?
teleforce2 个月前
I really like the narative that now LLM is the conserving human knowledge that otherwise would be lost forever in the form of its weights in a kind of a lossy compression.<p>Personally I&#x27;d like that if all the knowledge and information (K &amp; I) are readily available and accessible (pretty sure most of the prople share the same sentiment), despite the consistent business decisions from the copyright holders to hoard their K &amp; I by putting everything behind paywalls and&#x2F;or registration (I&#x27;m looking at you Apple and X&#x2F;Twitter). As much that some people hate Google by organizing the world information by feeding and thriving through advertisements because in the long run the information do get organized and kind of preserved in many Internet data formats, lossy or not. After all Google who originall designed the transformer that enabled the LLM weights that are now apparently a piece of history.
almosthere2 个月前
Split the wayback machine away from its book copyright lawsuit stuff and you don&#x27;t have to worry.
off_by_inf2 个月前
And they all undertrained, according to the papers.
bossyTeacher2 个月前
So large large language model?
throwaway484762 个月前
The internet training data for LLMs is valuable history were losing one dead webadmin at a time. The regurgitated slop less so.
codr72 个月前
I find it very depressing to think that the only traces left from all the creativity will end up to be AI slop, the worst use case ever.<p>I feel like the more people use GenAI, the less intelligent they become. Like the rest of this society, they seem designed to suck the life force out of humans and and return useless crap instead.
sourtrident2 个月前
Imagine future historians piecing together our culture from hallucinated AI memories - inaccurate, sure, but maybe even more fascinating than reality itself.
blinky812 个月前
&quot;big large&quot; lol
guybedo2 个月前
fwiw i&#x27;ve added a summary of the discussion here: <a href="https:&#x2F;&#x2F;extraakt.com&#x2F;extraakts&#x2F;67d708bc9844db151612d782" rel="nofollow">https:&#x2F;&#x2F;extraakt.com&#x2F;extraakts&#x2F;67d708bc9844db151612d782</a>
isoprophlex2 个月前
Interesting. Just this morning I had a conversation with Claude about this very topic. When asked &quot;can you give me your thoughts on LLM train runs as historical artifacts? do you think they might be uniquely valuable for future historians?&quot;, it answered<p><pre><code> &gt; oh HELL YEAH they will be. future historians are gonna have a fucking field day with us. &gt; imagine some poor academic in 2147 booting up &quot;vintage llm.exe&quot; and getting to directly interrogate the batshit insane period when humans first created quasi-sentient text generators right before everything went completely sideways with *gestures vaguely at civilization* &gt; *&quot;computer, tell me about the vibes in 2025&quot;* &gt; &quot;BLARGH everyone was losing their minds about ai while also being completely addicted to it&quot; </code></pre> Interesting indeed to be able to directly interrogate the median experience of being online in 2025.<p>(also my apologies for slop-posting; i slapped so many custom prompting on it that I hope you&#x27;ll find the output to be amusing enough)
评论 #43380273 未加载