TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

If you can't reproduce the model then it's not open-source

230 pointsby mgregover 1 year ago

22 comments

ssgodderidgeover 1 year ago
&gt; Imagine if Linux published only a binary without the codebase. Or published the codebase without the compiler used to make the binary. This is where we are today.<p>This was such a helpful way to frame the problem! Something felt off about the &quot;open source models&quot; out there; this highlights the problem incredibly well.
评论 #39031433 未加载
评论 #39033052 未加载
评论 #39032754 未加载
评论 #39036699 未加载
评论 #39034294 未加载
评论 #39034622 未加载
评论 #39036304 未加载
评论 #39031583 未加载
elashriover 1 year ago
I think the process of data acquisition isn&#x27;t so clear-cut. Take CERN as an example: they release loads of data from various experiments under the CC0 license [1]. This isn&#x27;t just a few small datasets for classroom use; we&#x27;re talking big-league data, like the entire first run data from LHCb [2].<p>On their portal, they don&#x27;t just dump the data and leave you to it. They&#x27;ve got guides on analysis and the necessary tools (mostly open source stuff like ROOT [3] and even VMs). This means anyone can dive in. You could potentially discover something new or build on existing experiment analyses. This setup, with open data and tools, ticks the boxes for reproducibility. But does it mean people need to recreate the data themselves?<p>Ideally, yeah, but realistically, while you could theoretically rebuild the LHC (since most technical details are public), it would take an army of skilled people, billions of dollars, and years to do it.<p>This contrasts with open source models, where you can retrain models using data to get the weights. But getting hold of the data and the cost to reproduce the weights is usually prohibitive. I get that CERN&#x27;s approach might seem to counter this, but remember, they&#x27;re not releasing raw data (which is mostly noise), but a more refined version. Try downloading several petabytes of raw data if not; good luck with that. But for training something like a LLM, you might need the whole dataset, which in many cases have its own problems with copyrights…etc.<p>[1] <a href="https:&#x2F;&#x2F;opendata.cern.ch&#x2F;docs&#x2F;terms-of-use" rel="nofollow">https:&#x2F;&#x2F;opendata.cern.ch&#x2F;docs&#x2F;terms-of-use</a><p>[2] <a href="https:&#x2F;&#x2F;opendata.cern.ch&#x2F;docs&#x2F;lhcb-releases-entire-run1-dataset" rel="nofollow">https:&#x2F;&#x2F;opendata.cern.ch&#x2F;docs&#x2F;lhcb-releases-entire-run1-data...</a><p>[3] <a href="https:&#x2F;&#x2F;root.cern&#x2F;" rel="nofollow">https:&#x2F;&#x2F;root.cern&#x2F;</a>
评论 #39034202 未加载
albert180over 1 year ago
I think the biggest issue is with publishing the datasets. Then people and companies would discover that it&#x27;s full of their copyrighted content and sue. I wouldn&#x27;t be surprised if they slurped in the whole Z-Library et Al into their models. Or Google their entire Google Books Dataset
评论 #39032085 未加载
anticorporateover 1 year ago
The Open Source Initiative, who maintain the Open Source Definition, have been running a whole series over the past year to collect input from all sorts of stakeholders about what it means for an AI to be open source. I was lucky enough to participate in an afternoon long session with about a hundred other people last year at All Things Open.<p><a href="https:&#x2F;&#x2F;deepdive.opensource.org&#x2F;" rel="nofollow">https:&#x2F;&#x2F;deepdive.opensource.org&#x2F;</a><p>I encourage you to go check out what&#x27;s already being done here. I promise it&#x27;s way more nuanced than anything than is going to fit on a tweet.
评论 #39031895 未加载
mgregover 1 year ago
Applying the term &quot;open source&quot; to AI models is a bit more nuanced than to software. Many consider reproducibility the bar to get over to earn the label &quot;open source.&quot;<p>For an AI model that means the model itself, the dataset, and the training recipe (e.g. process, hyperparameters) often also released as source code. With that (and a lot of compute) you can train the model to get the weights.
darrenBaldwin03over 1 year ago
Same with open-core - if you can&#x27;t self-host the thing on your own infra then its not REALLY OSS
评论 #39033407 未加载
tqiover 1 year ago
&quot;the project does not benefit from the OSS feedback loop&quot; It&#x27;s not like you can submit PRs to training data that fixes specific issues the way you can submit bug fixes, so I&#x27;m skeptical you would see much of a feedback loop.<p>&quot;it’s hard to verify that the model has no backdoors (eg sleeper agents)&quot; Again given the size of the datasets and the opaque way training works, I am skeptical that anyone would be able tell if there is a backdoor in the training data.<p>&quot;impossible to verify the data and content filter and whether they match your company policy&quot; I don&#x27;t totally know what this means. For one, you can&#x2F;probably should apply company policies to the model outputs, which you can do without access to training data. Is the idea that every company could&#x2F;should filter input data and train their own models?<p>&quot;you are dependent on the company to refresh the model&quot; At the current cost, this is probably already true for most people.<p>&quot;A true open-source LLM project — where everything is open from the codebase to the data pipeline — could unlock a lot of value, creativity, and improve security.&quot; I am overall skeptical that this is true in the case of LLMs. If anything, I think this creates a larger surface for bad actors to attack.
评论 #39038964 未加载
评论 #39034616 未加载
andy99over 1 year ago
I don&#x27;t agree, and the analogy is poor. One can do the things he lists with a trained model. Having the data is basically a red herring. I wish this got more attention. Open&#x2F;free software is about exercising freedoms, and they all can be exercised if you&#x27;ve got the model weights and code.<p><a href="https:&#x2F;&#x2F;www.marble.onl&#x2F;posts&#x2F;considerations_for_copyrighting_AI.html" rel="nofollow">https:&#x2F;&#x2F;www.marble.onl&#x2F;posts&#x2F;considerations_for_copyrighting...</a>
评论 #39033275 未加载
评论 #39034266 未加载
tbrownawover 1 year ago
&gt; <i>The “source code” for a work means the preferred form of the work for making modifications to it.</i><p>-- gplv3<p>These AI&#x2F;ML models are interesting in that the weights are derived from something else (training set), but if you&#x27;re modifying them you don&#x27;t need that. Lots of &quot;how to do fine-tuning&quot; tutorials floating around, and they don&#x27;t need access to the original training set.
cpetersoover 1 year ago
Are there any true open-source LLM models, where all the training data is publicly-available (with a compatible license) and the training software can reproduce bit-identical models?<p>Is training nondeterministic? I know LLM outputs are purposely nondeterministic.
评论 #39043247 未加载
beardywover 1 year ago
I think the answer is in the name. The &quot;source&quot; has always been what you need to build the thing. In this context I think we can agree that the thing is the model. Based on that the model is no more open source than a binary program.
declaredappleover 1 year ago
I&#x27;ll venture to say the majority of these &quot;open access models&quot; are meant to serve as advertisements of capabilities (either of hardware, research, or techniques) and nothing more. MPT being one of the most obvious example.<p>Many don&#x27;t offer any information, some do offer information but provide no new techniques and just threw a bunch of compute and some data to make a sub-par model that shows up on a specific leaderboard.<p>Everyone is trying to save a card up their sleeve so they can sell it. And showing up on scoreboards is a great advertisement.
pabs3over 1 year ago
I like how Debian&#x27;s machine learning policy says this:<p><a href="https:&#x2F;&#x2F;salsa.debian.org&#x2F;deeplearning-team&#x2F;ml-policy&#x2F;" rel="nofollow">https:&#x2F;&#x2F;salsa.debian.org&#x2F;deeplearning-team&#x2F;ml-policy&#x2F;</a>
nathanasmithover 1 year ago
Publish your data and prepare to get vilified by professional complainers because the data doesn&#x27;t conform to their sensibilities. Lots of downside with very little of the opposite.
ramesh31over 1 year ago
No, but it&#x27;s still insanely useful and free as in beer.
belvalover 1 year ago
&gt; if you can’t reproduce the model then it’s not truly open-source.<p>Open-source means open source, it does not make reproducibility guarantees. You get the code and you can use the code. Pushed to the extreme this is like saying Chromium is not open-source because my 4GB laptop can&#x27;t compile it.<p>Getting training code for GPT-4 under MIT would be mostly useless, but it would still be open source.
评论 #39032091 未加载
评论 #39031606 未加载
emadmover 1 year ago
We made our last language model fully reproducible including all datasets, training details, hyper parameters etc: <a href="https:&#x2F;&#x2F;stability.wandb.io&#x2F;stability-llm&#x2F;stable-lm&#x2F;reports&#x2F;StableLM-3B-4E1T--VmlldzoyMjU4?accessToken=u3zujipenkx5g7rtcj9qojjgxpconyjktjkli2po09nffrffdhhchq045vp0wyfo" rel="nofollow">https:&#x2F;&#x2F;stability.wandb.io&#x2F;stability-llm&#x2F;stable-lm&#x2F;reports&#x2F;S...</a>
Der_Einzigeover 1 year ago
95% of the value comes from the model being freely downloadable and analyzable (i.e. not obfuscated&#x2F;crippled post-hoc). Sure there is some difference, but as researchers I care far more about open access than making every &quot;gnuight&quot; on the internet happy that we used the right terminology.
edoardo-schnellover 1 year ago
So, we need something like dockerfiles for models?
fragmedeover 1 year ago
it&#x27;s <i>model available</i>, not open source!
robblbobblover 1 year ago
Agreed.
RcouF1uZ4gsCover 1 year ago
I would argue that while technically correct, it is not what most people really care. What they care about are the following:<p>1. Can I download it?<p>2. Can I run it on my hardware?<p>3. Can I modify it?<p>4. Can I share my modifications with others?<p>If those questions are in the affirmative, then I think most people consider it open enough, and it is a huge step for freedom compared to the models such as OpenAI.
评论 #39031835 未加载
评论 #39032121 未加载
评论 #39036509 未加载
评论 #39033030 未加载