TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Ask HN: Should OS licences evolve to consider training of AI models?

1 点作者 ColinEberhardt超过 2 年前
There has (quite rightly) been a lot of discussion on the internet about whether training Copilot on non-permissively licensed open source code is fair. There are similar debates to be had about training DALL-E and others on artworks.<p>For software, my feeling is that this is very much a grey area, so why not make it explicit?<p>Should we develop licences that are crystal clear around whether you are permitted to use this codebase for the purposes of training models?<p>(I vote &#x27;yes&#x27;)

2 条评论

smoldesu超过 2 年前
The ambiguity makes sense here. It sounds like you&#x27;re asking if people should be forbidden from training text models on the Linux kernel; I think this line is already drawn in the sand with GPL. If someone trains an AI model on the Linux kernel and use it to generate commercial code, that&#x27;s a potential license violation.<p>The lack of clear signalling around this topic leaves it ripe for OSS devs and enthusiasts to explore, but extremely scary for commercial entities to navigate. In other words, it&#x27;s working exactly as GPL should.
josephcsible超过 2 年前
No. Either such training is fair use, or it isn&#x27;t. If it is fair use, then it&#x27;s always allowed even if the license explicitly says it&#x27;s not. If it isn&#x27;t fair use, then Microsoft is already violating the licenses anyway, such as the GPL, by not making Copilot&#x27;s source available under the same license (and ditto for things like DALL-E, and also by violating the attribution clauses even of permissive licenses).