TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Wikipedia Will Survive A.I

2 点作者 akolbe超过 1 年前

1 comment

NoZebra120vClip超过 1 年前
Yeah, it&#x27;s a three-pronged threat, isn&#x27;t it?<p>I think this article adequately outlines those threats. You&#x27;ve got the question of content added to the project, the question of the projects&#x27; text being used to train LLMs, and you&#x27;ve got sources now that are using LLM content to write news articles and other stuff. And the article also mentions all the great uses of ML already incorporated into bots and the MediaWiki software, and the cloud platform.<p>I&#x27;ve already witnessed one fairly widespread case of an editor who began adding big swaths of LLM content to articles. They were caught, blocked, and reverted. Now we already deal with similar problems in the copyright field. There are often editors who violate copyright &quot;under the radar&quot; for years, and make many, many edits, and then they are caught, blocked, and a cleanup process is initiated. There&#x27;s a whole process for that, and it&#x27;s perpetually understaffed and backlogged. But the catching of violators tends to happen at a fairly good clip.<p>I think that if we can accurately catch copyright violators, then we should also be able to find LLM-only editors. Even though you can&#x27;t simply diff with an existing source, there are hallmarks, and eventually, they reveal themselves.<p>I&#x27;m not sure if keeping LLM content off Wikipedia will improve it, per se. A lot of editors are bad writers already, and also write outlandish stuff that&#x27;s unsourced. So it may be a wash. But we have to try.