TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

A community effort to extract structured information from Wikipedia

14 点作者 durana大约 16 年前

2 条评论

graemep大约 16 年前
I like Wikipedia less and less. It contains lots of errors and copied content (the latter including stuff from my own site). To be fair there are people who take out material that is in breach of copyright, but what often then happens is they re-phrase the content (not usually very well).<p>Why not use Google to look for sources that are further upstream.<p>Wikipedia is good in some areas (especially geeky stuff in general), but outside those areas, I tend to avoid it.
drewp大约 16 年前
freebase.com also extracts WP and can now return its results as rdf too.<p>dbpedia's RDF is somewhat easier to work with, but freebase results contain more sources than just WP, you can edit the data with a nice gui on freebase.com, and freebase spends a lot of effort tracking topics between updates. With dbpedia, I think you just get a snapshot of WP, even if that means your URIs from last week are dead.<p>dbpedia is open source, the freebase extractor is not.<p>The #2 use case example from dbpedia (use it to put WP data on your pages) is a big focus of freebase.com, and they have a bunch of tools to make that easy.