TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

KnowledgeNet: A Benchmark for Knowledge Base Population

33 点作者 miket超过 5 年前

5 条评论

miket超过 5 年前
When people think about using computers for Natural Language Processing, they often think about end-tasks like classification, translation, question answering, and models like BERT that model the statistical regularities in text. However, these tasks only measure indirectly how much the system has understood the meaning of the text, are largely unexplainable black boxes, and require reams of training data.<p>NLP is good enough that we can now explicitly measure how well a system reads text in terms of what knowledge is extracted from it. This task is called Knowledge Base Population, and we&#x27;ve released the first reproducible dataset called KnowledgeNet that measures this task, along with an open source state-of-the-art baseline.<p>Direct link to the Github repo: <a href="https:&#x2F;&#x2F;github.com&#x2F;diffbot&#x2F;knowledge-net" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;diffbot&#x2F;knowledge-net</a> EMNLP paper: <a href="https:&#x2F;&#x2F;www.aclweb.org&#x2F;anthology&#x2F;D19-1069.pdf" rel="nofollow">https:&#x2F;&#x2F;www.aclweb.org&#x2F;anthology&#x2F;D19-1069.pdf</a>
g82918超过 5 年前
Mostly an article pushing their benchmark and article: <a href="https:&#x2F;&#x2F;www.aclweb.org&#x2F;anthology&#x2F;D19-1069.pdf" rel="nofollow">https:&#x2F;&#x2F;www.aclweb.org&#x2F;anthology&#x2F;D19-1069.pdf</a>. In the article they compare existing benchmarks against a criteria they create to show their benchmark is the only one that features the things they say are important. All the others are somehow deficient by the totally objective metric they create.
bhl超过 5 年前
Reminds of a submission from a year ago on autogenerating knowledge base from articles from the web [1]. I think it&#x27;d be neat if Q&amp;A Nets and other techniques sufficed to the point where we would prefer using &quot;knowledge engines&quot; over search engines, like a generalized Wolfram Alpha.<p>[1] <a href="https:&#x2F;&#x2F;primer.ai&#x2F;blog&#x2F;quicksilver" rel="nofollow">https:&#x2F;&#x2F;primer.ai&#x2F;blog&#x2F;quicksilver</a>
nl超过 5 年前
<i>State-of-the-art models (using BERT) are far from achieving human performance (0.504 vs 0.822).</i><p>This is moderately surprising.<p>In question answering (QA) style tasks (SQUAD, SQUAD 2) we see state of the art models approach human performance. QA is similar to KBC in the sense that the answers are usually extracted from text in a similar way.<p>I&#x27;d imaging there is potential for fairly rapid improvement in this (Knowledge Base Population) task.
评论 #22071720 未加载
sdan超过 5 年前
Amazing! Love using Diffbot and although I&#x27;m not too deep into the NLP space yet, finding the relations of the text itself is a pretty important task.