TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Benchmarking LLMs against human expert-curated biomedical knowledge graphs

41 点作者 Al0neStar大约 1 年前

4 条评论

CraftingLinks大约 1 年前
Academic writing 101: The abstract is NOT meant to be written as a cliff-hanger!
评论 #39877440 未加载
nyrikki大约 1 年前
Due to the cliffhanger abstract, here is a part from the discussion that may help.<p>&gt; In our case, the manual curation of a proportion of triples revealed that Sherpa was able to extract more triples categorized as correct or partially correct. However, when compared to the manually curated gold standard, the performance of all automated tools remains subpar.
jmugan大约 1 年前
I didn&#x27;t see UMLS in the paper, but I&#x27;ve tried some of their human-created biomedical knowledge graphs, and they were too full of errors to be used. I imagine different ones have different levels of accuracy.
egberts1大约 1 年前
i was right; LLM needs two major components added before we can swan dive into humanistic aspect of medicine&#x2F;pyschology&#x2F;politics using a form of LLM.<p>1) weighting of each statement for probability of correctness and<p>2) citation for each source.