TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Mayo Clinic's secret weapon against AI hallucinations: Reverse RAG in action

187 pointsby ohjeez2 months ago

12 comments

mdp20212 months ago
The article is a week old, it was already submitted a few days ago, and the problem remains about finding some paper to shed more light into the practice.<p>A blog article* came out yesterday - but it is not immediately clear whether the author wrote what he understood, or whether he knows more.<p>But much perplexity remains: «summarization, which is what LLMs generally excel at» (original article); «The LLM ... reads the patient’s records ... and produces a summary or list of facts» (blog). This is possibly the beginning, and some of us will already be scared - as the summarization capabilities we experienced from LLMs were neither intelligent nor reliable. (...Or did new studies come up and determine that LLMs have become finally reliable, if not cognitively proficient, in summarization?)<p>* <a href="https:&#x2F;&#x2F;usmanshaheen.wordpress.com&#x2F;2025&#x2F;03&#x2F;14&#x2F;reverse-rag-reduce-hallucinations-and-errors-in-medical-genai-part-1&#x2F;" rel="nofollow">https:&#x2F;&#x2F;usmanshaheen.wordpress.com&#x2F;2025&#x2F;03&#x2F;14&#x2F;reverse-rag-re...</a>
评论 #43374646 未加载
wetherbeei2 months ago
Sounds similar to <a href="https:&#x2F;&#x2F;cloud.google.com&#x2F;generative-ai-app-builder&#x2F;docs&#x2F;check-grounding" rel="nofollow">https:&#x2F;&#x2F;cloud.google.com&#x2F;generative-ai-app-builder&#x2F;docs&#x2F;chec...</a>.<p>&quot;The check grounding API returns an overall support score of 0 to 1, which indicates how much the answer candidate agrees with the given facts. The response also includes citations to the facts supporting each claim in the answer candidate.<p>Perfect grounding requires that every claim in the answer candidate must be supported by one or more of the given facts. In other words, the claim is wholly entailed by the facts. If the claim is only partially entailed, it is not considered grounded.&quot;<p>There&#x27;s an example input and grounded output scores that shows how the model splits into claims, decides if the claim needs grounding, and the resulting entailment score for that claim in: <a href="https:&#x2F;&#x2F;cloud.google.com&#x2F;generative-ai-app-builder&#x2F;docs&#x2F;check-grounding#claim-level-score-response-examples" rel="nofollow">https:&#x2F;&#x2F;cloud.google.com&#x2F;generative-ai-app-builder&#x2F;docs&#x2F;chec...</a>
评论 #43381036 未加载
hn_throwaway_992 months ago
Can someone more versed in the field comment on whether this is just an ad or actually something unique or novel.<p>What they&#x27;re describing as &quot;reverse RAG&quot; sounds a lot to me like &quot;RAG with citations&quot;, which is a common technique. Am I misunderstanding?
评论 #43373411 未加载
评论 #43373189 未加载
评论 #43387216 未加载
评论 #43372010 未加载
评论 #43372440 未加载
评论 #43376024 未加载
评论 #43372536 未加载
评论 #43373819 未加载
评论 #43372672 未加载
评论 #43372275 未加载
评论 #43373541 未加载
natnat2 months ago
Can someone link to a real source for this? Like, a paper or something? This seems very interesting and important and I&#x27;d prefer to look at something less sketchy than venturebeat.com
评论 #43374044 未加载
ttpphd2 months ago
If LLMs were good at summarization, this wouldn&#x27;t be necessary. Turns out a stochastic model of language is not a summary in the way humans think of summaries. Thus all this extra faff.
评论 #43373879 未加载
beebaween2 months ago
Curious if anyone has attempted this in an open source context? Would be incredibly interested to see an example in the wild that can point back to pages of a PDF etc!
评论 #43339285 未加载
评论 #43338388 未加载
评论 #43339448 未加载
isoprophlex2 months ago
If only we could understand the actual mechanism involved in &quot;reverse RAG&quot;... was anyone able to find anything on this beyond the fuzzy details in tfa?
gloosx2 months ago
Hmm, 90 minutes of bureaucracy for practitioners every day, data can be extremely difficult to find and parse out, how can I augment the abilities and simplify the work of the physician?<p>Let&#x27;s smack a hallucinating LLM and try to figure out how to make it hallucinate less... Genius
mmooss2 months ago
&gt; A second LLM then scored how well the facts aligned with those sources, specifically if there was a causal relationship between the two.<p>What is &#x27;causal&#x27; about it? Maybe I&#x27;m reading one word too closely, but an accurate citation or summary isn&#x27;t a matter of cause and effect?
评论 #43373966 未加载
unification_fan2 months ago
That&#x27;s like trying to stop a hemorrhage with a band-aid<p>Daily reminder that traditional AI expert systems from the 60s have 0 problems with hallucinations by virtue of their own architecture<p>Why we aren&#x27;t building LLMs on top of ProbLog is a complete mystery to me (jk; it&#x27;s because 90% of the people who work in AI right now have never heard of it; because they got into the field through statistics instead of logic, and all they know is how to mash matrices together).<p>Clearly language by itself doesn&#x27;t cut it, you need some way to enforce logical rigor and capabilities such as backtracking if you care about getting an <i>explainable</i> answer out of the black box. Like we were doing 60 years ago before we suddenly forgot in favor of throwing teraflops at matrices.<p>If Prolog is Qt or, hell, even ncurses; then LLMs are basically Electron. They get the job done, but they&#x27;re horribly inefficient and they&#x27;re clearly not the best tool for the task. But inexperienced developers think that LLMs are this amazing oracle that solves every problem in the world, and so they throw LLMs at anything that vaguely looks like a problem.
评论 #43372411 未加载
评论 #43372090 未加载
评论 #43372214 未加载
评论 #43374872 未加载
评论 #43372096 未加载
评论 #43372542 未加载
评论 #43372367 未加载
评论 #43372419 未加载
评论 #43371997 未加载
tekacs2 months ago
This is very interesting, but it&#x27;s so perfect that the Mayo Clinic gets to use an algorithm called CURE, of all things.
评论 #43372689 未加载
评论 #43373496 未加载
htrp2 months ago
at that point it becomes a search problem?
评论 #43373108 未加载