TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Automated reasoning to remove LLM hallucinations

57 pointsby rustastra6 months ago

10 comments

lsy6 months ago
I find it hard to believe that anything like this will be feasible or effective beyond a certain level of complexity. It seems like a willful denial of the complexity and ambiguity of natural language, and I am not looking forward to some poor developer trying to reason their way out of a two-hundred-step paradox that was accidentally created.<p>And for a use-case simple enough for this system to work (e.g. regurgitate a policy), it seems like the LLM is unnecessary. After all, if your system can perfectly interpret the question and answer and see if this rule set applies, then you can likely just use the rule set to generate the answer rather than wasting resources with a giant language model.
评论 #42313921 未加载
评论 #42316496 未加载
评论 #42313990 未加载
Metricon6 months ago
This amuses me tremendously. I began programming in the early 1980s and quickly developed an interest in Artificial Intelligence. At the time there was a great interest in the advancement of AI by the introduction of &quot;Expert Systems&quot; (which would later play a part in the ‘Second AI Winter’).<p>What Amazon appears to have done here is use a transformers based neural network (aka LLM) to translate natural language into symbolic logic rules which are collectively used together in what could be identified as an Expert System.<p>Full Circle. Hilarious.<p>For reference to those on the younger side: The Computer Chronicles (1984) <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=_S3m0V_ZF_Q" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=_S3m0V_ZF_Q</a>
评论 #42313971 未加载
gibsonf16 months ago
If the automated reasoning worked, why would you need an LLM and its fabrications?
评论 #42314314 未加载
bloomingkales6 months ago
Just looking at this AWS workflow takes the joy out of programming for me.
评论 #42315845 未加载
pkoird6 months ago
I&#x27;ll say this again, any sufficiently advanced LLM is indistinguishable from Prolog.
评论 #42322823 未加载
majestik6 months ago
I hadn&#x27;t heard of Amazon Bedrock Guardrails before, but after reading about it, it seems similar to Nvidia NeMo Guardrails which I have heard of: <a href="https:&#x2F;&#x2F;docs.nvidia.com&#x2F;nemo&#x2F;guardrails&#x2F;introduction.html" rel="nofollow">https:&#x2F;&#x2F;docs.nvidia.com&#x2F;nemo&#x2F;guardrails&#x2F;introduction.html</a><p>The approaches seem very different though. I&#x27;m curious if anyone here has used either or both and can share feedback.
nl6 months ago
This is an interesting approach.<p>By constraining the field it is trying to solve it makes grounding the natural language question in a knowledge graph tractable.<p>An analogy is type inference in a computer language: it can&#x27;t solve every problem but it&#x27;s very useful much of the time (actually this is a lot more than an analogy because you can view a knowledge graph as an actual type system in some circumstances).
tomlockwood6 months ago
If this is necessary, LLMs have officially jumped the shark. And I do wonder how much of this &quot;necessary logic&quot; has already been added to ChatGPT and other platforms, where they&#x27;ve offloaded the creation of logic-based heuristics to Mechanical Turk participants, and like the old meme, AI unmasked is a bit of LLM and a tonne of IF, THEN statements.<p>I get the vibe VC money is being burned with promises of an AGI that may never eventuate and there&#x27;s no clear path to.
评论 #42314476 未加载
spartanatreyu6 months ago
Post title: Automated reasoning to remove LLM hallucinations<p>---<p>and yet, the paper that went around in March:<p>Paper Link: <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;2401.11817" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;2401.11817</a><p>Paper Title; Hallucination is Inevitable: An Innate Limitation of Large Language Models<p>---<p>Instead of trying to trick a bunch of people into thinking we can somehow ignore the flaws of post-LLM &quot;AI&quot; by also using the still flawed pre-LLM &quot;AI&quot;, why don&#x27;t we cut the salesman BS and just tell people not to use &quot;AI&quot; for the range of tasks it&#x27;s not suited for.
评论 #42314751 未加载
drew-y6 months ago
How does automation reasoning actually check a response against the set of rules without using ML? Wouldn&#x27;t it still need a language model to compare the response to the rule?
评论 #42313641 未加载