TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

DeepRAG: Thinking to retrieval step by step for large language models

191 点作者 fofoz3 个月前

3 条评论

mkw50533 个月前
This reminds me of the Agent Workflow Memory (AWM) paper [1], which also tries to find optimal decision paths for LLM-based agents but relies on in-context learning, whereas DeepRAG fine-tunes models to decide when to retrieve external knowledge.<p>I’ve been thinking about how modifying AWM to use fine-tuning or an external knowledge system (RAG) might work—capturing the ‘good’ workflows it discovers rather than relying purely on prompting.<p>[1] <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2409.07429" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2409.07429</a> - Agent Workflow Memory (Wang et al., 2024)
brunohaid3 个月前
Noice!<p>Does anyone have a good recommendation for a local dev setup that does something similar with available tools? Ie incorporates a bunch of PDFs (~10,000 pages of datasheets) and other docs, as well as a curl style importer?<p>Trying to wean myself off the next tech molochs, ideally with local functionality similar to OpenAIs Search + Reason, and gave up on Langchain during my first attempt 6 months ago.
评论 #42938121 未加载
评论 #42936683 未加载
评论 #42935700 未加载
评论 #42936759 未加载
评论 #42942761 未加载
jondwillis3 个月前
The title reads awkwardly to a native English speaker. A search of the PDF for &quot;latency&quot; returns one result, discussing how naive RAG can result in latency. What are the latency impacts and other trade-offs to achieve the claimed &quot;[improved] answer accuracy by 21.99%&quot;? Is there any way that I could replicate these results without having to write my own implementation?