TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Show HN: In-Browser Graph RAG with Kuzu-WASM and WebLLM

158 点作者 sdht02 个月前
We show the potential of modern, embedded graph databases in the browser by demonstrating a fully in-browser chatbot that can perform Graph RAG using Kuzu (the graph database we&#x27;re building) and WebLLM, a popular in-browser inference engine for LLMs. The post retrieves from the graph via a Text-to-Cypher pipeline that translates a user question into a Cypher query, and the LLM uses the retrieved results to synthesize a response. As LLMs get better, and WebGPU and Wasm64 become more widely adopted, we expect to be able to do more and more in the browser in combination with LLMs, so a lot of the performance limitations we see currently may not be as much of a problem in the future.<p>We will soon also be releasing a vector index as part of Kuzu that you can also use in the browser to build traditional RAG or Graph RAG that retrieves from both vectors and graphs. The system has come a long way since we open sourced it about 2 years ago, so please give us feedback about how it can be more useful!

10 条评论

willguest2 个月前
I absolutely love this. I make VR experiences that run on the ICP, which delivers wasm modules as smart contracts - I&#x27;ve been waiting for a combo of node-friendly, wasm deployable tools and webLLM. The ICP essentially facilitates self-hosting of data and provides consensus protocols for secure messaging and transactions.<p>This will make it super easy for me to add LLM functionality to existing webxr spaces, and I&#x27;m excited to see how an intelligent avatar or convo between them will play out. This is, very likely, the thing that will make this possible :)<p>If anyone wants to collab, or contribute in some way, I&#x27;m open to ideas and support. Search for &#x27;exeud&#x27; to find more info
评论 #43325678 未加载
esafak2 个月前
The example is not ideal for showcasing a graph analytics database because they could have used a traditional relational database to answer the same query, <i>Which of my contacts work at Google?</i>
评论 #43322132 未加载
评论 #43323540 未加载
nattaylor2 个月前
This is very cool. Kuzu has a ton of great blog content on all the ways they make Kuzu light and fast. WebLMM (or in the future chrome.ai.* etc) + embedded graph could make for some great UXes<p>At one time I thought I read that there was a project to embed Kuzu into DuckDB, but bringing a vector store natively into kuzu sounds even better.
评论 #43322704 未加载
mentalgear2 个月前
Nice! You might also want to check out Orama - which is also an open-source hybrid vector&#x2F;full text search engine for any js runtime.
canadiantim2 个月前
Could it be viable to have one or multiple kuzu databases per user? What’s the story like for backups with kuzu?<p>I saw you recently integrated FTS which is very exciting. I love everything about Kuzu and want to use it, but currently tempted to use Turso to allow for multiple sqlite dbs per user (eg one for each device).<p>Or would it be possible to use Kuzu to query data stored on sqlite?<p>Great work through and through tho. Really amazing to see the progress you’ve all made!
评论 #43457525 未加载
jasonthorsness2 个月前
Don&#x27;t the resource requirements from even small LLMs exclude most devices&#x2F;users from being able to use stuff like this?
评论 #43322681 未加载
nsonha2 个月前
Could someone please explain in-browser inference to me? So in the context of OpenAI usage (WebLLM github), this means I will send binary to OpenAI instead of text? And it will lower the cost and run faster?
评论 #43326892 未加载
DavidPP2 个月前
I&#x27;m new to the world of graph, and I just started building with SurrealDB in embedded mode.<p>If you don&#x27;t mind taking a few minutes, what are the main reasons to use Kuzu instead?
评论 #43327465 未加载
srameshc2 个月前
I heard about it for the first time, an embedable graph database Kuzu and even better the WASM mix and LLM.
itissid2 个月前
Since I already have a browser connected to the Internet where this would execute, could one have the option of transparently executing the webGPU + LLM in a cloud container communicating with the browser process?
评论 #43328911 未加载