TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Show HN: Bodhi App – Local LLM Inference

1 点作者 anagri3 个月前
Hi HN,<p>Bodhi App is an open-source local LLM inference solution that takes a different and simpler approach. Instead of re-inventing the wheel, it leverages existing, tried and tested ecosystem and solutions:<p>## Technical Architecture:<p>- llama.cpp as inference engine<p>- Rust&#x2F;Axum backend for type-safe API layer<p>- Tauri for multiplatform builds<p>- HuggingFace integration<p>- YAML based configurations and update at runtime (no restarts required)<p>- OpenAI&#x2F;Ollama API compatibility layer<p>## Key Technical Decisions:<p>1. No proprietary model format - directly use of GGUF files from HuggingFace<p>2. Opt-in Authentication, provides RBAC for team access<p>3. API design with proper authentication&#x2F;authorization<p>4. Built-in Swagger UI with complete OpenAPI specs<p>5. Built-in User guide<p># What Sets It Apart:<p>Designed with non-technical users in mind. So it comes a basic Web-based user interface, allowing users to get started quickly with their first AI-assistant conversation.<p>## Setup Wizard:<p>- App displays a setup wizard when run for first time<p>- Allows user to download popular models in a user friendly way<p>## Built-in Chat UI:<p>- Ships with a complete Chat UI<p>- Chat UI is simple enough for non-technical users to get started with their first AI-conversation<p>- Adapts to power users by providing complete control over request settings<p>- Supports realtime streaming response, markdown rendering, code rendering with syntax highlights<p>- Displays chat stats, request tokens, response tokens, token speed<p>- Allow copying of the AI-response etc.<p>## Built-in UI for Model + App Management + API access:<p>- Manage complete Model lifecycle from the UI<p>- Downloading models, deleting models<p>- Configuring models, request + inference server configurations using Model Alias yaml files<p>- Allows configuring for parallel processing of requests<p>- Configuring App Settings - chosing betwen CPU&#x2F;GPU, server idle time etc.<p>- API tokens for authenticated&#x2F;authorized access to APIs by 3rd party<p>## Tech for UI:<p>- Uses Nextjs, Tailwindcss, Shadcn to build powerful, responsive and user friendly UI<p>- Supports Dark&#x2F;Light mode<p>- Exported using config `output: &quot;export&quot;` to export the entire frontend as static html + javascript<p>- Served by the backend as static asset<p>- Thus no packaged nodejs server, reducing app size, complexity and compute<p># Links<p>Try it out: <a href="https:&#x2F;&#x2F;getbodhi.app&#x2F;" rel="nofollow">https:&#x2F;&#x2F;getbodhi.app&#x2F;</a><p>Source: <a href="https:&#x2F;&#x2F;github.com&#x2F;BodhiSearch&#x2F;BodhiApp">https:&#x2F;&#x2F;github.com&#x2F;BodhiSearch&#x2F;BodhiApp</a><p>Looking forward to technical feedback and discussions.

暂无评论

暂无评论