TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Ask HN: How are you managing LLM APIs in production?

3 点作者 lamroger将近 2 年前
Looks like LangChain has LangSmith but it&#x27;s in closed beta.<p>I saw a couple YC launches like Hegel AI.<p>I&#x27;m personally interested in deployments in small teams or teams with a lot of freedom to pick and choose their own tooling.

4 条评论

ianpurton将近 2 年前
I&#x27;m currently writing up a deployment architecture for LLM&#x27;s and the API question is answered here <a href="https:&#x2F;&#x2F;fine-tuna.com&#x2F;docs&#x2F;choosing-a-model&#x2F;model&#x2F;" rel="nofollow noreferrer">https:&#x2F;&#x2F;fine-tuna.com&#x2F;docs&#x2F;choosing-a-model&#x2F;model&#x2F;</a><p>Basically you can get a Docker container that will publish an Open AI API compatible end point. You can then choose the model that sits behind that API.<p>As deployment will be in Kuberenetes we will clusters with GPU resources to maxz out performance but we&#x27;re not there yet.
评论 #36943996 未加载
tab_jockey将近 2 年前
We built an AWS serverless app that handles:<p>- Configurable context and cases mapped to a RESTful API<p>- Multi-account and high throughput error handling<p>- DDB backed records of all requests and responses for evaluation, debugging &amp; training<p>- One-click devops deploy<p>Has helped us deploy and maintain LLM apps into production quite easily. Let me know if you would like access to the repo.
XGBoost将近 2 年前
Play around with langchain and then convert all of that into decent code. After a few prototypes, you&#x27;ll realize langchains or other pipelining are just for non-coders. You can architect elegant solutions yourself.
评论 #36966849 未加载
retrovrv将近 2 年前
Langsmith is broadly for tracing the chains - are you looking for prompt deployment solutions?
评论 #36966838 未加载