TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Ask HN: Best way to deploy an LLM for 100s of user

13 pointsby sujayk_3310 months ago
If you have any experience of deploying an LLM what resources&#x2F;tools did you use?<p>do you have a link to your project?<p>Anything will help.

3 comments

eschnou10 months ago
Could you detail what you mean by deploying LLMs ? Is it about integrating commercial LLMs in an enterprise context? Or running self-hosted LLM for a small company (e.g. Ollama + Ollama Web UI)? Or integrating an Agentic approach to existing software stack?
评论 #40882157 未加载
gardnr10 months ago
Not enough info.<p>Do they want near-realtime responses? Will they all hit it at the same time? Can you put some workloads in an overnight batch queue?
K0IN10 months ago
so for fast responses we usea a rtx4090 with vllm, but yeah it depends on your use case
评论 #40885806 未加载