TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Ask HN: Best platform for self-hosting LLM Models?

8 点作者 Kalpeshbhalekar大约 1 年前

3 条评论

verdverm大约 1 年前
This generally means running a GPU all the time. My personal preference is to use my preferred cloud (GCP).<p>FWIW, I&#x27;m using the VertexAI API rather than running an LLM all the time. They have data privacy in the ToS, so I&#x27;m not worried about them training on my data. It&#x27;s far cheaper and better than running a lower quality model myself. When I get around to some fine-tunings, they have options, but you can get pretty far with prompts, RAG, and agents
runjake大约 1 年前
Best platform would be hellishly expensive cloud compute or an expensive PC with a beefy, expensive Nvidia GPU, all of which generates a lot of heat.<p>I personally get a lot of mileage out of an M3 Max with 36gb memory.
fragmede大约 1 年前
the three choices are the cloud, a gaming rig, or a Mac. how much money do you have right now, how much money do you have for upkeep, how much time do you have for it, how much do you pay for power; it&#x27;s a broad question!<p>best is hard to definite so it depends on what you and your needs.