TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Ask HN: Where do I begin studying to run LLMs locally or in a private cloud?

34 点作者 bulldog13大约 2 年前
Where do I begin studying if my goal is to run LLMs locally or in a private cloud with solid software engineering practices? I am unsure of where to even begin.<p>Let&#x27;s say I want to:<p>0.) Find the locally run LLMs and identify which are applicable. 1.) Containerize the LLMs 2.) Use source control to capture changes to the LLM. Versioning output. 3.) Develop repeatable pipelines driven by APIs for sending data to it. 4.) Prompt engineering. 5.) Best ways to use langchain (or others) to make the system data-aware and agentic.<p>Any thoughts are appreciated.

2 条评论

smoldesu大约 2 年前
A lot of it will revolve around Nvidia hardware that you either own or rent. I&#x27;ve built CPU-accelerated AI bots on free VPSes before, but it&#x27;s slow and not a reflection of best-practices nor state-of-the-art inferencing. Right now, a lot of the meaningful &quot;private cloud&quot; AI stuff is built with extremely proprietary runtimes.
ftxbro大约 2 年前
ask GPT-4