TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

AskHN: What's a real world production use case for Ollama?

11 pointsby eclectic29over 1 year ago
Ollama is enjoying a LOT of hype, but I'm struggling to find a real world production use case for it. Where might I really want to use this? It's a wrapper around llama.cpp and makes it easier to download LLMs. Where might I want to download models in production like this? In production I would rather deploy thoroughly tested models. Plus, the model offloading capability on the fly is really not meant for production as it hinders performance. Thoughts?

4 comments

runjakeover 1 year ago
It’s not for production. Ollama’s documentation explicitly warns you not to use the API in production. It’s for messing around and experimenting with LLMs and different models.<p>If you don’t see the use of it, I recommend you use something else that suits your tastes more.
评论 #39312132 未加载
评论 #39319606 未加载
Tomteover 1 year ago
Not everything must be about commercial use. ollama is a great way to get from &quot;I have no idea what this LLM stuff is&quot; to &quot;cool, I‘m already talking to one?&quot;.
speedgooseover 1 year ago
We use it in our research group to run the largest LLMs quickly on a descent server. However, the people with the latest MacBooks don’t bother and prefer to use their laptops as they are fast enough.<p>We also throw money at Microsoft and OpenAI when we evaluate ChatGPT. But open LLMs are cheaper to play with if you already have the hardware. And more aligned to our research values.
yen223over 1 year ago
Didn&#x27;t know Ollama was used in production. I use Ollama + the Ollama Raycast plugin as it&#x27;s probably the quickest way to run an LLM locally