TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Deploy dedicated DeepSeek 32B on L40 GPUs ($8/hour)

19 pointsby wfalcon4 months ago

6 comments

woodr773 months ago
Everyone's saying I needed H100s for this. L40 is way easier for me to get my hands on. great news.
ashenWon4 months ago
Is this running ollama, vllm or sglang under the hood? Curious about these performance numbers.
lmilad4 months ago
How well does DeepSeek R1 handle generating long pieces of text with Qwen 32B?
tchaton844 months ago
Does it support largest Deepseek model ?
yewnork4 months ago
curious the performance / price tradeoffs between deepseek-r1 671b, 70b, 32b
neilbhatt4 months ago
nice, i can actually use my AWS start up creds