TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Gemma 2 on AWS Lambda with Llamafile

10 pointsby metaskills11 months ago

2 comments

metaskills11 months ago
A small experiment to see if we are there yet with highly virtualized CPU compute and Small Language Models (SLM). The answer is a resounding maybe, but most likely not. Huge thanks to Justine for her work on Llamafile supported by Mozilla. Hope folks find this R&D useful.
评论 #40898043 未加载
xhkkffbf11 months ago
This is great work. Has anyone used it enough to compare the lambda costs with the cost of running a comparable model on, say, OpenAI?