TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

ZML - High performance AI inference stack

36 点作者 msoad8 个月前

5 条评论

ismailmaj8 个月前
What would be the benefit of using ZML instead of relying on StableHLO/PJRT? Because the cost of porting models is for sure high.
评论 #41589461 未加载
onurcel8 个月前
First of all, great job! I think the inference will become more and more important.<p>That being said, I have a question regarding the ease of use. How difficult it is for someone with python&#x2F;c++ background to get used to zig and (re)write a model to use with zml?
评论 #41568948 未加载
评论 #41568884 未加载
hsjdhdvsk8 个月前
Hi ya! Want to say this looks awesome :) really interested in the sharded inference demo!!! You said it was experimental, is it in the examples folder at all?? (On phone atm, so apologies for not investigating further)
Palmik8 个月前
Given that the focus is performance, do you have any benchmarks to compare against the likes of TensoRT-LLM.
评论 #41576897 未加载
评论 #41576945 未加载
montyanderson8 个月前
my dreams have come true. hardware-agnostic ml primitives in a typed, compiled language.<p>my only question is: is zig stable enough to base such a project on?
评论 #41576970 未加载
评论 #41574100 未加载