TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Show HN: YPerf – Monitor LLM Inference API Performance

2 点作者 xjconlyme5 个月前
Our team operates several real-time AI applications, where both latency(TTFT) and throughput(TPS) are critical to most of our users. Unfortunately, nearly all of the major LLM APIs lack consistent stability.<p>To address this, I developed YPerf—a simple webpage designed to monitor the performance of inference APIs. I hope it helps you select better models and discover new trending ones as well.<p>The data is sourced from OpenRouter, an excellent provider that aggregates LLM API services.

1 comment

Oras5 个月前
Nice one. It would be great to use filtering. For example, I want to check the TPS of Llama 3.3 across multiple providers.
评论 #42493552 未加载