TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Show HN: A Vercel-like workflow for AI evals that makes sense

8 pointsby vikasnairabout 1 year ago

1 comment

vikasnairabout 1 year ago
Hi HN!<p>Most of us get how crucial AI evals are now. The thing is, almost all the eval platforms we&#x27;ve seen are clunky and don&#x27;t have any product cohesion. There&#x27;s too much manual setup and adaptation needed, which breaks developers&#x27; workflows.<p>That&#x27;s why we&#x27;re releasing a simpler workflow.<p>If you&#x27;re using GitHub, you only need to add two files to the repo (one config + one script). Then, connect your repo to Openlayer and define must-pass tests for your AI system. Once integrated, every commit triggers these tests automatically on Openlayer, ensuring continuous evaluation without extra effort.<p>We offer 100+ tests (and are always adding more), including custom tests. We&#x27;re language-agnostic, and you can customize the workflow using our CLI and REST API.<p>As a final note, you can leverage the same setup to monitor your live AI systems after you deploy them. It&#x27;s just a matter of setting some env vars in your staging&#x2F;prod environments, and your Openlayer tests will run on top of your live data and send alerts if they start failing.<p>Let us know what you think!