For those curious: Humanloop is a evals platform for building products with LLMs. We think of it as the platform for 'eval-driven development' needed for making AI products/features/experiences that work well<p>We learned three key things building evaluation tools for AI teams like Duolingo and Gusto:<p>- Most teams start by tweaking prompts without measuring impact<p>- Successful products establish clear quality metrics first<p>- Teams need both engineers and domain experts collaborating on prompts<p>One detail we cut from the post: the highest-performing teams treat prompts like versioned code, running automated eval suites before any production deployment. This catches most regressions before they reach users.