Hey HN! Long-time reader, occasional commenter here. I wanted to share something my team and I have been building to solve our own frustrations.<p>We've all seen the explosion of prompt engineering tools lately. While playing around in playgrounds is fun, when we tried to take our AI prompts to production, we hit a wall. I'm guessing many of you have experienced similar pain points.<p>We kept hitting questions nobody seemed to be answering: How do you version control thousands of prompts? How do you handle multiple production deployments? How do you scale from prototype to millions of requests per day? How do you collaborate across hundreds of engineers without stepping on each other's toes?<p>So we built Portkey's Prompt Engineering Studio - a complete toolkit designed specifically for productionizing AI prompts across 1600+ models.<p>Some technical details that make our approach different:<p>- High-performance infrastructure: We've deployed prompts as large as 500,000 tokens with production-level latency
- Git-like version control with instant rollbacks for prompt deployments
- Mustache templating system for parameterization and reusable snippets
- Publish/release flow with proper dev/staging/prod environments
- Real-time analytics tracking prompt performance, latency, and token usage
- Native integrations with Langchain, Llamaindex, and Promptfoo<p>The scaling capabilities have enabled some impressive use cases:<p>- A content company running 500+ prompts across 700+ websites
- A tech firm that cut deployment times from 3 days to near-instant
- Education platforms with hundreds of non-technical creators building AI workflows<p>Our platform has processed hundreds of millions of prompt completion requests already, with over 10,000 prompts deployed to production environments.<p>We think the HN community will especially appreciate our approach to bringing software engineering best practices to AI development!<p>You can try it yourself at prompt.new<p>I'd genuinely love to hear how others in the community are handling these challenges, what you think of our approach, or any other feedback you might have. This community has been invaluable in shaping how we think about developer tools.
This looks like a powerful step toward making prompt engineering more scalable and production-ready. The version control approach, along with staging environments and real-time analytics, seems particularly useful for teams handling high-volume AI workloads.<p>One question: How do you handle prompt drift over time? As models evolve, prompt effectiveness can degrade—do you provide any automated testing or monitoring to detect when a deployed prompt needs adjustment?<p>Looking forward to exploring Portkey’s capabilities.
> How do you version control thousands of prompts?<p>Kill me now.<p>> How do you collaborate across hundreds of engineers<p>What do you mean by that? This only targets a few big companies.<p>> A tech firm that cut deployment times from 3 days to near-instant<p>That's a process and maybe CI issue, I don't see how AI would improve any of that but I'll be gladly proven wrong.<p>> You can try it yourself at prompt.new<p>All I see is a login page from another company. Don't you have a web site with all those serious prompting you do?