TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Ask HN: How do you iterate on and manage prompts in production?

14 pointsby ivanpashenko11 months ago
I&#x27;m curious about how people handle managing and iterating on their prompts, especially for production-facing prompts that many users depend on.<p>- Where do you store your prompts? - Do you use version control? - How do you test prompts after editing? - Where do you store your test sets? - Do you evaluate results? If so, how? - Are you fine-tuning models like GPT-3.5 for better&#x2F;cheaper results?

2 comments

BWStearns11 months ago
We have a script and a input library that has a bunch of scoring dimensions and allows a head to head comparison of a new candidate prompt vs what&#x27;s in prod. It takes a configuration (prompt, which LLM to use, temperature etc.) and then gets run with all the various inputs and makes a json blob of the outputs for scoring.<p>Most of the score dimensions are deterministic but we&#x27;ve added some where we integrated an LLM to do the scoring (... which brings the new problem of scoring the scoring prompt!). We also do a manual scan of the outputs to sanity check. Not doing any fine tuning yet as we&#x27;re getting pretty good results with just prompting.
评论 #40667251 未加载
ivanpashenko11 months ago
&quot;7 likes &#x2F; no comments&quot; --&gt; should I read it as: people interested in others people experience, but have nothing to share about their own? - No prompt on production? - No testing or other routines about it yet?<p>Please share your current status :)