Hi HN! We’re been working hard on this low-code tool for rapid prompt discovery, robustness testing and LLM evaluation. We’ve just released documentation to help new users learn how to use it and what it can already do. Let us know what you think! :)
I think you should probably mention that its source is available! [0]<p>I don't personally have a need for this right now, but I can really see the use for the parameterised queries, as well as comparisons across models.<p>Thanks for your efforts!<p>0: <a href="https://github.com/ianarawjo/ChainForge">https://github.com/ianarawjo/ChainForge</a>
This looks excellent! It's a great interface for two things I'm struggling to make LlamaIndex do: explain and debug multi-step responses for agent flows, and cache queries aggressively. If I can work out how to hook it into my LlamaIndex-based pile, happy days.<p>Feature/guidance request: how to actually call functions, how to loop on responses to resolve multiple function calls. I've managed to mock a response to get_current_weather using this contraption: <a href="https://pasteboard.co/aO9BmHG5qsFt.png" rel="nofollow noreferrer">https://pasteboard.co/aO9BmHG5qsFt.png</a> . But it's messy and I can't see a way to actually evaluate function calls. And if I involve the Chat Turn node, the message sequences seem to get tangled with each other. Probably I'm holding it wrong!