Hi HN,<p>We're excited to open source our AI Proxy which supports Mistral, LLaMa2, OpenAI, Azure, Anthropic, and more through vanilla OpenAI SDKs. The proxy also supports configurable caching, API key management, and load balancing across multiple providers.<p>The proxy code also includes multiple deployment options: Cloudflare workers, Vercel, AWS Lambda, and plain-old Express. We're open sourcing this because our customers are starting to run their production workloads through it, and we believe that critical-path-of-production tools should be open source.<p>To play around with a hosted version of the proxy, you can simply set the base URL in your OpenAI libs to <a href="https://braintrustproxy.com/v1" rel="nofollow noreferrer">https://braintrustproxy.com/v1</a> (more detailed instructions in the repo[1] and our docs[2]).<p>We'd love your feedback -- on use cases we may not have thought of, models we should support, or other features that would be useful to include in this layer of abstraction.<p>[1]: <a href="https://github.com/braintrustdata/braintrust-proxy">https://github.com/braintrustdata/braintrust-proxy</a>
[2]: <a href="http://braintrustdata.com/docs/guides/proxy" rel="nofollow noreferrer">http://braintrustdata.com/docs/guides/proxy</a>