TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Launch HN: Release (YC W20) – Orchestrate AI Infrastructure and Applications

73 点作者 tommy_mcclung10 个月前
Hello Hacker News! We’re Erik, Tommy, and David, the founders of Release (<a href="https:&#x2F;&#x2F;release.ai&#x2F;" rel="nofollow">https:&#x2F;&#x2F;release.ai&#x2F;</a>). We launched on HN in 2020 (<a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=22486031">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=22486031</a>) after leaving TrueCar, where we managed a 300 person development team. Our original focus was making staging environments easier with ephemeral environments, but along the way AI applications started to emerge as an important and critical component of distributed applications. As we talked to customers using our original product, we realized we had built the underlying platform needed to address the needs of orchestrating AI applications and infrastructure. So here we are and we’re excited to share Release.ai with HN.<p>Here’s a video showcasing the platform and demonstrating how to easily manage new data and changes using the RAG stack of your choice: <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=-OdWRxMX1iA" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=-OdWRxMX1iA</a><p>If you want to try release.ai out, we’re offering a sandbox account with limited free GPU cycles so you can play around and get a feel for Release.ai: <a href="https:&#x2F;&#x2F;release.ai" rel="nofollow">https:&#x2F;&#x2F;release.ai</a>. We suggest playing around with some of the RAG AI templates and adding custom workflows like in the demo video. The sandbox comes with 5 free compute hours on an Amazon g5.2xlarge instance (A10 with 24GB VRAM, 8vCPUs and 32GB). You will also get 16 GB and 4vCPUs for cpu workloads such as web servers. You will be able to run an inference engine plus things like an api server, etc.<p>After the sandbox expires, you can switch to our free plan, which requires a credit card and associating an AWS&#x2F;GCP account with Release to manage the compute in your cloud account. The free account provides 100 free managed environment hours a month. If you never go over, you never pay us anything. If you do, our pricing is here: <a href="https:&#x2F;&#x2F;release.com&#x2F;pricing">https:&#x2F;&#x2F;release.com&#x2F;pricing</a>.<p>For those that like to read more, here’s the deeper background.<p>It’s clear that open source AI and AI privacy are going to be big. Yes, many developers are going to choose SaaS offerings like OpenAI to build their AI applications, but as open source frameworks and models improve, we’re seeing a shift to open source running on cloud. Security and privacy is a top concern of companies leveraging these SaaS solutions, which forces them to look at running infrastructure themselves. That’s where we hope to come in: we’ve built Release.ai so all your data, models and infrastructure stay in your cloud account and open source frameworks are first class citizens.<p>Orchestration - Integrating AI applications into a software development workflow and orchestrating their lifecycle is a new and different challenge than traditional web application development. Release also makes it possible to manage and integrate your web and AI apps using a single application and methodology.<p>To make orchestrating AI applications easier, we built a workflow engine that can create the complex workflows that AI applications require. For example, you can automate the redeployment of an AI inference server easily when underlying data changes using webhooks and our workflow engine.<p>Cost and expertise - Managing and scaling the hardware required to run AI workloads is hard and can be incredibly expensive. Release.ai lets you manage GPU compute resources across multiple clouds with different instance&#x2F;node groups for various jobs within a single admin interface. We use K8s under the covers to pull this off. With over 5 years of building and running K8s infrastructure our customers have told us this is how it should be done.<p>Getting started with AI frameworks is time consuming and requires some pretty in-depth expertise. We built out a library of AI templates (<a href="https:&#x2F;&#x2F;docs.release.com&#x2F;release.ai&#x2F;release.ai-templates">https:&#x2F;&#x2F;docs.release.com&#x2F;release.ai&#x2F;release.ai-templates</a>) using our Application Template format (which is kind of a super docker-compose: <a href="https:&#x2F;&#x2F;docs.release.com&#x2F;reference-documentation&#x2F;application-settings&#x2F;application-template">https:&#x2F;&#x2F;docs.release.com&#x2F;reference-documentation&#x2F;application...</a>) for common open source frameworks to make it easy to get started developing AI applications. Setting up and getting these frameworks running is a hassle, so we made it one click to launch and deploy.<p>We currently have over 20 templates including temples for RAG applications, fine tuning and useful tools like Juypter notebooks, Promptfoo, etc. We worked closely with Docker and Nvidia to support their frameworks: GenAI and Nvidia NEMO&#x2F;Nims. We plan to launch community templates soon after launch. If you have suggestions for more templates we should support, please let us know in the comments.<p>We’re thrilled to share Release.ai with you and would love to get your feedback. We hope you’ll try it out, and please let us know what you think!

17 条评论

JoeCortopassi10 个月前
I&#x27;ve noticed that while a bunch of developers have played with LLM&#x27;s for toy projects, few seem to have any actual experience taking it to prod in front of real users. I’ve personally had to do so for a few startups, and it&#x27;s like trying to nail Jell-O to a tree. Every random thing you change, from prompts to models, yields massively different&#x2F;unpredictable results.<p>I think because of this, a bunch of companies&#x2F;tools have tried to hop in this space and promised the world, but often times people are best served by just hitting OpenAI&#x2F;GPT directly, and jiggling the results until they get what they want. If you&#x27;re not comfortable doing that, there are even companies that do that for you, so you can just focus on the prompt itself.<p>So that being said, help me understand why I should be adding this whole system&#x2F;process to my workflow, versus just hitting OpenAI&#x2F;Anthropic&#x2F;Google directly?
评论 #41185326 未加载
评论 #41189927 未加载
BurritoKing10 个月前
This looks awesome, getting started with AI development is daunting and I really like how this focuses on integrating with a bunch of open source frameworks and then deploying them into your own cloud (I always prefer to run the infrastructure, it feels weird to rely on something that&#x27;s a complete black box).<p>The sandbox environment with free GPU hours is a cool way to try things out without a big commitment too. It&#x27;s nice seeing a product that genuinely seems to address the practical challenges of AI deployment. Looking forward to seeing how the platform develops!
评论 #41183982 未加载
bradhe10 个月前
Super interesting you guys have been working on this since 2020 if I&#x27;m reading the post title correctly? Would love to know the iterations you&#x27;ve gone through.
评论 #41185872 未加载
todd383410 个月前
This is very cool! I love seeing tooling targeting inference. I feel like stable diffusion and LLAMA have to be the primary use cases for these types of services. DALL-E is super lacking and GPT does actually start to get pretty expensive once you are using it in production.
michaelmior10 个月前
This looks cool, but I&#x27;m a little confused about the pricing model. It sounds like I&#x27;m paying you for every hour my jobs are running on my own infrastructure if I&#x27;m reading it right. That seems like a really odd way to price things if true.
评论 #41184473 未加载
评论 #41184375 未加载
the_pascal10 个月前
How does this compare to managed offerings like Google Gemini and AWS Bedrock? Thanks in advance and congratulations on the new product!!
评论 #41182560 未加载
mcsplat210 个月前
How do you hook up data to an environment? And what data sources do you support (Snowflake&#x2F;etc?)
评论 #41183785 未加载
mchiang10 个月前
This is cool. I&#x27;d like to give it a try. Press a button, and get GPU access to build apps on.
评论 #41183645 未加载
tommy_mcclung10 个月前
We might not have made it clear in the post how to signup for the sandbox. Just head to <a href="http:&#x2F;&#x2F;release.ai" rel="nofollow">http:&#x2F;&#x2F;release.ai</a> and click on &quot;Start Free Trial&quot;.
评论 #41192428 未加载
jakubmazanec10 个月前
Why do you have such generic name? It will make searching so much harder.
评论 #41183856 未加载
bluelightning2k10 个月前
Seems pretty vague. Something to do with half self-hosting open source LLMs with a proprietary docker-like template thing.<p>Am I on the right track?
评论 #41183954 未加载
评论 #41185152 未加载
nextworddev10 个月前
Pretty insane that these folks are YC 2020 and just now pivoted again. Shows how hard company building is.
评论 #41193992 未加载
sidcool10 个月前
Congrats on launching. Interesting pivot.
评论 #41183628 未加载
jodotgiff10 个月前
Great system, no complaints!!
drawnwren10 个月前
I&#x27;m pretty much exactly your target market. I run a kubernetes, docker, poetry devops hell at an ml startup and was curious how your product helped. You got about 2 minutes of my time. I scanned your website, I have no idea what you do or whether you fix my problem.<p>Not trying to be negative, but I think there may be a 30 second pitch to be made to people like me that isn&#x27;t made on your site.
评论 #41184954 未加载
billclerico10 个月前
congrats!!
评论 #41183591 未加载
bijutoha10 个月前
What has been the feedback from early users regarding the ease of transitioning from your original product focus to the current AI orchestration platform?
评论 #41194003 未加载