TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Show HN: Plexe – ML Models from a Prompt

124 点作者 vaibhavdubey972 天前
Hey HN! We’re Vaibhav and Marcello. We’re building Plexe (<a href="https:&#x2F;&#x2F;github.com&#x2F;plexe-ai&#x2F;plexe">https:&#x2F;&#x2F;github.com&#x2F;plexe-ai&#x2F;plexe</a>), an open-source agent that turns natural language task descriptions into trained ML models. Here’s a video walkthrough: <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=bUwCSglhcXY" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=bUwCSglhcXY</a>.<p>There are all kinds of uses for ML models that never get realized because the process of making them is messy and convoluted. You can spend months trying to find the data, clean it, experiment with models and deploy to production, only to find out that your project has been binned for taking so long. There are many tools for “automating” ML, but it still takes teams of ML experts to actually productionize something of value. And we can’t keep throwing LLMs at every ML problem. Why use a generic 10B parameter language model, if a logistic regression trained on your data could do the job better?<p>Our light-bulb moment was that we could use LLMs to generate task-specific ML models that would be trained on one’s own data. Thanks to the emergent reasoning ability of LLMs, it is now possible to create an agentic system that might automate most of the ML lifecycle.<p>A couple of months ago, we started developing a Python library that would let you define ML models on structured data using a description of the expected behaviour. Our initial implementation arranged potential solutions into a graph, using LLMs to write plans, implement them as code, and run the resulting training script. Using simple search algorithms, the system traversed the solution space to identify and package the best model.<p>However, we ran into several limitations, as the algorithm proved brittle under edge cases, and we kept having to put patches for every minor issue in the training process. We decided to rethink the approach, throw everything out, and rebuild the tool using an agentic approach prioritising generality and flexibility. What started as a single ML engineering agent turned into an agentic ML &quot;team&quot;, with all experiments tracked and logged using MLFlow.<p>Our current implementation uses the smolagents library to define an agent hierarchy. We mapped the functionality of our previous implementation to a set of specialized agents, such as an “ML scientist” that proposes solution plans, and so on. Each agent has specialized tools, instructions, and prompt templates. To facilitate cross-agent communication, we implemented a shared memory that enables objects (datasets, code snippets, etc) to be passed across agents indirectly by referencing keys in a registry. You can find a detailed write-up on how it works here: <a href="https:&#x2F;&#x2F;github.com&#x2F;plexe-ai&#x2F;plexe&#x2F;blob&#x2F;main&#x2F;docs&#x2F;architecture&#x2F;multi-agent-system.md">https:&#x2F;&#x2F;github.com&#x2F;plexe-ai&#x2F;plexe&#x2F;blob&#x2F;main&#x2F;docs&#x2F;architectur...</a><p>Plexe’s early release is focused on predictive problems over structured data, and can be used to build models such as forecasting player injury risk in high-intensity sports, product recommendations for an e-commerce marketplace, or predicting technical indicators for algorithmic trading. Here are some examples to get you started: <a href="https:&#x2F;&#x2F;github.com&#x2F;plexe-ai&#x2F;plexe&#x2F;tree&#x2F;main&#x2F;examples">https:&#x2F;&#x2F;github.com&#x2F;plexe-ai&#x2F;plexe&#x2F;tree&#x2F;main&#x2F;examples</a><p>To get it working on your data, you can dump any CSV, parquet, etc and Plexe uses what it needs from your dataset to figure out what features it should use. In the open-source tool, it only supports adding files right now but in our platform version, we&#x27;ll have support for integrating with Postgres where it pulls all available data based on an SQL query and dumps it into a parquet file for the agent to build models.<p>Next up, we’ll be tackling more of the ML project lifecycle: we’re currently working on adding a “feature engineering agent” that focuses on the complex data transformations that are often required for data to be ready for model training. If you&#x27;re interested, check Plexe out and let us know your thoughts!

14 条评论

Stiopa2 天前
Awesome work.<p>Only watched demo, but judging from the fact there are several agent-decided steps in the whole model generation process, I think it&#x27;d be useful for Plexe to ask the user in-between if they&#x27;re happy with the plan for the next steps, so it&#x27;s more interactive and not just a single, large one-shot.<p>E.g. telling the user what features the model plans to use, and the user being able to request any changes before that step is executed.<p>Also wanted to ask how you plan to scale to more advanced (case-specific) models? I see this as a quick and easy way to get the more trivial models working especially for less ML-experienced people, but am curious what would change for more complicated models or demanding users?
评论 #43907756 未加载
thefourthchime2 天前
This is a really interesting idea! I&#x27;ll be honest, it took me a minute to really get what it was doing. The GitHub page video doesn&#x27;t play with any audio, so it&#x27;s not clear what&#x27;s happening.<p>Once I watched the video, I think I have a better understanding. One thing I would like to see is more of a breakdown of how this solves a problem that just a big model itself wouldn&#x27;t.
评论 #43906629 未加载
Oras2 天前
I like the idea of trying multiple solutions.<p>Does it decide based on data if it should make its own ML model or fine-tune a relevant one?<p>Also, does it detect issues with the training data? When I was doing NLP ML models before LLMs, the tasks that took all my time were related to data cleaning, not the training or choosing the right approach.
评论 #43907266 未加载
dweinus2 天前
I don&#x27;t want to hate, what you built is really cool and should save time in a data scientist&#x27;s workflow, but... we did this. It won&#x27;t &quot;automate most of the ML lifecycle.&quot; Back in ~2018 &quot;autoML&quot; was all the rage. It failed because creating boilerplate and training models are not the hard parts of ML. The hard parts are evaluating data quality, seeking out new data, designing features, making appropriate choices to prevent leakage, designing evaluation appropriate to the business problem, and knowing how this will all interact with the model design choices.
评论 #43909518 未加载
评论 #43907160 未加载
fzysingularity2 天前
Is there a benchmark or eval for why this might be a better approach than actually modeling the problem? If you&#x27;re selling this a non-ML person, I get the draw. But you&#x27;d still have to show why using these LLMs would be better than training it with something simpler &#x2F; more lightweight.<p>That said, it&#x27;s likely that you&#x27;ll get good zero-shot performance, so the model building phase could benefit from fine-tuning the prompt given the dataset - instead of training the underlying model itself.
评论 #43907353 未加载
marinr11 天前
hey this is very cool I work at a bank and we are starting to look at something like this mainly to automate boilerplate code for experimentation and model training however we are a GCP shop, I might play with this over the weekend to see if i can add support for vertex.ai experiments.<p>Have you thought about extending this to cover the model development lifecycle and perhaps having agents to help with EDA, model selection, explanation and feature engineering? this is where we are seeing a lot of demand from users as well but we are starting out with experiment&#x2F; pipeline &#x2F; serving boilerplate.
评论 #43917601 未加载
vessenes2 天前
I like this a lot, thank you for building it.<p>Any review of smolagent? This combination of agents approach seems likely to be really useful in a lot of places, and I’m wondering if you liked it, loved it, hated it, …
评论 #43907046 未加载
评论 #43907035 未加载
gitroom1 天前
well i actually like when folks push old school ML instead of just LLM stuff everywhere. makes me feel like we&#x27;re not losing the basics
评论 #43912309 未加载
revskill2 天前
Instead of &quot;Attention is all we need&quot;, i expect an &quot;Intention is all we need&quot;.
评论 #43907227 未加载
drlobster2 天前
That&#x27;s great. Is there anyway to make it part of a scikit-learn compatible pipeline.?
评论 #43908578 未加载
MarcoDewey2 天前
I love that you all are doing real old school machine learning and not just LLM transformer based work!
评论 #43910227 未加载
srameshc2 天前
I am just trying to understand and an honest question: Are we getting a fine tuned model from the dataset ?
评论 #43909148 未加载
yu3zhou42 天前
Nice execution! I built a simpler version of it a year ago <a href="https:&#x2F;&#x2F;github.com&#x2F;jmaczan&#x2F;csv-to-ml">https:&#x2F;&#x2F;github.com&#x2F;jmaczan&#x2F;csv-to-ml</a> I hope you succeed with the product and push the automl forward
评论 #43907174 未加载
评论 #43907279 未加载
ratatoskrt2 天前
In my experience, humans are really bad at statistics and LLMs are even worse because they basically just mimic all the typical mistakes people make.
评论 #43906860 未加载