TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Show HN: I built an open-source data pipeline tool in Go

200 pointsby karakanb5 months ago
Every data pipeline job I had to tackle required quite a few components to set up:<p>- One tool to ingest data<p>- Another one to transform it<p>- If you wanted to run Python, set up an orchestrator<p>- If you need to check the data, a data quality tool<p>Let alone this being hard to set up and taking time, it is also pretty high-maintenance. I had to do a lot of infra work, and while this being billable hours for me I didn’t enjoy the work at all. For some parts of it, there were nice solutions like dbt, but in the end for an end-to-end workflow, it didn’t work. That’s why I decided to build an end-to-end solution that could take care of data ingestion, transformation, and Python stuff. Initially, it was just for our own usage, but in the end, we thought this could be a useful tool for everyone.<p>In its core, Bruin is a data framework that consists of a CLI application written in Golang, and a VS Code extension that supports it with a local UI.<p>Bruin supports quite a few stuff:<p>- Data ingestion using ingestr (<a href="https:&#x2F;&#x2F;github.com&#x2F;bruin-data&#x2F;ingestr">https:&#x2F;&#x2F;github.com&#x2F;bruin-data&#x2F;ingestr</a>)<p>- Data transformation in SQL &amp; Python, similar to dbt<p>- Python env management using uv<p>- Built-in data quality checks<p>- Secrets management<p>- Query validation &amp; SQL parsing<p>- Built-in templates for common scenarios, e.g. Shopify, Notion, Gorgias, BigQuery, etc<p>This means that you can write end-to-end pipelines within the same framework and get it running with a single command. You can run it on your own computer, on GitHub Actions, or in an EC2 instance somewhere. Using the templates, you can also have ready-to-go pipelines with modeled data for your data warehouse in seconds.<p>It includes an open-source VS Code extension as well, which allows working with the data pipelines locally, in a more visual way. The resulting changes are all in code, which means everything is version-controlled regardless, it just adds a nice layer.<p>Bruin can run SQL, Python, and data ingestion workflows, as well as quality checks. For Python stuff, we use the awesome (and it really is awesome!) uv under the hood, install dependencies in an isolated environment, and install and manage the Python versions locally, all in a cross-platform way. Then in order to manage data uploads to the data warehouse, it uses dlt under the hood to upload the data to the destination. It also uses Arrow’s memory-mapped files to easily access the data between the processes before uploading them to the destination.<p>We went with Golang because of its speed and strong concurrency primitives, but more importantly, I knew Go better than the other languages available to me and I enjoy writing Go, so there’s also that.<p>We had a small pool of beta testers for quite some time and I am really excited to launch Bruin CLI to the rest of the world and get feedback from you all. I know it is not often to build data tooling in Go but I believe we found ourselves in a nice spot in terms of features, speed, and stability.<p><a href="https:&#x2F;&#x2F;github.com&#x2F;bruin-data&#x2F;bruin">https:&#x2F;&#x2F;github.com&#x2F;bruin-data&#x2F;bruin</a><p>I’d love to hear your feedback and learn more about how we can make data pipelines easier and better to work with, looking forward to your thoughts!<p>Best, Burak

21 comments

NortySpock5 months ago
Interesting, I&#x27;ve been looking for a system &#x2F; tool that acknowledges that a dbt transformation pipeline tends to be joined-at-the-hip with the data ingestion mode....<p>As I read through the documentation, Do you have a mode in ingstr that lets you specify the maximum lateness of a file? (For late-arriving rows or files or backfills) I didn&#x27;t see it in my brief read through.<p><a href="https:&#x2F;&#x2F;bruin-data.github.io&#x2F;bruin&#x2F;assets&#x2F;ingestr.html" rel="nofollow">https:&#x2F;&#x2F;bruin-data.github.io&#x2F;bruin&#x2F;assets&#x2F;ingestr.html</a><p>Reminds me a bit of Benthos &#x2F; Bento &#x2F; RedPanda Connect (in a good way)<p>Interested to kick the tires on this (compared to, say, Python dlt)
评论 #42445720 未加载
peterm45 months ago
I&#x27;d absolutely love to love this.<p>Using dbt at $JOB, and building a custom dbt adapter for our legacy data repos, I&#x27;ve slowly developed a difficult relationship dbt&#x27;s internals and externals. Struggling with the way it (python) handles concurrency, threading, timeouts with long running (4hr+ jobs), and the like. Not to mention inconsistencies with the way it handles Jinja in config files vs SQL files. Also it&#x27;s lack of ingestion handling and VSCode&#x2F;editor support, which it seems like Bruin considers very well! Since starting poking around on the inside of dbt I&#x27;ve felt like Go or Rust would be a far more suitable platform for a pipeline building tool, and this looks to be going in a great direction, so congrats on the launch and best of luck with your cloud offering.<p>That being said, I tried starting the example bruin pipeline with duckdb on a current data project, and I&#x27;m having no luck getting the connection to appear with `bruin connections list` so nothing will run. So looks like I&#x27;m going to have to stick with dbt for now. Might be worth adding some more documentation around the .bruin.yml file; dbt has great documentation listing the purpose and layout of each file in the folder which is very helpful when trying to set things up.
评论 #42461109 未加载
评论 #42470223 未加载
jmccarthy5 months ago
Burak - one wish I&#x27;ve had recently is for a &quot;py data ecosystem compiler&quot;, specifically one which allows me to express structures and transformations in dbt and Ibis, but not rely on Python at runtime. [Go|Rust]+[DuckDB|chDB|DataFusion] for the runtime. Bruin seems very close to the mark! Following.
评论 #42444864 未加载
halfcat5 months ago
I always thought Hamilton [1] does a good job of giving enough visual hooks that draw you in.<p>I also noticed this pattern where library authors sometimes do a bit extra in terms of discussing and even promoting their competitors, and it makes me trust them more. A <i>“heres why ours is better and everyone else sucks …”</i> section always comes across as the infomercial character who is having quite a hard time peeling an apple to the point you wonder if this the first time they’ve used hands.<p>One thing wish for is a tool that’s essentially just Celery that doesn’t require a message broker (and can just use a database), and which is supported on Windows. There’s always a handful of edge cases where we’re pulling data from an old 32-bit system on Windows. And basically every system has some not-quite-ergonomic workaround that’s as much work as if you’d just built it yourself.<p>It seems like it’s just sending a JSON message over a queue or HTTP API and the worker receives it and runs the task. Maybe it’s way harder than I’m envisioning (but I don’t think so because I’ve already written most of it).<p>I guess that’s one thing I’m not clear on with Bruin, can I run workers if different physical locations and have them carry out the tasks in the right order? Or is this more of a centralized thing (meaning even if its K8s or Dask or Ray, those are all run in a cluster which happens to be distributed, but they’re all machines sitting in the same subnet, which isn’t the definition of a “distributed task” I’m going for.<p>[1] <a href="https:&#x2F;&#x2F;github.com&#x2F;DAGWorks-Inc&#x2F;hamilton">https:&#x2F;&#x2F;github.com&#x2F;DAGWorks-Inc&#x2F;hamilton</a>
评论 #42450179 未加载
thruflo5 months ago
It’s pretty remarkable what Bruin brings together into a single tool &#x2F; workflow.<p>If you’re doing data analytics in Python it’s well worth a look.
评论 #42444884 未加载
mushufasa5 months ago
Hi Burak, thanks for posting! We&#x27;re looking for a tool in this space and i&#x27;ll take a look.<p>Does Bruin support specifying and visualizing DAGs? I didn&#x27;t see that in the documentation via a quick look, but I thought to ask because you may use different terminology that can be a substitute.
评论 #42447267 未加载
评论 #42448735 未加载
评论 #42448909 未加载
alpb5 months ago
Congrats Burak, I can tell a lot of work has gone into this. If I may recommend, a comparison of this project with similar other&#x2F;state-of-the-art projects would be really good to have in your documentation set for others to understand how your approach differs from them.
评论 #42450047 未加载
havef5 months ago
Hi, Burak, it looks interesting. I was wondering, do you know about connect? Maybe you can take advantage of some of its ready-made components. In addition, it is also developed using Go<p>- <a href="https:&#x2F;&#x2F;docs.redpanda.com&#x2F;redpanda-connect&#x2F;home&#x2F;" rel="nofollow">https:&#x2F;&#x2F;docs.redpanda.com&#x2F;redpanda-connect&#x2F;home&#x2F;</a><p>- <a href="https:&#x2F;&#x2F;github.com&#x2F;redpanda-data&#x2F;connect">https:&#x2F;&#x2F;github.com&#x2F;redpanda-data&#x2F;connect</a>
评论 #42450043 未加载
JeffMcCune5 months ago
Congrats on the launch! Since this is Go have you considered using CUE or looked at their flow package? Curious how you see it relating or helping with data pipelines.
评论 #42444925 未加载
ellisv5 months ago
Direct link to the documentation:<p><a href="https:&#x2F;&#x2F;bruin-data.github.io&#x2F;bruin&#x2F;" rel="nofollow">https:&#x2F;&#x2F;bruin-data.github.io&#x2F;bruin&#x2F;</a>
gigatexal5 months ago
Ingestion with DLT likely would have given you more connections to things. Still very cool. I saw you talking about this on LinkedIn.
评论 #42470228 未加载
producthunter905 months ago
How does it handle scheduling or orchestrating pipeline runs? Do you integrate with tools like Airflow, or is there a built-in solution for that?
评论 #42443189 未加载
evalsock5 months ago
Do you have integration for ML orchestration to reuse bruin inside our existing pipeline?
wodenokoto5 months ago
That ingestr CLI you also developed and just casually reference seems very, very cool!
评论 #42450139 未加载
Multrex5 months ago
Why there is not MySQL integration? Will you plan to add it? MySQL is very popular.
评论 #42470237 未加载
sakshy145 months ago
I just used your getting started guide and it&#x27;s freaking amazing
评论 #42450140 未加载
kyt5 months ago
Why use this over Meltano?
评论 #42444164 未加载
评论 #42444775 未加载
uniquenamehere5 months ago
This looks cool! How would this compare to Benthos?
kakoni5 months ago
Is dlt part of bruin-stack?
评论 #42445797 未加载
drchaim5 months ago
&quot;Interesting, congrats! I&#x27;ve felt the same challenges but ended up using custom Python with dbt and DuckDB. I&#x27;ll take a look!&quot;
tony_francis5 months ago
How does this compare to ray data?
评论 #42444962 未加载