TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Apache iceberg the Hadoop of the modern-data-stack?

115 pointsby samrohn3 months ago

12 comments

hendiatris3 months ago
This is a huge challenge with Iceberg. I have found that there is substantial bang for your buck in tuning how parquet files are written, particularly in terms of row group size and column-level bloom filters. In addition to that, I make heavy use of the encoding options (dictionary&#x2F;RLE) while denormalizing data into as few files as possible. This has allowed me to rely on DuckDB for querying terabytes of data at low cost and acceptable performance.<p>What we are lacking now is tooling that gives you insight into how you should configure Iceberg. Does something like this exist? I have been looking for something that would show me the query plan that is developed from Iceberg metadata, but didn’t find anything. It would go a long way to showing where the bottleneck is for queries.
评论 #43280167 未加载
评论 #43287811 未加载
评论 #43286353 未加载
评论 #43279001 未加载
mritchie7123 months ago
This is a bit overblown.<p>Is Iceberg &quot;easy&quot; to set up? No.<p>Can you get set up in a week? Yes.<p>If you really need a datalake, spending a week setting it up is not so bad. We have a guide[0] here that will get you started in under an hour.<p>For smaller (e.g. under 10tb) data where you don&#x27;t need real-time, DuckDB is becoming a really solid option. Here&#x27;s on setup[1] we&#x27;ve played around with using Arrow Flight.<p>If you don&#x27;t want to mess with any of this, we[2] spin it all up for you.<p>0 - <a href="https:&#x2F;&#x2F;www.definite.app&#x2F;blog&#x2F;cloud-iceberg-duckdb-aws" rel="nofollow">https:&#x2F;&#x2F;www.definite.app&#x2F;blog&#x2F;cloud-iceberg-duckdb-aws</a><p>1 - <a href="https:&#x2F;&#x2F;www.definite.app&#x2F;blog&#x2F;duck-takes-flight" rel="nofollow">https:&#x2F;&#x2F;www.definite.app&#x2F;blog&#x2F;duck-takes-flight</a><p>2 - <a href="https:&#x2F;&#x2F;www.definite.app&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.definite.app&#x2F;</a>
评论 #43281725 未加载
评论 #43282853 未加载
评论 #43282669 未加载
simlevesque3 months ago
I&#x27;m working on an alternative Iceberg client to work better in write heavy use cases. Instead of many smaller files it writes on the same file until it&#x27;s 1mb in size but it gives it a new name. Then I update the manifest to the new filename and checksum. I keep old files on disk for 60 seconds to allow pending queries. I&#x27;m also working on auto compaction, when I have ten 1mb files I compact them, same with ten 10mb files, etc...<p>I feel like this could be a game changer for the ecosystem. It&#x27;s more cpu and network heavy for writes but the reads are always fast. And the writes are still faster than pyiceberg.<p>I want to hear opinions or how this could never work.
评论 #43282117 未加载
评论 #43281653 未加载
评论 #43281701 未加载
评论 #43282057 未加载
robertkoss3 months ago
This article is just shameless advertising for Estuary Flow, a company that the author is working for. &quot;Operational Maturity&quot;, as if Iceberg, Delta or Hudi are not mature. These are battle-tested frameworks that have been in production for years. The &quot;small files problem&quot; is not really a problem because every framework supports some way of compacting smaller files. Just run a nightly job that compacts the small files and you&#x27;re good 2 go.
评论 #43283635 未加载
Gasp0de3 months ago
Does anyone have a good alternative for storing large amounts of very small files that need to be individually queriable? We are dealing with a large amount of sensor readings that we need to be able to query on a per sensor basis and a timespan, and we are dealing with the problem mentioned in the article, that storing millions of small files in S3 is expensive.
评论 #43280367 未加载
评论 #43279492 未加载
评论 #43279563 未加载
评论 #43280433 未加载
评论 #43283081 未加载
评论 #43286236 未加载
alienreborn3 months ago
Better article (imo) on similar topic: <a href="https:&#x2F;&#x2F;www.dataengineeringweekly.com&#x2F;p&#x2F;is-apache-iceberg-the-new-hadoop" rel="nofollow">https:&#x2F;&#x2F;www.dataengineeringweekly.com&#x2F;p&#x2F;is-apache-iceberg-th...</a>
评论 #43281187 未加载
alexmorley3 months ago
Most of these issues will be ring true to lots of folk using Iceberg at the moment. But this does not:<p>&gt; Yet, competing table formats like Delta Lake and Hudi mirror this fragmentation. [ ... ] &gt; Just as Spark emerged as the dominant engine in the Hadoop ecosystem, a dominant table format and catalog may appear in the Iceberg era.<p>I think extremely few people are making bets on any other open source table format now - that consolidation has already happened in 2023-2024 (see e.g. Databricks who have their own competing format leaning heavily into iceberg; or adoption from all of the major data warehouse providers).
评论 #43279548 未加载
datax23 months ago
&quot;Hadoop’s meteoric rise led many organizations to implement it without understanding its complexities, often resulting in underutilized clusters or over-engineered architectures. Iceberg is walking a similar path.&quot;<p>This pain is too real, and too close to home. I&#x27;ve seen this outcome turn the entire business off of consuming their data via hadoop because it turns into a wasteland of delayed deliveries, broken datasets, op&#x27;s teams who cannot scale, and architects overselling too robust designs.<p>I&#x27;ve tried to scale down hadoop to the business user with visual etl tools like Alteryx, but there again compatibility between Alteryx and hadoop suck via ODBC connectors. I came from an AWS based stack into a poorly leapfrogged data stack and it&#x27;s hard not to pull my hair out between the business struggling to use it and infra + op&#x27;s not keeping up. Now these teams want to push to iceburg or big query while ignoring the mountains of tech debt they have created.<p>Don&#x27;t get me wrong Hadoop isn&#x27;t a bad idea, its just complex and a time suck, and unless you have time to dedicate to properly deploy these solutions which most business do not, your implementation will suffer, your business will suffer.<p>&quot;While the parallels to Hadoop are striking, we also have the opportunity to avoid its pitfalls.&quot; no one in IT learns from their failures unless they are writing the checks, most will flip before they feel the pain.
zhousun3 months ago
The only datastack iceberg (or lakehouse) will never replace is OLTP systems, for high-concurrency updates optimistic concurrency control &amp; object store is simply a no go.<p>Iceberg out-of-the-box is &quot;NOT&quot; good at streaming use cases, unlike formats like Hudi or Paimon, the table format does not have the concept of merge&#x2F; index. However, the beauty of iceberg is it is very unopinionated, so it is indeed possible to design an engine to stream write to iceberg. As far as I know this is how engines like Upsolver was implemented: 1. Have in-memory buffer to track incoming rows before flushing a version to iceberg (every 10s to a few minutes). 2. Build Indexing structure to write position deletes&#x2F; deletion vector instead of equality deletes. 3. The writer will all try to merge small files and optimize the table.<p>And stay tuned, we at <a href="https:&#x2F;&#x2F;www.mooncake.dev&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.mooncake.dev&#x2F;</a> are working on a solution to mirror a postgres table to iceberg, and keep them always up-to-date.
orthoxerox3 months ago
I think the complexity of Iceberg is overblown. It&#x27;s just a table format and it&#x27;s strictly better than the Hive-style &#x2F;schema&#x2F;table&#x2F;partition_key=partition_value&#x2F;one_of_many_files.parquet<p>It has a lot of knobs to fiddle with (more than Delta Lake, which tries very hard to come up with good defaults), but even if you don&#x27;t touch any of them, you already end up with tables that are as good as Hive&#x27;s, except now your writers don&#x27;t break your readers.<p>This is already a massive boon that lets you escape the rigidity of a timetable schedule for your data pipelines. Anything else you can come up with (switching your table to MOR and rewriting it as a separate step etc) is further improvements.
paulsutter3 months ago
Does this feel about 3x too verbose, like it’s generated?
评论 #43280207 未加载
评论 #43281400 未加载
theyinwhy3 months ago
What&#x27;s a good alternative? Google BigQuery?