TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Show HN: CSV GB+ by Data.olllo – Open and Process CSVs Locally

51 pointsby olllo25 days ago
I built CSV GB+ by Data.olllo, a local data tool that lets you open, clean, and export gigabyte-sized CSVs (even billions of rows) without writing code.<p>Most spreadsheet apps choke on big files. Coding in pandas or Polars works—but not everyone wants to write scripts just to filter or merge CSVs. CSV GB+ gives you a fast, point-and-click interface built on dual backends (memory-optimized or disk-backed) so you can process huge datasets offline.<p>Key Features: Handles massive CSVs with ease — merge, split, dedup, filter, batch export<p>Smart engine switch: disk-based &quot;V Core&quot; or RAM-based &quot;P Core&quot;<p>All processing is offline – no data upload or telemetry<p>Supports CSV, XLSX, JSON, DBF, Parquet and more<p>Designed for data pros, students, and privacy-conscious users<p>Register for 7-days free to pro try, pro versions remove row limits and unlock full features. I’m a solo dev building Data.olllo as a serious alternative to heavy coding or bloated enterprise tools.<p>Download for Windows: <a href="https:&#x2F;&#x2F;apps.microsoft.com&#x2F;detail&#x2F;9PFR86LCQPGS" rel="nofollow">https:&#x2F;&#x2F;apps.microsoft.com&#x2F;detail&#x2F;9PFR86LCQPGS</a><p>User Guide: <a href="https:&#x2F;&#x2F;olllo.top&#x2F;articles&#x2F;article-0-Data.olllo-UserGuide" rel="nofollow">https:&#x2F;&#x2F;olllo.top&#x2F;articles&#x2F;article-0-Data.olllo-UserGuide</a><p>Would love feedback! I’m actively improving it based on real use cases.

5 comments

TheTaytay25 days ago
Thank you for this. I find myself increasingly using CSVs (TSVs actually) as the data format of choice. I confess I wish this was written for Mac too, but I like the trend of (once again) moving data processing down to our super computers on our desk...
评论 #43987501 未加载
评论 #43989226 未加载
评论 #43993450 未加载
评论 #43987056 未加载
评论 #43989127 未加载
paddy_m25 days ago
Do you have a demo video?<p>What are you using for processing (polars)?<p>Marketing note: I&#x27;m sure you&#x27;re proud of P Core&#x2F;V Core, but that doesn&#x27;t matter to your users, it&#x27;s an implementation detail. At a maximum I&#x27;d write &quot;intelligent execution that scales from small files to large files&quot;.<p>As an implementation note, I would make it simple to operate on just the first 1000 (10k or 100k) rows so responses are super quick, then once the users are happy about the transform, make it a single click to operate on the entire file with a time estimate.<p>Another feature I&#x27;d like in this vein is execute on a small subset, then if you find an error with a larger subset, try to reduce the larger subset to a small quick to reproduce version. Especially for deduping.
评论 #43989242 未加载
评论 #43993458 未加载
评论 #43988003 未加载
bitbasher23 days ago
What are &quot;massive&quot; CSVs? I have CSVs in the terabytes that need to be deduped by a specific column. Can it handle that? What if I want to run a function on the column to normalize it before the deduping?
xnx25 days ago
Is this better than the free Tad (<a href="https:&#x2F;&#x2F;www.tadviewer.com&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.tadviewer.com&#x2F;</a>) which seems to do similar things for free?
评论 #43993463 未加载
评论 #43986053 未加载
评论 #43986152 未加载
dangerlibrary25 days ago
It is 2025 and CSVs still dominate data interchange between organizations.<p><a href="https:&#x2F;&#x2F;graydon2.dreamwidth.org&#x2F;193447.html" rel="nofollow">https:&#x2F;&#x2F;graydon2.dreamwidth.org&#x2F;193447.html</a>
评论 #43993468 未加载
评论 #43986128 未加载
评论 #43986145 未加载