Good article. As it mentions, itch.io client is open source: <a href="https://github.com/itchio/itch" rel="nofollow">https://github.com/itchio/itch</a><p>I wish GOG would also open up their client and release it cross platform. Or at least document their protocol, as they promised.
For some reason, I thought this is about the update step in games that happens once per 'tick', that is, the physics engine loop. It's about lossless compression and downloading update packages though. That's also fine with me.
Interesting, but I'd like more details on what's happening at the client.<p>Take Steam for example. For some games, downloading the update takes seconds, but calculating diffs and extracting takes 10-20 minutes. That's great for Valve, because little bandwidth is used, but terrible at the client side. On top of that, the update process slows the rest of the system almost to a halt, because of all the hard drive activity.<p>I can potentially see this mechanism making the same mistake.
Also relevant is Google's Courgette algorithm which is used to update Google Chrome - <a href="https://dev.chromium.org/developers/design-documents/software-updates-courgette" rel="nofollow">https://dev.chromium.org/developers/design-documents/softwar...</a> which is stated to be much better than bsdiff which is what itch's is using here.
Good article. As a note, I love how he uses hand drawn diagrams. I have yet to find any tool that allow me to draw diagrams as fast as I can do it on a piece of paper.
I enjoyed this article a lot.<p>One thing that seemed glossed over, so I'm not sure if it's obvious for their use case, is the trade-off between compress once, distribute many times.<p>When looking at how long it takes to compress vs transmit, the optimisation was done to make the sum of both as small as possible - minimise(time(compress) + time(transmit)).<p>Instead it seems like you want to do is - minimise(time(compress) + expected_transmissions * time(transmit))<p>For any reasonable number of distributed copies of a game, it seems like this time to transmit will quickly come to dominate the total time involved.<p>I suspect, however, that the time to compress grows extremely quickly, for not much gain in compression, so the potential improvement is probably tiny even if you expect to be transmitting to millions of clients.
The rsync example confuses me a little bit. If you add a single bit to the front, then all the bytes are shifted into different blocks and nearly none will hash to match. But if you add a single bit, rsync still performs well. Can someone explain why that difference from the explanation?<p>The problem also applies to the binary delta. Adding a prefix will shift everything forward causing a diff in everything.<p>Bsdiff solves this with the suffix sorting. But what does rsync do? Or am I just wrong that rsync still works well? In either case, I think the offset problem makes for a more interesting motivating example for bsdiff.
Amos has been doing an incredible work with the itch.io client. It improved not only the user experience, but also the developer experience as well.<p>The butler client is an incredible way of sending games to itch.io and I sincerely hope that Apple or Google had a good command line client as well.