My advice - go all the way into RDS territory.<p>This is a killer feature. But it is hard to adopt it by itself, while being on RDS, which is what most people end up doing.<p>If you can build a management layer that gives better-than-RDS features ...while allowing me to BYOB my own cloud, you are on your way to win.
Love the idea, but is this essentially then just a faster version of <i>pg_dump</i>?<p><pre><code> pg_dump -C -h remotehost -U remoteuser dbname | psql -h localhost -U localuser dbname</code></pre>
Question: Can this clone a database, but also apply certain operations to it?<p>I work at a small company, and currently we clone our production database to our dev machines for testing. However certain information in the database is sensitive and we don't want to include it on our dev machines. (This specific sensitive data is also stored in an encrypted format and the key is not included, but we'd still prefer it not included).<p>Basically I'd like to be able to clone the database, and run some SQL to replace the data in some tables with some mock data. But I can't think of an easy way to do this without cloning the database, starting a temporary one, run the sql, then clone it again - and distribute that final build to the devs.
This looks like open source Delphix using Postgres. It even seems to be using ZFS under the hood.<p>Any comments from people using it? I designed a CI system using Delphix at an old client and it was awesome!
Would it be practical to use this in an automated regression testing context to give each test its own copy of the database that it could mutate, check results, then throw away?
This helps a lot in developing and testing for postgres.<p>Maybe that's one of the tech which "John Carmack" meant when he said sophisticated tech is indistinguishable from magic.
Does such a thing exist for MySQL? If so I'd appreciate a link. We have weekly snapshots that we use for development, but it would be amazing to have a way to reset/throw away any changes quickly, especially to repeatedly test a specific use-case against some prod data.
Snowflake does something similar since it's a metadata store and cloning is mainly copying the metadata. It does scale with data volumes though so it doesn't seem like it's be as quick.<p>Really great idea here!
A bit tangential, but this reminded of an idea I had when I was learning about containerization and overlayfs. Would postgres work with data stored in an overlay and allow a similar workflow?