TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Ask HN: How do you backup 200TB across 15 servers onto a single 50TB server?

5 pointsby nanchover 11 years ago
I&#x27;m in a situation where I have a lot of data spread across lots of servers and I would like to be able to restore the data on a single server if it fails.<p>Suppose each server has less than 40TB.<p>The solution I&#x27;m thinking of is like RAID, but at a server-level; like a simple XOR of each server&#x27;s data onto a single server. In the case of a server failure, XOR each existing server and rebuild the data from the failed server.<p>Has anybody done this?

6 comments

Zenstover 11 years ago
Firstly unless you have a lot of duplication in that 200TB you can&#x27;t.<p>Two areas you need to ask yourself:<p>1) What is the daily amount of change per server and with that the total amount that you need to backup daily or whatever frequency you need). 2) What is the initial uniquiness in the data being backed up. Mostly core OS and system with standard same local copies of data tables?<p>With that you can work out what you are trying to do form a need perspective. May be case of local imaging servers and doing differential backups automated. But if compression and duplication are not the case and&#x2F;or high frequency of data then you will not be able to fit that data into a smaller area. You may only get 2:1 or less compression it is realy data centric and you need at least 4:1 and actualy better on all that data.<p>Look into the problem and define it better, with that a solution will be easier and may be case of proving that you can not do what you need may even need 400TB to back up those 200TB servers due to high volume data changes and need to be able to go back more than the last backup copy and may need to go back a month. Define the buisness needs, impact and with that the cost and budget as this is very much a case of working to a limitation when that limitation is not defined and may well be articifical and defeating in the objective to a cost that exceeds just adding more storage.<p>In short, define the problem better with regards to data and busness needs for backup -- hourly&#x2F;realtime daily weekly and the type of data and how dynamic that data is. May be case of easy to do as realy only 1TB of data that changes across all servers each week, just do not know.<p>Then do not go about doing yoru own backup solution from scratch, as something to replace what you have worth a look if you realy know what you are doing and you may not. But do not reinvent wheels, many options&#x2F;solution out there but until the true problem is defined then no easly solution can be given.<p>Heck do not even know if these are Windows, Linux, AIX, BSD or a combination of different servers and with that the options change. ALso factor in offsite aspect of backup as one electrical zapping from a storm or a fire could mess things up if that backupserver is sat next too the servers it is backing up. So many things to consider but above all sit down and define the problem in more detail as sadly the devil is always in the detail and better definition of the problem allows for easier solutions and over all less work as well as smarter work. It is a good leason too learn, just don&#x27;t do it the hard way and write a backup system backed system on a fag packet problem and design.
nwhover 11 years ago
Sounds like an outrageously dangerous place to roll your own system. What you&#x27;re describing is some sort of reed-solomon type data recovery, though I&#x27;ve absolutely no idea how that would work in practise. Wouldn&#x27;t you be spending all of your time updating the parity information?
评论 #6661265 未加载
tomfitzover 11 years ago
tahoe-lafs, <a href="https://tahoe-lafs.org/trac/tahoe-lafs" rel="nofollow">https:&#x2F;&#x2F;tahoe-lafs.org&#x2F;trac&#x2F;tahoe-lafs</a> , is &quot;an open source, secure, decentralized, fault-tolerant, peer-to-peer distributed data store and distributed file system.&quot;<p>It has the server-level RAID you desire. The FAQ states: ``You know how with RAID-5 you can lose any one drive and still recover? And there is also something called RAID-6 where you can lose any two drives and still recover. Erasure coding is the generalization of this pattern: you get to configure how many drives you could lose and still recover. You can choose how many drives (actually storage servers) will be used in total, from 1 to 256, and how many storage servers are required to recover all the data, from 1 to however many storage servers there are. We call the number of total servers N and the number required K, and we write the parameters as &quot;K-of-N&quot;.&#x27;&#x27;<p>It&#x27;s a very active open source project, with full-time contributors (I think?), funded by their commercial arm, <a href="https://leastauthority.com/" rel="nofollow">https:&#x2F;&#x2F;leastauthority.com&#x2F;</a> .
manglavover 11 years ago
Have you looked at FlexRAID? I am not too knowledgeable on the depths of this topic, but I know it can back up over SMB&#x2F;FTP. It&#x27;s a very interesting engine as well, and if you decide to roll your own, you might get some inspiration.<p><a href="http://www.flexraid.com/" rel="nofollow">http:&#x2F;&#x2F;www.flexraid.com&#x2F;</a>
评论 #6662111 未加载
chris_vaover 11 years ago
Your proposed scheme might work, but the odds of a system failure are extremely high (like, what if 2 servers go out instead of one?).<p>How hard is your constraint? Can you just use an aggressive Reed Solomon codec with a ~1.5x blowup factor (so, 100TB of extra space instead of 50TB), and spread the chunks around to each of the machines?
评论 #6662143 未加载
bobx11over 11 years ago
Hadoop is like raid for servers.