In the past, I had developed my own backup solution. Unfortunately, over time it didn't work out (mainly from changing systems, moving, using a laptop instead of a desktop, and maintaining it). However, I still like the idea of incremental backups as well as a mirrored version of my files (it saves space and lets me keep a history going back some number of days).
Now that I'm somewhat settled (and a little wiser), I decided to once more try my hand at a solid backup plan. This was mainly motivated by a recent reinstall of my wife's system (no lost data, just operating system upgrade). Since I don't have vast amounts of time on my hands, I didn't want to forward port my old solution to get it to work on Linux and Mac OS X, so I looked for new solutions. I recalled my advisor from Marquette mentioning rdiff-backup as what he put on his wife's machine during her dissertation days.
As it turns out, rdiff-backup does most of what I wanted out of my backup solution and, in fact, does it a little better. The main issue I had with my system was that it would periodically (monthly) take a snapshot of my home directory, after that it would periodically (weekly) build incremental diffs based off that snapshot. What this boils down to is that, if a catastrophic failure happens I would roll back to the most recent snapshot, then progress forward in time to the most recent incremental file. Not bad, but if you want better-than-weekly granularity it could be a lot of work. Obviously, I had scripted this part, but still it is wasted time. With rdiff-backup, it would be a single copy operation to restore to the most recent version. If you wanted older versions you could roll back through the incremental diffs (again, it is automated).
The other feature that I needed was the ability to remove backups/incremental data older than some time frame (monthly). Again, rdiff-backup gives me this ability at the command line. Other bonuses include the fact that it is cross-platform (via macports or most Linux repositories), written in Python, and not maintained by me!
With the basic service in place, it was time to make it automated. Again, linked off rdiff-backup's page is an article on how to do unattended backups. Besides the typical unattended SSH-keypair-without-a-passphrase and protecting-the-account steps, it introduced me to a new trick (which for some reason, despite having the knowledge on how to do it, never put together) using SSH config.
Host athena-backup Hostname athena.olympus User backups IdentityFile ~/.ssh/backups_rsa Compression yes Protocol 2
Now, if I try to
ssh athena-backup, it'll automatically use
the correct identity file and user name which saves me from having to
specify it on the command line (which you can't typically do with wrapper
functionality). More importantly, it doesn't break normal SSHing onto that
host since we made it a special host (that's the part I never put together,
even though I knew it was possible).
The next issue I never took that time to think about before was my having moved from desktop to laptop (thereby making 1:00am backups worthless sense the laptop isn't always on). Because rdiff-backup does a roll-back model instead of my roll-forward model, I decided to do hourly backups to my home machine, thus I'll likely catch at least one of these a day. But I'm not always at home! Getting around that is trivial, I'll just ping the backup server before trying. If it doesn't respond, I don't backup. This is done through:
$ ping -c1 -t1 $SERVER > /dev/null 2>&1
$SERVER is just the name of the backup server. It pings
the host once with a timeout of 1 second, if it succeeds the backup
continues; otherwise the script exits.
Of course, setting up the cronjob is as simple as:
0 */1 * * * $HOME/.crontab/rdiff-backup.sh
$HOME/.crontab/rdiff-backup.sh is the path to my rdiff-backup
Hopefully this time around the backup solution is more robust than before.