Merge branch 'master' of github.com:rtucker/backuppc-archive-s3

This commit is contained in:
Ryan Tucker 2011-09-20 22:35:54 -04:00
commit d07636a5ff
3 changed files with 158 additions and 1 deletions

19
COPYRIGHT Normal file
View file

@ -0,0 +1,19 @@
Copyright (c) 2009-2011 Ryan S. Tucker
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in
all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
THE SOFTWARE.

130
README.markdown Normal file
View file

@ -0,0 +1,130 @@
BackupPC_archiveHost_s3
=======================
This is a Python script that acts as an interface between
[BackupPC](http://backuppc.sourceforge.net/) and
[Amazon S3](http://aws.amazon.com/s3/). It uses BackupPC's
[archive function](http://backuppc.sourceforge.net/faq/BackupPC.html#archive_functions)
to extract a tarball and split it into chunks, like the normal archive
function. Then, the chunks are encrypted using gpg and transmitted to
S3 using [Boto](http://code.google.com/p/boto/).
Installation
------------
I wrote this script some years ago, and can't remember how to get it going.
But, here's going to be my best guess :-)
### Install the prerequisites
> You will need Python, [Boto](http://code.google.com/p/boto/), and a
> working BackupPC installation.
### Download and install this script
> Something like this seems like a good idea:
>
> cd /usr/local/src/
> git clone git://github.com/rtucker/backuppc-archive-s3.git
>
> Then create a link from `/usr/share/backuppc/bin/` to here:
>
> ln -s /usr/local/src/backuppc-archive-s3/BackupPC_archiveHost_s3 /usr/share/backuppc/bin/
### Configure this script
> Create a file in this directory called `secrets.py`, based upon the
> `secrets.py.orig` file. It should have your AWS Access and Shared keys,
> a passphrase that will be used to encrypt the tarballs, and, optionally,
> a path to a file that contains a maximum upload rate in kilobits per
> second:
>
> accesskey = 'ASDIASDVINASDVASsvblahblah'
> sharedkey = '889rv98rv8fmasmvasdvsdvasdv'
> gpgsymmetrickey = 'hunter2'
> speedfile = '/var/cache/speedlimit.txt'
>
> If you use the `speedfile` option, you can change this on the fly to
> limit upstream bandwidth usage during peak hours, etc.
### Configure BackupPC
> From the BackupPC configuration interface, go to `Edit Hosts` and add a
> new host, `archiveS3`, which looks like the existing `archive` host.
> Save this, select the `archives3` host, and then `Edit Config` for that
> host.
>
> Change the settings on each tab as follows:
>
>> #### Xfer
>> XferMethod: archive
>> ArchiveDest: /var/lib/backuppc/archives3
>> ArchiveComp: bzip2
>> ArchiveSplit: 500
>> ArchiveClientCmd: $Installdir/bin/BackupPC_archiveHost_s3 $tarCreatePath $splitpath $parpath $host $backupnumber $compression $compext $splitsize $archiveloc $parfile *
>>
>> #### Backup Settings
>> ClientTimeout: 720000
>
> That should be just about it. Note that `ArchiveDest` is where it will
> stage the tarballs before it uploads them; this must have enough disk
> space for your archive! `ArchiveSplit` is the size of each tar file,
> in megabytes; you may want to adjust this for your needs. Also, the
> `ArchiveClientCmd` is the default, except with the `_s3` added.
### Use it
> Go to the main page for the `archives3` host and click `Start Archive`.
> To start with, just tick the box next to the smallest backup you have,
> then `Archive selected hosts`. Go with the defaults (which look
> suspiciously like what you set on the Xfer tab, do they not? :-) and
> then `Start the Archive`.
>
> Watch syslog and hopefully everything will work.
backup-manager.py
-----------------
There is a companion script, `backup-manager.py`, that can be used to see
what's on S3. Run it with no arguments to get a listing of backups and
their ages, or use the `--help` argument to see what it can do.
The "crown jewel" of this whole system is the `script` command, which
produces a script that can be used to restore a backup. It uses S3's
[Query String Request Authentication](http://docs.amazonwebservices.com/AmazonS3/latest/dev/index.html?RESTAuthentication.html#RESTAuthenticationQueryStringAuth)
mechanism to generate temporary URLs to download each file required to
restore a backup.
Each night, from `cron`, I run a script:
#!/bin/sh
BACKUPMGR=/path/to/backup-manager.py
# Delete all backups older than 30 days.
$BACKUPMGR delete --age=30
# Create restore scripts, valid for one week, for all of my computers
cd /home/rtucker/Dropbox/RestoreScripts/
$BACKUPMGR --expire=604800 --host=gandalf script > restore_gandalf.sh
$BACKUPMGR --expire=604800 --host=witte script > restore_witte.sh
# etc, etc
# Output a list of what's on the server
$BACKUPMGR
The output of this is mailed to me, so I always know what's going on!
FAQs
----
* BackupPC is written in Perl. Why is this thing written in Python?
I know Python much better than I know Perl, so I wrote it in Python.
The good news is that BackupPC doesn't care, but it does mean this
probably won't be part of the BackupPC main distribution any time soon.
* Is this project dead?
You could say that. A lot of [my projects](https://github.com/rtucker/)
are one-off scripts that solve a very specific need I have, and I don't
put too much thought into making them useful for other people. This
script works for me and (sorta) meets my needs, so that's where it is.

View file

@ -150,7 +150,15 @@ def make_restore_script(backup, expire=86400):
def main(): def main():
# check command line options # check command line options
parser = optparse.OptionParser(usage="usage: %prog [options] list/delete/script") parser = optparse.OptionParser(
usage="usage: %prog [options] [list|delete|script]",
description="" +
"Companion maintenance script for BackupPC_archiveHost_s3. " +
"By default, it assumes the 'list' command, which displays all " +
"of the backups currently archived on S3. The 'delete' command " +
"is used to delete backups. The 'script' command produces a " +
"script that can be used to download and restore a backup."
)
parser.add_option("-H", "--host", dest="host", parser.add_option("-H", "--host", dest="host",
help="Name of backed-up host") help="Name of backed-up host")
parser.add_option("-b", "--backup-number", dest="backupnum", parser.add_option("-b", "--backup-number", dest="backupnum",