Bash script for automated backups with duplicity and Amazon's S3 online storage
Switch branches/tags
Nothing to show
Clone or download
Failed to load latest commit information.
.gitignore Initial commit ... Feb 12, 2010
AUTHORS Fixed a line in AUTHORS... Sep 3, 2013
CHANGELOG updated with version 0.5 info May 9, 2011
LICENSE initial commit ... Feb 12, 2010
README Oops, no markdown. May 6, 2012 Fixed typo, fixes #8 Mar 7, 2012


*Active version of script*:

NOTE (May 6th, 2012): I am no longer actively developing this script.  
*zertrin* has recently offered to take over for the time being.
I have not, myself, tested his fork or the merges he's made but more development 
is always better than less. If you have pull requests or changes, use his branch.
I will be quietly watching from the shadows. Good luck! -Damon


This bash script was designed to automate and simplify the remote backup
process of duplicity on Amazon S3.  After your script is configured, you can
easily backup, restore, verify and clean (either via cron or manually) your
data without having to remember lots of different command options and

Most importantly, you can easily backup the script and your gpg key in a
convenient passphrase-encrypted file.  This comes in in handy if/when your
machine ever does go belly up.

Optionally, you can set up an email address where the log file will be sent,
which is useful when the script is used via cron.

More information about this script avaiable at:

Latest version of the code is available at:


This script requires user configuration.  Instructions are in
the file itself and should be self-explanatory.  Be sure to replace all the
*foobar* values with your real ones.  Almost every value needs to be
configured in someway.

You can use multiple copies of the script with different settings for different
backup scenarios.  It is designed to run as a cron job and will log information
to a text file (including remote file sizes, if you have s3cmd installed).  Be
sure to make the script executable (chmod +x) before you hit the gas.


* duplicity
* gpg
* Amazon S3
* s3cmd (optional)
* mailx (optional)


* View help:

* Run an incremental backup:
	$ --backup

* Force a one-off full backup:
    $ --full

* Restore your entire backup:
	$ --restore
    You will be prompted for a restore directory

	$ --restore /home/user/restore-folder
    You can also provide a restore folder on the command line.

* Restore a specific file in the backup:
    $ --restore-file
    You will be prompted for a file to restore to the current directory

    $ --restore-file img/mom.jpg
    Restores the file img/mom.jpg to the current directory

    $ --restore-file img/mom.jpg /home/user/i-love-mom.jpg
    Restores the file img/mom.jpg to /home/user/i-love-mom.jpg

* List files in the remote archive
	$ --list-current-files

* Verify the backup
    $ --verify

* Backup the script and gpg key (for safekeeping)
    $ --backup-script


This script attempts to simplify the task of running a duplicity command; if you
are having any problems with the script the first step is to determine if the
script is generating an incorrect command or if duplicity itself is
causing your error.

To see exactly what is happening when you run dt-s3-backup, head to the bottom
of the user configuration portion of the script and uncomment the `ECHO=$(which
echo)` variable.  This will stop the script from running and will, instead,
output the generated command into your log file. You can then check to see if
what is being generated is causing an error or if it is duplicity causing you


* --restore-dir option
* restore files from a specific time period
* support multiple MTAs for the email alert feature

Thanks to Mario Santagiuliana and Razvan for their help.