aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorintrigeri <intrigeri@boum.org>2010-05-06 19:47:54 +0200
committerintrigeri <intrigeri@boum.org>2010-05-06 19:47:54 +0200
commit219a6eb958a6f8cc9dce34f9f87a95c02ef0c322 (patch)
treeb46fb1875c4aa123592ac823508af95a5697cc1b
parent6cf9c8988b137e42f0d29ade910d2a4f50372798 (diff)
downloadbackupninja-219a6eb958a6f8cc9dce34f9f87a95c02ef0c322.tar.gz
backupninja-219a6eb958a6f8cc9dce34f9f87a95c02ef0c322.tar.bz2
Better example.dup documentation. Thanks, Alster!
-rw-r--r--ChangeLog1
-rw-r--r--examples/example.dup24
-rw-r--r--handlers/dup.helper.in17
3 files changed, 29 insertions, 13 deletions
diff --git a/ChangeLog b/ChangeLog
index 3dfbed8..cfa0959 100644
--- a/ChangeLog
+++ b/ChangeLog
@@ -14,6 +14,7 @@ version 0.9.8 -- UNRELEASED
. Put archive directory (cache) into /var/backups/duplicity rather
than the default /root/.cache/duplicity, unless the user
has specified it (Closes: 580213)
+ . Better example.dup documentation. Thanks, Alster!
version 0.9.7 -- January 27, 2010
backupninja changes
diff --git a/examples/example.dup b/examples/example.dup
index 53039f5..2c0c41d 100644
--- a/examples/example.dup
+++ b/examples/example.dup
@@ -19,7 +19,8 @@
## Default:
# nicelevel = 0
-## test the connection? set to no to skip the test if the remote host is alive
+## test the connection? set to no to skip the test if the remote host is alive.
+## if 'desturl' is set below, 'testconnect' must be set to 'no' for now.
##
## Default:
# testconnect = yes
@@ -169,6 +170,7 @@ exclude = /home/*/.gnupg
## desturl = file:///usr/local/backup
## desturl = rsync://user@other.host//var/backup/bla
## desturl = s3+http://
+## desturl = ftp://myftpuser[:myftppassword]@ftp.example.org/remote/ftp/path
## the default value of this configuration option is not set:
##
## Default:
@@ -183,8 +185,9 @@ exclude = /home/*/.gnupg
# awsaccesskeyid =
# awssecretaccesskey =
-## bandwith limit, in kbit/s ; default is 0, i.e. no limit an example
-## setting would be:
+## bandwith limit, in kbit/s ; default is 0, i.e. no limit
+## if using 'desturl' above, 'bandwidthlimit' must not be set
+## an example setting of 128 kbps would be:
## bandwidthlimit = 128
##
## Default:
@@ -199,14 +202,18 @@ exclude = /home/*/.gnupg
## Default:
# sshoptions =
-## put the backups under this directory, this must be set!
+## put the backups under this destination directory
+## if using 'desturl' above, this must not be set
+## in all other cases, this must be set!
## an example setting would be:
## destdir = /backups
##
## Default:
# destdir =
-## the machine which will receive the backups, this must be set!
+## the machine which will receive the backups
+## if using 'desturl' above, this must not be set
+## in all other cases, this must be set!
## an example setting would be:
## desthost = backuphost
##
@@ -214,12 +221,11 @@ exclude = /home/*/.gnupg
# desthost =
## make the files owned by this user
-## note: you must be able to ssh backupuser@backhost
-## without specifying a password (if type = remote).
+## if using 'desturl' above, this must not be set
+## note: if using an SSH based transport and 'type' is set to 'remote', you must
+## be able to 'ssh backupuser@backuphost' without specifying a password.
## an example setting would be:
## destuser = backupuser
##
## Default:
# destuser =
-
-
diff --git a/handlers/dup.helper.in b/handlers/dup.helper.in
index 47f4469..a47064d 100644
--- a/handlers/dup.helper.in
+++ b/handlers/dup.helper.in
@@ -280,7 +280,8 @@ options = $dup_options
# default is 0, but set to 19 if you want to lower the priority.
nicelevel = $dup_nicelevel
-# default is yes. set to no to skip the test if the remote host is alive
+# default is yes. set to no to skip the test if the remote host is alive.
+# if 'desturl' is set below, 'testconnect' must be set to 'no' for now.
testconnect = $dup_testconnect
######################################################
@@ -429,6 +430,7 @@ keep = $dup_keep
#desturl = file:///usr/local/backup
#desturl = rsync://user@other.host//var/backup/bla
#desturl = s3+http://your_bucket
+#desturl = ftp://myftpuser[:myftppassword]@ftp.example.org/remote/ftp/path
# Amazon Web Services Access Key ID and Secret Access Key, needed for backups
# to S3 buckets.
@@ -436,6 +438,8 @@ keep = $dup_keep
#awssecretaccesskey = YOUR_AWS_SECRET_KEY
# bandwith limit, in kbit/s ; default is 0, i.e. no limit
+# if using 'desturl' above, 'bandwidthlimit' must not be set
+# an example setting of 128 kbps would be:
#bandwidthlimit = 128
bandwidthlimit = $dup_bandwidth
@@ -445,15 +449,20 @@ bandwidthlimit = $dup_bandwidth
#sshoptions = -o IdentityFile=/root/.ssh/id_dsa_duplicity
sshoptions = $dup_sshoptions
-# put the backups under this directory
+# put the backups under this destination directory
+# if using 'desturl' above, this must not be set
+# in all other cases, this must be set!
destdir = $dup_destdir
# the machine which will receive the backups
+# if using 'desturl' above, this must not be set
+# in all other cases, this must be set!
desthost = $dup_desthost
# make the files owned by this user
-# note: you must be able to ssh backupuser@backhost
-# without specifying a password (if type = remote).
+# if using 'desturl' above, this must not be set
+# note: if using an SSH based transport and 'type' is set to 'remote', you must
+# be able to 'ssh backupuser@backuphost' without specifying a password.
destuser = $dup_destuser
EOF