Qnap [ Duplicty ] [ 0.7.14.0 ] Encrypted bandwidth-efficient backup using the rsync algorithm

QoolBox

Représentant QNAP
2 Janvier 2014
10 577
165
153
50
France
www.qnap.com
duplicity-50f59634a9cde.png


Source : http://duplicity.nongnu.org

Download :

x86 version : http://www.positiv-it.fr/QNAP/APP/Duplicity_0.7.06_x86.qpkg.zip
x86 version : http://www.positiv-it.fr/QNAP/APP/Duplicity_0.7.07.1_x86.qpkg.zip

x86 version : http://www.positiv-it.fr/QNAP/APP/Duplicity_0.7.10.0_x86.qpkg.zip
x64 version : http://www.positiv-it.fr/QNAP/APP/Duplicity_0.7.10.0_x86_64.qpkg.zip

x64 version : http://www.positiv-it.fr/QNAP/APP/Duplicity_0.7.13.1_x86_64.qpkg.zip

x64 version : http://www.positiv-it.fr/QNAP/APP/Duplicity_0.7.14.0_x86_64.qpkg.zip

Note :

Duplicty and Duply automatically added to NAS PATH
all other binaries in /opt/Duplicity/bin ($PATH may need to be exported in some cases)

in this release : you can find in the path : Unison, Rsync 3.1.2, rdiff-backup, gnupg ...

About :

Duplicity backs directories by producing encrypted tar-format volumes and uploading them to a remote or local file server. Because duplicity uses librsync, the incremental archives are space efficient and only record the parts of files that have changed since the last backup. Because duplicity uses GnuPG to encrypt and/or sign these archives, they will be safe from spying and/or modification by the server.

The duplicity package also includes the rdiffdir utility. Rdiffdir is an extension of librsync's rdiff to directories---it can be used to produce signatures and deltas of directories as well as regular files. These signatures and deltas are in GNU tar format.

Code:
[~] # duplicity --help
Usage:
  duplicity [full|incremental] [options] source_dir target_url
  duplicity [restore] [options] source_url target_dir
  duplicity verify [options] source_url target_dir
  duplicity collection-status [options] target_url
  duplicity list-current-files [options] target_url
  duplicity cleanup [options] target_url
  duplicity remove-older-than time [options] target_url
  duplicity remove-all-but-n-full count [options] target_url
  duplicity remove-all-inc-of-but-n-full count [options] target_url

Backends and their URL formats:
  cf+http://container_name
  file:///some_dir
  ftp://user[:password]@other.host[:port]/some_dir
  ftps://user[:password]@other.host[:port]/some_dir
  hsi://user[:password]@other.host[:port]/some_dir
  imap://user[:password]@other.host[:port]/some_dir
  rsync://user[:password]@other.host[:port]::/module/some_dir
  rsync://user[:password]@other.host[:port]/relative_path
  rsync://user[:password]@other.host[:port]//absolute_path
  s3://other.host/bucket_name[/prefix]
  s3+http://bucket_name[/prefix]
  scp://user[:password]@other.host[:port]/some_dir
  ssh://user[:password]@other.host[:port]/some_dir
  swift://container_name
  tahoe://alias/directory
  webdav://user[:password]@other.host/some_dir
  webdavs://user[:password]@other.host/some_dir
  gdocs://user[:password]@other.host/some_dir
  pydrive://user@other.host/some_dir
  mega://user[:password]@other.host/some_dir
  copy://user[:password]@other.host/some_dir
  dpbx:///some_dir
  onedrive://some_dir
  azure://container_name
  b2://account_id[:application_key]@bucket_name/[some_dir/]

Commands:
  cleanup <target_url>
  collection-status <target_url>
  full <source_dir> <target_url>
  incr <source_dir> <target_url>
  list-current-files <target_url>
  restore <target_url> <source_dir>
  remove-older-than <time> <target_url>
  remove-all-but-n-full <count> <target_url>
  remove-all-inc-of-but-n-full <count> <target_url>
  verify <target_url> <source_dir>

Options:
  -h, --help            show this help message and exit
  --allow-source-mismatch
  --archive-dir=path
  --asynchronous-upload
  --compare-data
  --dry-run
  --encrypt-key=gpg-key-id
  --encrypt-secret-keyring=path
  --encrypt-sign-key=gpg-key-id
  --exclude=shell_pattern
  --exclude-device-files
  --exclude-filelist=filename
  --exclude-if-present=filename
  --exclude-other-filesystems
  --exclude-regexp=regular_expression
  --exclude-older-than=time
  --extra-clean
  --file-prefix=FILE_PREFIX
  --file-prefix-manifest=FILE_PREFIX_MANIFEST
  --file-prefix-archive=FILE_PREFIX_ARCHIVE
  --file-prefix-signature=FILE_PREFIX_SIGNATURE
  -r path, --file-to-restore=path
  --force
  --ftp-passive
  --ftp-regular
  --full-if-older-than=time
  --gio
  --gpg-binary=path
  --gpg-options=options
  --hidden-encrypt-key=gpg-key-id
  --ignore-errors
  --imap-full-address
  --imap-mailbox=imap_mailbox
  --include=shell_pattern
  --include-filelist=filename
  --include-regexp=regular_expression
  --log-fd=file_descriptor
  --log-file=filename
  --max-blocksize=number
  --name=backup name
  --no-encryption
  --no-compression
  --no-print-statistics
  --null-separator
  --num-retries=number
  --numeric-owner
  --old-filenames
  --par2-redundancy=number
  --par2-options=options
  --progress
  --progress-rate=number
  --pydevd
  --rename=RENAME
  -t time, --restore-time=time, --time=time
  --rsync-options=options
  --s3-european-buckets
  --s3-use-rrs
  --s3-use-ia
  --s3-use-new-style
  --s3-unencrypted-connection
  --s3-multipart-chunk-size=number
  --s3-multipart-max-procs=number
  --s3-multipart-max-timeout=number
  --s3-use-multiprocessing
  --s3-use-server-side-encryption
  --scp-command=command
  --sftp-command=command
  --cf-backend=pyrax|cloudfiles
  --short-filenames
  --sign-key=gpg-key-id
  --ssh-askpass
  --ssh-options=options
  --ssl-cacert-file=pem formatted bundle of certificate authorities
  --ssl-no-check-certificate
  --tempdir=path
  --timeout=seconds
  --time-separator=char
  --use-agent
  -v [0-9], --verbosity=[0-9]
  -V, --version
  --volsize=number
  --file-changed=path
 
Très intéressant, merci Stéphane !!

Sent from my A0001 using Tapatalk
 
bonjour,

comment faire ajouter ace_cli pour l'utilisation d'amazon cloud drive avec duplicity?