Hello community,

here is the log from the commit of package duply for openSUSE:Factory checked 
in at 2013-06-07 06:57:16
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Comparing /work/SRC/openSUSE:Factory/duply (Old)
 and      /work/SRC/openSUSE:Factory/.duply.new (New)
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Package is "duply"

Changes:
--------
--- /work/SRC/openSUSE:Factory/duply/duply.changes      2012-12-03 
09:32:25.000000000 +0100
+++ /work/SRC/openSUSE:Factory/.duply.new/duply.changes 2013-06-07 
06:57:17.000000000 +0200
@@ -1,0 +2,10 @@
+Tue Jun  4 22:27:01 UTC 2013 - [email protected]
+
+- Updated to version 1.5.10:
+  + Bugfix: exclude filter failed on ubuntu, mawk w/o posix char
+    class support.
+  + Bugfix: fix url_decoding generally and for python3.
+  + Bugfix 3609075: wrong script results in status line
+   (thx David Epping).
+
+-------------------------------------------------------------------

Old:
----
  duply_1.5.9.tgz

New:
----
  duply_1.5.10.tgz

++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Other differences:
------------------
++++++ duply.spec ++++++
--- /var/tmp/diff_new_pack.zlZkth/_old  2013-06-07 06:57:18.000000000 +0200
+++ /var/tmp/diff_new_pack.zlZkth/_new  2013-06-07 06:57:18.000000000 +0200
@@ -1,8 +1,8 @@
 #
 # spec file for package duply
 #
-# Copyright (c) 2012 SUSE LINUX Products GmbH, Nuernberg, Germany.
-# Copyright (c) 2011 Malcolm J Lewis <[email protected]>
+# Copyright (c) 2013 SUSE LINUX Products GmbH, Nuernberg, Germany.
+# Copyright (c) 2011-2013 Malcolm J Lewis <[email protected]>
 #
 # All modifications and additions to the file contributed by third parties
 # remain the property of their copyright owners, unless otherwise agreed
@@ -18,13 +18,13 @@
 
 
 Name:           duply
-Version:        1.5.9
+Version:        1.5.10
 Release:        0
 Summary:        A frontend for the mighty duplicity magic
 License:        GPL-2.0
 Group:          Productivity/Archiving/Compression
 Url:            http://duply.net/
-Source0:        
http://surfnet.dl.sourceforge.net/project/ftplicity/duply%20%28simple%20duplicity%29/1.5.x/duply_1.5.9.tgz
+Source0:        
http://surfnet.dl.sourceforge.net/project/ftplicity/duply%20%28simple%20duplicity%29/1.5.x/duply_1.5.10.tgz
 Requires:       duplicity
 BuildArch:      noarch
 BuildRoot:      %{_tmppath}/%{name}-%{version}-build

++++++ duply_1.5.9.tgz -> duply_1.5.10.tgz ++++++
diff -urN '--exclude=CVS' '--exclude=.cvsignore' '--exclude=.svn' 
'--exclude=.svnignore' old/duply_1.5.9/duply new/duply_1.5.10/duply
--- old/duply_1.5.9/duply       2012-11-22 18:16:27.000000000 +0100
+++ new/duply_1.5.10/duply      2013-03-26 17:45:30.000000000 +0100
@@ -38,6 +38,12 @@
 #  - import/export profile from/to .tgz function !!!
 #
 #  CHANGELOG:
+#  1.5.10 (26.03.2013)
+#  - minor indent and documentation fixes
+#  - bugfix: exclude filter failed on ubuntu, mawk w/o posix char class support
+#  - bugfix: fix url_decoding generally and for python3
+#  - bugfix 3609075: wrong script results in status line (thx David Epping)
+#
 #  1.5.9 (22.11.2012)
 #  - bugfix 3588926: filter --exclude* params for restore/fetch ate too much
 #  - restore/fetch now also ignores --include* or --exclude='foobar' 
@@ -300,7 +306,7 @@
 ME_LONG="$0"
 ME="$(basename $0)"
 ME_NAME="${ME%%.*}"
-ME_VERSION="1.5.9"
+ME_VERSION="1.5.10"
 ME_WEBSITE="http://duply.net";
 
 # default config values
@@ -578,27 +584,31 @@
 #   # for cloudfiles backend user id is CLOUDFILES_USERNAME, password is 
 #   # CLOUDFILES_APIKEY, you might need to set CLOUDFILES_AUTHURL manually
 #   cf+http://[user:password@]container_name
-#   file://[/absolute_]path
+#   file://[relative|/absolute]/local/path
 #   ftp[s]://user[:password]@other.host[:port]/some_dir
-#   gdocs://user[:password]@other.host/some_dir 
+#   hsi://user[:password]@other.host/some_dir
 #   hsi://user[:password]@other.host/some_dir
 #   imap[s]://user[:password]@host.com[/from_address_prefix]
-#   rsync://user[:password]@other.host[:port]::/module/some_dir
+#   rsync://user[:password]@host.com[:port]::[/]module/some_dir
 #   # rsync over ssh (only keyauth)
-#   rsync://[email protected][:port]/relative_path
-#   rsync://[email protected][:port]//absolute_path
+#   rsync://[email protected][:port]/[relative|/absolute]_path
 #   # for the s3 user/password are AWS_ACCESS_KEY_ID/AWS_SECRET_ACCESS_KEY
 #   s3://[user:password@]host/bucket_name[/prefix]
 #   s3+http://[user:password@]bucket_name[/prefix]
 #   # scp and sftp are aliases for the ssh backend
-#   ssh://user[:password]@other.host[:port]/some_dir
+#   ssh://user[:password]@other.host[:port]/[/]some_dir
 #   tahoe://alias/directory
+#   # for Ubuntu One set TARGET_PASS to oauth access token
+#   #   "consumer_key:consumer_secret:token:token_secret"
+#   # if non given credentials will be prompted for and one will be created
+#   u1://host_is_ignored/volume_path
+#   u1+http:///volume_path
 #   webdav[s]://user[:password]@other.host/some_dir
 # ATTENTION: characters other than A-Za-z0-9.-_.~ in user,password,path have 
 #            to be replaced by their url encoded pendants, see
 #            http://en.wikipedia.org/wiki/Url_encoding 
 #            if you define the credentials as TARGET_USER, TARGET_PASS below 
-#            duply will try to url_encode them for you if needed
+#            duply will try to url_encode them for you if need arises
 TARGET='${DEFAULT_TARGET}'
 # optionally the username/password can be defined as extra variables
 # setting them here _and_ in TARGET results in an error
@@ -832,17 +842,19 @@
 }
 
 function run_script { # run pre/post scripts
-  SCRIPT="$1"
+  local ERR=0
+  local SCRIPT="$1"
   if [ ! -z "$PREVIEW" ] ; then        
     echo $SCRIPT
   elif [ -r "$SCRIPT" ] ; then 
-       echo -n "Running '$SCRIPT' "
-       OUT=`. "$SCRIPT" 2>&1`; ERR=$?
-       [ $ERR -eq "0" ] && echo "- OK" || echo "- FAILED (code $ERR)"
-       echo -en ${OUT:+"Output: $OUT\n"} ;
+    echo -n "Running '$SCRIPT' "
+    OUT=`. "$SCRIPT" 2>&1`; ERR=$?
+    [ $ERR -eq "0" ] && echo "- OK" || echo "- FAILED (code $ERR)"
+    echo -en ${OUT:+"Output: $OUT\n"} ;
   else
     echo "Skipping n/a script '$SCRIPT'."
   fi
+  return $ERR
 }
 
 function run_cmd {
@@ -924,7 +936,7 @@
        ## in/exclude parameters are currently not supported on restores
        if [ "$cmd" = "fetch" ] || [ "$cmd" = "restore" ]; then
                # filter exclude params from fetch/restore
-               echo "$DUPL_PARAMS" | awk 
'{gsub(/--(ex|in)clude[a-z-]*(([[:space:]]+|=)[^-][[:graph:]]+)?/,"");print}'
+               echo "$DUPL_PARAMS" | awk '{gsub(/--(ex|in)clude[a-z-]*(([ 
\t]+|=)[^-][^ \t]+)?/,"");print}'
                return
        fi
        
@@ -983,7 +995,7 @@
        # date busybox with -d epoch -D %s
        date=$(date ${2:+-d $2 -D %s} ${1:++"$1"} 2> /dev/null) && \
                echo $date && return
-       ## some date commands do not support giving a time w/o setting it 
(irix,solaris,others?)
+       ## some date commands do not support giving a time w/o setting it 
systemwide (irix,solaris,others?)
        # python fallback
        date=$(python -c "import time;print 
time.strftime('${1:-$DEFAULTFORMAT}',time.localtime(${2}))" 2> /dev/null) && \
                echo $date && return
@@ -1031,7 +1043,11 @@
 
 function url_encode {
   # utilize python, silently do nothing on error - because no python no 
duplicity
-  OUT=$(python -c "import urllib; print urllib.${2}quote('$1')" 2>/dev/null ); 
ERR=$?
+  OUT=$(python -c "
+try: import urllib.request as urllib
+except ImportError: import urllib
+print(urllib.${2}quote('$1'));
+" 2>/dev/null ); ERR=$?
   [ "$ERR" -eq 0 ] && echo $OUT || echo $1
 }
 
@@ -1451,7 +1467,8 @@
   # ok here some exceptions:
   #   protocols that do not need passwords
   #   s3[+http] only needs password for write operations
-  if [ -n "$(tolower "${TARGET_URL_PROT}" | grep -e 
'^\(file\|tahoe\|ssh\|scp\|sftp\)://$')" ]; then
+  #   u1[+http] can ask for creds and create an oauth token
+  if [ -n "$(tolower "${TARGET_URL_PROT}" | grep -e 
'^\(file\|tahoe\|ssh\|scp\|sftp\|u1\(\+http\)\?\)://$')" ]; then
     : # all is well file/tahoe do not need passwords, ssh might use key auth
   elif [ -n "$(tolower "${TARGET_URL_PROT}" | grep -e '^s3\(\+http\)\?://$')" 
] && \
      [ -z "$(echo ${cmds} | grep -e '\(bkp\|incr\|full\|purge\|cleanup\)')" ]; 
then
@@ -1705,6 +1722,10 @@
 
 # command execution 
#####################################################################
 
+# urldecode url vars into plain text
+var_isset 'TARGET_URL_USER' && TARGET_URL_USER="$(url_decode 
"$TARGET_URL_USER")"
+var_isset 'TARGET_URL_PASS' && TARGET_URL_PASS="$(url_decode 
"$TARGET_URL_PASS")"
+
 # defined TARGET_USER&PASS vars replace their URL pendants 
 # (double defs already dealt with)
 var_isset 'TARGET_USER' && TARGET_URL_USER="$TARGET_USER"
@@ -1733,7 +1754,7 @@
                        BACKEND_PARAMS="$BACKEND_PARAMS CLOUDFILES_AUTHURL=$(qw 
"${CLOUDFILES_AUTHURL}")"
                fi
                ;;
-               'file'|'tahoe')
+       'file'|'tahoe')
                BACKEND_URL="${TARGET_URL_PROT}${TARGET_URL_HOSTPATH}"
                ;;
        'rsync')

-- 
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to