Module Name:    src
Committed By:   martin
Date:           Sun Dec  1 09:29:06 UTC 2013

Modified Files:
        src/external/lgpl3/gmp/dist/mpn/generic: div_qr_2.c divrem_2.c
            mod_1_1.c mod_1_2.c mod_1_3.c mod_1_4.c

Log Message:
Apply upstream fix for warnings seen on sparc64


To generate a diff of this commit:
cvs rdiff -u -r1.1.1.1 -r1.2 \
    src/external/lgpl3/gmp/dist/mpn/generic/div_qr_2.c
cvs rdiff -u -r1.2 -r1.3 src/external/lgpl3/gmp/dist/mpn/generic/divrem_2.c
cvs rdiff -u -r1.1.1.2 -r1.2 \
    src/external/lgpl3/gmp/dist/mpn/generic/mod_1_1.c \
    src/external/lgpl3/gmp/dist/mpn/generic/mod_1_2.c \
    src/external/lgpl3/gmp/dist/mpn/generic/mod_1_3.c \
    src/external/lgpl3/gmp/dist/mpn/generic/mod_1_4.c

Please note that diffs are not public domain; they are subject to the
copyright notices on the relevant files.

Modified files:

Index: src/external/lgpl3/gmp/dist/mpn/generic/div_qr_2.c
diff -u src/external/lgpl3/gmp/dist/mpn/generic/div_qr_2.c:1.1.1.1 src/external/lgpl3/gmp/dist/mpn/generic/div_qr_2.c:1.2
--- src/external/lgpl3/gmp/dist/mpn/generic/div_qr_2.c:1.1.1.1	Fri Nov 29 07:49:49 2013
+++ src/external/lgpl3/gmp/dist/mpn/generic/div_qr_2.c	Sun Dec  1 09:29:06 2013
@@ -130,10 +130,11 @@ along with the GNU MP Library.  If not, 
     umul_ppmm (_q1d,_q0, n2, di0);					\
     add_sssaaaa (_q3,_q2,_q1, _q2,_q1, _q2a,_q1d);			\
 									\
-    add_ssaaaa (r1, r0, n3, n2, 0, 1); /* FIXME: combine as in x86_64 asm */ \
+    add_ssaaaa (r1, r0, n3, n2, CNST_LIMB(0), CNST_LIMB(1));		\
+    				  /* FIXME: combine as in x86_64 asm */	\
 									\
     /* [q3,q2,q1,q0] += [n3,n3,n1,n0] */				\
-    add_csaac (_c, _q0, _q0, n0, 0);					\
+    add_csaac (_c, _q0, _q0, n0, CNST_LIMB(0));				\
     add_csaac (_c, _q1, _q1, n1, _c);					\
     add_csaac (_c, _q2, _q2, r0, _c);					\
     _q3 = _q3 + r1 + _c;						\
@@ -145,14 +146,14 @@ along with the GNU MP Library.  If not, 
 									\
     _mask = -(mp_limb_t) (r1 >= _q1 & (r1 > _q1 | r0 >= _q0));  /* (r1,r0) >= (q1,q0) */  \
     add_ssaaaa (r1, r0, r1, r0, d1 & _mask, d0 & _mask);		\
-    sub_ddmmss (_q3, _q2, _q3, _q2, 0, -_mask);				\
+    sub_ddmmss (_q3, _q2, _q3, _q2, CNST_LIMB(0), -_mask);		\
 									\
     if (UNLIKELY (r1 >= d1))						\
       {									\
 	if (r1 > d1 || r0 >= d0)					\
 	  {								\
 	    sub_ddmmss (r1, r0, r1, r0, d1, d0);			\
-	    add_ssaaaa (_q3, _q2, _q3, _q2, 0, 1);			\
+	    add_ssaaaa (_q3, _q2, _q3, _q2, CNST_LIMB(0), CNST_LIMB(1));\
 	  }								\
       }									\
     (q1) = _q3;								\

Index: src/external/lgpl3/gmp/dist/mpn/generic/divrem_2.c
diff -u src/external/lgpl3/gmp/dist/mpn/generic/divrem_2.c:1.2 src/external/lgpl3/gmp/dist/mpn/generic/divrem_2.c:1.3
--- src/external/lgpl3/gmp/dist/mpn/generic/divrem_2.c:1.2	Sat Nov 30 20:30:22 2013
+++ src/external/lgpl3/gmp/dist/mpn/generic/divrem_2.c	Sun Dec  1 09:29:06 2013
@@ -98,7 +98,7 @@ mpn_divrem_2 (mp_ptr qp, mp_size_t qxn,
       for (i = qxn - 1; i >= 0; i--)
 	{
 	  mp_limb_t q;
-	  udiv_qr_3by2 (q, r1, r0, r1, r0, 0UL, d1, d0, di.inv32);
+	  udiv_qr_3by2 (q, r1, r0, r1, r0, CNST_LIMB(0), d1, d0, di.inv32);
 	  qp[i] = q;
 	}
     }

Index: src/external/lgpl3/gmp/dist/mpn/generic/mod_1_1.c
diff -u src/external/lgpl3/gmp/dist/mpn/generic/mod_1_1.c:1.1.1.2 src/external/lgpl3/gmp/dist/mpn/generic/mod_1_1.c:1.2
--- src/external/lgpl3/gmp/dist/mpn/generic/mod_1_1.c:1.1.1.2	Fri Nov 29 07:49:49 2013
+++ src/external/lgpl3/gmp/dist/mpn/generic/mod_1_1.c	Sun Dec  1 09:29:06 2013
@@ -160,7 +160,7 @@ mpn_mod_1_1p_cps (mp_limb_t cps[4], mp_l
    *   B2modb = - b * bi;
    *   ASSERT (B2modb <= b);    // NB: equality iff b = B/2
    */
-  udiv_rnnd_preinv (B2modb, B1modb, 0, b, bi);
+  udiv_rnnd_preinv (B2modb, B1modb, CNST_LIMB(0), b, bi);
   cps[3] = B2modb >> cnt;
 }
 
Index: src/external/lgpl3/gmp/dist/mpn/generic/mod_1_2.c
diff -u src/external/lgpl3/gmp/dist/mpn/generic/mod_1_2.c:1.1.1.2 src/external/lgpl3/gmp/dist/mpn/generic/mod_1_2.c:1.2
--- src/external/lgpl3/gmp/dist/mpn/generic/mod_1_2.c:1.1.1.2	Fri Nov 29 07:49:49 2013
+++ src/external/lgpl3/gmp/dist/mpn/generic/mod_1_2.c	Sun Dec  1 09:29:06 2013
@@ -51,10 +51,10 @@ mpn_mod_1s_2p_cps (mp_limb_t cps[5], mp_
   ASSERT (B1modb <= b);		/* NB: not fully reduced mod b */
   cps[2] = B1modb >> cnt;
 
-  udiv_rnnd_preinv (B2modb, B1modb, 0, b, bi);
+  udiv_rnnd_preinv (B2modb, B1modb, CNST_LIMB(0), b, bi);
   cps[3] = B2modb >> cnt;
 
-  udiv_rnnd_preinv (B3modb, B2modb, 0, b, bi);
+  udiv_rnnd_preinv (B3modb, B2modb, CNST_LIMB(0), b, bi);
   cps[4] = B3modb >> cnt;
 
 #if WANT_ASSERT
Index: src/external/lgpl3/gmp/dist/mpn/generic/mod_1_3.c
diff -u src/external/lgpl3/gmp/dist/mpn/generic/mod_1_3.c:1.1.1.2 src/external/lgpl3/gmp/dist/mpn/generic/mod_1_3.c:1.2
--- src/external/lgpl3/gmp/dist/mpn/generic/mod_1_3.c:1.1.1.2	Fri Nov 29 07:49:49 2013
+++ src/external/lgpl3/gmp/dist/mpn/generic/mod_1_3.c	Sun Dec  1 09:29:06 2013
@@ -51,13 +51,13 @@ mpn_mod_1s_3p_cps (mp_limb_t cps[6], mp_
   ASSERT (B1modb <= b);		/* NB: not fully reduced mod b */
   cps[2] = B1modb >> cnt;
 
-  udiv_rnnd_preinv (B2modb, B1modb, 0, b, bi);
+  udiv_rnnd_preinv (B2modb, B1modb, CNST_LIMB(0), b, bi);
   cps[3] = B2modb >> cnt;
 
-  udiv_rnnd_preinv (B3modb, B2modb, 0, b, bi);
+  udiv_rnnd_preinv (B3modb, B2modb, CNST_LIMB(0), b, bi);
   cps[4] = B3modb >> cnt;
 
-  udiv_rnnd_preinv (B4modb, B3modb, 0, b, bi);
+  udiv_rnnd_preinv (B4modb, B3modb, CNST_LIMB(0), b, bi);
   cps[5] = B4modb >> cnt;
 
 #if WANT_ASSERT
Index: src/external/lgpl3/gmp/dist/mpn/generic/mod_1_4.c
diff -u src/external/lgpl3/gmp/dist/mpn/generic/mod_1_4.c:1.1.1.2 src/external/lgpl3/gmp/dist/mpn/generic/mod_1_4.c:1.2
--- src/external/lgpl3/gmp/dist/mpn/generic/mod_1_4.c:1.1.1.2	Fri Nov 29 07:49:49 2013
+++ src/external/lgpl3/gmp/dist/mpn/generic/mod_1_4.c	Sun Dec  1 09:29:06 2013
@@ -51,16 +51,16 @@ mpn_mod_1s_4p_cps (mp_limb_t cps[7], mp_
   ASSERT (B1modb <= b);		/* NB: not fully reduced mod b */
   cps[2] = B1modb >> cnt;
 
-  udiv_rnnd_preinv (B2modb, B1modb, 0, b, bi);
+  udiv_rnnd_preinv (B2modb, B1modb, CNST_LIMB(0), b, bi);
   cps[3] = B2modb >> cnt;
 
-  udiv_rnnd_preinv (B3modb, B2modb, 0, b, bi);
+  udiv_rnnd_preinv (B3modb, B2modb, CNST_LIMB(0), b, bi);
   cps[4] = B3modb >> cnt;
 
-  udiv_rnnd_preinv (B4modb, B3modb, 0, b, bi);
+  udiv_rnnd_preinv (B4modb, B3modb, CNST_LIMB(0), b, bi);
   cps[5] = B4modb >> cnt;
 
-  udiv_rnnd_preinv (B5modb, B4modb, 0, b, bi);
+  udiv_rnnd_preinv (B5modb, B4modb, CNST_LIMB(0), b, bi);
   cps[6] = B5modb >> cnt;
 
 #if WANT_ASSERT

Reply via email to