On Wednesday 24 December 2008 02:23:22 Jason Martin wrote:
> Thanks Jason & Bill,
>
> I'll try digging into those errors in two day (Christmas/Family
> obligations at the moment).  I really really dislike
> auto{conf|make|tool}, but I guess I need to embrace it... seems to be
> the only game in town.

The tag update has fixed the make speed , make tune errors


>
> --jason
>
> Jason Worth Martin
> Asst. Professor of Mathematics
> http://www.math.jmu.edu/~martin
>
> On Tue, Dec 23, 2008 at 7:51 PM,  <ja...@njkfrudils.plus.com> wrote:
> > On Wednesday 24 December 2008 00:31:50 Bill Hart wrote:
> >> On sage.math:
> >>
> >> cd tune
> >> make tune
> >>
> >> ./.libs/libspeed.a(gcd_bin.o): In function `__gmpn_ngcd_matrix_init':
> >> gcd_bin.c:(.text+0x0): multiple definition of `__gmpn_ngcd_matrix_init'
> >> gcd.o:gcd.c:(.text+0x0): first defined here
> >> ./.libs/libspeed.a(gcd_bin.o): In function `__gmpn_nhgcd_itch':
> >> gcd_bin.c:(.text+0x80): multiple definition of `__gmpn_nhgcd_itch'
> >> gcd.o:gcd.c:(.text+0x80): first defined here
> >> ./.libs/libspeed.a(gcd_bin.o): In function `__gmpn_nhgcd':
> >> gcd_bin.c:(.text+0xc4): multiple definition of `__gmpn_nhgcd'
> >> gcd.o:gcd.c:(.text+0xc4): first defined here
> >> ./.libs/libspeed.a(gcd_bin.o): In function `mpn_basic_gcd':
> >> gcd_bin.c:(.text+0x2ed): multiple definition of `mpn_basic_gcd'
> >> gcd.o:gcd.c:(.text+0x2ed): first defined here
> >
> > On my K8-linux same problem with make speed and tune
> >
> > ./configure && make && make check passes OK , and I got bored running
> > ./try mpn_gcd :)
> >
> > get this warning from the build though
> >
> > ngcd.c: In function 'mpn_ngcd':
> > ngcd.c:75: warning: implicit declaration of function 'mpn_basic_gcd'
> >
> >
> > gcc -std=gnu99 -DHAVE_CONFIG_H -I. -I.. -D__GMP_WITHIN_GMP -I.. -O2 -m64
> > -march=k8 -mtune=k8 -c gcd.c  -fPIC -DPIC -o .libs/gcd.o
> > gcd.c: In function 'mpz_rgcd':
> > gcd.c:167: warning: implicit declaration of function 'mpn_rgcd'
> > gcd.c: In function 'mpz_bgcd':
> > gcd.c:171: warning: implicit declaration of function 'mpn_bgcd'
> > gcd.c: In function 'mpz_sgcd':
> > gcd.c:175: warning: implicit declaration of function 'mpn_sgcd'
> > gcd.c: In function 'mpz_ngcd':
> > gcd.c:179: warning: implicit declaration of function 'mpn_ngcd'
> >
> >
> > ./configure  --enable-alloca=debug --enable-assert && make && make check
> > passes OK
> >
> > but
> >
> > ./configure  --enable-alloca=debug --enable-assert --enable-nails=2 &&
> > make && make check
> >
> > PASS: t-mul_i
> > PASS: t-tdiv
> > PASS: t-tdiv_ui
> > PASS: t-fdiv
> > PASS: t-fdiv_ui
> > PASS: t-cdiv_ui
> > nhgcd2.c:206: GNU MP assertion failed: h0 == h1
> > /bin/sh: line 4: 31599 Aborted                 ${dir}$tst
> > FAIL: t-gcd
> > PASS: t-gcd_ui
> > nhgcd2.c:206: GNU MP assertion failed: h0 == h1
> > /bin/sh: line 4: 31647 Aborted                 ${dir}$tst
> > FAIL: t-lcm
> > PASS: dive
> > PASS: dive_ui
> > PASS: t-sqrtrem
> > PASS: convert
> > PASS: io
> >
> >> Bill.
> >>
> >> 2008/12/23 Jason Martin <jason.worth.mar...@gmail.com>:
> >> > Attached are some edited versions of
> >> >
> >> > mpn/generic/gcd.c
> >> >
> >> > and
> >> >
> >> > mpn/generic/ngcd.c
> >> >
> >> > Drop them in, test them for correctness and speed.  Let me know what
> >> > breaks.  When everyone is happy, I'll check them in to svn
> >> >
> >> > --jason
> >> >
> >> > Jason Worth Martin
> >> > Asst. Professor of Mathematics
> >> > http://www.math.jmu.edu/~martin
> >> >
> >> >
> >> >
> >> > /* Schönhage's 1987 algorithm, reorganized into hgcd form */
> >> >
> >> > #include <stdio.h>  /* for NULL */
> >> >
> >> > #include "gmp.h"
> >> > #include "gmp-impl.h"
> >> > #include "longlong.h"
> >> >
> >> >
> >> > /* ******************************************************************
> >> >  *    Here we are including the original GMP version of mpn_gcd
> >> >  *    but we rename it as mpn_basic_gcd.  It needs to be available
> >> >  *    for the ngcd algorithm to call in the base case.
> >> >  *
> >> >  *  Preconditions [U = (up, usize) and V = (vp, vsize)]:
> >> >  *
> >> >  *   1.  V is odd.
> >> >  *   2.  numbits(U) >= numbits(V).
> >> >  *
> >> >  *   Both U and V are destroyed by the operation.  The result is left
> >> > at vp, *   and its size is returned.
> >> >  *
> >> >  *   Ken Weber (kwe...@mat.ufrgs.br, kwe...@mcs.kent.edu)
> >> >  *
> >> >  *   Funding for this work has been partially provided by Conselho
> >> >  *   Nacional de Desenvolvimento Cienti'fico e Tecnolo'gico (CNPq) do
> >> >  *   Brazil, Grant 301314194-2, and was done while I was a visiting
> >> >  *   reseacher in the Instituto de Matema'tica at Universidade Federal
> >> >  *   do Rio Grande do Sul (UFRGS).
> >> >  *
> >> >  *   Refer to K. Weber, The accelerated integer GCD algorithm, ACM
> >> >  *      Transactions on Mathematical Software, v. 21 (March), 1995,
> >> >  *      pp. 111-122.
> >> >  *
> >> >  * *****************************************************************/
> >> >
> >> >
> >> >
> >> > /* If MIN (usize, vsize) >= GCD_ACCEL_THRESHOLD, then the accelerated
> >> >   algorithm is used, otherwise the binary algorithm is used.  This may
> >> > be adjusted for different architectures.  */
> >> > #ifndef GCD_ACCEL_THRESHOLD
> >> > #define GCD_ACCEL_THRESHOLD 5
> >> > #endif
> >> >
> >> > /* When U and V differ in size by more than BMOD_THRESHOLD, the
> >> > accelerated algorithm reduces using the bmod operation.  Otherwise,
> >> > the k-ary reduction is used.  0 <= BMOD_THRESHOLD < GMP_NUMB_BITS.  */
> >> > enum
> >> >  {
> >> >    BMOD_THRESHOLD = GMP_NUMB_BITS/2
> >> >  };
> >> >
> >> >
> >> > /* Use binary algorithm to compute V <-- GCD (V, U) for usize, vsize
> >> > == 2. Both U and V must be odd.  */
> >> > static inline mp_size_t
> >> > gcd_2 (mp_ptr vp, mp_srcptr up)
> >> > {
> >> >  mp_limb_t u0, u1, v0, v1;
> >> >  mp_size_t vsize;
> >> >
> >> >  u0 = up[0];
> >> >  u1 = up[1];
> >> >  v0 = vp[0];
> >> >  v1 = vp[1];
> >> >
> >> >  while (u1 != v1 && u0 != v0)
> >> >    {
> >> >      unsigned long int r;
> >> >      if (u1 > v1)
> >> >        {
> >> >          u1 -= v1 + (u0 < v0);
> >> >          u0 = (u0 - v0) & GMP_NUMB_MASK;
> >> >          count_trailing_zeros (r, u0);
> >> >          u0 = ((u1 << (GMP_NUMB_BITS - r)) & GMP_NUMB_MASK) | (u0 >>
> >> > r); u1 >>= r;
> >> >        }
> >> >      else  /* u1 < v1.  */
> >> >        {
> >> >          v1 -= u1 + (v0 < u0);
> >> >          v0 = (v0 - u0) & GMP_NUMB_MASK;
> >> >          count_trailing_zeros (r, v0);
> >> >          v0 = ((v1 << (GMP_NUMB_BITS - r)) & GMP_NUMB_MASK) | (v0 >>
> >> > r); v1 >>= r;
> >> >        }
> >> >    }
> >> >
> >> >  vp[0] = v0, vp[1] = v1, vsize = 1 + (v1 != 0);
> >> >
> >> >  /* If U == V == GCD, done.  Otherwise, compute GCD (V, |U - V|).  */
> >> >  if (u1 == v1 && u0 == v0)
> >> >    return vsize;
> >> >
> >> >  v0 = (u0 == v0) ? (u1 > v1) ? u1-v1 : v1-u1 : (u0 > v0) ? u0-v0 :
> >> > v0-u0; vp[0] = mpn_gcd_1 (vp, vsize, v0);
> >> >
> >> >  return 1;
> >> > }
> >> >
> >> > /* The function find_a finds 0 < N < 2^GMP_NUMB_BITS such that there
> >> > exists 0 < |D| < 2^GMP_NUMB_BITS, and N == D * C mod
> >> > 2^(2*GMP_NUMB_BITS). In the reference article, D was computed along
> >> > with N, but it is better to compute D separately as D <-- N / C mod
> >> > 2^(GMP_NUMB_BITS + 1), treating the result as a twos' complement
> >> > signed integer.
> >> >
> >> >   Initialize N1 to C mod 2^(2*GMP_NUMB_BITS).  According to the
> >> > reference article, N2 should be initialized to 2^(2*GMP_NUMB_BITS),
> >> > but we use 2^(2*GMP_NUMB_BITS) - N1 to start the calculations within
> >> > double precision.  If N2 > N1 initially, the first iteration of the
> >> > while loop will swap them.  In all other situations, N1 >= N2 is
> >> > maintained.  */
> >> >
> >> > #if HAVE_NATIVE_mpn_gcd_finda
> >> > #define find_a(cp)  mpn_gcd_finda (cp)
> >> >
> >> > #else
> >> > static
> >> > #if ! defined (__i386__)
> >> > inline                          /* don't inline this for the x86 */
> >> > #endif
> >> > mp_limb_t
> >> > find_a (mp_srcptr cp)
> >> > {
> >> >  unsigned long int leading_zero_bits = 0;
> >> >
> >> >  mp_limb_t n1_l = cp[0];       /* N1 == n1_h * 2^GMP_NUMB_BITS + n1_l.
> >> > */ mp_limb_t n1_h = cp[1];
> >> >
> >> >  mp_limb_t n2_l = (-n1_l & GMP_NUMB_MASK);     /* N2 == n2_h *
> >> > 2^GMP_NUMB_BITS + n2_l.  */ mp_limb_t n2_h = (~n1_h & GMP_NUMB_MASK);
> >> >
> >> >  /* Main loop.  */
> >> >  while (n2_h != 0)             /* While N2 >= 2^GMP_NUMB_BITS.  */
> >> >    {
> >> >      /* N1 <-- N1 % N2.  */
> >> >      if (((GMP_NUMB_HIGHBIT >> leading_zero_bits) & n2_h) == 0)
> >> >        {
> >> >          unsigned long int i;
> >> >          count_leading_zeros (i, n2_h);
> >> >          i -= GMP_NAIL_BITS;
> >> >          i -= leading_zero_bits;
> >> >          leading_zero_bits += i;
> >> >          n2_h = ((n2_h << i) & GMP_NUMB_MASK) | (n2_l >>
> >> > (GMP_NUMB_BITS - i)); n2_l = (n2_l << i) & GMP_NUMB_MASK;
> >> >          do
> >> >            {
> >> >              if (n1_h > n2_h || (n1_h == n2_h && n1_l >= n2_l))
> >> >                {
> >> >                  n1_h -= n2_h + (n1_l < n2_l);
> >> >                  n1_l = (n1_l - n2_l) & GMP_NUMB_MASK;
> >> >                }
> >> >              n2_l = (n2_l >> 1) | ((n2_h << (GMP_NUMB_BITS - 1)) &
> >> > GMP_NUMB_MASK); n2_h >>= 1;
> >> >              i -= 1;
> >> >            }
> >> >          while (i != 0);
> >> >        }
> >> >      if (n1_h > n2_h || (n1_h == n2_h && n1_l >= n2_l))
> >> >        {
> >> >          n1_h -= n2_h + (n1_l < n2_l);
> >> >          n1_l = (n1_l - n2_l) & GMP_NUMB_MASK;
> >> >        }
> >> >
> >> >      MP_LIMB_T_SWAP (n1_h, n2_h);
> >> >      MP_LIMB_T_SWAP (n1_l, n2_l);
> >> >    }
> >> >
> >> >  return n2_l;
> >> > }
> >> > #endif
> >> >
> >> >
> >> > mp_size_t
> >> > mpn_basic_gcd (mp_ptr gp, mp_ptr up, mp_size_t usize, mp_ptr vp,
> >> > mp_size_t vsize) {
> >> >  mp_ptr orig_vp = vp;
> >> >  mp_size_t orig_vsize = vsize;
> >> >  int binary_gcd_ctr;           /* Number of times binary gcd will
> >> > execute.  */ mp_size_t scratch;
> >> >  mp_ptr tp;
> >> >  TMP_DECL;
> >> >
> >> >  ASSERT (usize >= 1);
> >> >  ASSERT (vsize >= 1);
> >> >  ASSERT (usize >= vsize);
> >> >  ASSERT (vp[0] & 1);
> >> >  ASSERT (up[usize - 1] != 0);
> >> >  ASSERT (vp[vsize - 1] != 0);
> >> > #if WANT_ASSERT
> >> >  if (usize == vsize)
> >> >    {
> >> >      int  uzeros, vzeros;
> >> >      count_leading_zeros (uzeros, up[usize - 1]);
> >> >      count_leading_zeros (vzeros, vp[vsize - 1]);
> >> >      ASSERT (uzeros <= vzeros);
> >> >    }
> >> > #endif
> >> >  ASSERT (! MPN_OVERLAP_P (up, usize, vp, vsize));
> >> >  ASSERT (MPN_SAME_OR_SEPARATE2_P (gp, vsize, up, usize));
> >> >  ASSERT (MPN_SAME_OR_SEPARATE2_P (gp, vsize, vp, vsize));
> >> >
> >> >  TMP_MARK;
> >> >
> >> >  /* Use accelerated algorithm if vsize is over GCD_ACCEL_THRESHOLD.
> >> >     Two EXTRA limbs for U and V are required for kary reduction.  */
> >> >  if (vsize >= GCD_ACCEL_THRESHOLD)
> >> >    {
> >> >      unsigned long int vbitsize, d;
> >> >      mp_ptr orig_up = up;
> >> >      mp_size_t orig_usize = usize;
> >> >      mp_ptr anchor_up = (mp_ptr) TMP_ALLOC ((usize + 2) *
> >> > BYTES_PER_MP_LIMB);
> >> >
> >> >      MPN_COPY (anchor_up, orig_up, usize);
> >> >      up = anchor_up;
> >> >
> >> >      count_leading_zeros (d, up[usize - 1]);
> >> >      d -= GMP_NAIL_BITS;
> >> >      d = usize * GMP_NUMB_BITS - d;
> >> >      count_leading_zeros (vbitsize, vp[vsize - 1]);
> >> >      vbitsize -= GMP_NAIL_BITS;
> >> >      vbitsize = vsize * GMP_NUMB_BITS - vbitsize;
> >> >      ASSERT (d >= vbitsize);
> >> >      d = d - vbitsize + 1;
> >> >
> >> >      /* Use bmod reduction to quickly discover whether V divides U. 
> >> > */ up[usize++] = 0;                          /* Insert leading zero.
> >> > */ mpn_bdivmod (up, up, usize, vp, vsize, d);
> >> >
> >> >      /* Now skip U/V mod 2^d and any low zero limbs.  */
> >> >      d /= GMP_NUMB_BITS, up += d, usize -= d;
> >> >      while (usize != 0 && up[0] == 0)
> >> >        up++, usize--;
> >> >
> >> >      if (usize == 0)                           /* GCD == ORIG_V.  */
> >> >        goto done;
> >> >
> >> >      vp = (mp_ptr) TMP_ALLOC ((vsize + 2) * BYTES_PER_MP_LIMB);
> >> >      MPN_COPY (vp, orig_vp, vsize);
> >> >
> >> >      do                                        /* Main loop.  */
> >> >        {
> >> >          /* mpn_com_n can't be used here because anchor_up and up may
> >> >             partially overlap */
> >> >          if ((up[usize - 1] & GMP_NUMB_HIGHBIT) != 0)  /* U < 0; take
> >> > twos' compl. */ {
> >> >              mp_size_t i;
> >> >              anchor_up[0] = -up[0] & GMP_NUMB_MASK;
> >> >              for (i = 1; i < usize; i++)
> >> >                anchor_up[i] = (~up[i] & GMP_NUMB_MASK);
> >> >              up = anchor_up;
> >> >            }
> >> >
> >> >          MPN_NORMALIZE_NOT_ZERO (up, usize);
> >> >
> >> >          if ((up[0] & 1) == 0)                 /* Result even; remove
> >> > twos. */ {
> >> >              unsigned int r;
> >> >              count_trailing_zeros (r, up[0]);
> >> >              mpn_rshift (anchor_up, up, usize, r);
> >> >              usize -= (anchor_up[usize - 1] == 0);
> >> >            }
> >> >          else if (anchor_up != up)
> >> >            MPN_COPY_INCR (anchor_up, up, usize);
> >> >
> >> >          MPN_PTR_SWAP (anchor_up,usize, vp,vsize);
> >> >          up = anchor_up;
> >> >
> >> >          if (vsize <= 2)               /* Kary can't handle < 2 limbs
> >> > and */ break;                      /* isn't efficient for == 2 limbs. 
> >> > */
> >> >
> >> >          d = vbitsize;
> >> >          count_leading_zeros (vbitsize, vp[vsize - 1]);
> >> >          vbitsize -= GMP_NAIL_BITS;
> >> >          vbitsize = vsize * GMP_NUMB_BITS - vbitsize;
> >> >          d = d - vbitsize + 1;
> >> >
> >> >          if (d > BMOD_THRESHOLD)       /* Bmod reduction.  */
> >> >            {
> >> >              up[usize++] = 0;
> >> >              mpn_bdivmod (up, up, usize, vp, vsize, d);
> >> >              d /= GMP_NUMB_BITS, up += d, usize -= d;
> >> >            }
> >> >          else                          /* Kary reduction.  */
> >> >            {
> >> >              mp_limb_t bp[2], cp[2];
> >> >
> >> >              /* C <-- V/U mod 2^(2*GMP_NUMB_BITS).  */
> >> >              {
> >> >                mp_limb_t u_inv, hi, lo;
> >> >                modlimb_invert (u_inv, up[0]);
> >> >                cp[0] = (vp[0] * u_inv) & GMP_NUMB_MASK;
> >> >                umul_ppmm (hi, lo, cp[0], up[0] << GMP_NAIL_BITS);
> >> >                lo >>= GMP_NAIL_BITS;
> >> >                cp[1] = (vp[1] - hi - cp[0] * up[1]) * u_inv &
> >> > GMP_NUMB_MASK; }
> >> >
> >> >              /* U <-- find_a (C)  *  U.  */
> >> >              up[usize] = mpn_mul_1 (up, up, usize, find_a (cp));
> >> >              usize++;
> >> >
> >> >              /* B <-- A/C == U/V mod 2^(GMP_NUMB_BITS + 1).
> >> >                  bp[0] <-- U/V mod 2^GMP_NUMB_BITS and
> >> >                  bp[1] <-- ( (U - bp[0] * V)/2^GMP_NUMB_BITS ) / V mod
> >> > 2
> >> >
> >> >                  Like V/U above, but simplified because only the low
> >> > bit of bp[1] is wanted. */
> >> >              {
> >> >                mp_limb_t  v_inv, hi, lo;
> >> >                modlimb_invert (v_inv, vp[0]);
> >> >                bp[0] = (up[0] * v_inv) & GMP_NUMB_MASK;
> >> >                umul_ppmm (hi, lo, bp[0], vp[0] << GMP_NAIL_BITS);
> >> >                lo >>= GMP_NAIL_BITS;
> >> >                bp[1] = (up[1] + hi + (bp[0] & vp[1])) & 1;
> >> >              }
> >> >
> >> >              up[usize++] = 0;
> >> >              if (bp[1] != 0)   /* B < 0: U <-- U + (-B)  * V.  */
> >> >                {
> >> >                   mp_limb_t c = mpn_addmul_1 (up, vp, vsize, -bp[0] &
> >> > GMP_NUMB_MASK); mpn_add_1 (up + vsize, up + vsize, usize - vsize, c);
> >> > } else              /* B >= 0:  U <-- U - B * V.  */ {
> >> >                  mp_limb_t b = mpn_submul_1 (up, vp, vsize, bp[0]);
> >> >                  mpn_sub_1 (up + vsize, up + vsize, usize - vsize, b);
> >> >                }
> >> >
> >> >              up += 2, usize -= 2;  /* At least two low limbs are zero.
> >> > */ }
> >> >
> >> >          /* Must remove low zero limbs before complementing.  */
> >> >          while (usize != 0 && up[0] == 0)
> >> >            up++, usize--;
> >> >        }
> >> >      while (usize != 0);
> >> >
> >> >      /* Compute GCD (ORIG_V, GCD (ORIG_U, V)).  Binary will execute
> >> > twice.  */ up = orig_up, usize = orig_usize;
> >> >      binary_gcd_ctr = 2;
> >> >    }
> >> >  else
> >> >    binary_gcd_ctr = 1;
> >> >
> >> >  scratch = MPN_NGCD_LEHMER_ITCH (vsize);
> >> >  if (usize + 1 > scratch)
> >> >    scratch = usize + 1;
> >> >
> >> >  tp = TMP_ALLOC_LIMBS (scratch);
> >> >
> >> >  /* Finish up with the binary algorithm.  Executes once or twice.  */
> >> >  for ( ; binary_gcd_ctr--; up = orig_vp, usize = orig_vsize)
> >> >    {
> >> > #if 1
> >> >      if (usize > vsize)
> >> >        {
> >> >          /* FIXME: Could use mpn_bdivmod. */
> >> >          mp_size_t rsize;
> >> >
> >> >          mpn_tdiv_qr (tp + vsize, tp, 0, up, usize, vp, vsize);
> >> >          rsize = vsize;
> >> >          MPN_NORMALIZE (tp, rsize);
> >> >          if (rsize == 0)
> >> >            continue;
> >> >
> >> >          MPN_COPY (up, tp, vsize);
> >> >        }
> >> >      vsize = mpn_ngcd_lehmer (vp, up, vp, vsize, tp);
> >> > #else
> >> >      if (usize > 2)            /* First make U close to V in size.  */
> >> >        {
> >> >          unsigned long int vbitsize, d;
> >> >          count_leading_zeros (d, up[usize - 1]);
> >> >          d -= GMP_NAIL_BITS;
> >> >          d = usize * GMP_NUMB_BITS - d;
> >> >          count_leading_zeros (vbitsize, vp[vsize - 1]);
> >> >          vbitsize -= GMP_NAIL_BITS;
> >> >          vbitsize = vsize * GMP_NUMB_BITS - vbitsize;
> >> >          d = d - vbitsize - 1;
> >> >          if (d != -(unsigned long int)1 && d > 2)
> >> >            {
> >> >              mpn_bdivmod (up, up, usize, vp, vsize, d);  /* Result >
> >> > 0. */ d /= (unsigned long int)GMP_NUMB_BITS, up += d, usize -= d; } }
> >> >
> >> >      /* Start binary GCD.  */
> >> >      do
> >> >        {
> >> >          mp_size_t zeros;
> >> >
> >> >          /* Make sure U is odd.  */
> >> >          MPN_NORMALIZE (up, usize);
> >> >          while (up[0] == 0)
> >> >            up += 1, usize -= 1;
> >> >          if ((up[0] & 1) == 0)
> >> >            {
> >> >              unsigned int r;
> >> >              count_trailing_zeros (r, up[0]);
> >> >              mpn_rshift (up, up, usize, r);
> >> >              usize -= (up[usize - 1] == 0);
> >> >            }
> >> >
> >> >          /* Keep usize >= vsize.  */
> >> >          if (usize < vsize)
> >> >            MPN_PTR_SWAP (up, usize, vp, vsize);
> >> >
> >> >          if (usize <= 2)                               /* Double
> >> > precision. */ {
> >> >              if (vsize == 1)
> >> >                vp[0] = mpn_gcd_1 (up, usize, vp[0]);
> >> >              else
> >> >                vsize = gcd_2 (vp, up);
> >> >              break;                                    /* Binary GCD
> >> > done.  */ }
> >> >
> >> >          /* Count number of low zero limbs of U - V.  */
> >> >          for (zeros = 0; up[zeros] == vp[zeros] && ++zeros != vsize; )
> >> >            continue;
> >> >
> >> >          /* If U < V, swap U and V; in any case, subtract V from U. 
> >> > */ if (zeros == vsize)                           /* Subtract done. */
> >> > up += zeros, usize -= zeros;
> >> >          else if (usize == vsize)
> >> >            {
> >> >              mp_size_t size = vsize;
> >> >              do
> >> >                size--;
> >> >              while (up[size] == vp[size]);
> >> >              if (up[size] < vp[size])                  /* usize ==
> >> > vsize. */ MP_PTR_SWAP (up, vp);
> >> >              up += zeros, usize = size + 1 - zeros;
> >> >              mpn_sub_n (up, up, vp + zeros, usize);
> >> >            }
> >> >          else
> >> >            {
> >> >              mp_size_t size = vsize - zeros;
> >> >              up += zeros, usize -= zeros;
> >> >              if (mpn_sub_n (up, up, vp + zeros, size))
> >> >                {
> >> >                  while (up[size] == 0)                 /* Propagate
> >> > borrow. */ up[size++] = -(mp_limb_t)1;
> >> >                  up[size] -= 1;
> >> >                }
> >> >            }
> >> >        }
> >> >      while (usize);                                    /* End binary
> >> > GCD. */ #endif
> >> >    }
> >> >
> >> > done:
> >> >  if (vp != gp)
> >> >    MPN_COPY_INCR (gp, vp, vsize);
> >> >  TMP_FREE;
> >> >  return vsize;
> >> > }
> >> >
> >> >
> >> >
> >> > /* ******************************************************************
> >> >  *     END of original GMP mpn_gcd
> >> >  * *****************************************************************/
> >> >
> >> >
> >> >
> >> >
> >> >
> >> > /* For input of size n, matrix elements are of size at most ceil(n/2)
> >> >   - 1, but we need one limb extra. */
> >> >
> >> > void
> >> > mpn_ngcd_matrix_init (struct ngcd_matrix *M, mp_size_t n, mp_ptr p)
> >> > {
> >> >  mp_size_t s = (n+1)/2;
> >> >  M->alloc = s;
> >> >  M->n = 1;
> >> >  MPN_ZERO (p, 4 * s);
> >> >  M->p[0][0] = p;
> >> >  M->p[0][1] = p + s;
> >> >  M->p[1][0] = p + 2 * s;
> >> >  M->p[1][1] = p + 3 * s;
> >> >  M->tp = p + 4*s;
> >> >
> >> >  M->p[0][0][0] = M->p[1][1][0] = 1;
> >> > }
> >> >
> >> > #define NHGCD_BASE_ITCH MPN_NGCD_STEP_ITCH
> >> >
> >> > /* Reduces a,b until |a-b| fits in n/2 + 1 limbs. Constructs matrix M
> >> >   with elements of size at most (n+1)/2 - 1. Returns new size of a,
> >> >   b, or zero if no reduction is possible. */
> >> > static mp_size_t
> >> > nhgcd_base (mp_ptr ap, mp_ptr bp, mp_size_t n,
> >> >            struct ngcd_matrix *M, mp_ptr tp)
> >> > {
> >> >  mp_size_t s = n/2 + 1;
> >> >  mp_size_t nn;
> >> >
> >> >  ASSERT (n > s);
> >> >  ASSERT (ap[n-1] > 0 || bp[n-1] > 0);
> >> >
> >> >  nn = mpn_ngcd_step (n, ap, bp, s, M, tp);
> >> >  if (!nn)
> >> >    return 0;
> >> >
> >> >  for (;;)
> >> >    {
> >> >      n = nn;
> >> >      ASSERT (n > s);
> >> >      nn = mpn_ngcd_step (n, ap, bp, s, M, tp);
> >> >      if (!nn )
> >> >        return n;
> >> >    }
> >> > }
> >> >
> >> > /* Size analysis for nhgcd:
> >> >
> >> >   For the recursive calls, we have n1 <= ceil(n / 2). Then the
> >> >   storage need is determined by the storage for the recursive call
> >> >   computing M1, and ngcd_matrix_adjust and ngcd_matrix_mul calls that
> >> > use M1 (after this, the storage needed for M1 can be recycled).
> >> >
> >> >   Let S(r) denote the required storage. For M1 we need 5 * ceil(n1/2)
> >> >   = 5 * ceil(n/4), and for the ngcd_matrix_adjust call, we need n + 2.
> >> > In total, 5 * ceil(n/4) + n + 2 <= 9 ceil(n/4) + 2.
> >> >
> >> >   For the recursive call, we need S(n1) = S(ceil(n/2)).
> >> >
> >> >   S(n) <= 9*ceil(n/4) + 2 + S(ceil(n/2))
> >> >        <= 9*(ceil(n/4) + ... + ceil(n/2^(1+k))) + 2k + S(ceil(n/2^k))
> >> >        <= 9*(2 ceil(n/4) + k) + 2k + S(n/2^k)
> >> >        <= 18 ceil(n/4) + 11k + S(n/2^k)
> >> >
> >> > */
> >> >
> >> > mp_size_t
> >> > mpn_nhgcd_itch (mp_size_t n)
> >> > {
> >> >  unsigned k;
> >> >  mp_size_t nn;
> >> >
> >> >  /* Inefficient way to almost compute
> >> >     log_2(n/NHGCD_BASE_THRESHOLD) */
> >> >  for (k = 0, nn = n;
> >> >       ABOVE_THRESHOLD (nn, NHGCD_THRESHOLD);
> >> >       nn = (nn + 1) / 2)
> >> >    k++;
> >> >
> >> >  if (k == 0)
> >> >    return NHGCD_BASE_ITCH (n);
> >> >
> >> >  return 18 * ((n+3) / 4) + 11 * k
> >> >    + NHGCD_BASE_ITCH (NHGCD_THRESHOLD);
> >> > }
> >> >
> >> > /* Reduces a,b until |a-b| fits in n/2 + 1 limbs. Constructs matrix M
> >> >   with elements of size at most (n+1)/2 - 1. Returns new size of a,
> >> >   b, or zero if no reduction is possible. */
> >> >
> >> > mp_size_t
> >> > mpn_nhgcd (mp_ptr ap, mp_ptr bp, mp_size_t n,
> >> >           struct ngcd_matrix *M, mp_ptr tp)
> >> > {
> >> >  mp_size_t s = n/2 + 1;
> >> >  mp_size_t n2 = (3*n)/4 + 1;
> >> >
> >> >  mp_size_t p, nn;
> >> >  unsigned count;
> >> >  int success = 0;
> >> >
> >> >  ASSERT (n > s);
> >> >  ASSERT ((ap[n-1] | bp[n-1]) > 0);
> >> >
> >> >  ASSERT ((n+1)/2 - 1 < M->alloc);
> >> >
> >> >  if (BELOW_THRESHOLD (n, NHGCD_THRESHOLD))
> >> >    return nhgcd_base (ap, bp, n, M, tp);
> >> >
> >> >  p = n/2;
> >> >  nn = mpn_nhgcd (ap + p, bp + p, n - p, M, tp);
> >> >  if (nn > 0)
> >> >    {
> >> >      /* Needs 2*(p + M->n) <= 2*(floor(n/2) + ceil(n/2) - 1)
> >> >         = 2 (n - 1) */
> >> >      n = mpn_ngcd_matrix_adjust (M, p + nn, ap, bp, p, tp);
> >> >      success = 1;
> >> >    }
> >> >  count = 0;
> >> >  while (n > n2)
> >> >    {
> >> >      count++;
> >> >      /* Needs n + 1 storage */
> >> >      nn = mpn_ngcd_step (n, ap, bp, s, M, tp);
> >> >      if (!nn)
> >> >        return success ? n : 0;
> >> >      n = nn;
> >> >      success = 1;
> >> >    }
> >> >
> >> >  if (n > s + 2)
> >> >    {
> >> >      struct ngcd_matrix M1;
> >> >      mp_size_t scratch;
> >> >
> >> >      p = 2*s - n + 1;
> >> >      scratch = MPN_NGCD_MATRIX_INIT_ITCH (n-p);
> >> >
> >> >      mpn_ngcd_matrix_init(&M1, n - p, tp);
> >> >      nn = mpn_nhgcd (ap + p, bp + p, n - p, &M1, tp + scratch);
> >> >      if (nn > 0)
> >> >        {
> >> >          /* Needs 2 (p + M->n) <= 2 (2*s - n2 + 1 + n2 - s - 1)
> >> >             = 2*s <= 2*(floor(n/2) + 1) <= n + 2. */
> >> >          n = mpn_ngcd_matrix_adjust (&M1, p + nn, ap, bp, p, tp +
> >> > scratch); /* Needs M->n <= n2 - s - 1 < n/4 */
> >> >          mpn_ngcd_matrix_mul (M, &M1, tp + scratch);
> >> >          success = 1;
> >> >        }
> >> >    }
> >> >
> >> >  /* FIXME: This really is the base case */
> >> >  for (count = 0;; count++)
> >> >    {
> >> >      /* Needs s+3 < n */
> >> >      nn = mpn_ngcd_step (n, ap, bp, s, M, tp);
> >> >      if (!nn)
> >> >        return success ? n : 0;
> >> >
> >> >      n = nn;
> >> >      success = 1;
> >> >    }
> >> > }
> >> >
> >> > #define EVEN_P(x) (((x) & 1) == 0)
> >> >
> >> > mp_size_t
> >> > mpn_gcd (mp_ptr gp, mp_ptr ap, mp_size_t an, mp_ptr bp, mp_size_t n)
> >> > {
> >> >  mp_size_t init_scratch;
> >> >  mp_size_t scratch;
> >> >  mp_ptr tp;
> >> >  TMP_DECL;
> >> >
> >> >  ASSERT (an >= n);
> >> >
> >> >  if (BELOW_THRESHOLD (n, NGCD_THRESHOLD))
> >> >    return mpn_basic_gcd (gp, ap, an, bp, n);
> >> >
> >> >  init_scratch = MPN_NGCD_MATRIX_INIT_ITCH ((n+1)/2);
> >> >  scratch = mpn_nhgcd_itch ((n+1)/2);
> >> >
> >> >  /* Space needed for mpn_ngcd_matrix_adjust */
> >> >  if (scratch < 2*n)
> >> >    scratch = 2*n;
> >> >
> >> >  TMP_MARK;
> >> >
> >> >  if (an + 1 > init_scratch + scratch)
> >> >    tp = TMP_ALLOC_LIMBS (an + 1);
> >> >  else
> >> >    tp = TMP_ALLOC_LIMBS (init_scratch + scratch);
> >> >
> >> >  if (an > n)
> >> >    {
> >> >      mp_ptr rp = tp;
> >> >      mp_ptr qp = rp + n;
> >> >
> >> >      mpn_tdiv_qr (qp, rp, 0, ap, an, bp, n);
> >> >      MPN_COPY (ap, rp, n);
> >> >      an = n;
> >> >      MPN_NORMALIZE (ap, an);
> >> >      if (an == 0)
> >> >        {
> >> >          MPN_COPY (gp, bp, n);
> >> >          TMP_FREE;
> >> >          return n;
> >> >        }
> >> >    }
> >> >
> >> >  while (ABOVE_THRESHOLD (n, NGCD_THRESHOLD))
> >> >    {
> >> >      struct ngcd_matrix M;
> >> >      mp_size_t p = n/2;
> >> >      mp_size_t nn;
> >> >
> >> >      mpn_ngcd_matrix_init (&M, n - p, tp);
> >> >      nn = mpn_nhgcd (ap + p, bp + p, n - p, &M, tp + init_scratch);
> >> >      if (nn > 0)
> >> >        /* Needs 2*(p + M->n) <= 2*(floor(n/2) + ceil(n/2) - 1)
> >> >           = 2(n-1) */
> >> >        n = mpn_ngcd_matrix_adjust (&M, p + nn, ap, bp, p, tp +
> >> > init_scratch);
> >> >
> >> >      else
> >> >        {
> >> >          mp_size_t gn;
> >> >          n = mpn_ngcd_subdiv_step (gp, &gn, ap, bp, n, tp);
> >> >          if (n == 0)
> >> >            {
> >> >              TMP_FREE;
> >> >              return gn;
> >> >            }
> >> >        }
> >> >    }
> >> >
> >> >  ASSERT (ap[n-1] > 0 || bp[n-1] > 0);
> >> > #if 0
> >> >  /* FIXME: We may want to use lehmer on some systems. */
> >> >  n = mpn_ngcd_lehmer (gp, ap, bp, n, tp);
> >> >
> >> >  TMP_FREE;
> >> >  return n;
> >> > #endif
> >> >
> >> >  if (ap[n-1] < bp[n-1])
> >> >    MP_PTR_SWAP (ap, bp);
> >> >
> >> >  an = n;
> >> >  MPN_NORMALIZE (bp, n);
> >> >
> >> >  if (n == 0)
> >> >    {
> >> >      MPN_COPY (gp, ap, an);
> >> >      TMP_FREE;
> >> >      return an;
> >> >    }
> >> >
> >> >  if (EVEN_P (bp[0]))
> >> >    {
> >> >      /* Then a must be odd (since the calling convention implies that
> >> >         there's no common factor of 2) */
> >> >      ASSERT (!EVEN_P (ap[0]));
> >> >
> >> >      while (bp[0] == 0)
> >> >        {
> >> >          bp++;
> >> >          n--;
> >> >        }
> >> >
> >> >      if (EVEN_P(bp[0]))
> >> >        {
> >> >          int count;
> >> >          count_trailing_zeros (count, bp[0]);
> >> >          ASSERT_NOCARRY (mpn_rshift (bp, bp, n, count));
> >> >          n -= (bp[n-1] == 0);
> >> >        }
> >> >    }
> >> >
> >> >  TMP_FREE;
> >> >  return mpn_basic_gcd (gp, ap, an, bp, n);
> >> > }
> >> >
> >> > /* Schönhage's 1987 algorithm, reorganized into hgcd form */
> >> >
> >> > #include <stdio.h>  /* for NULL */
> >> >
> >> > #include "gmp.h"
> >> > #include "gmp-impl.h"
> >> > #include "longlong.h"
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> > /* For input of size n, matrix elements are of size at most ceil(n/2)
> >> >   - 1, but we need one limb extra. */
> >> >
> >> > void
> >> > mpn_ngcd_matrix_init (struct ngcd_matrix *M, mp_size_t n, mp_ptr p);
> >> >
> >> > #define NHGCD_BASE_ITCH MPN_NGCD_STEP_ITCH
> >> >
> >> > /* Reduces a,b until |a-b| fits in n/2 + 1 limbs. Constructs matrix M
> >> >   with elements of size at most (n+1)/2 - 1. Returns new size of a,
> >> >   b, or zero if no reduction is possible. */
> >> > static mp_size_t
> >> > nhgcd_base (mp_ptr ap, mp_ptr bp, mp_size_t n,
> >> >            struct ngcd_matrix *M, mp_ptr tp);
> >> >
> >> > /* Size analysis for nhgcd:
> >> >
> >> >   For the recursive calls, we have n1 <= ceil(n / 2). Then the
> >> >   storage need is determined by the storage for the recursive call
> >> >   computing M1, and ngcd_matrix_adjust and ngcd_matrix_mul calls that
> >> > use M1 (after this, the storage needed for M1 can be recycled).
> >> >
> >> >   Let S(r) denote the required storage. For M1 we need 5 * ceil(n1/2)
> >> >   = 5 * ceil(n/4), and for the ngcd_matrix_adjust call, we need n + 2.
> >> > In total, 5 * ceil(n/4) + n + 2 <= 9 ceil(n/4) + 2.
> >> >
> >> >   For the recursive call, we need S(n1) = S(ceil(n/2)).
> >> >
> >> >   S(n) <= 9*ceil(n/4) + 2 + S(ceil(n/2))
> >> >        <= 9*(ceil(n/4) + ... + ceil(n/2^(1+k))) + 2k + S(ceil(n/2^k))
> >> >        <= 9*(2 ceil(n/4) + k) + 2k + S(n/2^k)
> >> >        <= 18 ceil(n/4) + 11k + S(n/2^k)
> >> >
> >> > */
> >> >
> >> > mp_size_t
> >> > mpn_nhgcd_itch (mp_size_t n);
> >> >
> >> >
> >> > /* Reduces a,b until |a-b| fits in n/2 + 1 limbs. Constructs matrix M
> >> >   with elements of size at most (n+1)/2 - 1. Returns new size of a,
> >> >   b, or zero if no reduction is possible. */
> >> >
> >> > mp_size_t
> >> > mpn_nhgcd (mp_ptr ap, mp_ptr bp, mp_size_t n,
> >> >           struct ngcd_matrix *M, mp_ptr tp);
> >> >
> >> >
> >> > #define EVEN_P(x) (((x) & 1) == 0)
> >> >
> >> > mp_size_t
> >> > mpn_ngcd (mp_ptr gp, mp_ptr ap, mp_size_t an, mp_ptr bp, mp_size_t n)
> >> > {
> >> >  mp_size_t init_scratch;
> >> >  mp_size_t scratch;
> >> >  mp_ptr tp;
> >> >  TMP_DECL;
> >> >
> >> >  ASSERT (an >= n);
> >> >
> >> >  if (BELOW_THRESHOLD (n, NGCD_THRESHOLD))
> >> >    return mpn_basic_gcd (gp, ap, an, bp, n);
> >> >
> >> >  init_scratch = MPN_NGCD_MATRIX_INIT_ITCH ((n+1)/2);
> >> >  scratch = mpn_nhgcd_itch ((n+1)/2);
> >> >
> >> >  /* Space needed for mpn_ngcd_matrix_adjust */
> >> >  if (scratch < 2*n)
> >> >    scratch = 2*n;
> >> >
> >> >  TMP_MARK;
> >> >
> >> >  if (an + 1 > init_scratch + scratch)
> >> >    tp = TMP_ALLOC_LIMBS (an + 1);
> >> >  else
> >> >    tp = TMP_ALLOC_LIMBS (init_scratch + scratch);
> >> >
> >> >  if (an > n)
> >> >    {
> >> >      mp_ptr rp = tp;
> >> >      mp_ptr qp = rp + n;
> >> >
> >> >      mpn_tdiv_qr (qp, rp, 0, ap, an, bp, n);
> >> >      MPN_COPY (ap, rp, n);
> >> >      an = n;
> >> >      MPN_NORMALIZE (ap, an);
> >> >      if (an == 0)
> >> >        {
> >> >          MPN_COPY (gp, bp, n);
> >> >          TMP_FREE;
> >> >          return n;
> >> >        }
> >> >    }
> >> >
> >> >  while (ABOVE_THRESHOLD (n, NGCD_THRESHOLD))
> >> >    {
> >> >      struct ngcd_matrix M;
> >> >      mp_size_t p = n/2;
> >> >      mp_size_t nn;
> >> >
> >> >      mpn_ngcd_matrix_init (&M, n - p, tp);
> >> >      nn = mpn_nhgcd (ap + p, bp + p, n - p, &M, tp + init_scratch);
> >> >      if (nn > 0)
> >> >        /* Needs 2*(p + M->n) <= 2*(floor(n/2) + ceil(n/2) - 1)
> >> >           = 2(n-1) */
> >> >        n = mpn_ngcd_matrix_adjust (&M, p + nn, ap, bp, p, tp +
> >> > init_scratch);
> >> >
> >> >      else
> >> >        {
> >> >          mp_size_t gn;
> >> >          n = mpn_ngcd_subdiv_step (gp, &gn, ap, bp, n, tp);
> >> >          if (n == 0)
> >> >            {
> >> >              TMP_FREE;
> >> >              return gn;
> >> >            }
> >> >        }
> >> >    }
> >> >
> >> >  ASSERT (ap[n-1] > 0 || bp[n-1] > 0);
> >> > #if 0
> >> >  /* FIXME: We may want to use lehmer on some systems. */
> >> >  n = mpn_ngcd_lehmer (gp, ap, bp, n, tp);
> >> >
> >> >  TMP_FREE;
> >> >  return n;
> >> > #endif
> >> >
> >> >  if (ap[n-1] < bp[n-1])
> >> >    MP_PTR_SWAP (ap, bp);
> >> >
> >> >  an = n;
> >> >  MPN_NORMALIZE (bp, n);
> >> >
> >> >  if (n == 0)
> >> >    {
> >> >      MPN_COPY (gp, ap, an);
> >> >      TMP_FREE;
> >> >      return an;
> >> >    }
> >> >
> >> >  if (EVEN_P (bp[0]))
> >> >    {
> >> >      /* Then a must be odd (since the calling convention implies that
> >> >         there's no common factor of 2) */
> >> >      ASSERT (!EVEN_P (ap[0]));
> >> >
> >> >      while (bp[0] == 0)
> >> >        {
> >> >          bp++;
> >> >          n--;
> >> >        }
> >> >
> >> >      if (EVEN_P(bp[0]))
> >> >        {
> >> >          int count;
> >> >          count_trailing_zeros (count, bp[0]);
> >> >          ASSERT_NOCARRY (mpn_rshift (bp, bp, n, count));
> >> >          n -= (bp[n-1] == 0);
> >> >        }
> >> >    }
> >> >
> >> >  TMP_FREE;
> >> >  return mpn_basic_gcd (gp, ap, an, bp, n);
> >> > }
>
> 


--~--~---------~--~----~------------~-------~--~----~
You received this message because you are subscribed to the Google Groups 
"mpir-devel" group.
To post to this group, send email to mpir-devel@googlegroups.com
To unsubscribe from this group, send email to 
mpir-devel+unsubscr...@googlegroups.com
For more options, visit this group at 
http://groups.google.com/group/mpir-devel?hl=en
-~----------~----~----~----~------~----~------~--~---

Reply via email to