mpir/mpn/generic/hgcd2.c
2014-02-27 21:35:04 +00:00

438 lines
8.9 KiB
C

/* hgcd2.c
THE FUNCTIONS IN THIS FILE ARE INTERNAL WITH MUTABLE INTERFACES. IT IS ONLY
SAFE TO REACH THEM THROUGH DOCUMENTED INTERFACES. IN FACT, IT IS ALMOST
GUARANTEED THAT THEY'LL CHANGE OR DISAPPEAR IN A FUTURE GNU MP RELEASE.
Copyright 1996, 1998, 2000, 2001, 2002, 2003, 2004, 2008, 2012 Free Software
Foundation, Inc.
This file is part of the GNU MP Library.
The GNU MP Library is free software; you can redistribute it and/or modify
it under the terms of the GNU Lesser General Public License as published by
the Free Software Foundation; either version 3 of the License, or (at your
option) any later version.
The GNU MP Library is distributed in the hope that it will be useful, but
WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public
License for more details.
You should have received a copy of the GNU Lesser General Public License
along with the GNU MP Library. If not, see http://www.gnu.org/licenses/. */
#include "mpir.h"
#include "gmp-impl.h"
#include "longlong.h"
#if GMP_NAIL_BITS == 0
/* Copied from the old mpn/generic/gcdext.c, and modified slightly to return
the remainder. */
/* Single-limb division optimized for small quotients. */
static inline mp_limb_t
div1 (mp_ptr rp,
mp_limb_t n0,
mp_limb_t d0)
{
mp_limb_t q = 0;
if ((mp_limb_signed_t) n0 < 0)
{
int cnt;
for (cnt = 1; (mp_limb_signed_t) d0 >= 0; cnt++)
{
d0 = d0 << 1;
}
q = 0;
while (cnt)
{
q <<= 1;
if (n0 >= d0)
{
n0 = n0 - d0;
q |= 1;
}
d0 = d0 >> 1;
cnt--;
}
}
else
{
int cnt;
for (cnt = 0; n0 >= d0; cnt++)
{
d0 = d0 << 1;
}
q = 0;
while (cnt)
{
d0 = d0 >> 1;
q <<= 1;
if (n0 >= d0)
{
n0 = n0 - d0;
q |= 1;
}
cnt--;
}
}
*rp = n0;
return q;
}
/* Two-limb division optimized for small quotients. */
static inline mp_limb_t
div2 (mp_ptr rp,
mp_limb_t nh, mp_limb_t nl,
mp_limb_t dh, mp_limb_t dl)
{
mp_limb_t q = 0;
if ((mp_limb_signed_t) nh < 0)
{
int cnt;
for (cnt = 1; (mp_limb_signed_t) dh >= 0; cnt++)
{
dh = (dh << 1) | (dl >> (GMP_LIMB_BITS - 1));
dl = dl << 1;
}
while (cnt)
{
q <<= 1;
if (nh > dh || (nh == dh && nl >= dl))
{
sub_ddmmss (nh, nl, nh, nl, dh, dl);
q |= 1;
}
dl = (dh << (GMP_LIMB_BITS - 1)) | (dl >> 1);
dh = dh >> 1;
cnt--;
}
}
else
{
int cnt;
for (cnt = 0; nh > dh || (nh == dh && nl >= dl); cnt++)
{
dh = (dh << 1) | (dl >> (GMP_LIMB_BITS - 1));
dl = dl << 1;
}
while (cnt)
{
dl = (dh << (GMP_LIMB_BITS - 1)) | (dl >> 1);
dh = dh >> 1;
q <<= 1;
if (nh > dh || (nh == dh && nl >= dl))
{
sub_ddmmss (nh, nl, nh, nl, dh, dl);
q |= 1;
}
cnt--;
}
}
rp[0] = nl;
rp[1] = nh;
return q;
}
#if 0
/* This div2 uses less branches, but it seems to nevertheless be
slightly slower than the above code. */
static inline mp_limb_t
div2 (mp_ptr rp,
mp_limb_t nh, mp_limb_t nl,
mp_limb_t dh, mp_limb_t dl)
{
mp_limb_t q = 0;
int ncnt;
int dcnt;
count_leading_zeros (ncnt, nh);
count_leading_zeros (dcnt, dh);
dcnt -= ncnt;
dh = (dh << dcnt) + (-(dcnt > 0) & (dl >> (GMP_LIMB_BITS - dcnt)));
dl <<= dcnt;
do
{
mp_limb_t bit;
q <<= 1;
if (UNLIKELY (nh == dh))
bit = (nl >= dl);
else
bit = (nh > dh);
q |= bit;
sub_ddmmss (nh, nl, nh, nl, (-bit) & dh, (-bit) & dl);
dl = (dh << (GMP_LIMB_BITS - 1)) | (dl >> 1);
dh = dh >> 1;
}
while (dcnt--);
rp[0] = nl;
rp[1] = nh;
return q;
}
#endif
#else /* GMP_NAIL_BITS != 0 */
/* Check all functions for nail support. */
/* hgcd2 should be defined to take inputs including nail bits, and
produce a matrix with elements also including nail bits. This is
necessary, for the matrix elements to be useful with mpn_mul_1,
mpn_addmul_1 and friends. */
#error Not implemented
#endif /* GMP_NAIL_BITS != 0 */
/* Reduces a,b until |a-b| (almost) fits in one limb + 1 bit. Constructs
matrix M. Returns 1 if we make progress, i.e. can perform at least
one subtraction. Otherwise returns zero. */
/* FIXME: Possible optimizations:
The div2 function starts with checking the most significant bit of
the numerator. We can maintained normalized operands here, call
hgcd with normalized operands only, which should make the code
simpler and possibly faster.
Experiment with table lookups on the most significant bits.
This function is also a candidate for assembler implementation.
*/
int
mpn_hgcd2 (mp_limb_t ah, mp_limb_t al, mp_limb_t bh, mp_limb_t bl,
struct hgcd_matrix1 *M)
{
mp_limb_t u00, u01, u10, u11;
if (ah < 2 || bh < 2)
return 0;
if (ah > bh || (ah == bh && al > bl))
{
sub_ddmmss (ah, al, ah, al, bh, bl);
if (ah < 2)
return 0;
u00 = u01 = u11 = 1;
u10 = 0;
}
else
{
sub_ddmmss (bh, bl, bh, bl, ah, al);
if (bh < 2)
return 0;
u00 = u10 = u11 = 1;
u01 = 0;
}
if (ah < bh)
goto subtract_a;
for (;;)
{
ASSERT (ah >= bh);
if (ah == bh)
goto done;
if (ah < (CNST_LIMB(1) << (GMP_LIMB_BITS / 2)))
{
ah = (ah << (GMP_LIMB_BITS / 2) ) + (al >> (GMP_LIMB_BITS / 2));
bh = (bh << (GMP_LIMB_BITS / 2) ) + (bl >> (GMP_LIMB_BITS / 2));
break;
}
/* Subtract a -= q b, and multiply M from the right by (1 q ; 0
1), affecting the second column of M. */
ASSERT (ah > bh);
sub_ddmmss (ah, al, ah, al, bh, bl);
if (ah < 2)
goto done;
if (ah <= bh)
{
/* Use q = 1 */
u01 += u00;
u11 += u10;
}
else
{
mp_limb_t r[2];
mp_limb_t q = div2 (r, ah, al, bh, bl);
al = r[0]; ah = r[1];
if (ah < 2)
{
/* A is too small, but q is correct. */
u01 += q * u00;
u11 += q * u10;
goto done;
}
q++;
u01 += q * u00;
u11 += q * u10;
}
subtract_a:
ASSERT (bh >= ah);
if (ah == bh)
goto done;
if (bh < (CNST_LIMB(1) << (GMP_LIMB_BITS / 2)))
{
ah = (ah << (GMP_LIMB_BITS / 2) ) + (al >> (GMP_LIMB_BITS / 2));
bh = (bh << (GMP_LIMB_BITS / 2) ) + (bl >> (GMP_LIMB_BITS / 2));
goto subtract_a1;
}
/* Subtract b -= q a, and multiply M from the right by (1 0 ; q
1), affecting the first column of M. */
sub_ddmmss (bh, bl, bh, bl, ah, al);
if (bh < 2)
goto done;
if (bh <= ah)
{
/* Use q = 1 */
u00 += u01;
u10 += u11;
}
else
{
mp_limb_t r[2];
mp_limb_t q = div2 (r, bh, bl, ah, al);
bl = r[0]; bh = r[1];
if (bh < 2)
{
/* B is too small, but q is correct. */
u00 += q * u01;
u10 += q * u11;
goto done;
}
q++;
u00 += q * u01;
u10 += q * u11;
}
}
/* NOTE: Since we discard the least significant half limb, we don't
get a truly maximal M (corresponding to |a - b| <
2^{GMP_LIMB_BITS +1}). */
/* Single precision loop */
for (;;)
{
ASSERT (ah >= bh);
ah -= bh;
if (ah < (CNST_LIMB (1) << (GMP_LIMB_BITS / 2 + 1)))
break;
if (ah <= bh)
{
/* Use q = 1 */
u01 += u00;
u11 += u10;
}
else
{
mp_limb_t r;
mp_limb_t q = div1 (&r, ah, bh);
ah = r;
if (ah < (CNST_LIMB(1) << (GMP_LIMB_BITS / 2 + 1)))
{
/* A is too small, but q is correct. */
u01 += q * u00;
u11 += q * u10;
break;
}
q++;
u01 += q * u00;
u11 += q * u10;
}
subtract_a1:
ASSERT (bh >= ah);
bh -= ah;
if (bh < (CNST_LIMB (1) << (GMP_LIMB_BITS / 2 + 1)))
break;
if (bh <= ah)
{
/* Use q = 1 */
u00 += u01;
u10 += u11;
}
else
{
mp_limb_t r;
mp_limb_t q = div1 (&r, bh, ah);
bh = r;
if (bh < (CNST_LIMB(1) << (GMP_LIMB_BITS / 2 + 1)))
{
/* B is too small, but q is correct. */
u00 += q * u01;
u10 += q * u11;
break;
}
q++;
u00 += q * u01;
u10 += q * u11;
}
}
done:
M->u[0][0] = u00; M->u[0][1] = u01;
M->u[1][0] = u10; M->u[1][1] = u11;
return 1;
}
/* Sets (r;b) = (a;b) M, with M = (u00, u01; u10, u11). Vector must
* have space for n + 1 limbs. Uses three buffers to avoid a copy*/
mp_size_t
mpn_hgcd_mul_matrix1_vector (const struct hgcd_matrix1 *M,
mp_ptr rp, mp_srcptr ap, mp_ptr bp, mp_size_t n)
{
mp_limb_t ah, bh;
/* Compute (r,b) <-- (u00 a + u10 b, u01 a + u11 b) as
r = u00 * a
r += u10 * b
b *= u11
b += u01 * a
*/
#if HAVE_NATIVE_mpn_addaddmul_1msb0
ah = mpn_addaddmul_1msb0 (rp, ap, bp, n, M->u[0][0], M->u[1][0]);
bh = mpn_addaddmul_1msb0 (bp, bp, ap, n, M->u[1][1], M->u[0][1]);
#else
ah = mpn_mul_1 (rp, ap, n, M->u[0][0]);
ah += mpn_addmul_1 (rp, bp, n, M->u[1][0]);
bh = mpn_mul_1 (bp, bp, n, M->u[1][1]);
bh += mpn_addmul_1 (bp, ap, n, M->u[0][1]);
#endif
rp[n] = ah;
bp[n] = bh;
n += (ah | bh) > 0;
return n;
}