mpir/mpz/fac_ui.c
jasonmoxham e554550755 for file in $(find -name \*.c ) ; do sed -e "s/#include \"gmp\.h\"/#include \"mpir.h\"/g" $file > temp ; mv temp $file ; done
for file in $(find -name \*.h ) ; do sed -e "s/#include \"gmp\.h\"/#include \"mpir.h\"/g" $file > temp ; mv temp $file ; done
for file in $(find -name \*.cc) ; do sed -e "s/#include \"gmp\.h\"/#include \"mpir.h\"/g" $file > temp ; mv temp $file ; done
2009-02-12 10:24:24 +00:00

399 lines
11 KiB
C

/* mpz_fac_ui(result, n) -- Set RESULT to N!.
Copyright 1991, 1993, 1994, 1995, 2000, 2001, 2002, 2003 Free Software
Foundation, Inc.
This file is part of the GNU MP Library.
The GNU MP Library is free software; you can redistribute it and/or modify
it under the terms of the GNU Lesser General Public License as published by
the Free Software Foundation; either version 2.1 of the License, or (at your
option) any later version.
The GNU MP Library is distributed in the hope that it will be useful, but
WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public
License for more details.
You should have received a copy of the GNU Lesser General Public License
along with the GNU MP Library; see the file COPYING.LIB. If not, write to
the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston,
MA 02110-1301, USA. */
#include "mpir.h"
#include "gmp-impl.h"
#include "longlong.h"
#include "fac_ui.h"
static void odd_product _PROTO ((unsigned long low, unsigned long high, mpz_t * st));
static void ap_product_small _PROTO ((mpz_t ret, mp_limb_t start, mp_limb_t step, unsigned long count, unsigned long nm));
/* must be >=2 */
#define APCONST 5
/* for single non-zero limb */
#define MPZ_SET_1_NZ(z,n) \
do { \
mpz_ptr __z = (z); \
ASSERT ((n) != 0); \
PTR(__z)[0] = (n); \
SIZ(__z) = 1; \
} while (0)
/* for src>0 and n>0 */
#define MPZ_MUL_1_POS(dst,src,n) \
do { \
mpz_ptr __dst = (dst); \
mpz_srcptr __src = (src); \
mp_size_t __size = SIZ(__src); \
mp_ptr __dst_p; \
mp_limb_t __c; \
\
ASSERT (__size > 0); \
ASSERT ((n) != 0); \
\
MPZ_REALLOC (__dst, __size+1); \
__dst_p = PTR(__dst); \
\
__c = mpn_mul_1 (__dst_p, PTR(__src), __size, n); \
__dst_p[__size] = __c; \
SIZ(__dst) = __size + (__c != 0); \
} while (0)
#if BITS_PER_ULONG == BITS_PER_MP_LIMB
#define BSWAP_ULONG(x,y) BSWAP_LIMB(x,y)
#endif
/* We used to have a case here for limb==2*long, doing a BSWAP_LIMB followed
by a shift down to get the high part. But it provoked incorrect code
from "HP aC++/ANSI C B3910B A.05.52 [Sep 05 2003]" in ILP32 mode. This
case would have been nice for gcc ia64 where BSWAP_LIMB is a mux1, but we
can get that directly muxing a 4-byte ulong if it matters enough. */
#if ! defined (BSWAP_ULONG)
#define BSWAP_ULONG(dst, src) \
do { \
unsigned long __bswapl_src = (src); \
unsigned long __bswapl_dst = 0; \
int __i; \
for (__i = 0; __i < sizeof(unsigned long); __i++) \
{ \
__bswapl_dst = (__bswapl_dst << 8) | (__bswapl_src & 0xFF); \
__bswapl_src >>= 8; \
} \
(dst) = __bswapl_dst; \
} while (0)
#endif
/* x is bit reverse of y */
/* Note the divides below are all exact */
#define BITREV_ULONG(x,y) \
do { \
unsigned long __dst; \
BSWAP_ULONG(__dst,y); \
__dst = ((__dst>>4)&(ULONG_MAX/17)) | ((__dst<<4)&((ULONG_MAX/17)*16)); \
__dst = ((__dst>>2)&(ULONG_MAX/5) ) | ((__dst<<2)&((ULONG_MAX/5)*4) ); \
__dst = ((__dst>>1)&(ULONG_MAX/3) ) | ((__dst<<1)&((ULONG_MAX/3)*2) ); \
(x) = __dst; \
} while(0)
/* above could be improved if cpu has a nibble/bit swap/muxing instruction */
/* above code is serialized, possible to write as a big parallel expression */
void
mpz_fac_ui (mpz_ptr x, unsigned long n)
{
unsigned long z, stt;
int i, j;
mpz_t t1, st[8 * sizeof (unsigned long) + 1 - APCONST];
mp_limb_t d[4];
static const mp_limb_t table[] = { ONE_LIMB_FACTORIAL_TABLE };
if (n < numberof (table))
{
MPZ_SET_1_NZ (x, table[n]);
return;
}
/* NOTE : MUST have n>=3 here */
ASSERT (n >= 3);
/* for estimating the alloc sizes the calculation of these formula's is not
exact and also the formulas are only approximations, also we ignore
the few "side" calculations, correct allocation seems to speed up the
small sizes better, having very little effect on the large sizes */
/* estimate space for stack entries see below
number of bits for n! is
(1+log_2(2*pi)/2)-n*log_2(exp(1))+(n+1/2)*log_2(n)=
2.325748065-n*1.442695041+(n+0.5)*log_2(n) */
umul_ppmm (d[1], d[0], (mp_limb_t) n, (mp_limb_t) FAC2OVERE);
/* d[1] is 2n/e, d[0] ignored */
count_leading_zeros (z, d[1]);
z = GMP_LIMB_BITS - z - 1; /* z=floor(log_2(2n/e)) */
umul_ppmm (d[1], d[0], (mp_limb_t) n, (mp_limb_t) z);
/* d=n*floor(log_2(2n/e)) */
d[0] = (d[0] >> 2) | (d[1] << (GMP_LIMB_BITS - 2));
d[1] >>= 2;
/* d=n*floor(log_2(2n/e))/4 */
z = d[0] + 1; /* have to ignore any overflow */
/* so z is the number of bits wanted for st[0] */
if (n <= ((unsigned long) 1) << (APCONST))
{
mpz_realloc2 (x, 4 * z);
ap_product_small (x, CNST_LIMB(2), CNST_LIMB(1), n - 1, 4L);
return;
}
if (n <= ((unsigned long) 1) << (APCONST + 1))
{ /* use n!=odd(1,n)*(n/2)!*2^(n/2) */
mpz_init2 (t1, 2 * z);
mpz_realloc2 (x, 4 * z);
ap_product_small (x, CNST_LIMB(2), CNST_LIMB(1), n / 2 - 1, 4L);
ap_product_small (t1, CNST_LIMB(3), CNST_LIMB(2), (n - 1) / 2, 4L);
mpz_mul (x, x, t1);
mpz_clear (t1);
mpz_mul_2exp (x, x, n / 2);
return;
}
if (n <= ((unsigned long) 1) << (APCONST + 2))
{
/* use n!=C_2(1,n/2)^2*C_2(n/2,n)*(n/4)!*2^(n/2+n/4) all int divs
so need (BITS_IN_N-APCONST+1)=(APCONST+3-APCONST+1)=4 stack entries */
mpz_init2 (t1, 2 * z);
mpz_realloc2 (x, 4 * z);
for (i = 0; i < 4; i++)
{
mpz_init2 (st[i], z);
z >>= 1;
}
odd_product (1, n / 2, st);
mpz_set (x, st[0]);
odd_product (n / 2, n, st);
mpz_mul (x, x, x);
ASSERT (n / 4 <= FACMUL4 + 6);
ap_product_small (t1, CNST_LIMB(2), CNST_LIMB(1), n / 4 - 1, 4L);
/* must have 2^APCONST odd numbers max */
mpz_mul (t1, t1, st[0]);
for (i = 0; i < 4; i++)
mpz_clear (st[i]);
mpz_mul (x, x, t1);
mpz_clear (t1);
mpz_mul_2exp (x, x, n / 2 + n / 4);
return;
}
count_leading_zeros (stt, (mp_limb_t) n);
stt = GMP_LIMB_BITS - stt + 1 - APCONST;
for (i = 0; i < (signed long) stt; i++)
{
mpz_init2 (st[i], z);
z >>= 1;
}
count_leading_zeros (z, (mp_limb_t) (n / 3));
/* find z st 2^z>n/3 range for z is 1 <= z <= 8 * sizeof(unsigned long)-1 */
z = GMP_LIMB_BITS - z;
/*
n! = 2^e * PRODUCT_{i=0}^{i=z-1} C_2( n/2^{i+1}, n/2^i )^{i+1}
where 2^e || n! 3.2^z>n C_2(a,b)=PRODUCT of odd z such that a<z<=b
*/
mpz_init_set_ui (t1, 1);
for (j = 8 * sizeof (unsigned long) / 2; j != 0; j >>= 1)
{
MPZ_SET_1_NZ (x, 1);
for (i = 8 * sizeof (unsigned long) - j; i >= j; i -= 2 * j)
if ((signed long) z >= i)
{
odd_product (n >> i, n >> (i - 1), st);
/* largest odd product when j=i=1 then we have
odd_product(n/2,n,st) which is approx (2n/e)^(n/4)
so log_base2(largest oddproduct)=n*log_base2(2n/e)/4
number of bits is n*log_base2(2n/e)/4+1 */
if (i != j)
mpz_pow_ui (st[0], st[0], i / j);
mpz_mul (x, x, st[0]);
}
if ((signed long) z >= j && j != 1)
{
mpz_mul (t1, t1, x);
mpz_mul (t1, t1, t1);
}
}
for (i = 0; i < (signed long) stt; i++)
mpz_clear (st[i]);
mpz_mul (x, x, t1);
mpz_clear (t1);
popc_limb (i, (mp_limb_t) n);
mpz_mul_2exp (x, x, n - i);
return;
}
/* start,step are mp_limb_t although they will fit in unsigned long */
static void
ap_product_small (mpz_t ret, mp_limb_t start, mp_limb_t step,
unsigned long count, unsigned long nm)
{
unsigned long a;
mp_limb_t b;
ASSERT (count <= (((unsigned long) 1) << APCONST));
/* count can never be zero ? check this and remove test below */
if (count == 0)
{
MPZ_SET_1_NZ (ret, 1);
return;
}
if (count == 1)
{
MPZ_SET_1_NZ (ret, start);
return;
}
switch (nm)
{
case 1:
MPZ_SET_1_NZ (ret, start);
b = start + step;
for (a = 0; a < count - 1; b += step, a++)
MPZ_MUL_1_POS (ret, ret, b);
return;
case 2:
MPZ_SET_1_NZ (ret, start * (start + step));
if (count == 2)
return;
for (b = start + 2 * step, a = count / 2 - 1; a != 0;
a--, b += 2 * step)
MPZ_MUL_1_POS (ret, ret, b * (b + step));
if (count % 2 == 1)
MPZ_MUL_1_POS (ret, ret, b);
return;
case 3:
if (count == 2)
{
MPZ_SET_1_NZ (ret, start * (start + step));
return;
}
MPZ_SET_1_NZ (ret, start * (start + step) * (start + 2 * step));
if (count == 3)
return;
for (b = start + 3 * step, a = count / 3 - 1; a != 0;
a--, b += 3 * step)
MPZ_MUL_1_POS (ret, ret, b * (b + step) * (b + 2 * step));
if (count % 3 == 2)
b = b * (b + step);
if (count % 3 != 0)
MPZ_MUL_1_POS (ret, ret, b);
return;
default: /* ie nm=4 */
if (count == 2)
{
MPZ_SET_1_NZ (ret, start * (start + step));
return;
}
if (count == 3)
{
MPZ_SET_1_NZ (ret, start * (start + step) * (start + 2 * step));
return;
}
MPZ_SET_1_NZ (ret,
start * (start + step) * (start + 2 * step) * (start +
3 * step));
if (count == 4)
return;
for (b = start + 4 * step, a = count / 4 - 1; a != 0;
a--, b += 4 * step)
MPZ_MUL_1_POS (ret, ret,
b * (b + step) * (b + 2 * step) * (b + 3 * step));
if (count % 4 == 2)
b = b * (b + step);
if (count % 4 == 3)
b = b * (b + step) * (b + 2 * step);
if (count % 4 != 0)
MPZ_MUL_1_POS (ret, ret, b);
return;
}
}
/* return value in st[0]
odd_product(l,h)=sqrt((h/e)^h/(l/e)^l) using Stirling approx and e=exp(1)
so st[0] needs enough bits for above, st[1] needs half these bits and
st[2] needs 1/4 of these bits etc */
static void
odd_product (unsigned long low, unsigned long high, mpz_t * st)
{
unsigned long stc = 1, stn = 0, n, y, mask, a, nm = 1;
signed long z;
low++;
if (low % 2 == 0)
low++;
if (high == 0)
high = 1;
if (high % 2 == 0)
high--;
/* must have high>=low ? check this and remove test below */
if (high < low)
{
MPZ_SET_1_NZ (st[0], 1);
return;
}
if (high == low)
{
MPZ_SET_1_NZ (st[0], low);
return;
}
if (high <= FACMUL2 + 2)
{
nm = 2;
if (high <= FACMUL3 + 4)
{
nm = 3;
if (high <= FACMUL4 + 6)
nm = 4;
}
}
high = (high - low) / 2 + 1; /* high is now count,high<=2^(BITS_PER_ULONG-1) */
if (high <= (((unsigned long) 1) << APCONST))
{
ap_product_small (st[0], (mp_limb_t) low, CNST_LIMB(2), high, nm);
return;
}
count_leading_zeros (n, (mp_limb_t) high);
/* assumes clz above is LIMB based not NUMB based */
n = GMP_LIMB_BITS - n - APCONST;
mask = (((unsigned long) 1) << n);
a = mask << 1;
mask--;
/* have 2^(BITS_IN_N-APCONST) iterations so need
(BITS_IN_N-APCONST+1) stack entries */
for (z = mask; z >= 0; z--)
{
BITREV_ULONG (y, z);
y >>= (BITS_PER_ULONG - n);
ap_product_small (st[stn],
(mp_limb_t) (low + 2 * ((~y) & mask)), (mp_limb_t) a,
(high + y) >> n, nm);
ASSERT (((high + y) >> n) <= (((unsigned long) 1) << APCONST));
stn++;
y = stc++;
while ((y & 1) == 0)
{
mpz_mul (st[stn - 2], st[stn - 2], st[stn - 1]);
stn--;
y >>= 1;
}
}
ASSERT (stn == 1);
return;
}