a197a2d3eb
Removed directories for no longer supported architectures.
132 lines
3.1 KiB
C
132 lines
3.1 KiB
C
/* mpn_addmul_1 -- multiply the N long limb vector pointed to by UP by VL,
|
|
add the N least significant limbs of the product to the limb vector
|
|
pointed to by RP. Return the most significant limb of the product,
|
|
adjusted for carry-out from the addition.
|
|
|
|
Copyright 1992, 1993, 1994, 1996, 2000, 2002, 2004 Free Software Foundation,
|
|
Inc.
|
|
|
|
This file is part of the GNU MP Library.
|
|
|
|
The GNU MP Library is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU Lesser General Public License as published by
|
|
the Free Software Foundation; either version 2.1 of the License, or (at your
|
|
option) any later version.
|
|
|
|
The GNU MP Library is distributed in the hope that it will be useful, but
|
|
WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
|
|
or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public
|
|
License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public License
|
|
along with the GNU MP Library; see the file COPYING.LIB. If not, write to
|
|
the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston,
|
|
MA 02110-1301, USA. */
|
|
|
|
#include "gmp.h"
|
|
#include "gmp-impl.h"
|
|
#include "longlong.h"
|
|
|
|
|
|
#if GMP_NAIL_BITS == 0
|
|
|
|
mp_limb_t
|
|
mpn_addmul_1 (mp_ptr rp, mp_srcptr up, mp_size_t n, mp_limb_t vl)
|
|
{
|
|
mp_limb_t ul, cl, hpl, lpl, rl;
|
|
|
|
ASSERT (n >= 1);
|
|
ASSERT (MPN_SAME_OR_SEPARATE_P (rp, up, n));
|
|
|
|
cl = 0;
|
|
do
|
|
{
|
|
ul = *up++;
|
|
umul_ppmm (hpl, lpl, ul, vl);
|
|
|
|
lpl += cl;
|
|
cl = (lpl < cl) + hpl;
|
|
|
|
rl = *rp;
|
|
lpl = rl + lpl;
|
|
cl += lpl < rl;
|
|
*rp++ = lpl;
|
|
}
|
|
while (--n != 0);
|
|
|
|
return cl;
|
|
}
|
|
|
|
#endif
|
|
|
|
#if GMP_NAIL_BITS == 1
|
|
|
|
mp_limb_t
|
|
mpn_addmul_1 (mp_ptr rp, mp_srcptr up, mp_size_t n, mp_limb_t vl)
|
|
{
|
|
mp_limb_t shifted_vl, ul, rl, lpl, hpl, prev_hpl, cl, xl, c1, c2, c3;
|
|
|
|
ASSERT (n >= 1);
|
|
ASSERT (MPN_SAME_OR_SEPARATE_P (rp, up, n));
|
|
ASSERT_MPN (rp, n);
|
|
ASSERT_MPN (up, n);
|
|
ASSERT_LIMB (vl);
|
|
|
|
shifted_vl = vl << GMP_NAIL_BITS;
|
|
cl = 0;
|
|
prev_hpl = 0;
|
|
do
|
|
{
|
|
ul = *up++;
|
|
rl = *rp;
|
|
umul_ppmm (hpl, lpl, ul, shifted_vl);
|
|
lpl >>= GMP_NAIL_BITS;
|
|
ADDC_LIMB (c1, xl, prev_hpl, lpl);
|
|
ADDC_LIMB (c2, xl, xl, rl);
|
|
ADDC_LIMB (c3, xl, xl, cl);
|
|
cl = c1 + c2 + c3;
|
|
*rp++ = xl;
|
|
prev_hpl = hpl;
|
|
}
|
|
while (--n != 0);
|
|
|
|
return prev_hpl + cl;
|
|
}
|
|
|
|
#endif
|
|
|
|
#if GMP_NAIL_BITS >= 2
|
|
|
|
mp_limb_t
|
|
mpn_addmul_1 (mp_ptr rp, mp_srcptr up, mp_size_t n, mp_limb_t vl)
|
|
{
|
|
mp_limb_t shifted_vl, ul, rl, lpl, hpl, prev_hpl, xw, cl, xl;
|
|
|
|
ASSERT (n >= 1);
|
|
ASSERT (MPN_SAME_OR_SEPARATE_P (rp, up, n));
|
|
ASSERT_MPN (rp, n);
|
|
ASSERT_MPN (up, n);
|
|
ASSERT_LIMB (vl);
|
|
|
|
shifted_vl = vl << GMP_NAIL_BITS;
|
|
cl = 0;
|
|
prev_hpl = 0;
|
|
do
|
|
{
|
|
ul = *up++;
|
|
rl = *rp;
|
|
umul_ppmm (hpl, lpl, ul, shifted_vl);
|
|
lpl >>= GMP_NAIL_BITS;
|
|
xw = prev_hpl + lpl + rl + cl;
|
|
cl = xw >> GMP_NUMB_BITS;
|
|
xl = xw & GMP_NUMB_MASK;
|
|
*rp++ = xl;
|
|
prev_hpl = hpl;
|
|
}
|
|
while (--n != 0);
|
|
|
|
return prev_hpl + cl;
|
|
}
|
|
|
|
#endif
|