| /src/external/lgpl3/gmp/dist/mpn/generic/ |
| toom43_mul.c | 100 #define bsm2 (pp + n + 1) /* n+1 */ macro 116 /* Compute bs2 and bsm2. */ 127 mpn_add_n_sub_n (bs2, bsm2, b1d, b0b2, n+1); 132 mpn_add_n_sub_n (bs2, bsm2, b0b2, b1d, n+1); 138 mpn_sub_n (bsm2, b1d, b0b2, n+1); 143 mpn_sub_n (bsm2, b0b2, b1d, n+1); 185 ASSERT (bsm2[n] <= 4); 191 mpn_mul_n (vm2, asm2, bsm2, n+1); /* W2 */ 217 #undef bsm2 macro
|
| toom52_mul.c | 103 #define bsm2 (pp + n + 1) /* n+1 */ macro 161 /* Compute bs2 and bsm2, recycling bs1 and bsm1. bs2=bs1+b1; bsm2=bsm1-b1 */ 165 bsm2[n] = mpn_add (bsm2, bsm1, n, b1, t); 170 bsm2[n] = 0; 175 mpn_sub_n (bsm2, b1, bsm1, n); 180 mpn_sub_n (bsm2, bsm1, b1, n); 187 mpn_sub_n (bsm2, b1, bsm1, t); 188 MPN_ZERO (bsm2 + t, n - t) 240 #undef bsm2 macro [all...] |
| toom53_mul.c | 71 mp_ptr bs1, bsm1, bs2, bsm2, bsh; local 104 bsm2 = tmp; tmp += n + 1; 170 /* Compute bs2 and bsm2. */ 191 ASSERT_NOCARRY (mpn_add_n_sub_n (bs2, bsm2, gp, bs2, n+1)); 196 ASSERT_NOCARRY (mpn_add_n_sub_n (bs2, bsm2, bs2, gp, n+1)); 201 ASSERT_NOCARRY (mpn_sub_n (bsm2, gp, bs2, n+1)); 206 ASSERT_NOCARRY (mpn_sub_n (bsm2, bs2, gp, n+1)); 237 ASSERT (bsm2[n] <= 4); 254 mpn_mul_n (vm2, asm2, bsm2, n + 1); /* vm2, 2n+1 limbs */
|
| toom62_mul.c | 71 mp_ptr bs1, bsm1, bs2, bsm2, bsh; local 104 bsm2 = TMP_SALLOC_LIMBS (n + 1); 190 /* Compute bs2 and bsm2. Recycling bs1 and bsm1; bs2=bs1+b1, bsm2 = 195 bsm2[n] = mpn_add (bsm2, bsm1, n, b1, t); 205 ASSERT_NOCARRY (mpn_sub_n (bsm2, b1, bsm1, t)); 206 MPN_ZERO (bsm2 + t, n + 1 - t); 211 ASSERT_NOCARRY (mpn_sub (bsm2, bsm1, n, b1, t)); 212 bsm2[n] = 0 [all...] |