HomeSort by: relevance | last modified time | path
    Searched defs:u0 (Results 1 - 25 of 44) sorted by relevancy

1 2

  /src/external/lgpl3/gmp/dist/mpn/generic/
addmul_1.c 44 mp_limb_t u0, crec, c, p1, p0, r0; local
52 u0 = *up++;
53 umul_ppmm (p1, p0, u0, v0);
81 mp_limb_t shifted_v0, u0, r0, p0, p1, prev_p1, crec, xl, c1, c2, c3; local
94 u0 = *up++;
96 umul_ppmm (p1, p0, u0, shifted_v0);
117 mp_limb_t shifted_v0, u0, r0, p0, p1, prev_p1, xw, crec, xl; local
130 u0 = *up++;
132 umul_ppmm (p1, p0, u0, shifted_v0);
div_qr_1n_pi1.c 192 mp_limb_t u0, u2; local
215 u0 = up[n-1]; /* Early read, to allow qp == up. */
218 add_mssaaaa (u2, u1, u0, u0, up[n-2], p1, p0);
234 * | 1 | (conditional on u0 + u2 B2 carry)
248 ADDC_LIMB (cy, u0, u0, u2 & B2);
249 u0 -= (-cy) & d;
256 add_mssaaaa (u2, u1, u0, u0, up[j], p1, p0)
    [all...]
submul_1.c 43 mp_limb_t u0, crec, c, p1, p0, r0; local
51 u0 = *up++;
52 umul_ppmm (p1, p0, u0, v0);
80 mp_limb_t shifted_v0, u0, r0, p0, p1, prev_p1, cl, xl, c1, c2, c3; local
93 u0 = *up++;
95 umul_ppmm (p1, p0, u0, shifted_v0);
116 mp_limb_t shifted_v0, u0, r0, p0, p1, prev_p1, xw, cl, xl; local
129 u0 = *up++;
131 umul_ppmm (p1, p0, u0, shifted_v0);
matrix22_mul.c 119 The scheduling uses two temporaries U0 and U1 to store products, and
134 mp_ptr s0, t0, u0, u1; local
138 u0 = tp; tp += rn + mn + 1;
141 MUL (u0, r1, rn, m2, mn); /* u5 = s5 * t6 */
169 MUL (u1, r0, rn, m0, mn); /* u0 = s0 * t0 */
170 r0[rn+mn] = mpn_add_n (r0, u0, u1, rn + mn);
171 ASSERT (r0[rn+mn] < 2); /* u0 + u5 */
207 u0[rn+mn] = 0;
210 r3s = abs_sub_n (r3, u0, r3, rn + mn + 1);
214 ASSERT_NOCARRY (mpn_add_n (r3, r3, u0, rn + mn + 1))
    [all...]
div_qr_1n_pi2.c 122 mp_limb_t r, u2, u1, u0; local
153 u0 = up[0];
155 /* Dividend in {r,u1,u0} */
160 add_sssaaaa (r,q2b,q1b, q2b,q1b, u1,u0);
170 r = u0 - q2 * d0;
div_qr_1u_pi2.c 120 mp_limb_t r, u2, u1, u0; local
149 u0 = (up[1] << cnt) | (up[0] >> (GMP_NUMB_BITS - cnt));
151 /* Dividend in {r,u1,u0} */
156 add_sssaaaa (r,q2b,q1b, q2b,q1b, u1,u0);
166 r = u0 - q2 * d0;
197 u0 = (up[1] << cnt);
199 /* Dividend in {r,u1,u0} */
204 add_sssaaaa (r,q2b,q1b, q2b,q1b, u1,u0);
214 r = u0 - q2 * d0;
gcdext_1.c 229 a = u0 A + v0 B
234 mp_limb_signed_t u0 = 1; local
258 u0 -= q * u1;
267 *up = u0;
271 u1 -= q * u0;
gcdext_lehmer.c 58 /* Must return the smallest cofactor, +u1 or -u0 */
59 MPN_CMP (c, ctx->u0, ctx->u1, un);
60 ASSERT (c != 0 || (un == 1 && ctx->u0[0] == 1 && ctx->u1[0] == 1));
65 up = d ? ctx->u0 : ctx->u1;
75 mp_ptr u0 = ctx->u0; local
81 MP_PTR_SWAP (u0, u1);
85 /* Update u0 += q * u1 */
92 cy = mpn_add_n (u0, u0, u1, un)
169 mp_ptr u0; local
    [all...]
gcd.c 226 mp_limb_t u0, u1, v0, v1; local
229 u0 = up[0];
236 *gp = mpn_gcd_11 (u0, v0 >> cnt);
258 g = mpn_gcd_22 (u1, u0, v1, v0);
  /src/external/lgpl3/mpfr/dist/src/
set.c 136 mp_limb_t u0 = up[0]; local
138 /* it suffices to round (u0, inex) */
139 rb = u0 & (MPFR_LIMB_ONE << (sh - 1));
140 sb = (u0 & mask) ^ rb;
141 sp[0] = u0 & ~mask;
147 /* we need to round (u1, u0, inex) */
159 thus we need to subtract something from (u0, rb, sb):
ai.c 335 mpfr_t u0, u1; local
426 mpfr_init2 (u0, wprec); mpfr_init2 (u1, wprec);
439 mpfr_set_prec (u0, wprec); mpfr_set_prec (u1, wprec);
442 mpfr_set_ui (u0, 1, MPFR_RNDN);
464 mpfr_set_ui (u0, 9, MPFR_RNDN);
465 mpfr_cbrt (u0, u0, MPFR_RNDN);
466 mpfr_mul (u0, u0, temp2, MPFR_RNDN);
467 mpfr_ui_div (u0, 1, u0, MPFR_RNDN); /* u0 = 1/( Gamma (2/3)*9^(1/3) ) *
    [all...]
sqrt.c 80 mp_limb_t u0, r0, rb, sb, mask = MPFR_LIMB_MASK(sh); local
86 u0 = MPFR_MANT(u)[0];
89 u0 >>= 1;
96 u0*2^GMP_NUMB_BITS */
97 __gmpfr_sqrt_limb_approx (r0, u0);
106 Since r0 <= sqrt(2^64*u0) <= r0 + 7, as soon as sqrt(2^64*u0)>=2^63+7,
107 which happens for u0>=2^62+8, then r0 >= 2^63.
108 It thus remains to check that for 2^62 <= u0 <= 2^62+7,
109 __gmpfr_sqrt_limb_approx (r0, u0) gives r0 >= 2^63, which is indee
225 mp_limb_t u0, r0, rb, sb, low; local
    [all...]
  /src/common/lib/libc/quad/
muldi3.c 55 * u = 2^n u1 * u0 (n = number of bits in `u_int', usu. 32)
63 * uv = 2^2n u1 v1 + 2^n u1 v0 + 2^n v1 u0 + u0 v0
64 * = 2^2n u1 v1 + 2^n (u1 v0 + v1 u0) + u0 v0
67 * and add 2^n u0 v0 to the last term and subtract it from the middle.
71 * (2^n) (u1 v0 - u1 v1 + u0 v1 - u0 v0) +
72 * (2^n + 1) (u0 v0)
77 * (2^n) (u1 - u0) (v0 - v1) + [(u1-u0)... = mid
115 #define u0 macro
170 #undef u0 macro
195 u_int u1, u0, v1, v0, udiff, vdiff, high, mid, low; local
    [all...]
  /src/external/gpl3/gcc/dist/libgcc/config/epiphany/
divsi3.c 37 fu u0, u1, u2, u1b, u2b; local
46 u0.i = 0x40000000;
47 u1b.i = u2b.i = u0.i;
50 u1.i = a | u0.i;
57 u2.i = b | u0.i;
modsi3.c 35 fu u0, u1, u2, u1b, u2b; local
44 u0.i = 0x40000000;
45 u1b.i = u2b.i = u0.i;
48 u1.i = a | u0.i;
55 u2.i = b | u0.i;
udivsi3-float.c 34 fu u0, u1, u2, u1b, u2b; local
46 u0.f = a;
49 u0.f = (int) a;
53 u0.i = (a >> 8) - 0x00800000 + 0x3f800000 + (31 << 23);
59 u0.i = (t - c);
63 s0 = u0.i + 1 /* Compensate for rounding errors. */
udivsi3.c 34 fu u0, u1, u2, u1b, u2b; local
40 u0.i = 0x40000000;
41 u1b.i = u2b.i = u0.i;
44 u1.i = a | u0.i;
51 u2.i = b | u0.i;
umodsi3.c 34 fu u0, u1, u2, u1b, u2b; local
40 u0.i = 0x40000000;
41 u1b.i = u2b.i = u0.i;
44 u1.i = a | u0.i;
51 u2.i = b | u0.i;
  /src/external/gpl3/gcc.old/dist/libgcc/config/epiphany/
divsi3.c 37 fu u0, u1, u2, u1b, u2b; local
46 u0.i = 0x40000000;
47 u1b.i = u2b.i = u0.i;
50 u1.i = a | u0.i;
57 u2.i = b | u0.i;
modsi3.c 35 fu u0, u1, u2, u1b, u2b; local
44 u0.i = 0x40000000;
45 u1b.i = u2b.i = u0.i;
48 u1.i = a | u0.i;
55 u2.i = b | u0.i;
udivsi3-float.c 34 fu u0, u1, u2, u1b, u2b; local
46 u0.f = a;
49 u0.f = (int) a;
53 u0.i = (a >> 8) - 0x00800000 + 0x3f800000 + (31 << 23);
59 u0.i = (t - c);
63 s0 = u0.i + 1 /* Compensate for rounding errors. */
udivsi3.c 34 fu u0, u1, u2, u1b, u2b; local
40 u0.i = 0x40000000;
41 u1b.i = u2b.i = u0.i;
44 u1.i = a | u0.i;
51 u2.i = b | u0.i;
umodsi3.c 34 fu u0, u1, u2, u1b, u2b; local
40 u0.i = 0x40000000;
41 u1b.i = u2b.i = u0.i;
44 u1.i = a | u0.i;
51 u2.i = b | u0.i;
  /src/lib/libm/noieee_src/
n_j1.c 102 * U(z) = u0 + u1*z + ... + u4*z^4
196 static const double u0[5] = { variable
275 u = u0[0]+z*(u0[1]+z*(u0[2]+z*(u0[3]+z*u0[4])));
  /src/sys/crypto/aes/
aes_bear.c 399 uint32_t t0, t1, t2, t3, u0, u1, u2, u3; local
450 u0 = t0; u1 = t1; u2 = t2; u3 = t3;
451 aesbear_xts_update(&u0, &u1, &u2, &u3);
458 q[2*0 + 1] = le32dec(in + 16 + 4*0) ^ u0;
473 le32enc(out + 16 + 4*0, q[2*0 + 1] ^ u0);
479 t0 = u0; t1 = u1; t2 = u2; t3 = u3;
503 uint32_t t0, t1, t2, t3, u0, u1, u2, u3; local
554 u0 = t0; u1 = t1; u2 = t2; u3 = t3;
555 aesbear_xts_update(&u0, &u1, &u2, &u3);
562 q[2*0 + 1] = le32dec(in + 16 + 4*0) ^ u0;
    [all...]

Completed in 36 milliseconds

1 2