Home | History | Annotate | Download | only in generic

Lines Matching defs:u0

58 	  /* Must return the smallest cofactor, +u1 or -u0 */
59 MPN_CMP (c, ctx->u0, ctx->u1, un);
60 ASSERT (c != 0 || (un == 1 && ctx->u0[0] == 1 && ctx->u1[0] == 1));
65 up = d ? ctx->u0 : ctx->u1;
75 mp_ptr u0 = ctx->u0;
81 MP_PTR_SWAP (u0, u1);
85 /* Update u0 += q * u1 */
92 cy = mpn_add_n (u0, u0, u1, un);
94 cy = mpn_addmul_1 (u0, u1, un, q);
107 /* Should always have u1n == un here, and u1 >= u0. The
108 reason is that we alternate adding u0 to u1 and u1 to u0
126 cy = mpn_add (u0, tp, u1n, u0, un);
131 cy = mpn_add (u0, u0, un, tp, u1n);
134 u0[un] = cy;
153 * M = (v0, v1 ; u0, u1)
157 * M^{-1} = (u1, -v1; -u0, v0)
162 * b = -u0 A (mod B)
169 mp_ptr u0;
174 u0 = tp; tp += ualloc;
227 un = mpn_hgcd_mul_matrix1_vector(&M, u2, u0, u1, un);
228 MP_PTR_SWAP (u0, u2);
235 ctx.u0 = u0;
256 /* Which cofactor to return now? Candidates are +u1 and -u0,
263 MPN_CMP (c, u0, u1, un);
264 ASSERT (c != 0 || (un == 1 && u0[0] == 1 && u1[0] == 1));
267 MPN_NORMALIZE (u0, un);
268 MPN_COPY (up, u0, un);
288 /* Set up = u u1 - v u0. Keep track of size, un grows by one or
294 MPN_NORMALIZE (u0, un);
295 MPN_COPY (up, u0, un);
321 vh = mpn_addmul_1 (up, u0, un, v);