longlong.h revision 1.1.1.1.8.1 1 1.1 mrg /* longlong.h -- definitions for mixed size 32/64 bit arithmetic.
2 1.1 mrg
3 1.1 mrg Copyright 1991, 1992, 1993, 1994, 1996, 1997, 1999, 2000, 2001, 2002, 2003,
4 1.1.1.1.8.1 tls 2004, 2005, 2007, 2008, 2009, 2011, 2012 Free Software Foundation, Inc.
5 1.1 mrg
6 1.1 mrg This file is free software; you can redistribute it and/or modify it under the
7 1.1 mrg terms of the GNU Lesser General Public License as published by the Free
8 1.1 mrg Software Foundation; either version 3 of the License, or (at your option) any
9 1.1 mrg later version.
10 1.1 mrg
11 1.1 mrg This file is distributed in the hope that it will be useful, but WITHOUT ANY
12 1.1 mrg WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A
13 1.1 mrg PARTICULAR PURPOSE. See the GNU Lesser General Public License for more
14 1.1 mrg details.
15 1.1 mrg
16 1.1 mrg You should have received a copy of the GNU Lesser General Public License
17 1.1 mrg along with this file. If not, see http://www.gnu.org/licenses/. */
18 1.1 mrg
19 1.1 mrg /* You have to define the following before including this file:
20 1.1 mrg
21 1.1 mrg UWtype -- An unsigned type, default type for operations (typically a "word")
22 1.1 mrg UHWtype -- An unsigned type, at least half the size of UWtype
23 1.1 mrg UDWtype -- An unsigned type, at least twice as large a UWtype
24 1.1 mrg W_TYPE_SIZE -- size in bits of UWtype
25 1.1 mrg
26 1.1 mrg SItype, USItype -- Signed and unsigned 32 bit types
27 1.1 mrg DItype, UDItype -- Signed and unsigned 64 bit types
28 1.1 mrg
29 1.1 mrg On a 32 bit machine UWtype should typically be USItype;
30 1.1 mrg on a 64 bit machine, UWtype should typically be UDItype.
31 1.1 mrg
32 1.1 mrg Optionally, define:
33 1.1 mrg
34 1.1 mrg LONGLONG_STANDALONE -- Avoid code that needs machine-dependent support files
35 1.1 mrg NO_ASM -- Disable inline asm
36 1.1 mrg
37 1.1 mrg
38 1.1 mrg CAUTION! Using this version of longlong.h outside of GMP is not safe. You
39 1.1 mrg need to include gmp.h and gmp-impl.h, or certain things might not work as
40 1.1 mrg expected.
41 1.1 mrg */
42 1.1 mrg
43 1.1 mrg #define __BITS4 (W_TYPE_SIZE / 4)
44 1.1 mrg #define __ll_B ((UWtype) 1 << (W_TYPE_SIZE / 2))
45 1.1 mrg #define __ll_lowpart(t) ((UWtype) (t) & (__ll_B - 1))
46 1.1 mrg #define __ll_highpart(t) ((UWtype) (t) >> (W_TYPE_SIZE / 2))
47 1.1 mrg
48 1.1 mrg /* This is used to make sure no undesirable sharing between different libraries
49 1.1 mrg that use this file takes place. */
50 1.1 mrg #ifndef __MPN
51 1.1 mrg #define __MPN(x) __##x
52 1.1 mrg #endif
53 1.1 mrg
54 1.1 mrg /* Define auxiliary asm macros.
55 1.1 mrg
56 1.1 mrg 1) umul_ppmm(high_prod, low_prod, multiplier, multiplicand) multiplies two
57 1.1 mrg UWtype integers MULTIPLIER and MULTIPLICAND, and generates a two UWtype
58 1.1 mrg word product in HIGH_PROD and LOW_PROD.
59 1.1 mrg
60 1.1 mrg 2) __umulsidi3(a,b) multiplies two UWtype integers A and B, and returns a
61 1.1 mrg UDWtype product. This is just a variant of umul_ppmm.
62 1.1 mrg
63 1.1 mrg 3) udiv_qrnnd(quotient, remainder, high_numerator, low_numerator,
64 1.1 mrg denominator) divides a UDWtype, composed by the UWtype integers
65 1.1 mrg HIGH_NUMERATOR and LOW_NUMERATOR, by DENOMINATOR and places the quotient
66 1.1 mrg in QUOTIENT and the remainder in REMAINDER. HIGH_NUMERATOR must be less
67 1.1 mrg than DENOMINATOR for correct operation. If, in addition, the most
68 1.1 mrg significant bit of DENOMINATOR must be 1, then the pre-processor symbol
69 1.1 mrg UDIV_NEEDS_NORMALIZATION is defined to 1.
70 1.1 mrg
71 1.1 mrg 4) sdiv_qrnnd(quotient, remainder, high_numerator, low_numerator,
72 1.1 mrg denominator). Like udiv_qrnnd but the numbers are signed. The quotient
73 1.1 mrg is rounded towards 0.
74 1.1 mrg
75 1.1 mrg 5) count_leading_zeros(count, x) counts the number of zero-bits from the
76 1.1 mrg msb to the first non-zero bit in the UWtype X. This is the number of
77 1.1 mrg steps X needs to be shifted left to set the msb. Undefined for X == 0,
78 1.1 mrg unless the symbol COUNT_LEADING_ZEROS_0 is defined to some value.
79 1.1 mrg
80 1.1 mrg 6) count_trailing_zeros(count, x) like count_leading_zeros, but counts
81 1.1 mrg from the least significant end.
82 1.1 mrg
83 1.1 mrg 7) add_ssaaaa(high_sum, low_sum, high_addend_1, low_addend_1,
84 1.1 mrg high_addend_2, low_addend_2) adds two UWtype integers, composed by
85 1.1 mrg HIGH_ADDEND_1 and LOW_ADDEND_1, and HIGH_ADDEND_2 and LOW_ADDEND_2
86 1.1 mrg respectively. The result is placed in HIGH_SUM and LOW_SUM. Overflow
87 1.1 mrg (i.e. carry out) is not stored anywhere, and is lost.
88 1.1 mrg
89 1.1 mrg 8) sub_ddmmss(high_difference, low_difference, high_minuend, low_minuend,
90 1.1 mrg high_subtrahend, low_subtrahend) subtracts two two-word UWtype integers,
91 1.1 mrg composed by HIGH_MINUEND_1 and LOW_MINUEND_1, and HIGH_SUBTRAHEND_2 and
92 1.1 mrg LOW_SUBTRAHEND_2 respectively. The result is placed in HIGH_DIFFERENCE
93 1.1 mrg and LOW_DIFFERENCE. Overflow (i.e. carry out) is not stored anywhere,
94 1.1 mrg and is lost.
95 1.1 mrg
96 1.1 mrg If any of these macros are left undefined for a particular CPU,
97 1.1 mrg C macros are used.
98 1.1 mrg
99 1.1 mrg
100 1.1 mrg Notes:
101 1.1 mrg
102 1.1 mrg For add_ssaaaa the two high and two low addends can both commute, but
103 1.1 mrg unfortunately gcc only supports one "%" commutative in each asm block.
104 1.1 mrg This has always been so but is only documented in recent versions
105 1.1 mrg (eg. pre-release 3.3). Having two or more "%"s can cause an internal
106 1.1 mrg compiler error in certain rare circumstances.
107 1.1 mrg
108 1.1 mrg Apparently it was only the last "%" that was ever actually respected, so
109 1.1 mrg the code has been updated to leave just that. Clearly there's a free
110 1.1 mrg choice whether high or low should get it, if there's a reason to favour
111 1.1 mrg one over the other. Also obviously when the constraints on the two
112 1.1 mrg operands are identical there's no benefit to the reloader in any "%" at
113 1.1 mrg all.
114 1.1 mrg
115 1.1 mrg */
116 1.1 mrg
117 1.1 mrg /* The CPUs come in alphabetical order below.
118 1.1 mrg
119 1.1 mrg Please add support for more CPUs here, or improve the current support
120 1.1 mrg for the CPUs below! */
121 1.1 mrg
122 1.1 mrg
123 1.1 mrg /* count_leading_zeros_gcc_clz is count_leading_zeros implemented with gcc
124 1.1 mrg 3.4 __builtin_clzl or __builtin_clzll, according to our limb size.
125 1.1 mrg Similarly count_trailing_zeros_gcc_ctz using __builtin_ctzl or
126 1.1 mrg __builtin_ctzll.
127 1.1 mrg
128 1.1 mrg These builtins are only used when we check what code comes out, on some
129 1.1 mrg chips they're merely libgcc calls, where we will instead want an inline
130 1.1 mrg in that case (either asm or generic C).
131 1.1 mrg
132 1.1 mrg These builtins are better than an asm block of the same insn, since an
133 1.1 mrg asm block doesn't give gcc any information about scheduling or resource
134 1.1 mrg usage. We keep an asm block for use on prior versions of gcc though.
135 1.1 mrg
136 1.1 mrg For reference, __builtin_ffs existed in gcc prior to __builtin_clz, but
137 1.1 mrg it's not used (for count_leading_zeros) because it generally gives extra
138 1.1 mrg code to ensure the result is 0 when the input is 0, which we don't need
139 1.1 mrg or want. */
140 1.1 mrg
141 1.1 mrg #ifdef _LONG_LONG_LIMB
142 1.1 mrg #define count_leading_zeros_gcc_clz(count,x) \
143 1.1 mrg do { \
144 1.1 mrg ASSERT ((x) != 0); \
145 1.1 mrg (count) = __builtin_clzll (x); \
146 1.1 mrg } while (0)
147 1.1 mrg #else
148 1.1 mrg #define count_leading_zeros_gcc_clz(count,x) \
149 1.1 mrg do { \
150 1.1 mrg ASSERT ((x) != 0); \
151 1.1 mrg (count) = __builtin_clzl (x); \
152 1.1 mrg } while (0)
153 1.1 mrg #endif
154 1.1 mrg
155 1.1 mrg #ifdef _LONG_LONG_LIMB
156 1.1 mrg #define count_trailing_zeros_gcc_ctz(count,x) \
157 1.1 mrg do { \
158 1.1 mrg ASSERT ((x) != 0); \
159 1.1 mrg (count) = __builtin_ctzll (x); \
160 1.1 mrg } while (0)
161 1.1 mrg #else
162 1.1 mrg #define count_trailing_zeros_gcc_ctz(count,x) \
163 1.1 mrg do { \
164 1.1 mrg ASSERT ((x) != 0); \
165 1.1 mrg (count) = __builtin_ctzl (x); \
166 1.1 mrg } while (0)
167 1.1 mrg #endif
168 1.1 mrg
169 1.1 mrg
170 1.1 mrg /* FIXME: The macros using external routines like __MPN(count_leading_zeros)
171 1.1 mrg don't need to be under !NO_ASM */
172 1.1 mrg #if ! defined (NO_ASM)
173 1.1 mrg
174 1.1 mrg #if defined (__alpha) && W_TYPE_SIZE == 64
175 1.1 mrg /* Most alpha-based machines, except Cray systems. */
176 1.1 mrg #if defined (__GNUC__)
177 1.1 mrg #if __GMP_GNUC_PREREQ (3,3)
178 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \
179 1.1 mrg do { \
180 1.1 mrg UDItype __m0 = (m0), __m1 = (m1); \
181 1.1 mrg (ph) = __builtin_alpha_umulh (__m0, __m1); \
182 1.1 mrg (pl) = __m0 * __m1; \
183 1.1 mrg } while (0)
184 1.1 mrg #else
185 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \
186 1.1 mrg do { \
187 1.1 mrg UDItype __m0 = (m0), __m1 = (m1); \
188 1.1 mrg __asm__ ("umulh %r1,%2,%0" \
189 1.1 mrg : "=r" (ph) \
190 1.1 mrg : "%rJ" (m0), "rI" (m1)); \
191 1.1 mrg (pl) = __m0 * __m1; \
192 1.1 mrg } while (0)
193 1.1 mrg #endif
194 1.1 mrg #define UMUL_TIME 18
195 1.1 mrg #else /* ! __GNUC__ */
196 1.1 mrg #include <machine/builtins.h>
197 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \
198 1.1 mrg do { \
199 1.1 mrg UDItype __m0 = (m0), __m1 = (m1); \
200 1.1 mrg (ph) = __UMULH (m0, m1); \
201 1.1 mrg (pl) = __m0 * __m1; \
202 1.1 mrg } while (0)
203 1.1 mrg #endif
204 1.1 mrg #ifndef LONGLONG_STANDALONE
205 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
206 1.1 mrg do { UWtype __di; \
207 1.1 mrg __di = __MPN(invert_limb) (d); \
208 1.1 mrg udiv_qrnnd_preinv (q, r, n1, n0, d, __di); \
209 1.1 mrg } while (0)
210 1.1 mrg #define UDIV_PREINV_ALWAYS 1
211 1.1 mrg #define UDIV_NEEDS_NORMALIZATION 1
212 1.1 mrg #define UDIV_TIME 220
213 1.1 mrg #endif /* LONGLONG_STANDALONE */
214 1.1 mrg
215 1.1 mrg /* clz_tab is required in all configurations, since mpn/alpha/cntlz.asm
216 1.1 mrg always goes into libgmp.so, even when not actually used. */
217 1.1 mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
218 1.1 mrg
219 1.1 mrg #if defined (__GNUC__) && HAVE_HOST_CPU_alpha_CIX
220 1.1 mrg #define count_leading_zeros(COUNT,X) \
221 1.1 mrg __asm__("ctlz %1,%0" : "=r"(COUNT) : "r"(X))
222 1.1 mrg #define count_trailing_zeros(COUNT,X) \
223 1.1 mrg __asm__("cttz %1,%0" : "=r"(COUNT) : "r"(X))
224 1.1 mrg #endif /* clz/ctz using cix */
225 1.1 mrg
226 1.1 mrg #if ! defined (count_leading_zeros) \
227 1.1 mrg && defined (__GNUC__) && ! defined (LONGLONG_STANDALONE)
228 1.1 mrg /* ALPHA_CMPBGE_0 gives "cmpbge $31,src,dst", ie. test src bytes == 0.
229 1.1 mrg "$31" is written explicitly in the asm, since an "r" constraint won't
230 1.1 mrg select reg 31. There seems no need to worry about "r31" syntax for cray,
231 1.1 mrg since gcc itself (pre-release 3.4) emits just $31 in various places. */
232 1.1 mrg #define ALPHA_CMPBGE_0(dst, src) \
233 1.1 mrg do { asm ("cmpbge $31, %1, %0" : "=r" (dst) : "r" (src)); } while (0)
234 1.1 mrg /* Zero bytes are turned into bits with cmpbge, a __clz_tab lookup counts
235 1.1 mrg them, locating the highest non-zero byte. A second __clz_tab lookup
236 1.1 mrg counts the leading zero bits in that byte, giving the result. */
237 1.1 mrg #define count_leading_zeros(count, x) \
238 1.1 mrg do { \
239 1.1 mrg UWtype __clz__b, __clz__c, __clz__x = (x); \
240 1.1 mrg ALPHA_CMPBGE_0 (__clz__b, __clz__x); /* zero bytes */ \
241 1.1 mrg __clz__b = __clz_tab [(__clz__b >> 1) ^ 0x7F]; /* 8 to 1 byte */ \
242 1.1 mrg __clz__b = __clz__b * 8 - 7; /* 57 to 1 shift */ \
243 1.1 mrg __clz__x >>= __clz__b; \
244 1.1 mrg __clz__c = __clz_tab [__clz__x]; /* 8 to 1 bit */ \
245 1.1 mrg __clz__b = 65 - __clz__b; \
246 1.1 mrg (count) = __clz__b - __clz__c; \
247 1.1 mrg } while (0)
248 1.1 mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
249 1.1 mrg #endif /* clz using cmpbge */
250 1.1 mrg
251 1.1 mrg #if ! defined (count_leading_zeros) && ! defined (LONGLONG_STANDALONE)
252 1.1 mrg #if HAVE_ATTRIBUTE_CONST
253 1.1.1.1.8.1 tls long __MPN(count_leading_zeros) (UDItype) __attribute__ ((const));
254 1.1 mrg #else
255 1.1.1.1.8.1 tls long __MPN(count_leading_zeros) (UDItype);
256 1.1 mrg #endif
257 1.1 mrg #define count_leading_zeros(count, x) \
258 1.1 mrg ((count) = __MPN(count_leading_zeros) (x))
259 1.1 mrg #endif /* clz using mpn */
260 1.1 mrg #endif /* __alpha */
261 1.1 mrg
262 1.1.1.1.8.1 tls #if defined (__AVR) && W_TYPE_SIZE == 8
263 1.1.1.1.8.1 tls #define umul_ppmm(ph, pl, m0, m1) \
264 1.1.1.1.8.1 tls do { \
265 1.1.1.1.8.1 tls unsigned short __p = (unsigned short) (m0) * (m1); \
266 1.1.1.1.8.1 tls (ph) = __p >> 8; \
267 1.1.1.1.8.1 tls (pl) = __p; \
268 1.1.1.1.8.1 tls } while (0)
269 1.1.1.1.8.1 tls #endif /* AVR */
270 1.1.1.1.8.1 tls
271 1.1 mrg #if defined (_CRAY) && W_TYPE_SIZE == 64
272 1.1 mrg #include <intrinsics.h>
273 1.1 mrg #define UDIV_PREINV_ALWAYS 1
274 1.1 mrg #define UDIV_NEEDS_NORMALIZATION 1
275 1.1 mrg #define UDIV_TIME 220
276 1.1.1.1.8.1 tls long __MPN(count_leading_zeros) (UDItype);
277 1.1 mrg #define count_leading_zeros(count, x) \
278 1.1 mrg ((count) = _leadz ((UWtype) (x)))
279 1.1 mrg #if defined (_CRAYIEEE) /* I.e., Cray T90/ieee, T3D, and T3E */
280 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \
281 1.1 mrg do { \
282 1.1 mrg UDItype __m0 = (m0), __m1 = (m1); \
283 1.1 mrg (ph) = _int_mult_upper (m0, m1); \
284 1.1 mrg (pl) = __m0 * __m1; \
285 1.1 mrg } while (0)
286 1.1 mrg #ifndef LONGLONG_STANDALONE
287 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
288 1.1 mrg do { UWtype __di; \
289 1.1 mrg __di = __MPN(invert_limb) (d); \
290 1.1 mrg udiv_qrnnd_preinv (q, r, n1, n0, d, __di); \
291 1.1 mrg } while (0)
292 1.1 mrg #endif /* LONGLONG_STANDALONE */
293 1.1 mrg #endif /* _CRAYIEEE */
294 1.1 mrg #endif /* _CRAY */
295 1.1 mrg
296 1.1 mrg #if defined (__ia64) && W_TYPE_SIZE == 64
297 1.1 mrg /* This form encourages gcc (pre-release 3.4 at least) to emit predicated
298 1.1 mrg "sub r=r,r" and "sub r=r,r,1", giving a 2 cycle latency. The generic
299 1.1 mrg code using "al<bl" arithmetically comes out making an actual 0 or 1 in a
300 1.1 mrg register, which takes an extra cycle. */
301 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
302 1.1 mrg do { \
303 1.1 mrg UWtype __x; \
304 1.1 mrg __x = (al) - (bl); \
305 1.1 mrg if ((al) < (bl)) \
306 1.1 mrg (sh) = (ah) - (bh) - 1; \
307 1.1 mrg else \
308 1.1 mrg (sh) = (ah) - (bh); \
309 1.1 mrg (sl) = __x; \
310 1.1 mrg } while (0)
311 1.1 mrg #if defined (__GNUC__) && ! defined (__INTEL_COMPILER)
312 1.1 mrg /* Do both product parts in assembly, since that gives better code with
313 1.1 mrg all gcc versions. Some callers will just use the upper part, and in
314 1.1 mrg that situation we waste an instruction, but not any cycles. */
315 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \
316 1.1 mrg __asm__ ("xma.hu %0 = %2, %3, f0\n\txma.l %1 = %2, %3, f0" \
317 1.1 mrg : "=&f" (ph), "=f" (pl) \
318 1.1 mrg : "f" (m0), "f" (m1))
319 1.1 mrg #define UMUL_TIME 14
320 1.1 mrg #define count_leading_zeros(count, x) \
321 1.1 mrg do { \
322 1.1 mrg UWtype _x = (x), _y, _a, _c; \
323 1.1 mrg __asm__ ("mux1 %0 = %1, @rev" : "=r" (_y) : "r" (_x)); \
324 1.1 mrg __asm__ ("czx1.l %0 = %1" : "=r" (_a) : "r" (-_y | _y)); \
325 1.1 mrg _c = (_a - 1) << 3; \
326 1.1 mrg _x >>= _c; \
327 1.1 mrg if (_x >= 1 << 4) \
328 1.1 mrg _x >>= 4, _c += 4; \
329 1.1 mrg if (_x >= 1 << 2) \
330 1.1 mrg _x >>= 2, _c += 2; \
331 1.1 mrg _c += _x >> 1; \
332 1.1 mrg (count) = W_TYPE_SIZE - 1 - _c; \
333 1.1 mrg } while (0)
334 1.1 mrg /* similar to what gcc does for __builtin_ffs, but 0 based rather than 1
335 1.1 mrg based, and we don't need a special case for x==0 here */
336 1.1 mrg #define count_trailing_zeros(count, x) \
337 1.1 mrg do { \
338 1.1 mrg UWtype __ctz_x = (x); \
339 1.1 mrg __asm__ ("popcnt %0 = %1" \
340 1.1 mrg : "=r" (count) \
341 1.1 mrg : "r" ((__ctz_x-1) & ~__ctz_x)); \
342 1.1 mrg } while (0)
343 1.1 mrg #endif
344 1.1 mrg #if defined (__INTEL_COMPILER)
345 1.1 mrg #include <ia64intrin.h>
346 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \
347 1.1 mrg do { \
348 1.1 mrg UWtype _m0 = (m0), _m1 = (m1); \
349 1.1 mrg ph = _m64_xmahu (_m0, _m1, 0); \
350 1.1 mrg pl = _m0 * _m1; \
351 1.1 mrg } while (0)
352 1.1 mrg #endif
353 1.1 mrg #ifndef LONGLONG_STANDALONE
354 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
355 1.1 mrg do { UWtype __di; \
356 1.1 mrg __di = __MPN(invert_limb) (d); \
357 1.1 mrg udiv_qrnnd_preinv (q, r, n1, n0, d, __di); \
358 1.1 mrg } while (0)
359 1.1 mrg #define UDIV_PREINV_ALWAYS 1
360 1.1 mrg #define UDIV_NEEDS_NORMALIZATION 1
361 1.1 mrg #endif
362 1.1 mrg #define UDIV_TIME 220
363 1.1 mrg #endif
364 1.1 mrg
365 1.1 mrg
366 1.1 mrg #if defined (__GNUC__)
367 1.1 mrg
368 1.1 mrg /* We sometimes need to clobber "cc" with gcc2, but that would not be
369 1.1 mrg understood by gcc1. Use cpp to avoid major code duplication. */
370 1.1 mrg #if __GNUC__ < 2
371 1.1 mrg #define __CLOBBER_CC
372 1.1 mrg #define __AND_CLOBBER_CC
373 1.1 mrg #else /* __GNUC__ >= 2 */
374 1.1 mrg #define __CLOBBER_CC : "cc"
375 1.1 mrg #define __AND_CLOBBER_CC , "cc"
376 1.1 mrg #endif /* __GNUC__ < 2 */
377 1.1 mrg
378 1.1 mrg #if (defined (__a29k__) || defined (_AM29K)) && W_TYPE_SIZE == 32
379 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
380 1.1 mrg __asm__ ("add %1,%4,%5\n\taddc %0,%2,%3" \
381 1.1 mrg : "=r" (sh), "=&r" (sl) \
382 1.1 mrg : "r" (ah), "rI" (bh), "%r" (al), "rI" (bl))
383 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
384 1.1 mrg __asm__ ("sub %1,%4,%5\n\tsubc %0,%2,%3" \
385 1.1 mrg : "=r" (sh), "=&r" (sl) \
386 1.1 mrg : "r" (ah), "rI" (bh), "r" (al), "rI" (bl))
387 1.1 mrg #define umul_ppmm(xh, xl, m0, m1) \
388 1.1 mrg do { \
389 1.1 mrg USItype __m0 = (m0), __m1 = (m1); \
390 1.1 mrg __asm__ ("multiplu %0,%1,%2" \
391 1.1 mrg : "=r" (xl) \
392 1.1 mrg : "r" (__m0), "r" (__m1)); \
393 1.1 mrg __asm__ ("multmu %0,%1,%2" \
394 1.1 mrg : "=r" (xh) \
395 1.1 mrg : "r" (__m0), "r" (__m1)); \
396 1.1 mrg } while (0)
397 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
398 1.1 mrg __asm__ ("dividu %0,%3,%4" \
399 1.1 mrg : "=r" (q), "=q" (r) \
400 1.1 mrg : "1" (n1), "r" (n0), "r" (d))
401 1.1 mrg #define count_leading_zeros(count, x) \
402 1.1 mrg __asm__ ("clz %0,%1" \
403 1.1 mrg : "=r" (count) \
404 1.1 mrg : "r" (x))
405 1.1 mrg #define COUNT_LEADING_ZEROS_0 32
406 1.1 mrg #endif /* __a29k__ */
407 1.1 mrg
408 1.1 mrg #if defined (__arc__)
409 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
410 1.1 mrg __asm__ ("add.f\t%1, %4, %5\n\tadc\t%0, %2, %3" \
411 1.1 mrg : "=r" (sh), \
412 1.1 mrg "=&r" (sl) \
413 1.1 mrg : "r" ((USItype) (ah)), \
414 1.1 mrg "rIJ" ((USItype) (bh)), \
415 1.1 mrg "%r" ((USItype) (al)), \
416 1.1 mrg "rIJ" ((USItype) (bl)))
417 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
418 1.1 mrg __asm__ ("sub.f\t%1, %4, %5\n\tsbc\t%0, %2, %3" \
419 1.1 mrg : "=r" (sh), \
420 1.1 mrg "=&r" (sl) \
421 1.1 mrg : "r" ((USItype) (ah)), \
422 1.1 mrg "rIJ" ((USItype) (bh)), \
423 1.1 mrg "r" ((USItype) (al)), \
424 1.1 mrg "rIJ" ((USItype) (bl)))
425 1.1 mrg #endif
426 1.1 mrg
427 1.1.1.1.8.1 tls #if defined (__arm__) && !defined (__thumb__) && W_TYPE_SIZE == 32
428 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
429 1.1 mrg __asm__ ("adds\t%1, %4, %5\n\tadc\t%0, %2, %3" \
430 1.1 mrg : "=r" (sh), "=&r" (sl) \
431 1.1 mrg : "r" (ah), "rI" (bh), "%r" (al), "rI" (bl) __CLOBBER_CC)
432 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
433 1.1 mrg do { \
434 1.1 mrg if (__builtin_constant_p (al)) \
435 1.1 mrg { \
436 1.1 mrg if (__builtin_constant_p (ah)) \
437 1.1 mrg __asm__ ("rsbs\t%1, %5, %4\n\trsc\t%0, %3, %2" \
438 1.1 mrg : "=r" (sh), "=&r" (sl) \
439 1.1 mrg : "rI" (ah), "r" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \
440 1.1 mrg else \
441 1.1 mrg __asm__ ("rsbs\t%1, %5, %4\n\tsbc\t%0, %2, %3" \
442 1.1 mrg : "=r" (sh), "=&r" (sl) \
443 1.1 mrg : "r" (ah), "rI" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \
444 1.1 mrg } \
445 1.1 mrg else if (__builtin_constant_p (ah)) \
446 1.1 mrg { \
447 1.1 mrg if (__builtin_constant_p (bl)) \
448 1.1 mrg __asm__ ("subs\t%1, %4, %5\n\trsc\t%0, %3, %2" \
449 1.1 mrg : "=r" (sh), "=&r" (sl) \
450 1.1 mrg : "rI" (ah), "r" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \
451 1.1 mrg else \
452 1.1 mrg __asm__ ("rsbs\t%1, %5, %4\n\trsc\t%0, %3, %2" \
453 1.1 mrg : "=r" (sh), "=&r" (sl) \
454 1.1 mrg : "rI" (ah), "r" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \
455 1.1 mrg } \
456 1.1 mrg else if (__builtin_constant_p (bl)) \
457 1.1 mrg { \
458 1.1 mrg if (__builtin_constant_p (bh)) \
459 1.1 mrg __asm__ ("subs\t%1, %4, %5\n\tsbc\t%0, %2, %3" \
460 1.1 mrg : "=r" (sh), "=&r" (sl) \
461 1.1 mrg : "r" (ah), "rI" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \
462 1.1 mrg else \
463 1.1 mrg __asm__ ("subs\t%1, %4, %5\n\trsc\t%0, %3, %2" \
464 1.1 mrg : "=r" (sh), "=&r" (sl) \
465 1.1 mrg : "rI" (ah), "r" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \
466 1.1 mrg } \
467 1.1 mrg else /* only bh might be a constant */ \
468 1.1 mrg __asm__ ("subs\t%1, %4, %5\n\tsbc\t%0, %2, %3" \
469 1.1 mrg : "=r" (sh), "=&r" (sl) \
470 1.1 mrg : "r" (ah), "rI" (bh), "r" (al), "rI" (bl) __CLOBBER_CC);\
471 1.1 mrg } while (0)
472 1.1 mrg #if 1 || defined (__arm_m__) /* `M' series has widening multiply support */
473 1.1 mrg #define umul_ppmm(xh, xl, a, b) \
474 1.1 mrg __asm__ ("umull %0,%1,%2,%3" : "=&r" (xl), "=&r" (xh) : "r" (a), "r" (b))
475 1.1 mrg #define UMUL_TIME 5
476 1.1 mrg #define smul_ppmm(xh, xl, a, b) \
477 1.1 mrg __asm__ ("smull %0,%1,%2,%3" : "=&r" (xl), "=&r" (xh) : "r" (a), "r" (b))
478 1.1 mrg #ifndef LONGLONG_STANDALONE
479 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
480 1.1 mrg do { UWtype __di; \
481 1.1 mrg __di = __MPN(invert_limb) (d); \
482 1.1 mrg udiv_qrnnd_preinv (q, r, n1, n0, d, __di); \
483 1.1 mrg } while (0)
484 1.1 mrg #define UDIV_PREINV_ALWAYS 1
485 1.1 mrg #define UDIV_NEEDS_NORMALIZATION 1
486 1.1 mrg #define UDIV_TIME 70
487 1.1 mrg #endif /* LONGLONG_STANDALONE */
488 1.1 mrg #else
489 1.1 mrg #define umul_ppmm(xh, xl, a, b) \
490 1.1 mrg __asm__ ("%@ Inlined umul_ppmm\n" \
491 1.1 mrg " mov %|r0, %2, lsr #16\n" \
492 1.1 mrg " mov %|r2, %3, lsr #16\n" \
493 1.1 mrg " bic %|r1, %2, %|r0, lsl #16\n" \
494 1.1 mrg " bic %|r2, %3, %|r2, lsl #16\n" \
495 1.1 mrg " mul %1, %|r1, %|r2\n" \
496 1.1 mrg " mul %|r2, %|r0, %|r2\n" \
497 1.1 mrg " mul %|r1, %0, %|r1\n" \
498 1.1 mrg " mul %0, %|r0, %0\n" \
499 1.1 mrg " adds %|r1, %|r2, %|r1\n" \
500 1.1 mrg " addcs %0, %0, #65536\n" \
501 1.1 mrg " adds %1, %1, %|r1, lsl #16\n" \
502 1.1 mrg " adc %0, %0, %|r1, lsr #16" \
503 1.1 mrg : "=&r" (xh), "=r" (xl) \
504 1.1 mrg : "r" (a), "r" (b) \
505 1.1 mrg : "r0", "r1", "r2")
506 1.1 mrg #define UMUL_TIME 20
507 1.1 mrg #ifndef LONGLONG_STANDALONE
508 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
509 1.1 mrg do { UWtype __r; \
510 1.1 mrg (q) = __MPN(udiv_qrnnd) (&__r, (n1), (n0), (d)); \
511 1.1 mrg (r) = __r; \
512 1.1 mrg } while (0)
513 1.1.1.1.8.1 tls extern UWtype __MPN(udiv_qrnnd) (UWtype *, UWtype, UWtype, UWtype);
514 1.1 mrg #define UDIV_TIME 200
515 1.1 mrg #endif /* LONGLONG_STANDALONE */
516 1.1 mrg #endif
517 1.1.1.1.8.1 tls /* This is a bizarre test, but GCC doesn't define useful common symbol. */
518 1.1.1.1.8.1 tls #if defined (__ARM_ARCH_5__) || defined (__ARM_ARCH_5T__) || \
519 1.1.1.1.8.1 tls defined (__ARM_ARCH_5E__) || defined (__ARM_ARCH_5TE__)|| \
520 1.1.1.1.8.1 tls defined (__ARM_ARCH_6__) || defined (__ARM_ARCH_6J__) || \
521 1.1.1.1.8.1 tls defined (__ARM_ARCH_6K__) || defined (__ARM_ARCH_6Z__) || \
522 1.1.1.1.8.1 tls defined (__ARM_ARCH_6ZK__)|| defined (__ARM_ARCH_6T2__)|| \
523 1.1.1.1.8.1 tls defined (__ARM_ARCH_6M__) || defined (__ARM_ARCH_7__) || \
524 1.1.1.1.8.1 tls defined (__ARM_ARCH_7A__) || defined (__ARM_ARCH_7R__) || \
525 1.1.1.1.8.1 tls defined (__ARM_ARCH_7M__) || defined (__ARM_ARCH_7EM__)
526 1.1 mrg #define count_leading_zeros(count, x) \
527 1.1 mrg __asm__ ("clz\t%0, %1" : "=r" (count) : "r" (x))
528 1.1 mrg #define COUNT_LEADING_ZEROS_0 32
529 1.1 mrg #endif
530 1.1 mrg #endif /* __arm__ */
531 1.1 mrg
532 1.1.1.1.8.1 tls #if defined (__aarch64__) && W_TYPE_SIZE == 64
533 1.1.1.1.8.1 tls /* FIXME: Extend the immediate range for the low word by using both
534 1.1.1.1.8.1 tls ADDS and SUBS, since they set carry in the same way. */
535 1.1.1.1.8.1 tls #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
536 1.1.1.1.8.1 tls __asm__ ("adds\t%1, %x4, %5\n\tadc\t%0, %x2, %x3" \
537 1.1.1.1.8.1 tls : "=r" (sh), "=&r" (sl) \
538 1.1.1.1.8.1 tls : "rZ" (ah), "rZ" (bh), "%r" (al), "rI" (bl) __CLOBBER_CC)
539 1.1.1.1.8.1 tls #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
540 1.1.1.1.8.1 tls __asm__ ("subs\t%1, %x4, %5\n\tsbc\t%0, %x2, %x3" \
541 1.1.1.1.8.1 tls : "=r,r" (sh), "=&r,&r" (sl) \
542 1.1.1.1.8.1 tls : "rZ,rZ" (ah), "rZ,rZ" (bh), "r,Z" (al), "rI,r" (bl) __CLOBBER_CC)
543 1.1.1.1.8.1 tls #define umul_ppmm(ph, pl, m0, m1) \
544 1.1.1.1.8.1 tls do { \
545 1.1.1.1.8.1 tls UDItype __m0 = (m0), __m1 = (m1); \
546 1.1.1.1.8.1 tls __asm__ ("umulh\t%0, %1, %2" : "=r" (ph) : "r" (m0), "r" (m1)); \
547 1.1.1.1.8.1 tls (pl) = __m0 * __m1; \
548 1.1.1.1.8.1 tls } while (0)
549 1.1.1.1.8.1 tls #define count_leading_zeros(count, x) \
550 1.1.1.1.8.1 tls __asm__ ("clz\t%0, %1" : "=r" (count) : "r" (x))
551 1.1.1.1.8.1 tls #define COUNT_LEADING_ZEROS_0 64
552 1.1.1.1.8.1 tls #endif /* __aarch64__ */
553 1.1.1.1.8.1 tls
554 1.1 mrg #if defined (__clipper__) && W_TYPE_SIZE == 32
555 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
556 1.1 mrg ({union {UDItype __ll; \
557 1.1 mrg struct {USItype __l, __h;} __i; \
558 1.1 mrg } __x; \
559 1.1 mrg __asm__ ("mulwux %2,%0" \
560 1.1 mrg : "=r" (__x.__ll) \
561 1.1 mrg : "%0" ((USItype)(u)), "r" ((USItype)(v))); \
562 1.1 mrg (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
563 1.1 mrg #define smul_ppmm(w1, w0, u, v) \
564 1.1 mrg ({union {DItype __ll; \
565 1.1 mrg struct {SItype __l, __h;} __i; \
566 1.1 mrg } __x; \
567 1.1 mrg __asm__ ("mulwx %2,%0" \
568 1.1 mrg : "=r" (__x.__ll) \
569 1.1 mrg : "%0" ((SItype)(u)), "r" ((SItype)(v))); \
570 1.1 mrg (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
571 1.1 mrg #define __umulsidi3(u, v) \
572 1.1 mrg ({UDItype __w; \
573 1.1 mrg __asm__ ("mulwux %2,%0" \
574 1.1 mrg : "=r" (__w) : "%0" ((USItype)(u)), "r" ((USItype)(v))); \
575 1.1 mrg __w; })
576 1.1 mrg #endif /* __clipper__ */
577 1.1 mrg
578 1.1 mrg /* Fujitsu vector computers. */
579 1.1 mrg #if defined (__uxp__) && W_TYPE_SIZE == 32
580 1.1 mrg #define umul_ppmm(ph, pl, u, v) \
581 1.1 mrg do { \
582 1.1 mrg union {UDItype __ll; \
583 1.1 mrg struct {USItype __h, __l;} __i; \
584 1.1 mrg } __x; \
585 1.1 mrg __asm__ ("mult.lu %1,%2,%0" : "=r" (__x.__ll) : "%r" (u), "rK" (v));\
586 1.1 mrg (ph) = __x.__i.__h; \
587 1.1 mrg (pl) = __x.__i.__l; \
588 1.1 mrg } while (0)
589 1.1 mrg #define smul_ppmm(ph, pl, u, v) \
590 1.1 mrg do { \
591 1.1 mrg union {UDItype __ll; \
592 1.1 mrg struct {USItype __h, __l;} __i; \
593 1.1 mrg } __x; \
594 1.1 mrg __asm__ ("mult.l %1,%2,%0" : "=r" (__x.__ll) : "%r" (u), "rK" (v)); \
595 1.1 mrg (ph) = __x.__i.__h; \
596 1.1 mrg (pl) = __x.__i.__l; \
597 1.1 mrg } while (0)
598 1.1 mrg #endif
599 1.1 mrg
600 1.1 mrg #if defined (__gmicro__) && W_TYPE_SIZE == 32
601 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
602 1.1 mrg __asm__ ("add.w %5,%1\n\taddx %3,%0" \
603 1.1 mrg : "=g" (sh), "=&g" (sl) \
604 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \
605 1.1 mrg "%1" ((USItype)(al)), "g" ((USItype)(bl)))
606 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
607 1.1 mrg __asm__ ("sub.w %5,%1\n\tsubx %3,%0" \
608 1.1 mrg : "=g" (sh), "=&g" (sl) \
609 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \
610 1.1 mrg "1" ((USItype)(al)), "g" ((USItype)(bl)))
611 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \
612 1.1 mrg __asm__ ("mulx %3,%0,%1" \
613 1.1 mrg : "=g" (ph), "=r" (pl) \
614 1.1 mrg : "%0" ((USItype)(m0)), "g" ((USItype)(m1)))
615 1.1 mrg #define udiv_qrnnd(q, r, nh, nl, d) \
616 1.1 mrg __asm__ ("divx %4,%0,%1" \
617 1.1 mrg : "=g" (q), "=r" (r) \
618 1.1 mrg : "1" ((USItype)(nh)), "0" ((USItype)(nl)), "g" ((USItype)(d)))
619 1.1 mrg #define count_leading_zeros(count, x) \
620 1.1 mrg __asm__ ("bsch/1 %1,%0" \
621 1.1 mrg : "=g" (count) : "g" ((USItype)(x)), "0" ((USItype)0))
622 1.1 mrg #endif
623 1.1 mrg
624 1.1 mrg #if defined (__hppa) && W_TYPE_SIZE == 32
625 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
626 1.1 mrg __asm__ ("add%I5 %5,%r4,%1\n\taddc %r2,%r3,%0" \
627 1.1 mrg : "=r" (sh), "=&r" (sl) \
628 1.1 mrg : "rM" (ah), "rM" (bh), "%rM" (al), "rI" (bl))
629 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
630 1.1 mrg __asm__ ("sub%I4 %4,%r5,%1\n\tsubb %r2,%r3,%0" \
631 1.1 mrg : "=r" (sh), "=&r" (sl) \
632 1.1 mrg : "rM" (ah), "rM" (bh), "rI" (al), "rM" (bl))
633 1.1 mrg #if defined (_PA_RISC1_1)
634 1.1 mrg #define umul_ppmm(wh, wl, u, v) \
635 1.1 mrg do { \
636 1.1 mrg union {UDItype __ll; \
637 1.1 mrg struct {USItype __h, __l;} __i; \
638 1.1 mrg } __x; \
639 1.1 mrg __asm__ ("xmpyu %1,%2,%0" : "=*f" (__x.__ll) : "*f" (u), "*f" (v)); \
640 1.1 mrg (wh) = __x.__i.__h; \
641 1.1 mrg (wl) = __x.__i.__l; \
642 1.1 mrg } while (0)
643 1.1 mrg #define UMUL_TIME 8
644 1.1 mrg #define UDIV_TIME 60
645 1.1 mrg #else
646 1.1 mrg #define UMUL_TIME 40
647 1.1 mrg #define UDIV_TIME 80
648 1.1 mrg #endif
649 1.1 mrg #define count_leading_zeros(count, x) \
650 1.1 mrg do { \
651 1.1 mrg USItype __tmp; \
652 1.1 mrg __asm__ ( \
653 1.1 mrg "ldi 1,%0\n" \
654 1.1 mrg " extru,= %1,15,16,%%r0 ; Bits 31..16 zero?\n" \
655 1.1 mrg " extru,tr %1,15,16,%1 ; No. Shift down, skip add.\n" \
656 1.1 mrg " ldo 16(%0),%0 ; Yes. Perform add.\n" \
657 1.1 mrg " extru,= %1,23,8,%%r0 ; Bits 15..8 zero?\n" \
658 1.1 mrg " extru,tr %1,23,8,%1 ; No. Shift down, skip add.\n" \
659 1.1 mrg " ldo 8(%0),%0 ; Yes. Perform add.\n" \
660 1.1 mrg " extru,= %1,27,4,%%r0 ; Bits 7..4 zero?\n" \
661 1.1 mrg " extru,tr %1,27,4,%1 ; No. Shift down, skip add.\n" \
662 1.1 mrg " ldo 4(%0),%0 ; Yes. Perform add.\n" \
663 1.1 mrg " extru,= %1,29,2,%%r0 ; Bits 3..2 zero?\n" \
664 1.1 mrg " extru,tr %1,29,2,%1 ; No. Shift down, skip add.\n" \
665 1.1 mrg " ldo 2(%0),%0 ; Yes. Perform add.\n" \
666 1.1 mrg " extru %1,30,1,%1 ; Extract bit 1.\n" \
667 1.1 mrg " sub %0,%1,%0 ; Subtract it.\n" \
668 1.1 mrg : "=r" (count), "=r" (__tmp) : "1" (x)); \
669 1.1 mrg } while (0)
670 1.1 mrg #endif /* hppa */
671 1.1 mrg
672 1.1 mrg /* These macros are for ABI=2.0w. In ABI=2.0n they can't be used, since GCC
673 1.1 mrg (3.2) puts longlong into two adjacent 32-bit registers. Presumably this
674 1.1 mrg is just a case of no direct support for 2.0n but treating it like 1.0. */
675 1.1 mrg #if defined (__hppa) && W_TYPE_SIZE == 64 && ! defined (_LONG_LONG_LIMB)
676 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
677 1.1 mrg __asm__ ("add%I5 %5,%r4,%1\n\tadd,dc %r2,%r3,%0" \
678 1.1 mrg : "=r" (sh), "=&r" (sl) \
679 1.1 mrg : "rM" (ah), "rM" (bh), "%rM" (al), "rI" (bl))
680 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
681 1.1 mrg __asm__ ("sub%I4 %4,%r5,%1\n\tsub,db %r2,%r3,%0" \
682 1.1 mrg : "=r" (sh), "=&r" (sl) \
683 1.1 mrg : "rM" (ah), "rM" (bh), "rI" (al), "rM" (bl))
684 1.1 mrg #endif /* hppa */
685 1.1 mrg
686 1.1 mrg #if (defined (__i370__) || defined (__s390__) || defined (__mvs__)) && W_TYPE_SIZE == 32
687 1.1.1.1.8.1 tls #if defined (__zarch__) || defined (HAVE_HOST_CPU_s390_zarch)
688 1.1.1.1.8.1 tls #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
689 1.1.1.1.8.1 tls do { \
690 1.1.1.1.8.1 tls /* if (__builtin_constant_p (bl)) \
691 1.1.1.1.8.1 tls __asm__ ("alfi\t%1,%o5\n\talcr\t%0,%3" \
692 1.1.1.1.8.1 tls : "=r" (sh), "=&r" (sl) \
693 1.1.1.1.8.1 tls : "0" (ah), "r" (bh), "%1" (al), "n" (bl) __CLOBBER_CC);\
694 1.1.1.1.8.1 tls else \
695 1.1.1.1.8.1 tls */ __asm__ ("alr\t%1,%5\n\talcr\t%0,%3" \
696 1.1.1.1.8.1 tls : "=r" (sh), "=&r" (sl) \
697 1.1.1.1.8.1 tls : "0" (ah), "r" (bh), "%1" (al), "r" (bl)__CLOBBER_CC); \
698 1.1.1.1.8.1 tls } while (0)
699 1.1.1.1.8.1 tls #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
700 1.1.1.1.8.1 tls do { \
701 1.1.1.1.8.1 tls /* if (__builtin_constant_p (bl)) \
702 1.1.1.1.8.1 tls __asm__ ("slfi\t%1,%o5\n\tslbr\t%0,%3" \
703 1.1.1.1.8.1 tls : "=r" (sh), "=&r" (sl) \
704 1.1.1.1.8.1 tls : "0" (ah), "r" (bh), "1" (al), "n" (bl) __CLOBBER_CC); \
705 1.1.1.1.8.1 tls else \
706 1.1.1.1.8.1 tls */ __asm__ ("slr\t%1,%5\n\tslbr\t%0,%3" \
707 1.1.1.1.8.1 tls : "=r" (sh), "=&r" (sl) \
708 1.1.1.1.8.1 tls : "0" (ah), "r" (bh), "1" (al), "r" (bl) __CLOBBER_CC); \
709 1.1.1.1.8.1 tls } while (0)
710 1.1.1.1.8.1 tls #if __GMP_GNUC_PREREQ (4,5)
711 1.1.1.1.8.1 tls #define umul_ppmm(xh, xl, m0, m1) \
712 1.1.1.1.8.1 tls do { \
713 1.1.1.1.8.1 tls union {UDItype __ll; \
714 1.1.1.1.8.1 tls struct {USItype __h, __l;} __i; \
715 1.1.1.1.8.1 tls } __x; \
716 1.1.1.1.8.1 tls __x.__ll = (UDItype) (m0) * (UDItype) (m1); \
717 1.1.1.1.8.1 tls (xh) = __x.__i.__h; (xl) = __x.__i.__l; \
718 1.1.1.1.8.1 tls } while (0)
719 1.1.1.1.8.1 tls #else
720 1.1.1.1.8.1 tls #if 0
721 1.1.1.1.8.1 tls /* FIXME: this fails if gcc knows about the 64-bit registers. Use only
722 1.1.1.1.8.1 tls with a new enough processor pretending we have 32-bit registers. */
723 1.1.1.1.8.1 tls #define umul_ppmm(xh, xl, m0, m1) \
724 1.1.1.1.8.1 tls do { \
725 1.1.1.1.8.1 tls union {UDItype __ll; \
726 1.1.1.1.8.1 tls struct {USItype __h, __l;} __i; \
727 1.1.1.1.8.1 tls } __x; \
728 1.1.1.1.8.1 tls __asm__ ("mlr\t%0,%2" \
729 1.1.1.1.8.1 tls : "=r" (__x.__ll) \
730 1.1.1.1.8.1 tls : "%0" (m0), "r" (m1)); \
731 1.1.1.1.8.1 tls (xh) = __x.__i.__h; (xl) = __x.__i.__l; \
732 1.1.1.1.8.1 tls } while (0)
733 1.1.1.1.8.1 tls #else
734 1.1.1.1.8.1 tls #define umul_ppmm(xh, xl, m0, m1) \
735 1.1.1.1.8.1 tls do { \
736 1.1.1.1.8.1 tls /* When we have 64-bit regs and gcc is aware of that, we cannot simply use
737 1.1.1.1.8.1 tls DImode for the product, since that would be allocated to a single 64-bit
738 1.1.1.1.8.1 tls register, whereas mlr uses the low 32-bits of an even-odd register pair.
739 1.1.1.1.8.1 tls */ \
740 1.1.1.1.8.1 tls register USItype __r0 __asm__ ("0"); \
741 1.1.1.1.8.1 tls register USItype __r1 __asm__ ("1") = (m0); \
742 1.1.1.1.8.1 tls __asm__ ("mlr\t%0,%3" \
743 1.1.1.1.8.1 tls : "=r" (__r0), "=r" (__r1) \
744 1.1.1.1.8.1 tls : "r" (__r1), "r" (m1)); \
745 1.1.1.1.8.1 tls (xh) = __r0; (xl) = __r1; \
746 1.1.1.1.8.1 tls } while (0)
747 1.1.1.1.8.1 tls #endif /* if 0 */
748 1.1.1.1.8.1 tls #endif
749 1.1.1.1.8.1 tls #if 0
750 1.1.1.1.8.1 tls /* FIXME: this fails if gcc knows about the 64-bit registers. Use only
751 1.1.1.1.8.1 tls with a new enough processor pretending we have 32-bit registers. */
752 1.1.1.1.8.1 tls #define udiv_qrnnd(q, r, n1, n0, d) \
753 1.1.1.1.8.1 tls do { \
754 1.1.1.1.8.1 tls union {UDItype __ll; \
755 1.1.1.1.8.1 tls struct {USItype __h, __l;} __i; \
756 1.1.1.1.8.1 tls } __x; \
757 1.1.1.1.8.1 tls __x.__i.__h = n1; __x.__i.__l = n0; \
758 1.1.1.1.8.1 tls __asm__ ("dlr\t%0,%2" \
759 1.1.1.1.8.1 tls : "=r" (__x.__ll) \
760 1.1.1.1.8.1 tls : "0" (__x.__ll), "r" (d)); \
761 1.1.1.1.8.1 tls (q) = __x.__i.__l; (r) = __x.__i.__h; \
762 1.1.1.1.8.1 tls } while (0)
763 1.1.1.1.8.1 tls #else
764 1.1.1.1.8.1 tls #define udiv_qrnnd(q, r, n1, n0, d) \
765 1.1.1.1.8.1 tls do { \
766 1.1.1.1.8.1 tls register USItype __r0 __asm__ ("0") = (n1); \
767 1.1.1.1.8.1 tls register USItype __r1 __asm__ ("1") = (n0); \
768 1.1.1.1.8.1 tls __asm__ ("dlr\t%0,%4" \
769 1.1.1.1.8.1 tls : "=r" (__r0), "=r" (__r1) \
770 1.1.1.1.8.1 tls : "r" (__r0), "r" (__r1), "r" (d)); \
771 1.1.1.1.8.1 tls (q) = __r1; (r) = __r0; \
772 1.1.1.1.8.1 tls } while (0)
773 1.1.1.1.8.1 tls #endif /* if 0 */
774 1.1.1.1.8.1 tls #else /* if __zarch__ */
775 1.1.1.1.8.1 tls /* FIXME: this fails if gcc knows about the 64-bit registers. */
776 1.1.1.1.8.1 tls #define smul_ppmm(xh, xl, m0, m1) \
777 1.1 mrg do { \
778 1.1 mrg union {DItype __ll; \
779 1.1 mrg struct {USItype __h, __l;} __i; \
780 1.1 mrg } __x; \
781 1.1.1.1.8.1 tls __asm__ ("mr\t%0,%2" \
782 1.1.1.1.8.1 tls : "=r" (__x.__ll) \
783 1.1.1.1.8.1 tls : "%0" (m0), "r" (m1)); \
784 1.1 mrg (xh) = __x.__i.__h; (xl) = __x.__i.__l; \
785 1.1 mrg } while (0)
786 1.1.1.1.8.1 tls /* FIXME: this fails if gcc knows about the 64-bit registers. */
787 1.1.1.1.8.1 tls #define sdiv_qrnnd(q, r, n1, n0, d) \
788 1.1 mrg do { \
789 1.1 mrg union {DItype __ll; \
790 1.1 mrg struct {USItype __h, __l;} __i; \
791 1.1 mrg } __x; \
792 1.1 mrg __x.__i.__h = n1; __x.__i.__l = n0; \
793 1.1.1.1.8.1 tls __asm__ ("dr\t%0,%2" \
794 1.1 mrg : "=r" (__x.__ll) \
795 1.1 mrg : "0" (__x.__ll), "r" (d)); \
796 1.1 mrg (q) = __x.__i.__l; (r) = __x.__i.__h; \
797 1.1 mrg } while (0)
798 1.1.1.1.8.1 tls #endif /* if __zarch__ */
799 1.1.1.1.8.1 tls #endif
800 1.1.1.1.8.1 tls
801 1.1.1.1.8.1 tls #if defined (__s390x__) && W_TYPE_SIZE == 64
802 1.1.1.1.8.1 tls /* We need to cast operands with register constraints, otherwise their types
803 1.1.1.1.8.1 tls will be assumed to be SImode by gcc. For these machines, such operations
804 1.1.1.1.8.1 tls will insert a value into the low 32 bits, and leave the high 32 bits with
805 1.1.1.1.8.1 tls garbage. */
806 1.1.1.1.8.1 tls #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
807 1.1.1.1.8.1 tls do { \
808 1.1.1.1.8.1 tls __asm__ ("algr\t%1,%5\n\talcgr\t%0,%3" \
809 1.1.1.1.8.1 tls : "=r" (sh), "=&r" (sl) \
810 1.1.1.1.8.1 tls : "0" ((UDItype)(ah)), "r" ((UDItype)(bh)), \
811 1.1.1.1.8.1 tls "%1" ((UDItype)(al)), "r" ((UDItype)(bl)) __CLOBBER_CC); \
812 1.1.1.1.8.1 tls } while (0)
813 1.1.1.1.8.1 tls #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
814 1.1.1.1.8.1 tls do { \
815 1.1.1.1.8.1 tls __asm__ ("slgr\t%1,%5\n\tslbgr\t%0,%3" \
816 1.1.1.1.8.1 tls : "=r" (sh), "=&r" (sl) \
817 1.1.1.1.8.1 tls : "0" ((UDItype)(ah)), "r" ((UDItype)(bh)), \
818 1.1.1.1.8.1 tls "1" ((UDItype)(al)), "r" ((UDItype)(bl)) __CLOBBER_CC); \
819 1.1.1.1.8.1 tls } while (0)
820 1.1.1.1.8.1 tls #define umul_ppmm(xh, xl, m0, m1) \
821 1.1.1.1.8.1 tls do { \
822 1.1.1.1.8.1 tls union {unsigned int __attribute__ ((mode(TI))) __ll; \
823 1.1.1.1.8.1 tls struct {UDItype __h, __l;} __i; \
824 1.1.1.1.8.1 tls } __x; \
825 1.1.1.1.8.1 tls __asm__ ("mlgr\t%0,%2" \
826 1.1.1.1.8.1 tls : "=r" (__x.__ll) \
827 1.1.1.1.8.1 tls : "%0" ((UDItype)(m0)), "r" ((UDItype)(m1))); \
828 1.1.1.1.8.1 tls (xh) = __x.__i.__h; (xl) = __x.__i.__l; \
829 1.1.1.1.8.1 tls } while (0)
830 1.1.1.1.8.1 tls #define udiv_qrnnd(q, r, n1, n0, d) \
831 1.1.1.1.8.1 tls do { \
832 1.1.1.1.8.1 tls union {unsigned int __attribute__ ((mode(TI))) __ll; \
833 1.1.1.1.8.1 tls struct {UDItype __h, __l;} __i; \
834 1.1.1.1.8.1 tls } __x; \
835 1.1.1.1.8.1 tls __x.__i.__h = n1; __x.__i.__l = n0; \
836 1.1.1.1.8.1 tls __asm__ ("dlgr\t%0,%2" \
837 1.1.1.1.8.1 tls : "=r" (__x.__ll) \
838 1.1.1.1.8.1 tls : "0" (__x.__ll), "r" ((UDItype)(d))); \
839 1.1.1.1.8.1 tls (q) = __x.__i.__l; (r) = __x.__i.__h; \
840 1.1.1.1.8.1 tls } while (0)
841 1.1.1.1.8.1 tls #if 0 /* FIXME: Enable for z10 (?) */
842 1.1.1.1.8.1 tls #define count_leading_zeros(cnt, x) \
843 1.1.1.1.8.1 tls do { \
844 1.1.1.1.8.1 tls union {unsigned int __attribute__ ((mode(TI))) __ll; \
845 1.1.1.1.8.1 tls struct {UDItype __h, __l;} __i; \
846 1.1.1.1.8.1 tls } __clr_cnt; \
847 1.1.1.1.8.1 tls __asm__ ("flogr\t%0,%1" \
848 1.1.1.1.8.1 tls : "=r" (__clr_cnt.__ll) \
849 1.1.1.1.8.1 tls : "r" (x) __CLOBBER_CC); \
850 1.1.1.1.8.1 tls (cnt) = __clr_cnt.__i.__h; \
851 1.1.1.1.8.1 tls } while (0)
852 1.1.1.1.8.1 tls #endif
853 1.1 mrg #endif
854 1.1 mrg
855 1.1 mrg #if (defined (__i386__) || defined (__i486__)) && W_TYPE_SIZE == 32
856 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
857 1.1 mrg __asm__ ("addl %5,%k1\n\tadcl %3,%k0" \
858 1.1 mrg : "=r" (sh), "=&r" (sl) \
859 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \
860 1.1 mrg "%1" ((USItype)(al)), "g" ((USItype)(bl)))
861 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
862 1.1 mrg __asm__ ("subl %5,%k1\n\tsbbl %3,%k0" \
863 1.1 mrg : "=r" (sh), "=&r" (sl) \
864 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \
865 1.1 mrg "1" ((USItype)(al)), "g" ((USItype)(bl)))
866 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
867 1.1 mrg __asm__ ("mull %3" \
868 1.1 mrg : "=a" (w0), "=d" (w1) \
869 1.1 mrg : "%0" ((USItype)(u)), "rm" ((USItype)(v)))
870 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, dx) /* d renamed to dx avoiding "=d" */\
871 1.1 mrg __asm__ ("divl %4" /* stringification in K&R C */ \
872 1.1 mrg : "=a" (q), "=d" (r) \
873 1.1 mrg : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "rm" ((USItype)(dx)))
874 1.1 mrg
875 1.1 mrg #if HAVE_HOST_CPU_i586 || HAVE_HOST_CPU_pentium || HAVE_HOST_CPU_pentiummmx
876 1.1 mrg /* Pentium bsrl takes between 10 and 72 cycles depending where the most
877 1.1 mrg significant 1 bit is, hence the use of the following alternatives. bsfl
878 1.1 mrg is slow too, between 18 and 42 depending where the least significant 1
879 1.1 mrg bit is, so let the generic count_trailing_zeros below make use of the
880 1.1 mrg count_leading_zeros here too. */
881 1.1 mrg
882 1.1 mrg #if HAVE_HOST_CPU_pentiummmx && ! defined (LONGLONG_STANDALONE)
883 1.1 mrg /* The following should be a fixed 14 or 15 cycles, but possibly plus an L1
884 1.1 mrg cache miss reading from __clz_tab. For P55 it's favoured over the float
885 1.1 mrg below so as to avoid mixing MMX and x87, since the penalty for switching
886 1.1 mrg between the two is about 100 cycles.
887 1.1 mrg
888 1.1 mrg The asm block sets __shift to -3 if the high 24 bits are clear, -2 for
889 1.1 mrg 16, -1 for 8, or 0 otherwise. This could be written equivalently as
890 1.1 mrg follows, but as of gcc 2.95.2 it results in conditional jumps.
891 1.1 mrg
892 1.1 mrg __shift = -(__n < 0x1000000);
893 1.1 mrg __shift -= (__n < 0x10000);
894 1.1 mrg __shift -= (__n < 0x100);
895 1.1 mrg
896 1.1 mrg The middle two sbbl and cmpl's pair, and with luck something gcc
897 1.1 mrg generates might pair with the first cmpl and the last sbbl. The "32+1"
898 1.1 mrg constant could be folded into __clz_tab[], but it doesn't seem worth
899 1.1 mrg making a different table just for that. */
900 1.1 mrg
901 1.1 mrg #define count_leading_zeros(c,n) \
902 1.1 mrg do { \
903 1.1 mrg USItype __n = (n); \
904 1.1 mrg USItype __shift; \
905 1.1 mrg __asm__ ("cmpl $0x1000000, %1\n" \
906 1.1 mrg "sbbl %0, %0\n" \
907 1.1 mrg "cmpl $0x10000, %1\n" \
908 1.1 mrg "sbbl $0, %0\n" \
909 1.1 mrg "cmpl $0x100, %1\n" \
910 1.1 mrg "sbbl $0, %0\n" \
911 1.1 mrg : "=&r" (__shift) : "r" (__n)); \
912 1.1 mrg __shift = __shift*8 + 24 + 1; \
913 1.1 mrg (c) = 32 + 1 - __shift - __clz_tab[__n >> __shift]; \
914 1.1 mrg } while (0)
915 1.1 mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
916 1.1 mrg #define COUNT_LEADING_ZEROS_0 31 /* n==0 indistinguishable from n==1 */
917 1.1 mrg
918 1.1 mrg #else /* ! pentiummmx || LONGLONG_STANDALONE */
919 1.1 mrg /* The following should be a fixed 14 cycles or so. Some scheduling
920 1.1 mrg opportunities should be available between the float load/store too. This
921 1.1 mrg sort of code is used in gcc 3 for __builtin_ffs (with "n&-n") and is
922 1.1 mrg apparently suggested by the Intel optimizing manual (don't know exactly
923 1.1 mrg where). gcc 2.95 or up will be best for this, so the "double" is
924 1.1 mrg correctly aligned on the stack. */
925 1.1 mrg #define count_leading_zeros(c,n) \
926 1.1 mrg do { \
927 1.1 mrg union { \
928 1.1 mrg double d; \
929 1.1 mrg unsigned a[2]; \
930 1.1 mrg } __u; \
931 1.1 mrg ASSERT ((n) != 0); \
932 1.1 mrg __u.d = (UWtype) (n); \
933 1.1 mrg (c) = 0x3FF + 31 - (__u.a[1] >> 20); \
934 1.1 mrg } while (0)
935 1.1 mrg #define COUNT_LEADING_ZEROS_0 (0x3FF + 31)
936 1.1 mrg #endif /* pentiummx */
937 1.1 mrg
938 1.1 mrg #else /* ! pentium */
939 1.1 mrg
940 1.1 mrg #if __GMP_GNUC_PREREQ (3,4) /* using bsrl */
941 1.1 mrg #define count_leading_zeros(count,x) count_leading_zeros_gcc_clz(count,x)
942 1.1 mrg #endif /* gcc clz */
943 1.1 mrg
944 1.1 mrg /* On P6, gcc prior to 3.0 generates a partial register stall for
945 1.1 mrg __cbtmp^31, due to using "xorb $31" instead of "xorl $31", the former
946 1.1 mrg being 1 code byte smaller. "31-__cbtmp" is a workaround, probably at the
947 1.1 mrg cost of one extra instruction. Do this for "i386" too, since that means
948 1.1 mrg generic x86. */
949 1.1 mrg #if ! defined (count_leading_zeros) && __GNUC__ < 3 \
950 1.1 mrg && (HAVE_HOST_CPU_i386 \
951 1.1 mrg || HAVE_HOST_CPU_i686 \
952 1.1 mrg || HAVE_HOST_CPU_pentiumpro \
953 1.1 mrg || HAVE_HOST_CPU_pentium2 \
954 1.1 mrg || HAVE_HOST_CPU_pentium3)
955 1.1 mrg #define count_leading_zeros(count, x) \
956 1.1 mrg do { \
957 1.1 mrg USItype __cbtmp; \
958 1.1 mrg ASSERT ((x) != 0); \
959 1.1 mrg __asm__ ("bsrl %1,%0" : "=r" (__cbtmp) : "rm" ((USItype)(x))); \
960 1.1 mrg (count) = 31 - __cbtmp; \
961 1.1 mrg } while (0)
962 1.1 mrg #endif /* gcc<3 asm bsrl */
963 1.1 mrg
964 1.1 mrg #ifndef count_leading_zeros
965 1.1 mrg #define count_leading_zeros(count, x) \
966 1.1 mrg do { \
967 1.1 mrg USItype __cbtmp; \
968 1.1 mrg ASSERT ((x) != 0); \
969 1.1 mrg __asm__ ("bsrl %1,%0" : "=r" (__cbtmp) : "rm" ((USItype)(x))); \
970 1.1 mrg (count) = __cbtmp ^ 31; \
971 1.1 mrg } while (0)
972 1.1 mrg #endif /* asm bsrl */
973 1.1 mrg
974 1.1 mrg #if __GMP_GNUC_PREREQ (3,4) /* using bsfl */
975 1.1 mrg #define count_trailing_zeros(count,x) count_trailing_zeros_gcc_ctz(count,x)
976 1.1 mrg #endif /* gcc ctz */
977 1.1 mrg
978 1.1 mrg #ifndef count_trailing_zeros
979 1.1 mrg #define count_trailing_zeros(count, x) \
980 1.1 mrg do { \
981 1.1 mrg ASSERT ((x) != 0); \
982 1.1 mrg __asm__ ("bsfl %1,%k0" : "=r" (count) : "rm" ((USItype)(x))); \
983 1.1 mrg } while (0)
984 1.1 mrg #endif /* asm bsfl */
985 1.1 mrg
986 1.1 mrg #endif /* ! pentium */
987 1.1 mrg
988 1.1 mrg #ifndef UMUL_TIME
989 1.1 mrg #define UMUL_TIME 10
990 1.1 mrg #endif
991 1.1 mrg #ifndef UDIV_TIME
992 1.1 mrg #define UDIV_TIME 40
993 1.1 mrg #endif
994 1.1 mrg #endif /* 80x86 */
995 1.1 mrg
996 1.1 mrg #if defined (__amd64__) && W_TYPE_SIZE == 64
997 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
998 1.1 mrg __asm__ ("addq %5,%q1\n\tadcq %3,%q0" \
999 1.1 mrg : "=r" (sh), "=&r" (sl) \
1000 1.1 mrg : "0" ((UDItype)(ah)), "rme" ((UDItype)(bh)), \
1001 1.1 mrg "%1" ((UDItype)(al)), "rme" ((UDItype)(bl)))
1002 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1003 1.1 mrg __asm__ ("subq %5,%q1\n\tsbbq %3,%q0" \
1004 1.1 mrg : "=r" (sh), "=&r" (sl) \
1005 1.1 mrg : "0" ((UDItype)(ah)), "rme" ((UDItype)(bh)), \
1006 1.1 mrg "1" ((UDItype)(al)), "rme" ((UDItype)(bl)))
1007 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1008 1.1 mrg __asm__ ("mulq %3" \
1009 1.1 mrg : "=a" (w0), "=d" (w1) \
1010 1.1 mrg : "%0" ((UDItype)(u)), "rm" ((UDItype)(v)))
1011 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, dx) /* d renamed to dx avoiding "=d" */\
1012 1.1 mrg __asm__ ("divq %4" /* stringification in K&R C */ \
1013 1.1 mrg : "=a" (q), "=d" (r) \
1014 1.1 mrg : "0" ((UDItype)(n0)), "1" ((UDItype)(n1)), "rm" ((UDItype)(dx)))
1015 1.1 mrg /* bsrq destination must be a 64-bit register, hence UDItype for __cbtmp. */
1016 1.1 mrg #define count_leading_zeros(count, x) \
1017 1.1 mrg do { \
1018 1.1 mrg UDItype __cbtmp; \
1019 1.1 mrg ASSERT ((x) != 0); \
1020 1.1 mrg __asm__ ("bsrq %1,%0" : "=r" (__cbtmp) : "rm" ((UDItype)(x))); \
1021 1.1 mrg (count) = __cbtmp ^ 63; \
1022 1.1 mrg } while (0)
1023 1.1 mrg /* bsfq destination must be a 64-bit register, "%q0" forces this in case
1024 1.1 mrg count is only an int. */
1025 1.1 mrg #define count_trailing_zeros(count, x) \
1026 1.1 mrg do { \
1027 1.1 mrg ASSERT ((x) != 0); \
1028 1.1 mrg __asm__ ("bsfq %1,%q0" : "=r" (count) : "rm" ((UDItype)(x))); \
1029 1.1 mrg } while (0)
1030 1.1 mrg #endif /* x86_64 */
1031 1.1 mrg
1032 1.1 mrg #if defined (__i860__) && W_TYPE_SIZE == 32
1033 1.1 mrg #define rshift_rhlc(r,h,l,c) \
1034 1.1 mrg __asm__ ("shr %3,r0,r0\;shrd %1,%2,%0" \
1035 1.1 mrg "=r" (r) : "r" (h), "r" (l), "rn" (c))
1036 1.1 mrg #endif /* i860 */
1037 1.1 mrg
1038 1.1 mrg #if defined (__i960__) && W_TYPE_SIZE == 32
1039 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1040 1.1 mrg __asm__ ("cmpo 1,0\;addc %5,%4,%1\;addc %3,%2,%0" \
1041 1.1 mrg : "=r" (sh), "=&r" (sl) \
1042 1.1 mrg : "dI" (ah), "dI" (bh), "%dI" (al), "dI" (bl))
1043 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1044 1.1 mrg __asm__ ("cmpo 0,0\;subc %5,%4,%1\;subc %3,%2,%0" \
1045 1.1 mrg : "=r" (sh), "=&r" (sl) \
1046 1.1 mrg : "dI" (ah), "dI" (bh), "dI" (al), "dI" (bl))
1047 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1048 1.1 mrg ({union {UDItype __ll; \
1049 1.1 mrg struct {USItype __l, __h;} __i; \
1050 1.1 mrg } __x; \
1051 1.1 mrg __asm__ ("emul %2,%1,%0" \
1052 1.1 mrg : "=d" (__x.__ll) : "%dI" (u), "dI" (v)); \
1053 1.1 mrg (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
1054 1.1 mrg #define __umulsidi3(u, v) \
1055 1.1 mrg ({UDItype __w; \
1056 1.1 mrg __asm__ ("emul %2,%1,%0" : "=d" (__w) : "%dI" (u), "dI" (v)); \
1057 1.1 mrg __w; })
1058 1.1 mrg #define udiv_qrnnd(q, r, nh, nl, d) \
1059 1.1 mrg do { \
1060 1.1 mrg union {UDItype __ll; \
1061 1.1 mrg struct {USItype __l, __h;} __i; \
1062 1.1 mrg } __nn; \
1063 1.1 mrg __nn.__i.__h = (nh); __nn.__i.__l = (nl); \
1064 1.1 mrg __asm__ ("ediv %d,%n,%0" \
1065 1.1 mrg : "=d" (__rq.__ll) : "dI" (__nn.__ll), "dI" (d)); \
1066 1.1 mrg (r) = __rq.__i.__l; (q) = __rq.__i.__h; \
1067 1.1 mrg } while (0)
1068 1.1 mrg #define count_leading_zeros(count, x) \
1069 1.1 mrg do { \
1070 1.1 mrg USItype __cbtmp; \
1071 1.1 mrg __asm__ ("scanbit %1,%0" : "=r" (__cbtmp) : "r" (x)); \
1072 1.1 mrg (count) = __cbtmp ^ 31; \
1073 1.1 mrg } while (0)
1074 1.1 mrg #define COUNT_LEADING_ZEROS_0 (-32) /* sic */
1075 1.1 mrg #if defined (__i960mx) /* what is the proper symbol to test??? */
1076 1.1 mrg #define rshift_rhlc(r,h,l,c) \
1077 1.1 mrg do { \
1078 1.1 mrg union {UDItype __ll; \
1079 1.1 mrg struct {USItype __l, __h;} __i; \
1080 1.1 mrg } __nn; \
1081 1.1 mrg __nn.__i.__h = (h); __nn.__i.__l = (l); \
1082 1.1 mrg __asm__ ("shre %2,%1,%0" : "=d" (r) : "dI" (__nn.__ll), "dI" (c)); \
1083 1.1 mrg }
1084 1.1 mrg #endif /* i960mx */
1085 1.1 mrg #endif /* i960 */
1086 1.1 mrg
1087 1.1 mrg #if (defined (__mc68000__) || defined (__mc68020__) || defined(mc68020) \
1088 1.1 mrg || defined (__m68k__) || defined (__mc5200__) || defined (__mc5206e__) \
1089 1.1 mrg || defined (__mc5307__)) && W_TYPE_SIZE == 32
1090 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1091 1.1 mrg __asm__ ("add%.l %5,%1\n\taddx%.l %3,%0" \
1092 1.1 mrg : "=d" (sh), "=&d" (sl) \
1093 1.1 mrg : "0" ((USItype)(ah)), "d" ((USItype)(bh)), \
1094 1.1 mrg "%1" ((USItype)(al)), "g" ((USItype)(bl)))
1095 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1096 1.1 mrg __asm__ ("sub%.l %5,%1\n\tsubx%.l %3,%0" \
1097 1.1 mrg : "=d" (sh), "=&d" (sl) \
1098 1.1 mrg : "0" ((USItype)(ah)), "d" ((USItype)(bh)), \
1099 1.1 mrg "1" ((USItype)(al)), "g" ((USItype)(bl)))
1100 1.1 mrg /* The '020, '030, '040 and CPU32 have 32x32->64 and 64/32->32q-32r. */
1101 1.1 mrg #if defined (__mc68020__) || defined(mc68020) \
1102 1.1 mrg || defined (__mc68030__) || defined (mc68030) \
1103 1.1 mrg || defined (__mc68040__) || defined (mc68040) \
1104 1.1 mrg || defined (__mcpu32__) || defined (mcpu32) \
1105 1.1 mrg || defined (__NeXT__)
1106 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1107 1.1 mrg __asm__ ("mulu%.l %3,%1:%0" \
1108 1.1 mrg : "=d" (w0), "=d" (w1) \
1109 1.1 mrg : "%0" ((USItype)(u)), "dmi" ((USItype)(v)))
1110 1.1 mrg #define UMUL_TIME 45
1111 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
1112 1.1 mrg __asm__ ("divu%.l %4,%1:%0" \
1113 1.1 mrg : "=d" (q), "=d" (r) \
1114 1.1 mrg : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "dmi" ((USItype)(d)))
1115 1.1 mrg #define UDIV_TIME 90
1116 1.1 mrg #define sdiv_qrnnd(q, r, n1, n0, d) \
1117 1.1 mrg __asm__ ("divs%.l %4,%1:%0" \
1118 1.1 mrg : "=d" (q), "=d" (r) \
1119 1.1 mrg : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "dmi" ((USItype)(d)))
1120 1.1 mrg #else /* for other 68k family members use 16x16->32 multiplication */
1121 1.1 mrg #define umul_ppmm(xh, xl, a, b) \
1122 1.1 mrg do { USItype __umul_tmp1, __umul_tmp2; \
1123 1.1 mrg __asm__ ("| Inlined umul_ppmm\n" \
1124 1.1 mrg " move%.l %5,%3\n" \
1125 1.1 mrg " move%.l %2,%0\n" \
1126 1.1 mrg " move%.w %3,%1\n" \
1127 1.1 mrg " swap %3\n" \
1128 1.1 mrg " swap %0\n" \
1129 1.1 mrg " mulu%.w %2,%1\n" \
1130 1.1 mrg " mulu%.w %3,%0\n" \
1131 1.1 mrg " mulu%.w %2,%3\n" \
1132 1.1 mrg " swap %2\n" \
1133 1.1 mrg " mulu%.w %5,%2\n" \
1134 1.1 mrg " add%.l %3,%2\n" \
1135 1.1 mrg " jcc 1f\n" \
1136 1.1 mrg " add%.l %#0x10000,%0\n" \
1137 1.1 mrg "1: move%.l %2,%3\n" \
1138 1.1 mrg " clr%.w %2\n" \
1139 1.1 mrg " swap %2\n" \
1140 1.1 mrg " swap %3\n" \
1141 1.1 mrg " clr%.w %3\n" \
1142 1.1 mrg " add%.l %3,%1\n" \
1143 1.1 mrg " addx%.l %2,%0\n" \
1144 1.1 mrg " | End inlined umul_ppmm" \
1145 1.1 mrg : "=&d" (xh), "=&d" (xl), \
1146 1.1 mrg "=d" (__umul_tmp1), "=&d" (__umul_tmp2) \
1147 1.1 mrg : "%2" ((USItype)(a)), "d" ((USItype)(b))); \
1148 1.1 mrg } while (0)
1149 1.1 mrg #define UMUL_TIME 100
1150 1.1 mrg #define UDIV_TIME 400
1151 1.1 mrg #endif /* not mc68020 */
1152 1.1 mrg /* The '020, '030, '040 and '060 have bitfield insns.
1153 1.1 mrg GCC 3.4 defines __mc68020__ when in CPU32 mode, check for __mcpu32__ to
1154 1.1 mrg exclude bfffo on that chip (bitfield insns not available). */
1155 1.1 mrg #if (defined (__mc68020__) || defined (mc68020) \
1156 1.1 mrg || defined (__mc68030__) || defined (mc68030) \
1157 1.1 mrg || defined (__mc68040__) || defined (mc68040) \
1158 1.1 mrg || defined (__mc68060__) || defined (mc68060) \
1159 1.1 mrg || defined (__NeXT__)) \
1160 1.1 mrg && ! defined (__mcpu32__)
1161 1.1 mrg #define count_leading_zeros(count, x) \
1162 1.1 mrg __asm__ ("bfffo %1{%b2:%b2},%0" \
1163 1.1 mrg : "=d" (count) \
1164 1.1 mrg : "od" ((USItype) (x)), "n" (0))
1165 1.1 mrg #define COUNT_LEADING_ZEROS_0 32
1166 1.1 mrg #endif
1167 1.1 mrg #endif /* mc68000 */
1168 1.1 mrg
1169 1.1 mrg #if defined (__m88000__) && W_TYPE_SIZE == 32
1170 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1171 1.1 mrg __asm__ ("addu.co %1,%r4,%r5\n\taddu.ci %0,%r2,%r3" \
1172 1.1 mrg : "=r" (sh), "=&r" (sl) \
1173 1.1 mrg : "rJ" (ah), "rJ" (bh), "%rJ" (al), "rJ" (bl))
1174 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1175 1.1 mrg __asm__ ("subu.co %1,%r4,%r5\n\tsubu.ci %0,%r2,%r3" \
1176 1.1 mrg : "=r" (sh), "=&r" (sl) \
1177 1.1 mrg : "rJ" (ah), "rJ" (bh), "rJ" (al), "rJ" (bl))
1178 1.1 mrg #define count_leading_zeros(count, x) \
1179 1.1 mrg do { \
1180 1.1 mrg USItype __cbtmp; \
1181 1.1 mrg __asm__ ("ff1 %0,%1" : "=r" (__cbtmp) : "r" (x)); \
1182 1.1 mrg (count) = __cbtmp ^ 31; \
1183 1.1 mrg } while (0)
1184 1.1 mrg #define COUNT_LEADING_ZEROS_0 63 /* sic */
1185 1.1 mrg #if defined (__m88110__)
1186 1.1 mrg #define umul_ppmm(wh, wl, u, v) \
1187 1.1 mrg do { \
1188 1.1 mrg union {UDItype __ll; \
1189 1.1 mrg struct {USItype __h, __l;} __i; \
1190 1.1 mrg } __x; \
1191 1.1 mrg __asm__ ("mulu.d %0,%1,%2" : "=r" (__x.__ll) : "r" (u), "r" (v)); \
1192 1.1 mrg (wh) = __x.__i.__h; \
1193 1.1 mrg (wl) = __x.__i.__l; \
1194 1.1 mrg } while (0)
1195 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
1196 1.1 mrg ({union {UDItype __ll; \
1197 1.1 mrg struct {USItype __h, __l;} __i; \
1198 1.1 mrg } __x, __q; \
1199 1.1 mrg __x.__i.__h = (n1); __x.__i.__l = (n0); \
1200 1.1 mrg __asm__ ("divu.d %0,%1,%2" \
1201 1.1 mrg : "=r" (__q.__ll) : "r" (__x.__ll), "r" (d)); \
1202 1.1 mrg (r) = (n0) - __q.__l * (d); (q) = __q.__l; })
1203 1.1 mrg #define UMUL_TIME 5
1204 1.1 mrg #define UDIV_TIME 25
1205 1.1 mrg #else
1206 1.1 mrg #define UMUL_TIME 17
1207 1.1 mrg #define UDIV_TIME 150
1208 1.1 mrg #endif /* __m88110__ */
1209 1.1 mrg #endif /* __m88000__ */
1210 1.1 mrg
1211 1.1 mrg #if defined (__mips) && W_TYPE_SIZE == 32
1212 1.1.1.1.8.1 tls #if __GMP_GNUC_PREREQ (4,4) || defined(__clang__)
1213 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1214 1.1 mrg do { \
1215 1.1 mrg UDItype __ll = (UDItype)(u) * (v); \
1216 1.1 mrg w1 = __ll >> 32; \
1217 1.1 mrg w0 = __ll; \
1218 1.1 mrg } while (0)
1219 1.1 mrg #endif
1220 1.1 mrg #if !defined (umul_ppmm) && __GMP_GNUC_PREREQ (2,7)
1221 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1222 1.1 mrg __asm__ ("multu %2,%3" : "=l" (w0), "=h" (w1) : "d" (u), "d" (v))
1223 1.1 mrg #endif
1224 1.1 mrg #if !defined (umul_ppmm)
1225 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1226 1.1 mrg __asm__ ("multu %2,%3\n\tmflo %0\n\tmfhi %1" \
1227 1.1 mrg : "=d" (w0), "=d" (w1) : "d" (u), "d" (v))
1228 1.1 mrg #endif
1229 1.1 mrg #define UMUL_TIME 10
1230 1.1 mrg #define UDIV_TIME 100
1231 1.1 mrg #endif /* __mips */
1232 1.1 mrg
1233 1.1 mrg #if (defined (__mips) && __mips >= 3) && W_TYPE_SIZE == 64
1234 1.1.1.1.8.1 tls #if __GMP_GNUC_PREREQ (4,4) || defined(__clang__)
1235 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1236 1.1 mrg do { \
1237 1.1 mrg typedef unsigned int __ll_UTItype __attribute__((mode(TI))); \
1238 1.1 mrg __ll_UTItype __ll = (__ll_UTItype)(u) * (v); \
1239 1.1 mrg w1 = __ll >> 64; \
1240 1.1 mrg w0 = __ll; \
1241 1.1 mrg } while (0)
1242 1.1 mrg #endif
1243 1.1 mrg #if !defined (umul_ppmm) && __GMP_GNUC_PREREQ (2,7)
1244 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1245 1.1 mrg __asm__ ("dmultu %2,%3" : "=l" (w0), "=h" (w1) : "d" (u), "d" (v))
1246 1.1 mrg #endif
1247 1.1 mrg #if !defined (umul_ppmm)
1248 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1249 1.1 mrg __asm__ ("dmultu %2,%3\n\tmflo %0\n\tmfhi %1" \
1250 1.1 mrg : "=d" (w0), "=d" (w1) : "d" (u), "d" (v))
1251 1.1 mrg #endif
1252 1.1 mrg #define UMUL_TIME 20
1253 1.1 mrg #define UDIV_TIME 140
1254 1.1 mrg #endif /* __mips */
1255 1.1 mrg
1256 1.1 mrg #if defined (__mmix__) && W_TYPE_SIZE == 64
1257 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1258 1.1 mrg __asm__ ("MULU %0,%2,%3" : "=r" (w0), "=z" (w1) : "r" (u), "r" (v))
1259 1.1 mrg #endif
1260 1.1 mrg
1261 1.1 mrg #if defined (__ns32000__) && W_TYPE_SIZE == 32
1262 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1263 1.1 mrg ({union {UDItype __ll; \
1264 1.1 mrg struct {USItype __l, __h;} __i; \
1265 1.1 mrg } __x; \
1266 1.1 mrg __asm__ ("meid %2,%0" \
1267 1.1 mrg : "=g" (__x.__ll) \
1268 1.1 mrg : "%0" ((USItype)(u)), "g" ((USItype)(v))); \
1269 1.1 mrg (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
1270 1.1 mrg #define __umulsidi3(u, v) \
1271 1.1 mrg ({UDItype __w; \
1272 1.1 mrg __asm__ ("meid %2,%0" \
1273 1.1 mrg : "=g" (__w) \
1274 1.1 mrg : "%0" ((USItype)(u)), "g" ((USItype)(v))); \
1275 1.1 mrg __w; })
1276 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
1277 1.1 mrg ({union {UDItype __ll; \
1278 1.1 mrg struct {USItype __l, __h;} __i; \
1279 1.1 mrg } __x; \
1280 1.1 mrg __x.__i.__h = (n1); __x.__i.__l = (n0); \
1281 1.1 mrg __asm__ ("deid %2,%0" \
1282 1.1 mrg : "=g" (__x.__ll) \
1283 1.1 mrg : "0" (__x.__ll), "g" ((USItype)(d))); \
1284 1.1 mrg (r) = __x.__i.__l; (q) = __x.__i.__h; })
1285 1.1 mrg #define count_trailing_zeros(count,x) \
1286 1.1 mrg do { \
1287 1.1 mrg __asm__ ("ffsd %2,%0" \
1288 1.1 mrg : "=r" (count) \
1289 1.1 mrg : "0" ((USItype) 0), "r" ((USItype) (x))); \
1290 1.1 mrg } while (0)
1291 1.1 mrg #endif /* __ns32000__ */
1292 1.1 mrg
1293 1.1 mrg /* In the past we had a block of various #defines tested
1294 1.1 mrg _ARCH_PPC - AIX
1295 1.1 mrg _ARCH_PWR - AIX
1296 1.1 mrg __powerpc__ - gcc
1297 1.1 mrg __POWERPC__ - BEOS
1298 1.1 mrg __ppc__ - Darwin
1299 1.1 mrg PPC - old gcc, GNU/Linux, SysV
1300 1.1 mrg The plain PPC test was not good for vxWorks, since PPC is defined on all
1301 1.1 mrg CPUs there (eg. m68k too), as a constant one is expected to compare
1302 1.1 mrg CPU_FAMILY against.
1303 1.1 mrg
1304 1.1 mrg At any rate, this was pretty unattractive and a bit fragile. The use of
1305 1.1 mrg HAVE_HOST_CPU_FAMILY is designed to cut through it all and be sure of
1306 1.1 mrg getting the desired effect.
1307 1.1 mrg
1308 1.1 mrg ENHANCE-ME: We should test _IBMR2 here when we add assembly support for
1309 1.1 mrg the system vendor compilers. (Is that vendor compilers with inline asm,
1310 1.1 mrg or what?) */
1311 1.1 mrg
1312 1.1 mrg #if (HAVE_HOST_CPU_FAMILY_power || HAVE_HOST_CPU_FAMILY_powerpc) \
1313 1.1 mrg && W_TYPE_SIZE == 32
1314 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1315 1.1 mrg do { \
1316 1.1 mrg if (__builtin_constant_p (bh) && (bh) == 0) \
1317 1.1.1.1.8.1 tls __asm__ ("add%I4c %1,%3,%4\n\taddze %0,%2" \
1318 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (ah), "%r" (al), "rI" (bl));\
1319 1.1 mrg else if (__builtin_constant_p (bh) && (bh) == ~(USItype) 0) \
1320 1.1.1.1.8.1 tls __asm__ ("add%I4c %1,%3,%4\n\taddme %0,%2" \
1321 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (ah), "%r" (al), "rI" (bl));\
1322 1.1 mrg else \
1323 1.1.1.1.8.1 tls __asm__ ("add%I5c %1,%4,%5\n\tadde %0,%2,%3" \
1324 1.1 mrg : "=r" (sh), "=&r" (sl) \
1325 1.1 mrg : "r" (ah), "r" (bh), "%r" (al), "rI" (bl)); \
1326 1.1 mrg } while (0)
1327 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1328 1.1 mrg do { \
1329 1.1 mrg if (__builtin_constant_p (ah) && (ah) == 0) \
1330 1.1.1.1.8.1 tls __asm__ ("subf%I3c %1,%4,%3\n\tsubfze %0,%2" \
1331 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl));\
1332 1.1 mrg else if (__builtin_constant_p (ah) && (ah) == ~(USItype) 0) \
1333 1.1.1.1.8.1 tls __asm__ ("subf%I3c %1,%4,%3\n\tsubfme %0,%2" \
1334 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl));\
1335 1.1 mrg else if (__builtin_constant_p (bh) && (bh) == 0) \
1336 1.1.1.1.8.1 tls __asm__ ("subf%I3c %1,%4,%3\n\taddme %0,%2" \
1337 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "r" (bl));\
1338 1.1 mrg else if (__builtin_constant_p (bh) && (bh) == ~(USItype) 0) \
1339 1.1.1.1.8.1 tls __asm__ ("subf%I3c %1,%4,%3\n\taddze %0,%2" \
1340 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "r" (bl));\
1341 1.1 mrg else \
1342 1.1.1.1.8.1 tls __asm__ ("subf%I4c %1,%5,%4\n\tsubfe %0,%3,%2" \
1343 1.1 mrg : "=r" (sh), "=&r" (sl) \
1344 1.1 mrg : "r" (ah), "r" (bh), "rI" (al), "r" (bl)); \
1345 1.1 mrg } while (0)
1346 1.1 mrg #define count_leading_zeros(count, x) \
1347 1.1.1.1.8.1 tls __asm__ ("cntlzw %0,%1" : "=r" (count) : "r" (x))
1348 1.1 mrg #define COUNT_LEADING_ZEROS_0 32
1349 1.1 mrg #if HAVE_HOST_CPU_FAMILY_powerpc
1350 1.1.1.1.8.1 tls #if __GMP_GNUC_PREREQ (4,4) || defined(__clang__)
1351 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1352 1.1 mrg do { \
1353 1.1 mrg UDItype __ll = (UDItype)(u) * (v); \
1354 1.1 mrg w1 = __ll >> 32; \
1355 1.1 mrg w0 = __ll; \
1356 1.1 mrg } while (0)
1357 1.1 mrg #endif
1358 1.1 mrg #if !defined (umul_ppmm)
1359 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \
1360 1.1 mrg do { \
1361 1.1 mrg USItype __m0 = (m0), __m1 = (m1); \
1362 1.1 mrg __asm__ ("mulhwu %0,%1,%2" : "=r" (ph) : "%r" (m0), "r" (m1)); \
1363 1.1 mrg (pl) = __m0 * __m1; \
1364 1.1 mrg } while (0)
1365 1.1 mrg #endif
1366 1.1 mrg #define UMUL_TIME 15
1367 1.1 mrg #define smul_ppmm(ph, pl, m0, m1) \
1368 1.1 mrg do { \
1369 1.1 mrg SItype __m0 = (m0), __m1 = (m1); \
1370 1.1 mrg __asm__ ("mulhw %0,%1,%2" : "=r" (ph) : "%r" (m0), "r" (m1)); \
1371 1.1 mrg (pl) = __m0 * __m1; \
1372 1.1 mrg } while (0)
1373 1.1 mrg #define SMUL_TIME 14
1374 1.1 mrg #define UDIV_TIME 120
1375 1.1 mrg #else
1376 1.1 mrg #define UMUL_TIME 8
1377 1.1 mrg #define smul_ppmm(xh, xl, m0, m1) \
1378 1.1 mrg __asm__ ("mul %0,%2,%3" : "=r" (xh), "=q" (xl) : "r" (m0), "r" (m1))
1379 1.1 mrg #define SMUL_TIME 4
1380 1.1 mrg #define sdiv_qrnnd(q, r, nh, nl, d) \
1381 1.1 mrg __asm__ ("div %0,%2,%4" : "=r" (q), "=q" (r) : "r" (nh), "1" (nl), "r" (d))
1382 1.1 mrg #define UDIV_TIME 100
1383 1.1 mrg #endif
1384 1.1 mrg #endif /* 32-bit POWER architecture variants. */
1385 1.1 mrg
1386 1.1 mrg /* We should test _IBMR2 here when we add assembly support for the system
1387 1.1 mrg vendor compilers. */
1388 1.1 mrg #if HAVE_HOST_CPU_FAMILY_powerpc && W_TYPE_SIZE == 64
1389 1.1 mrg #if !defined (_LONG_LONG_LIMB)
1390 1.1 mrg /* _LONG_LONG_LIMB is ABI=mode32 where adde operates on 32-bit values. So
1391 1.1 mrg use adde etc only when not _LONG_LONG_LIMB. */
1392 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1393 1.1 mrg do { \
1394 1.1 mrg if (__builtin_constant_p (bh) && (bh) == 0) \
1395 1.1.1.1.8.1 tls __asm__ ("add%I4c %1,%3,%4\n\taddze %0,%2" \
1396 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (ah), "%r" (al), "rI" (bl));\
1397 1.1 mrg else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0) \
1398 1.1.1.1.8.1 tls __asm__ ("add%I4c %1,%3,%4\n\taddme %0,%2" \
1399 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (ah), "%r" (al), "rI" (bl));\
1400 1.1 mrg else \
1401 1.1.1.1.8.1 tls __asm__ ("add%I5c %1,%4,%5\n\tadde %0,%2,%3" \
1402 1.1 mrg : "=r" (sh), "=&r" (sl) \
1403 1.1 mrg : "r" (ah), "r" (bh), "%r" (al), "rI" (bl)); \
1404 1.1 mrg } while (0)
1405 1.1 mrg /* We use "*rI" for the constant operand here, since with just "I", gcc barfs.
1406 1.1 mrg This might seem strange, but gcc folds away the dead code late. */
1407 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1408 1.1 mrg do { \
1409 1.1 mrg if (__builtin_constant_p (bl) && bl > -0x8000 && bl <= 0x8000) { \
1410 1.1 mrg if (__builtin_constant_p (ah) && (ah) == 0) \
1411 1.1.1.1.8.1 tls __asm__ ("addic %1,%3,%4\n\tsubfze %0,%2" \
1412 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "*rI" (-bl)); \
1413 1.1 mrg else if (__builtin_constant_p (ah) && (ah) == ~(UDItype) 0) \
1414 1.1.1.1.8.1 tls __asm__ ("addic %1,%3,%4\n\tsubfme %0,%2" \
1415 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "*rI" (-bl)); \
1416 1.1 mrg else if (__builtin_constant_p (bh) && (bh) == 0) \
1417 1.1.1.1.8.1 tls __asm__ ("addic %1,%3,%4\n\taddme %0,%2" \
1418 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "*rI" (-bl)); \
1419 1.1 mrg else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0) \
1420 1.1.1.1.8.1 tls __asm__ ("addic %1,%3,%4\n\taddze %0,%2" \
1421 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "*rI" (-bl)); \
1422 1.1 mrg else \
1423 1.1.1.1.8.1 tls __asm__ ("addic %1,%4,%5\n\tsubfe %0,%3,%2" \
1424 1.1 mrg : "=r" (sh), "=&r" (sl) \
1425 1.1 mrg : "r" (ah), "r" (bh), "rI" (al), "*rI" (-bl)); \
1426 1.1 mrg } else { \
1427 1.1 mrg if (__builtin_constant_p (ah) && (ah) == 0) \
1428 1.1.1.1.8.1 tls __asm__ ("subf%I3c %1,%4,%3\n\tsubfze %0,%2" \
1429 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl)); \
1430 1.1 mrg else if (__builtin_constant_p (ah) && (ah) == ~(UDItype) 0) \
1431 1.1.1.1.8.1 tls __asm__ ("subf%I3c %1,%4,%3\n\tsubfme %0,%2" \
1432 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl)); \
1433 1.1 mrg else if (__builtin_constant_p (bh) && (bh) == 0) \
1434 1.1.1.1.8.1 tls __asm__ ("subf%I3c %1,%4,%3\n\taddme %0,%2" \
1435 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "r" (bl)); \
1436 1.1 mrg else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0) \
1437 1.1.1.1.8.1 tls __asm__ ("subf%I3c %1,%4,%3\n\taddze %0,%2" \
1438 1.1 mrg : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "r" (bl)); \
1439 1.1 mrg else \
1440 1.1.1.1.8.1 tls __asm__ ("subf%I4c %1,%5,%4\n\tsubfe %0,%3,%2" \
1441 1.1 mrg : "=r" (sh), "=&r" (sl) \
1442 1.1 mrg : "r" (ah), "r" (bh), "rI" (al), "r" (bl)); \
1443 1.1 mrg } \
1444 1.1 mrg } while (0)
1445 1.1 mrg #endif /* ! _LONG_LONG_LIMB */
1446 1.1 mrg #define count_leading_zeros(count, x) \
1447 1.1 mrg __asm__ ("cntlzd %0,%1" : "=r" (count) : "r" (x))
1448 1.1 mrg #define COUNT_LEADING_ZEROS_0 64
1449 1.1.1.1.8.1 tls #if 0 && __GMP_GNUC_PREREQ (4,4) /* Disable, this results in libcalls! */
1450 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1451 1.1 mrg do { \
1452 1.1 mrg typedef unsigned int __ll_UTItype __attribute__((mode(TI))); \
1453 1.1 mrg __ll_UTItype __ll = (__ll_UTItype)(u) * (v); \
1454 1.1 mrg w1 = __ll >> 64; \
1455 1.1 mrg w0 = __ll; \
1456 1.1 mrg } while (0)
1457 1.1 mrg #endif
1458 1.1 mrg #if !defined (umul_ppmm)
1459 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \
1460 1.1 mrg do { \
1461 1.1 mrg UDItype __m0 = (m0), __m1 = (m1); \
1462 1.1 mrg __asm__ ("mulhdu %0,%1,%2" : "=r" (ph) : "%r" (m0), "r" (m1)); \
1463 1.1 mrg (pl) = __m0 * __m1; \
1464 1.1 mrg } while (0)
1465 1.1 mrg #endif
1466 1.1 mrg #define UMUL_TIME 15
1467 1.1 mrg #define smul_ppmm(ph, pl, m0, m1) \
1468 1.1 mrg do { \
1469 1.1 mrg DItype __m0 = (m0), __m1 = (m1); \
1470 1.1 mrg __asm__ ("mulhd %0,%1,%2" : "=r" (ph) : "%r" (m0), "r" (m1)); \
1471 1.1 mrg (pl) = __m0 * __m1; \
1472 1.1 mrg } while (0)
1473 1.1 mrg #define SMUL_TIME 14 /* ??? */
1474 1.1 mrg #define UDIV_TIME 120 /* ??? */
1475 1.1 mrg #endif /* 64-bit PowerPC. */
1476 1.1 mrg
1477 1.1 mrg #if defined (__pyr__) && W_TYPE_SIZE == 32
1478 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1479 1.1 mrg __asm__ ("addw %5,%1\n\taddwc %3,%0" \
1480 1.1 mrg : "=r" (sh), "=&r" (sl) \
1481 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \
1482 1.1 mrg "%1" ((USItype)(al)), "g" ((USItype)(bl)))
1483 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1484 1.1 mrg __asm__ ("subw %5,%1\n\tsubwb %3,%0" \
1485 1.1 mrg : "=r" (sh), "=&r" (sl) \
1486 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \
1487 1.1 mrg "1" ((USItype)(al)), "g" ((USItype)(bl)))
1488 1.1 mrg /* This insn works on Pyramids with AP, XP, or MI CPUs, but not with SP. */
1489 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1490 1.1 mrg ({union {UDItype __ll; \
1491 1.1 mrg struct {USItype __h, __l;} __i; \
1492 1.1 mrg } __x; \
1493 1.1 mrg __asm__ ("movw %1,%R0\n\tuemul %2,%0" \
1494 1.1 mrg : "=&r" (__x.__ll) \
1495 1.1 mrg : "g" ((USItype) (u)), "g" ((USItype)(v))); \
1496 1.1 mrg (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
1497 1.1 mrg #endif /* __pyr__ */
1498 1.1 mrg
1499 1.1 mrg #if defined (__ibm032__) /* RT/ROMP */ && W_TYPE_SIZE == 32
1500 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1501 1.1 mrg __asm__ ("a %1,%5\n\tae %0,%3" \
1502 1.1 mrg : "=r" (sh), "=&r" (sl) \
1503 1.1 mrg : "0" ((USItype)(ah)), "r" ((USItype)(bh)), \
1504 1.1 mrg "%1" ((USItype)(al)), "r" ((USItype)(bl)))
1505 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1506 1.1 mrg __asm__ ("s %1,%5\n\tse %0,%3" \
1507 1.1 mrg : "=r" (sh), "=&r" (sl) \
1508 1.1 mrg : "0" ((USItype)(ah)), "r" ((USItype)(bh)), \
1509 1.1 mrg "1" ((USItype)(al)), "r" ((USItype)(bl)))
1510 1.1 mrg #define smul_ppmm(ph, pl, m0, m1) \
1511 1.1 mrg __asm__ ( \
1512 1.1 mrg "s r2,r2\n" \
1513 1.1 mrg " mts r10,%2\n" \
1514 1.1 mrg " m r2,%3\n" \
1515 1.1 mrg " m r2,%3\n" \
1516 1.1 mrg " m r2,%3\n" \
1517 1.1 mrg " m r2,%3\n" \
1518 1.1 mrg " m r2,%3\n" \
1519 1.1 mrg " m r2,%3\n" \
1520 1.1 mrg " m r2,%3\n" \
1521 1.1 mrg " m r2,%3\n" \
1522 1.1 mrg " m r2,%3\n" \
1523 1.1 mrg " m r2,%3\n" \
1524 1.1 mrg " m r2,%3\n" \
1525 1.1 mrg " m r2,%3\n" \
1526 1.1 mrg " m r2,%3\n" \
1527 1.1 mrg " m r2,%3\n" \
1528 1.1 mrg " m r2,%3\n" \
1529 1.1 mrg " m r2,%3\n" \
1530 1.1 mrg " cas %0,r2,r0\n" \
1531 1.1 mrg " mfs r10,%1" \
1532 1.1 mrg : "=r" (ph), "=r" (pl) \
1533 1.1 mrg : "%r" ((USItype)(m0)), "r" ((USItype)(m1)) \
1534 1.1 mrg : "r2")
1535 1.1 mrg #define UMUL_TIME 20
1536 1.1 mrg #define UDIV_TIME 200
1537 1.1 mrg #define count_leading_zeros(count, x) \
1538 1.1 mrg do { \
1539 1.1 mrg if ((x) >= 0x10000) \
1540 1.1 mrg __asm__ ("clz %0,%1" \
1541 1.1 mrg : "=r" (count) : "r" ((USItype)(x) >> 16)); \
1542 1.1 mrg else \
1543 1.1 mrg { \
1544 1.1 mrg __asm__ ("clz %0,%1" \
1545 1.1 mrg : "=r" (count) : "r" ((USItype)(x))); \
1546 1.1 mrg (count) += 16; \
1547 1.1 mrg } \
1548 1.1 mrg } while (0)
1549 1.1 mrg #endif /* RT/ROMP */
1550 1.1 mrg
1551 1.1.1.1.8.1 tls #if (defined (__SH2__) || defined (__SH3__) || defined (__SH4__)) && W_TYPE_SIZE == 32
1552 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1553 1.1 mrg __asm__ ("dmulu.l %2,%3\n\tsts macl,%1\n\tsts mach,%0" \
1554 1.1 mrg : "=r" (w1), "=r" (w0) : "r" (u), "r" (v) : "macl", "mach")
1555 1.1 mrg #define UMUL_TIME 5
1556 1.1 mrg #endif
1557 1.1 mrg
1558 1.1 mrg #if defined (__sparc__) && W_TYPE_SIZE == 32
1559 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1560 1.1 mrg __asm__ ("addcc %r4,%5,%1\n\taddx %r2,%3,%0" \
1561 1.1 mrg : "=r" (sh), "=&r" (sl) \
1562 1.1 mrg : "rJ" (ah), "rI" (bh),"%rJ" (al), "rI" (bl) \
1563 1.1 mrg __CLOBBER_CC)
1564 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1565 1.1 mrg __asm__ ("subcc %r4,%5,%1\n\tsubx %r2,%3,%0" \
1566 1.1 mrg : "=r" (sh), "=&r" (sl) \
1567 1.1 mrg : "rJ" (ah), "rI" (bh), "rJ" (al), "rI" (bl) \
1568 1.1 mrg __CLOBBER_CC)
1569 1.1 mrg /* FIXME: When gcc -mcpu=v9 is used on solaris, gcc/config/sol2-sld-64.h
1570 1.1 mrg doesn't define anything to indicate that to us, it only sets __sparcv8. */
1571 1.1 mrg #if defined (__sparc_v9__) || defined (__sparcv9)
1572 1.1 mrg /* Perhaps we should use floating-point operations here? */
1573 1.1 mrg #if 0
1574 1.1 mrg /* Triggers a bug making mpz/tests/t-gcd.c fail.
1575 1.1 mrg Perhaps we simply need explicitly zero-extend the inputs? */
1576 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1577 1.1 mrg __asm__ ("mulx %2,%3,%%g1; srl %%g1,0,%1; srlx %%g1,32,%0" : \
1578 1.1 mrg "=r" (w1), "=r" (w0) : "r" (u), "r" (v) : "g1")
1579 1.1 mrg #else
1580 1.1 mrg /* Use v8 umul until above bug is fixed. */
1581 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1582 1.1 mrg __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v))
1583 1.1 mrg #endif
1584 1.1 mrg /* Use a plain v8 divide for v9. */
1585 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
1586 1.1 mrg do { \
1587 1.1 mrg USItype __q; \
1588 1.1 mrg __asm__ ("mov %1,%%y;nop;nop;nop;udiv %2,%3,%0" \
1589 1.1 mrg : "=r" (__q) : "r" (n1), "r" (n0), "r" (d)); \
1590 1.1 mrg (r) = (n0) - __q * (d); \
1591 1.1 mrg (q) = __q; \
1592 1.1 mrg } while (0)
1593 1.1 mrg #else
1594 1.1 mrg #if defined (__sparc_v8__) /* gcc normal */ \
1595 1.1 mrg || defined (__sparcv8) /* gcc solaris */ \
1596 1.1 mrg || HAVE_HOST_CPU_supersparc
1597 1.1 mrg /* Don't match immediate range because, 1) it is not often useful,
1598 1.1 mrg 2) the 'I' flag thinks of the range as a 13 bit signed interval,
1599 1.1 mrg while we want to match a 13 bit interval, sign extended to 32 bits,
1600 1.1 mrg but INTERPRETED AS UNSIGNED. */
1601 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1602 1.1 mrg __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v))
1603 1.1 mrg #define UMUL_TIME 5
1604 1.1 mrg
1605 1.1 mrg #if HAVE_HOST_CPU_supersparc
1606 1.1 mrg #define UDIV_TIME 60 /* SuperSPARC timing */
1607 1.1 mrg #else
1608 1.1 mrg /* Don't use this on SuperSPARC because its udiv only handles 53 bit
1609 1.1 mrg dividends and will trap to the kernel for the rest. */
1610 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
1611 1.1 mrg do { \
1612 1.1 mrg USItype __q; \
1613 1.1 mrg __asm__ ("mov %1,%%y;nop;nop;nop;udiv %2,%3,%0" \
1614 1.1 mrg : "=r" (__q) : "r" (n1), "r" (n0), "r" (d)); \
1615 1.1 mrg (r) = (n0) - __q * (d); \
1616 1.1 mrg (q) = __q; \
1617 1.1 mrg } while (0)
1618 1.1 mrg #define UDIV_TIME 25
1619 1.1 mrg #endif /* HAVE_HOST_CPU_supersparc */
1620 1.1 mrg
1621 1.1 mrg #else /* ! __sparc_v8__ */
1622 1.1 mrg #if defined (__sparclite__)
1623 1.1 mrg /* This has hardware multiply but not divide. It also has two additional
1624 1.1 mrg instructions scan (ffs from high bit) and divscc. */
1625 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1626 1.1 mrg __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v))
1627 1.1 mrg #define UMUL_TIME 5
1628 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
1629 1.1 mrg __asm__ ("! Inlined udiv_qrnnd\n" \
1630 1.1 mrg " wr %%g0,%2,%%y ! Not a delayed write for sparclite\n" \
1631 1.1 mrg " tst %%g0\n" \
1632 1.1 mrg " divscc %3,%4,%%g1\n" \
1633 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1634 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1635 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1636 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1637 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1638 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1639 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1640 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1641 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1642 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1643 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1644 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1645 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1646 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1647 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1648 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1649 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1650 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1651 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1652 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1653 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1654 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1655 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1656 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1657 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1658 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1659 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1660 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1661 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1662 1.1 mrg " divscc %%g1,%4,%%g1\n" \
1663 1.1 mrg " divscc %%g1,%4,%0\n" \
1664 1.1 mrg " rd %%y,%1\n" \
1665 1.1 mrg " bl,a 1f\n" \
1666 1.1 mrg " add %1,%4,%1\n" \
1667 1.1 mrg "1: ! End of inline udiv_qrnnd" \
1668 1.1 mrg : "=r" (q), "=r" (r) : "r" (n1), "r" (n0), "rI" (d) \
1669 1.1 mrg : "%g1" __AND_CLOBBER_CC)
1670 1.1 mrg #define UDIV_TIME 37
1671 1.1 mrg #define count_leading_zeros(count, x) \
1672 1.1 mrg __asm__ ("scan %1,1,%0" : "=r" (count) : "r" (x))
1673 1.1 mrg /* Early sparclites return 63 for an argument of 0, but they warn that future
1674 1.1 mrg implementations might change this. Therefore, leave COUNT_LEADING_ZEROS_0
1675 1.1 mrg undefined. */
1676 1.1 mrg #endif /* __sparclite__ */
1677 1.1 mrg #endif /* __sparc_v8__ */
1678 1.1 mrg #endif /* __sparc_v9__ */
1679 1.1 mrg /* Default to sparc v7 versions of umul_ppmm and udiv_qrnnd. */
1680 1.1 mrg #ifndef umul_ppmm
1681 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1682 1.1 mrg __asm__ ("! Inlined umul_ppmm\n" \
1683 1.1 mrg " wr %%g0,%2,%%y ! SPARC has 0-3 delay insn after a wr\n" \
1684 1.1 mrg " sra %3,31,%%g2 ! Don't move this insn\n" \
1685 1.1 mrg " and %2,%%g2,%%g2 ! Don't move this insn\n" \
1686 1.1 mrg " andcc %%g0,0,%%g1 ! Don't move this insn\n" \
1687 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1688 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1689 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1690 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1691 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1692 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1693 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1694 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1695 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1696 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1697 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1698 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1699 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1700 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1701 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1702 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1703 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1704 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1705 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1706 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1707 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1708 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1709 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1710 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1711 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1712 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1713 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1714 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1715 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1716 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1717 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1718 1.1 mrg " mulscc %%g1,%3,%%g1\n" \
1719 1.1 mrg " mulscc %%g1,0,%%g1\n" \
1720 1.1 mrg " add %%g1,%%g2,%0\n" \
1721 1.1 mrg " rd %%y,%1" \
1722 1.1 mrg : "=r" (w1), "=r" (w0) : "%rI" (u), "r" (v) \
1723 1.1 mrg : "%g1", "%g2" __AND_CLOBBER_CC)
1724 1.1 mrg #define UMUL_TIME 39 /* 39 instructions */
1725 1.1 mrg #endif
1726 1.1 mrg #ifndef udiv_qrnnd
1727 1.1 mrg #ifndef LONGLONG_STANDALONE
1728 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
1729 1.1 mrg do { UWtype __r; \
1730 1.1 mrg (q) = __MPN(udiv_qrnnd) (&__r, (n1), (n0), (d)); \
1731 1.1 mrg (r) = __r; \
1732 1.1 mrg } while (0)
1733 1.1.1.1.8.1 tls extern UWtype __MPN(udiv_qrnnd) (UWtype *, UWtype, UWtype, UWtype);
1734 1.1 mrg #ifndef UDIV_TIME
1735 1.1 mrg #define UDIV_TIME 140
1736 1.1 mrg #endif
1737 1.1 mrg #endif /* LONGLONG_STANDALONE */
1738 1.1 mrg #endif /* udiv_qrnnd */
1739 1.1 mrg #endif /* __sparc__ */
1740 1.1 mrg
1741 1.1 mrg #if defined (__sparc__) && W_TYPE_SIZE == 64
1742 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1743 1.1 mrg __asm__ ( \
1744 1.1 mrg "addcc %r4,%5,%1\n" \
1745 1.1 mrg " addccc %r6,%7,%%g0\n" \
1746 1.1 mrg " addc %r2,%3,%0" \
1747 1.1 mrg : "=r" (sh), "=&r" (sl) \
1748 1.1 mrg : "rJ" (ah), "rI" (bh), "%rJ" (al), "rI" (bl), \
1749 1.1 mrg "%rJ" ((al) >> 32), "rI" ((bl) >> 32) \
1750 1.1 mrg __CLOBBER_CC)
1751 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1752 1.1 mrg __asm__ ( \
1753 1.1 mrg "subcc %r4,%5,%1\n" \
1754 1.1 mrg " subccc %r6,%7,%%g0\n" \
1755 1.1 mrg " subc %r2,%3,%0" \
1756 1.1 mrg : "=r" (sh), "=&r" (sl) \
1757 1.1 mrg : "rJ" (ah), "rI" (bh), "rJ" (al), "rI" (bl), \
1758 1.1 mrg "rJ" ((al) >> 32), "rI" ((bl) >> 32) \
1759 1.1 mrg __CLOBBER_CC)
1760 1.1 mrg #endif
1761 1.1 mrg
1762 1.1.1.1.8.1 tls #if (defined (__vax) || defined (__vax__)) && W_TYPE_SIZE == 32
1763 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1764 1.1 mrg __asm__ ("addl2 %5,%1\n\tadwc %3,%0" \
1765 1.1 mrg : "=g" (sh), "=&g" (sl) \
1766 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \
1767 1.1 mrg "%1" ((USItype)(al)), "g" ((USItype)(bl)))
1768 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1769 1.1 mrg __asm__ ("subl2 %5,%1\n\tsbwc %3,%0" \
1770 1.1 mrg : "=g" (sh), "=&g" (sl) \
1771 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \
1772 1.1 mrg "1" ((USItype)(al)), "g" ((USItype)(bl)))
1773 1.1 mrg #define smul_ppmm(xh, xl, m0, m1) \
1774 1.1 mrg do { \
1775 1.1 mrg union {UDItype __ll; \
1776 1.1 mrg struct {USItype __l, __h;} __i; \
1777 1.1 mrg } __x; \
1778 1.1 mrg USItype __m0 = (m0), __m1 = (m1); \
1779 1.1 mrg __asm__ ("emul %1,%2,$0,%0" \
1780 1.1 mrg : "=g" (__x.__ll) : "g" (__m0), "g" (__m1)); \
1781 1.1 mrg (xh) = __x.__i.__h; (xl) = __x.__i.__l; \
1782 1.1 mrg } while (0)
1783 1.1 mrg #define sdiv_qrnnd(q, r, n1, n0, d) \
1784 1.1 mrg do { \
1785 1.1 mrg union {DItype __ll; \
1786 1.1 mrg struct {SItype __l, __h;} __i; \
1787 1.1 mrg } __x; \
1788 1.1 mrg __x.__i.__h = n1; __x.__i.__l = n0; \
1789 1.1 mrg __asm__ ("ediv %3,%2,%0,%1" \
1790 1.1 mrg : "=g" (q), "=g" (r) : "g" (__x.__ll), "g" (d)); \
1791 1.1 mrg } while (0)
1792 1.1 mrg #if 0
1793 1.1 mrg /* FIXME: This instruction appears to be unimplemented on some systems (vax
1794 1.1 mrg 8800 maybe). */
1795 1.1 mrg #define count_trailing_zeros(count,x) \
1796 1.1 mrg do { \
1797 1.1 mrg __asm__ ("ffs 0, 31, %1, %0" \
1798 1.1 mrg : "=g" (count) \
1799 1.1 mrg : "g" ((USItype) (x))); \
1800 1.1 mrg } while (0)
1801 1.1 mrg #endif
1802 1.1.1.1.8.1 tls #endif /* vax */
1803 1.1 mrg
1804 1.1 mrg #if defined (__z8000__) && W_TYPE_SIZE == 16
1805 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1806 1.1 mrg __asm__ ("add %H1,%H5\n\tadc %H0,%H3" \
1807 1.1 mrg : "=r" (sh), "=&r" (sl) \
1808 1.1 mrg : "0" ((unsigned int)(ah)), "r" ((unsigned int)(bh)), \
1809 1.1 mrg "%1" ((unsigned int)(al)), "rQR" ((unsigned int)(bl)))
1810 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1811 1.1 mrg __asm__ ("sub %H1,%H5\n\tsbc %H0,%H3" \
1812 1.1 mrg : "=r" (sh), "=&r" (sl) \
1813 1.1 mrg : "0" ((unsigned int)(ah)), "r" ((unsigned int)(bh)), \
1814 1.1 mrg "1" ((unsigned int)(al)), "rQR" ((unsigned int)(bl)))
1815 1.1 mrg #define umul_ppmm(xh, xl, m0, m1) \
1816 1.1 mrg do { \
1817 1.1 mrg union {long int __ll; \
1818 1.1 mrg struct {unsigned int __h, __l;} __i; \
1819 1.1 mrg } __x; \
1820 1.1 mrg unsigned int __m0 = (m0), __m1 = (m1); \
1821 1.1 mrg __asm__ ("mult %S0,%H3" \
1822 1.1 mrg : "=r" (__x.__i.__h), "=r" (__x.__i.__l) \
1823 1.1 mrg : "%1" (m0), "rQR" (m1)); \
1824 1.1 mrg (xh) = __x.__i.__h; (xl) = __x.__i.__l; \
1825 1.1 mrg (xh) += ((((signed int) __m0 >> 15) & __m1) \
1826 1.1 mrg + (((signed int) __m1 >> 15) & __m0)); \
1827 1.1 mrg } while (0)
1828 1.1 mrg #endif /* __z8000__ */
1829 1.1 mrg
1830 1.1 mrg #endif /* __GNUC__ */
1831 1.1 mrg
1832 1.1 mrg #endif /* NO_ASM */
1833 1.1 mrg
1834 1.1 mrg
1835 1.1.1.1.8.1 tls /* FIXME: "sidi" here is highly doubtful, should sometimes be "diti". */
1836 1.1 mrg #if !defined (umul_ppmm) && defined (__umulsidi3)
1837 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \
1838 1.1 mrg { \
1839 1.1 mrg UDWtype __ll = __umulsidi3 (m0, m1); \
1840 1.1 mrg ph = (UWtype) (__ll >> W_TYPE_SIZE); \
1841 1.1 mrg pl = (UWtype) __ll; \
1842 1.1 mrg }
1843 1.1 mrg #endif
1844 1.1 mrg
1845 1.1 mrg #if !defined (__umulsidi3)
1846 1.1 mrg #define __umulsidi3(u, v) \
1847 1.1 mrg ({UWtype __hi, __lo; \
1848 1.1 mrg umul_ppmm (__hi, __lo, u, v); \
1849 1.1 mrg ((UDWtype) __hi << W_TYPE_SIZE) | __lo; })
1850 1.1 mrg #endif
1851 1.1 mrg
1852 1.1 mrg
1853 1.1 mrg /* Use mpn_umul_ppmm or mpn_udiv_qrnnd functions, if they exist. The "_r"
1854 1.1 mrg forms have "reversed" arguments, meaning the pointer is last, which
1855 1.1 mrg sometimes allows better parameter passing, in particular on 64-bit
1856 1.1 mrg hppa. */
1857 1.1 mrg
1858 1.1 mrg #define mpn_umul_ppmm __MPN(umul_ppmm)
1859 1.1.1.1.8.1 tls extern UWtype mpn_umul_ppmm (UWtype *, UWtype, UWtype);
1860 1.1 mrg
1861 1.1 mrg #if ! defined (umul_ppmm) && HAVE_NATIVE_mpn_umul_ppmm \
1862 1.1 mrg && ! defined (LONGLONG_STANDALONE)
1863 1.1 mrg #define umul_ppmm(wh, wl, u, v) \
1864 1.1 mrg do { \
1865 1.1 mrg UWtype __umul_ppmm__p0; \
1866 1.1 mrg (wh) = mpn_umul_ppmm (&__umul_ppmm__p0, (UWtype) (u), (UWtype) (v)); \
1867 1.1 mrg (wl) = __umul_ppmm__p0; \
1868 1.1 mrg } while (0)
1869 1.1 mrg #endif
1870 1.1 mrg
1871 1.1 mrg #define mpn_umul_ppmm_r __MPN(umul_ppmm_r)
1872 1.1.1.1.8.1 tls extern UWtype mpn_umul_ppmm_r (UWtype, UWtype, UWtype *);
1873 1.1 mrg
1874 1.1 mrg #if ! defined (umul_ppmm) && HAVE_NATIVE_mpn_umul_ppmm_r \
1875 1.1 mrg && ! defined (LONGLONG_STANDALONE)
1876 1.1 mrg #define umul_ppmm(wh, wl, u, v) \
1877 1.1 mrg do { \
1878 1.1 mrg UWtype __umul_ppmm__p0; \
1879 1.1 mrg (wh) = mpn_umul_ppmm_r ((UWtype) (u), (UWtype) (v), &__umul_ppmm__p0); \
1880 1.1 mrg (wl) = __umul_ppmm__p0; \
1881 1.1 mrg } while (0)
1882 1.1 mrg #endif
1883 1.1 mrg
1884 1.1 mrg #define mpn_udiv_qrnnd __MPN(udiv_qrnnd)
1885 1.1.1.1.8.1 tls extern UWtype mpn_udiv_qrnnd (UWtype *, UWtype, UWtype, UWtype);
1886 1.1 mrg
1887 1.1 mrg #if ! defined (udiv_qrnnd) && HAVE_NATIVE_mpn_udiv_qrnnd \
1888 1.1 mrg && ! defined (LONGLONG_STANDALONE)
1889 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
1890 1.1 mrg do { \
1891 1.1 mrg UWtype __udiv_qrnnd__r; \
1892 1.1 mrg (q) = mpn_udiv_qrnnd (&__udiv_qrnnd__r, \
1893 1.1 mrg (UWtype) (n1), (UWtype) (n0), (UWtype) d); \
1894 1.1 mrg (r) = __udiv_qrnnd__r; \
1895 1.1 mrg } while (0)
1896 1.1 mrg #endif
1897 1.1 mrg
1898 1.1 mrg #define mpn_udiv_qrnnd_r __MPN(udiv_qrnnd_r)
1899 1.1.1.1.8.1 tls extern UWtype mpn_udiv_qrnnd_r (UWtype, UWtype, UWtype, UWtype *);
1900 1.1 mrg
1901 1.1 mrg #if ! defined (udiv_qrnnd) && HAVE_NATIVE_mpn_udiv_qrnnd_r \
1902 1.1 mrg && ! defined (LONGLONG_STANDALONE)
1903 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \
1904 1.1 mrg do { \
1905 1.1 mrg UWtype __udiv_qrnnd__r; \
1906 1.1 mrg (q) = mpn_udiv_qrnnd_r ((UWtype) (n1), (UWtype) (n0), (UWtype) d, \
1907 1.1 mrg &__udiv_qrnnd__r); \
1908 1.1 mrg (r) = __udiv_qrnnd__r; \
1909 1.1 mrg } while (0)
1910 1.1 mrg #endif
1911 1.1 mrg
1912 1.1 mrg
1913 1.1 mrg /* If this machine has no inline assembler, use C macros. */
1914 1.1 mrg
1915 1.1 mrg #if !defined (add_ssaaaa)
1916 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
1917 1.1 mrg do { \
1918 1.1 mrg UWtype __x; \
1919 1.1 mrg __x = (al) + (bl); \
1920 1.1 mrg (sh) = (ah) + (bh) + (__x < (al)); \
1921 1.1 mrg (sl) = __x; \
1922 1.1 mrg } while (0)
1923 1.1 mrg #endif
1924 1.1 mrg
1925 1.1 mrg #if !defined (sub_ddmmss)
1926 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
1927 1.1 mrg do { \
1928 1.1 mrg UWtype __x; \
1929 1.1 mrg __x = (al) - (bl); \
1930 1.1 mrg (sh) = (ah) - (bh) - ((al) < (bl)); \
1931 1.1 mrg (sl) = __x; \
1932 1.1 mrg } while (0)
1933 1.1 mrg #endif
1934 1.1 mrg
1935 1.1 mrg /* If we lack umul_ppmm but have smul_ppmm, define umul_ppmm in terms of
1936 1.1 mrg smul_ppmm. */
1937 1.1 mrg #if !defined (umul_ppmm) && defined (smul_ppmm)
1938 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1939 1.1 mrg do { \
1940 1.1 mrg UWtype __w1; \
1941 1.1 mrg UWtype __xm0 = (u), __xm1 = (v); \
1942 1.1 mrg smul_ppmm (__w1, w0, __xm0, __xm1); \
1943 1.1 mrg (w1) = __w1 + (-(__xm0 >> (W_TYPE_SIZE - 1)) & __xm1) \
1944 1.1 mrg + (-(__xm1 >> (W_TYPE_SIZE - 1)) & __xm0); \
1945 1.1 mrg } while (0)
1946 1.1 mrg #endif
1947 1.1 mrg
1948 1.1 mrg /* If we still don't have umul_ppmm, define it using plain C.
1949 1.1 mrg
1950 1.1 mrg For reference, when this code is used for squaring (ie. u and v identical
1951 1.1 mrg expressions), gcc recognises __x1 and __x2 are the same and generates 3
1952 1.1 mrg multiplies, not 4. The subsequent additions could be optimized a bit,
1953 1.1 mrg but the only place GMP currently uses such a square is mpn_sqr_basecase,
1954 1.1 mrg and chips obliged to use this generic C umul will have plenty of worse
1955 1.1 mrg performance problems than a couple of extra instructions on the diagonal
1956 1.1 mrg of sqr_basecase. */
1957 1.1 mrg
1958 1.1 mrg #if !defined (umul_ppmm)
1959 1.1 mrg #define umul_ppmm(w1, w0, u, v) \
1960 1.1 mrg do { \
1961 1.1 mrg UWtype __x0, __x1, __x2, __x3; \
1962 1.1 mrg UHWtype __ul, __vl, __uh, __vh; \
1963 1.1 mrg UWtype __u = (u), __v = (v); \
1964 1.1 mrg \
1965 1.1 mrg __ul = __ll_lowpart (__u); \
1966 1.1 mrg __uh = __ll_highpart (__u); \
1967 1.1 mrg __vl = __ll_lowpart (__v); \
1968 1.1 mrg __vh = __ll_highpart (__v); \
1969 1.1 mrg \
1970 1.1 mrg __x0 = (UWtype) __ul * __vl; \
1971 1.1 mrg __x1 = (UWtype) __ul * __vh; \
1972 1.1 mrg __x2 = (UWtype) __uh * __vl; \
1973 1.1 mrg __x3 = (UWtype) __uh * __vh; \
1974 1.1 mrg \
1975 1.1 mrg __x1 += __ll_highpart (__x0);/* this can't give carry */ \
1976 1.1 mrg __x1 += __x2; /* but this indeed can */ \
1977 1.1 mrg if (__x1 < __x2) /* did we get it? */ \
1978 1.1 mrg __x3 += __ll_B; /* yes, add it in the proper pos. */ \
1979 1.1 mrg \
1980 1.1 mrg (w1) = __x3 + __ll_highpart (__x1); \
1981 1.1 mrg (w0) = (__x1 << W_TYPE_SIZE/2) + __ll_lowpart (__x0); \
1982 1.1 mrg } while (0)
1983 1.1 mrg #endif
1984 1.1 mrg
1985 1.1 mrg /* If we don't have smul_ppmm, define it using umul_ppmm (which surely will
1986 1.1 mrg exist in one form or another. */
1987 1.1 mrg #if !defined (smul_ppmm)
1988 1.1 mrg #define smul_ppmm(w1, w0, u, v) \
1989 1.1 mrg do { \
1990 1.1 mrg UWtype __w1; \
1991 1.1 mrg UWtype __xm0 = (u), __xm1 = (v); \
1992 1.1 mrg umul_ppmm (__w1, w0, __xm0, __xm1); \
1993 1.1 mrg (w1) = __w1 - (-(__xm0 >> (W_TYPE_SIZE - 1)) & __xm1) \
1994 1.1 mrg - (-(__xm1 >> (W_TYPE_SIZE - 1)) & __xm0); \
1995 1.1 mrg } while (0)
1996 1.1 mrg #endif
1997 1.1 mrg
1998 1.1 mrg /* Define this unconditionally, so it can be used for debugging. */
1999 1.1 mrg #define __udiv_qrnnd_c(q, r, n1, n0, d) \
2000 1.1 mrg do { \
2001 1.1 mrg UWtype __d1, __d0, __q1, __q0, __r1, __r0, __m; \
2002 1.1 mrg \
2003 1.1 mrg ASSERT ((d) != 0); \
2004 1.1 mrg ASSERT ((n1) < (d)); \
2005 1.1 mrg \
2006 1.1 mrg __d1 = __ll_highpart (d); \
2007 1.1 mrg __d0 = __ll_lowpart (d); \
2008 1.1 mrg \
2009 1.1 mrg __q1 = (n1) / __d1; \
2010 1.1 mrg __r1 = (n1) - __q1 * __d1; \
2011 1.1 mrg __m = __q1 * __d0; \
2012 1.1 mrg __r1 = __r1 * __ll_B | __ll_highpart (n0); \
2013 1.1 mrg if (__r1 < __m) \
2014 1.1 mrg { \
2015 1.1 mrg __q1--, __r1 += (d); \
2016 1.1 mrg if (__r1 >= (d)) /* i.e. we didn't get carry when adding to __r1 */\
2017 1.1 mrg if (__r1 < __m) \
2018 1.1 mrg __q1--, __r1 += (d); \
2019 1.1 mrg } \
2020 1.1 mrg __r1 -= __m; \
2021 1.1 mrg \
2022 1.1 mrg __q0 = __r1 / __d1; \
2023 1.1 mrg __r0 = __r1 - __q0 * __d1; \
2024 1.1 mrg __m = __q0 * __d0; \
2025 1.1 mrg __r0 = __r0 * __ll_B | __ll_lowpart (n0); \
2026 1.1 mrg if (__r0 < __m) \
2027 1.1 mrg { \
2028 1.1 mrg __q0--, __r0 += (d); \
2029 1.1 mrg if (__r0 >= (d)) \
2030 1.1 mrg if (__r0 < __m) \
2031 1.1 mrg __q0--, __r0 += (d); \
2032 1.1 mrg } \
2033 1.1 mrg __r0 -= __m; \
2034 1.1 mrg \
2035 1.1 mrg (q) = __q1 * __ll_B | __q0; \
2036 1.1 mrg (r) = __r0; \
2037 1.1 mrg } while (0)
2038 1.1 mrg
2039 1.1 mrg /* If the processor has no udiv_qrnnd but sdiv_qrnnd, go through
2040 1.1 mrg __udiv_w_sdiv (defined in libgcc or elsewhere). */
2041 1.1 mrg #if !defined (udiv_qrnnd) && defined (sdiv_qrnnd)
2042 1.1 mrg #define udiv_qrnnd(q, r, nh, nl, d) \
2043 1.1 mrg do { \
2044 1.1 mrg UWtype __r; \
2045 1.1 mrg (q) = __MPN(udiv_w_sdiv) (&__r, nh, nl, d); \
2046 1.1 mrg (r) = __r; \
2047 1.1 mrg } while (0)
2048 1.1.1.1.8.1 tls __GMP_DECLSPEC UWtype __MPN(udiv_w_sdiv) (UWtype *, UWtype, UWtype, UWtype);
2049 1.1 mrg #endif
2050 1.1 mrg
2051 1.1 mrg /* If udiv_qrnnd was not defined for this processor, use __udiv_qrnnd_c. */
2052 1.1 mrg #if !defined (udiv_qrnnd)
2053 1.1 mrg #define UDIV_NEEDS_NORMALIZATION 1
2054 1.1 mrg #define udiv_qrnnd __udiv_qrnnd_c
2055 1.1 mrg #endif
2056 1.1 mrg
2057 1.1 mrg #if !defined (count_leading_zeros)
2058 1.1 mrg #define count_leading_zeros(count, x) \
2059 1.1 mrg do { \
2060 1.1 mrg UWtype __xr = (x); \
2061 1.1 mrg UWtype __a; \
2062 1.1 mrg \
2063 1.1 mrg if (W_TYPE_SIZE == 32) \
2064 1.1 mrg { \
2065 1.1 mrg __a = __xr < ((UWtype) 1 << 2*__BITS4) \
2066 1.1 mrg ? (__xr < ((UWtype) 1 << __BITS4) ? 1 : __BITS4 + 1) \
2067 1.1 mrg : (__xr < ((UWtype) 1 << 3*__BITS4) ? 2*__BITS4 + 1 \
2068 1.1 mrg : 3*__BITS4 + 1); \
2069 1.1 mrg } \
2070 1.1 mrg else \
2071 1.1 mrg { \
2072 1.1 mrg for (__a = W_TYPE_SIZE - 8; __a > 0; __a -= 8) \
2073 1.1 mrg if (((__xr >> __a) & 0xff) != 0) \
2074 1.1 mrg break; \
2075 1.1 mrg ++__a; \
2076 1.1 mrg } \
2077 1.1 mrg \
2078 1.1 mrg (count) = W_TYPE_SIZE + 1 - __a - __clz_tab[__xr >> __a]; \
2079 1.1 mrg } while (0)
2080 1.1 mrg /* This version gives a well-defined value for zero. */
2081 1.1 mrg #define COUNT_LEADING_ZEROS_0 (W_TYPE_SIZE - 1)
2082 1.1 mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
2083 1.1.1.1.8.1 tls #define COUNT_LEADING_ZEROS_SLOW
2084 1.1 mrg #endif
2085 1.1 mrg
2086 1.1 mrg /* clz_tab needed by mpn/x86/pentium/mod_1.asm in a fat binary */
2087 1.1 mrg #if HAVE_HOST_CPU_FAMILY_x86 && WANT_FAT_BINARY
2088 1.1 mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
2089 1.1 mrg #endif
2090 1.1 mrg
2091 1.1 mrg #ifdef COUNT_LEADING_ZEROS_NEED_CLZ_TAB
2092 1.1.1.1.8.1 tls extern const unsigned char __GMP_DECLSPEC __clz_tab[129];
2093 1.1 mrg #endif
2094 1.1 mrg
2095 1.1 mrg #if !defined (count_trailing_zeros)
2096 1.1.1.1.8.1 tls #if !defined (COUNT_LEADING_ZEROS_SLOW)
2097 1.1.1.1.8.1 tls /* Define count_trailing_zeros using an asm count_leading_zeros. */
2098 1.1.1.1.8.1 tls #define count_trailing_zeros(count, x) \
2099 1.1 mrg do { \
2100 1.1 mrg UWtype __ctz_x = (x); \
2101 1.1 mrg UWtype __ctz_c; \
2102 1.1 mrg ASSERT (__ctz_x != 0); \
2103 1.1 mrg count_leading_zeros (__ctz_c, __ctz_x & -__ctz_x); \
2104 1.1 mrg (count) = W_TYPE_SIZE - 1 - __ctz_c; \
2105 1.1 mrg } while (0)
2106 1.1.1.1.8.1 tls #else
2107 1.1.1.1.8.1 tls /* Define count_trailing_zeros in plain C, assuming small counts are common.
2108 1.1.1.1.8.1 tls We use clz_tab without ado, since the C count_leading_zeros above will have
2109 1.1.1.1.8.1 tls pulled it in. */
2110 1.1.1.1.8.1 tls #define count_trailing_zeros(count, x) \
2111 1.1.1.1.8.1 tls do { \
2112 1.1.1.1.8.1 tls UWtype __ctz_x = (x); \
2113 1.1.1.1.8.1 tls int __ctz_c; \
2114 1.1.1.1.8.1 tls \
2115 1.1.1.1.8.1 tls if (LIKELY ((__ctz_x & 0xff) != 0)) \
2116 1.1.1.1.8.1 tls (count) = __clz_tab[__ctz_x & -__ctz_x] - 2; \
2117 1.1.1.1.8.1 tls else \
2118 1.1.1.1.8.1 tls { \
2119 1.1.1.1.8.1 tls for (__ctz_c = 8 - 2; __ctz_c < W_TYPE_SIZE - 2; __ctz_c += 8) \
2120 1.1.1.1.8.1 tls { \
2121 1.1.1.1.8.1 tls __ctz_x >>= 8; \
2122 1.1.1.1.8.1 tls if (LIKELY ((__ctz_x & 0xff) != 0)) \
2123 1.1.1.1.8.1 tls break; \
2124 1.1.1.1.8.1 tls } \
2125 1.1.1.1.8.1 tls \
2126 1.1.1.1.8.1 tls (count) = __ctz_c + __clz_tab[__ctz_x & -__ctz_x]; \
2127 1.1.1.1.8.1 tls } \
2128 1.1.1.1.8.1 tls } while (0)
2129 1.1.1.1.8.1 tls #endif
2130 1.1 mrg #endif
2131 1.1 mrg
2132 1.1 mrg #ifndef UDIV_NEEDS_NORMALIZATION
2133 1.1 mrg #define UDIV_NEEDS_NORMALIZATION 0
2134 1.1 mrg #endif
2135 1.1 mrg
2136 1.1 mrg /* Whether udiv_qrnnd is actually implemented with udiv_qrnnd_preinv, and
2137 1.1 mrg that hence the latter should always be used. */
2138 1.1 mrg #ifndef UDIV_PREINV_ALWAYS
2139 1.1 mrg #define UDIV_PREINV_ALWAYS 0
2140 1.1 mrg #endif
2141 1.1 mrg
2142 1.1 mrg /* Give defaults for UMUL_TIME and UDIV_TIME. */
2143 1.1 mrg #ifndef UMUL_TIME
2144 1.1 mrg #define UMUL_TIME 1
2145 1.1 mrg #endif
2146 1.1 mrg
2147 1.1 mrg #ifndef UDIV_TIME
2148 1.1 mrg #define UDIV_TIME UMUL_TIME
2149 1.1 mrg #endif
2150