1 1.1 mrg /* longlong.h -- definitions for mixed size 32/64 bit arithmetic. 2 1.1 mrg 3 1.4 mrg Copyright 1991-1994, 1996, 1997, 1999-2005, 2007-2009, 2011-2020 Free Software 4 1.3 mrg Foundation, Inc. 5 1.1 mrg 6 1.3 mrg This file is part of the GNU MP Library. 7 1.1 mrg 8 1.3 mrg The GNU MP Library is free software; you can redistribute it and/or modify 9 1.3 mrg it under the terms of either: 10 1.3 mrg 11 1.3 mrg * the GNU Lesser General Public License as published by the Free 12 1.3 mrg Software Foundation; either version 3 of the License, or (at your 13 1.3 mrg option) any later version. 14 1.3 mrg 15 1.3 mrg or 16 1.3 mrg 17 1.3 mrg * the GNU General Public License as published by the Free Software 18 1.3 mrg Foundation; either version 2 of the License, or (at your option) any 19 1.3 mrg later version. 20 1.3 mrg 21 1.3 mrg or both in parallel, as here. 22 1.3 mrg 23 1.3 mrg The GNU MP Library is distributed in the hope that it will be useful, but 24 1.3 mrg WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY 25 1.3 mrg or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 26 1.3 mrg for more details. 27 1.3 mrg 28 1.3 mrg You should have received copies of the GNU General Public License and the 29 1.3 mrg GNU Lesser General Public License along with the GNU MP Library. If not, 30 1.3 mrg see https://www.gnu.org/licenses/. */ 31 1.1 mrg 32 1.1 mrg /* You have to define the following before including this file: 33 1.1 mrg 34 1.1 mrg UWtype -- An unsigned type, default type for operations (typically a "word") 35 1.1 mrg UHWtype -- An unsigned type, at least half the size of UWtype 36 1.1 mrg UDWtype -- An unsigned type, at least twice as large a UWtype 37 1.1 mrg W_TYPE_SIZE -- size in bits of UWtype 38 1.1 mrg 39 1.1 mrg SItype, USItype -- Signed and unsigned 32 bit types 40 1.1 mrg DItype, UDItype -- Signed and unsigned 64 bit types 41 1.1 mrg 42 1.1 mrg On a 32 bit machine UWtype should typically be USItype; 43 1.1 mrg on a 64 bit machine, UWtype should typically be UDItype. 44 1.1 mrg 45 1.1 mrg Optionally, define: 46 1.1 mrg 47 1.1 mrg LONGLONG_STANDALONE -- Avoid code that needs machine-dependent support files 48 1.1 mrg NO_ASM -- Disable inline asm 49 1.1 mrg 50 1.1 mrg 51 1.1 mrg CAUTION! Using this version of longlong.h outside of GMP is not safe. You 52 1.1 mrg need to include gmp.h and gmp-impl.h, or certain things might not work as 53 1.1 mrg expected. 54 1.1 mrg */ 55 1.1 mrg 56 1.1 mrg #define __BITS4 (W_TYPE_SIZE / 4) 57 1.1 mrg #define __ll_B ((UWtype) 1 << (W_TYPE_SIZE / 2)) 58 1.1 mrg #define __ll_lowpart(t) ((UWtype) (t) & (__ll_B - 1)) 59 1.1 mrg #define __ll_highpart(t) ((UWtype) (t) >> (W_TYPE_SIZE / 2)) 60 1.1 mrg 61 1.1 mrg /* This is used to make sure no undesirable sharing between different libraries 62 1.1 mrg that use this file takes place. */ 63 1.1 mrg #ifndef __MPN 64 1.1 mrg #define __MPN(x) __##x 65 1.1 mrg #endif 66 1.1 mrg 67 1.1 mrg /* Define auxiliary asm macros. 68 1.1 mrg 69 1.1 mrg 1) umul_ppmm(high_prod, low_prod, multiplier, multiplicand) multiplies two 70 1.1 mrg UWtype integers MULTIPLIER and MULTIPLICAND, and generates a two UWtype 71 1.1 mrg word product in HIGH_PROD and LOW_PROD. 72 1.1 mrg 73 1.1 mrg 2) __umulsidi3(a,b) multiplies two UWtype integers A and B, and returns a 74 1.1 mrg UDWtype product. This is just a variant of umul_ppmm. 75 1.1 mrg 76 1.1 mrg 3) udiv_qrnnd(quotient, remainder, high_numerator, low_numerator, 77 1.1 mrg denominator) divides a UDWtype, composed by the UWtype integers 78 1.1 mrg HIGH_NUMERATOR and LOW_NUMERATOR, by DENOMINATOR and places the quotient 79 1.1 mrg in QUOTIENT and the remainder in REMAINDER. HIGH_NUMERATOR must be less 80 1.1 mrg than DENOMINATOR for correct operation. If, in addition, the most 81 1.1 mrg significant bit of DENOMINATOR must be 1, then the pre-processor symbol 82 1.1 mrg UDIV_NEEDS_NORMALIZATION is defined to 1. 83 1.1 mrg 84 1.1 mrg 4) sdiv_qrnnd(quotient, remainder, high_numerator, low_numerator, 85 1.1 mrg denominator). Like udiv_qrnnd but the numbers are signed. The quotient 86 1.1 mrg is rounded towards 0. 87 1.1 mrg 88 1.1 mrg 5) count_leading_zeros(count, x) counts the number of zero-bits from the 89 1.1 mrg msb to the first non-zero bit in the UWtype X. This is the number of 90 1.1 mrg steps X needs to be shifted left to set the msb. Undefined for X == 0, 91 1.1 mrg unless the symbol COUNT_LEADING_ZEROS_0 is defined to some value. 92 1.1 mrg 93 1.1 mrg 6) count_trailing_zeros(count, x) like count_leading_zeros, but counts 94 1.1 mrg from the least significant end. 95 1.1 mrg 96 1.1 mrg 7) add_ssaaaa(high_sum, low_sum, high_addend_1, low_addend_1, 97 1.1 mrg high_addend_2, low_addend_2) adds two UWtype integers, composed by 98 1.1 mrg HIGH_ADDEND_1 and LOW_ADDEND_1, and HIGH_ADDEND_2 and LOW_ADDEND_2 99 1.1 mrg respectively. The result is placed in HIGH_SUM and LOW_SUM. Overflow 100 1.1 mrg (i.e. carry out) is not stored anywhere, and is lost. 101 1.1 mrg 102 1.1 mrg 8) sub_ddmmss(high_difference, low_difference, high_minuend, low_minuend, 103 1.1 mrg high_subtrahend, low_subtrahend) subtracts two two-word UWtype integers, 104 1.1 mrg composed by HIGH_MINUEND_1 and LOW_MINUEND_1, and HIGH_SUBTRAHEND_2 and 105 1.1 mrg LOW_SUBTRAHEND_2 respectively. The result is placed in HIGH_DIFFERENCE 106 1.1 mrg and LOW_DIFFERENCE. Overflow (i.e. carry out) is not stored anywhere, 107 1.1 mrg and is lost. 108 1.1 mrg 109 1.1 mrg If any of these macros are left undefined for a particular CPU, 110 1.1 mrg C macros are used. 111 1.1 mrg 112 1.1 mrg 113 1.1 mrg Notes: 114 1.1 mrg 115 1.1 mrg For add_ssaaaa the two high and two low addends can both commute, but 116 1.1 mrg unfortunately gcc only supports one "%" commutative in each asm block. 117 1.1 mrg This has always been so but is only documented in recent versions 118 1.1 mrg (eg. pre-release 3.3). Having two or more "%"s can cause an internal 119 1.1 mrg compiler error in certain rare circumstances. 120 1.1 mrg 121 1.1 mrg Apparently it was only the last "%" that was ever actually respected, so 122 1.1 mrg the code has been updated to leave just that. Clearly there's a free 123 1.1 mrg choice whether high or low should get it, if there's a reason to favour 124 1.1 mrg one over the other. Also obviously when the constraints on the two 125 1.1 mrg operands are identical there's no benefit to the reloader in any "%" at 126 1.1 mrg all. 127 1.1 mrg 128 1.1 mrg */ 129 1.1 mrg 130 1.1 mrg /* The CPUs come in alphabetical order below. 131 1.1 mrg 132 1.1 mrg Please add support for more CPUs here, or improve the current support 133 1.1 mrg for the CPUs below! */ 134 1.1 mrg 135 1.1 mrg 136 1.1 mrg /* count_leading_zeros_gcc_clz is count_leading_zeros implemented with gcc 137 1.1 mrg 3.4 __builtin_clzl or __builtin_clzll, according to our limb size. 138 1.1 mrg Similarly count_trailing_zeros_gcc_ctz using __builtin_ctzl or 139 1.1 mrg __builtin_ctzll. 140 1.1 mrg 141 1.1 mrg These builtins are only used when we check what code comes out, on some 142 1.1 mrg chips they're merely libgcc calls, where we will instead want an inline 143 1.1 mrg in that case (either asm or generic C). 144 1.1 mrg 145 1.1 mrg These builtins are better than an asm block of the same insn, since an 146 1.1 mrg asm block doesn't give gcc any information about scheduling or resource 147 1.1 mrg usage. We keep an asm block for use on prior versions of gcc though. 148 1.1 mrg 149 1.1 mrg For reference, __builtin_ffs existed in gcc prior to __builtin_clz, but 150 1.1 mrg it's not used (for count_leading_zeros) because it generally gives extra 151 1.1 mrg code to ensure the result is 0 when the input is 0, which we don't need 152 1.1 mrg or want. */ 153 1.1 mrg 154 1.1 mrg #ifdef _LONG_LONG_LIMB 155 1.3 mrg #define count_leading_zeros_gcc_clz(count,x) \ 156 1.3 mrg do { \ 157 1.3 mrg ASSERT ((x) != 0); \ 158 1.3 mrg (count) = __builtin_clzll (x); \ 159 1.1 mrg } while (0) 160 1.1 mrg #else 161 1.3 mrg #define count_leading_zeros_gcc_clz(count,x) \ 162 1.3 mrg do { \ 163 1.3 mrg ASSERT ((x) != 0); \ 164 1.3 mrg (count) = __builtin_clzl (x); \ 165 1.1 mrg } while (0) 166 1.1 mrg #endif 167 1.1 mrg 168 1.1 mrg #ifdef _LONG_LONG_LIMB 169 1.3 mrg #define count_trailing_zeros_gcc_ctz(count,x) \ 170 1.3 mrg do { \ 171 1.3 mrg ASSERT ((x) != 0); \ 172 1.3 mrg (count) = __builtin_ctzll (x); \ 173 1.1 mrg } while (0) 174 1.1 mrg #else 175 1.3 mrg #define count_trailing_zeros_gcc_ctz(count,x) \ 176 1.3 mrg do { \ 177 1.3 mrg ASSERT ((x) != 0); \ 178 1.3 mrg (count) = __builtin_ctzl (x); \ 179 1.1 mrg } while (0) 180 1.1 mrg #endif 181 1.1 mrg 182 1.1 mrg 183 1.1 mrg /* FIXME: The macros using external routines like __MPN(count_leading_zeros) 184 1.1 mrg don't need to be under !NO_ASM */ 185 1.1 mrg #if ! defined (NO_ASM) 186 1.1 mrg 187 1.1 mrg #if defined (__alpha) && W_TYPE_SIZE == 64 188 1.1 mrg /* Most alpha-based machines, except Cray systems. */ 189 1.1 mrg #if defined (__GNUC__) 190 1.1 mrg #if __GMP_GNUC_PREREQ (3,3) 191 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \ 192 1.1 mrg do { \ 193 1.1 mrg UDItype __m0 = (m0), __m1 = (m1); \ 194 1.1 mrg (ph) = __builtin_alpha_umulh (__m0, __m1); \ 195 1.1 mrg (pl) = __m0 * __m1; \ 196 1.1 mrg } while (0) 197 1.1 mrg #else 198 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \ 199 1.1 mrg do { \ 200 1.1 mrg UDItype __m0 = (m0), __m1 = (m1); \ 201 1.1 mrg __asm__ ("umulh %r1,%2,%0" \ 202 1.1 mrg : "=r" (ph) \ 203 1.3 mrg : "%rJ" (__m0), "rI" (__m1)); \ 204 1.1 mrg (pl) = __m0 * __m1; \ 205 1.1 mrg } while (0) 206 1.1 mrg #endif 207 1.1 mrg #else /* ! __GNUC__ */ 208 1.1 mrg #include <machine/builtins.h> 209 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \ 210 1.1 mrg do { \ 211 1.1 mrg UDItype __m0 = (m0), __m1 = (m1); \ 212 1.3 mrg (ph) = __UMULH (__m0, __m1); \ 213 1.1 mrg (pl) = __m0 * __m1; \ 214 1.1 mrg } while (0) 215 1.1 mrg #endif 216 1.1 mrg #ifndef LONGLONG_STANDALONE 217 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 218 1.1 mrg do { UWtype __di; \ 219 1.1 mrg __di = __MPN(invert_limb) (d); \ 220 1.1 mrg udiv_qrnnd_preinv (q, r, n1, n0, d, __di); \ 221 1.1 mrg } while (0) 222 1.1 mrg #define UDIV_PREINV_ALWAYS 1 223 1.1 mrg #define UDIV_NEEDS_NORMALIZATION 1 224 1.1 mrg #endif /* LONGLONG_STANDALONE */ 225 1.1 mrg 226 1.1 mrg /* clz_tab is required in all configurations, since mpn/alpha/cntlz.asm 227 1.1 mrg always goes into libgmp.so, even when not actually used. */ 228 1.1 mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB 229 1.1 mrg 230 1.1 mrg #if defined (__GNUC__) && HAVE_HOST_CPU_alpha_CIX 231 1.1 mrg #define count_leading_zeros(COUNT,X) \ 232 1.1 mrg __asm__("ctlz %1,%0" : "=r"(COUNT) : "r"(X)) 233 1.1 mrg #define count_trailing_zeros(COUNT,X) \ 234 1.1 mrg __asm__("cttz %1,%0" : "=r"(COUNT) : "r"(X)) 235 1.1 mrg #endif /* clz/ctz using cix */ 236 1.1 mrg 237 1.3 mrg #if ! defined (count_leading_zeros) \ 238 1.1 mrg && defined (__GNUC__) && ! defined (LONGLONG_STANDALONE) 239 1.1 mrg /* ALPHA_CMPBGE_0 gives "cmpbge $31,src,dst", ie. test src bytes == 0. 240 1.1 mrg "$31" is written explicitly in the asm, since an "r" constraint won't 241 1.1 mrg select reg 31. There seems no need to worry about "r31" syntax for cray, 242 1.3 mrg since gcc itself (pre-release 3.4) emits just $31 in various places. */ 243 1.3 mrg #define ALPHA_CMPBGE_0(dst, src) \ 244 1.1 mrg do { asm ("cmpbge $31, %1, %0" : "=r" (dst) : "r" (src)); } while (0) 245 1.1 mrg /* Zero bytes are turned into bits with cmpbge, a __clz_tab lookup counts 246 1.1 mrg them, locating the highest non-zero byte. A second __clz_tab lookup 247 1.1 mrg counts the leading zero bits in that byte, giving the result. */ 248 1.3 mrg #define count_leading_zeros(count, x) \ 249 1.3 mrg do { \ 250 1.3 mrg UWtype __clz__b, __clz__c, __clz__x = (x); \ 251 1.3 mrg ALPHA_CMPBGE_0 (__clz__b, __clz__x); /* zero bytes */ \ 252 1.3 mrg __clz__b = __clz_tab [(__clz__b >> 1) ^ 0x7F]; /* 8 to 1 byte */ \ 253 1.3 mrg __clz__b = __clz__b * 8 - 7; /* 57 to 1 shift */ \ 254 1.3 mrg __clz__x >>= __clz__b; \ 255 1.3 mrg __clz__c = __clz_tab [__clz__x]; /* 8 to 1 bit */ \ 256 1.3 mrg __clz__b = 65 - __clz__b; \ 257 1.3 mrg (count) = __clz__b - __clz__c; \ 258 1.1 mrg } while (0) 259 1.1 mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB 260 1.1 mrg #endif /* clz using cmpbge */ 261 1.1 mrg 262 1.1 mrg #if ! defined (count_leading_zeros) && ! defined (LONGLONG_STANDALONE) 263 1.1 mrg #if HAVE_ATTRIBUTE_CONST 264 1.2 joerg long __MPN(count_leading_zeros) (UDItype) __attribute__ ((const)); 265 1.1 mrg #else 266 1.2 joerg long __MPN(count_leading_zeros) (UDItype); 267 1.1 mrg #endif 268 1.1 mrg #define count_leading_zeros(count, x) \ 269 1.1 mrg ((count) = __MPN(count_leading_zeros) (x)) 270 1.1 mrg #endif /* clz using mpn */ 271 1.1 mrg #endif /* __alpha */ 272 1.1 mrg 273 1.2 joerg #if defined (__AVR) && W_TYPE_SIZE == 8 274 1.2 joerg #define umul_ppmm(ph, pl, m0, m1) \ 275 1.2 joerg do { \ 276 1.2 joerg unsigned short __p = (unsigned short) (m0) * (m1); \ 277 1.2 joerg (ph) = __p >> 8; \ 278 1.2 joerg (pl) = __p; \ 279 1.2 joerg } while (0) 280 1.2 joerg #endif /* AVR */ 281 1.2 joerg 282 1.1 mrg #if defined (_CRAY) && W_TYPE_SIZE == 64 283 1.1 mrg #include <intrinsics.h> 284 1.1 mrg #define UDIV_PREINV_ALWAYS 1 285 1.1 mrg #define UDIV_NEEDS_NORMALIZATION 1 286 1.2 joerg long __MPN(count_leading_zeros) (UDItype); 287 1.1 mrg #define count_leading_zeros(count, x) \ 288 1.1 mrg ((count) = _leadz ((UWtype) (x))) 289 1.1 mrg #if defined (_CRAYIEEE) /* I.e., Cray T90/ieee, T3D, and T3E */ 290 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \ 291 1.1 mrg do { \ 292 1.1 mrg UDItype __m0 = (m0), __m1 = (m1); \ 293 1.3 mrg (ph) = _int_mult_upper (__m0, __m1); \ 294 1.1 mrg (pl) = __m0 * __m1; \ 295 1.1 mrg } while (0) 296 1.1 mrg #ifndef LONGLONG_STANDALONE 297 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 298 1.1 mrg do { UWtype __di; \ 299 1.1 mrg __di = __MPN(invert_limb) (d); \ 300 1.1 mrg udiv_qrnnd_preinv (q, r, n1, n0, d, __di); \ 301 1.1 mrg } while (0) 302 1.1 mrg #endif /* LONGLONG_STANDALONE */ 303 1.1 mrg #endif /* _CRAYIEEE */ 304 1.1 mrg #endif /* _CRAY */ 305 1.1 mrg 306 1.1 mrg #if defined (__ia64) && W_TYPE_SIZE == 64 307 1.1 mrg /* This form encourages gcc (pre-release 3.4 at least) to emit predicated 308 1.1 mrg "sub r=r,r" and "sub r=r,r,1", giving a 2 cycle latency. The generic 309 1.1 mrg code using "al<bl" arithmetically comes out making an actual 0 or 1 in a 310 1.1 mrg register, which takes an extra cycle. */ 311 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 312 1.3 mrg do { \ 313 1.3 mrg UWtype __x; \ 314 1.3 mrg __x = (al) - (bl); \ 315 1.3 mrg if ((al) < (bl)) \ 316 1.3 mrg (sh) = (ah) - (bh) - 1; \ 317 1.3 mrg else \ 318 1.3 mrg (sh) = (ah) - (bh); \ 319 1.3 mrg (sl) = __x; \ 320 1.1 mrg } while (0) 321 1.1 mrg #if defined (__GNUC__) && ! defined (__INTEL_COMPILER) 322 1.1 mrg /* Do both product parts in assembly, since that gives better code with 323 1.1 mrg all gcc versions. Some callers will just use the upper part, and in 324 1.1 mrg that situation we waste an instruction, but not any cycles. */ 325 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \ 326 1.1 mrg __asm__ ("xma.hu %0 = %2, %3, f0\n\txma.l %1 = %2, %3, f0" \ 327 1.1 mrg : "=&f" (ph), "=f" (pl) \ 328 1.1 mrg : "f" (m0), "f" (m1)) 329 1.1 mrg #define count_leading_zeros(count, x) \ 330 1.1 mrg do { \ 331 1.1 mrg UWtype _x = (x), _y, _a, _c; \ 332 1.1 mrg __asm__ ("mux1 %0 = %1, @rev" : "=r" (_y) : "r" (_x)); \ 333 1.1 mrg __asm__ ("czx1.l %0 = %1" : "=r" (_a) : "r" (-_y | _y)); \ 334 1.1 mrg _c = (_a - 1) << 3; \ 335 1.1 mrg _x >>= _c; \ 336 1.1 mrg if (_x >= 1 << 4) \ 337 1.1 mrg _x >>= 4, _c += 4; \ 338 1.1 mrg if (_x >= 1 << 2) \ 339 1.1 mrg _x >>= 2, _c += 2; \ 340 1.1 mrg _c += _x >> 1; \ 341 1.1 mrg (count) = W_TYPE_SIZE - 1 - _c; \ 342 1.1 mrg } while (0) 343 1.1 mrg /* similar to what gcc does for __builtin_ffs, but 0 based rather than 1 344 1.1 mrg based, and we don't need a special case for x==0 here */ 345 1.1 mrg #define count_trailing_zeros(count, x) \ 346 1.1 mrg do { \ 347 1.1 mrg UWtype __ctz_x = (x); \ 348 1.1 mrg __asm__ ("popcnt %0 = %1" \ 349 1.1 mrg : "=r" (count) \ 350 1.1 mrg : "r" ((__ctz_x-1) & ~__ctz_x)); \ 351 1.1 mrg } while (0) 352 1.1 mrg #endif 353 1.1 mrg #if defined (__INTEL_COMPILER) 354 1.1 mrg #include <ia64intrin.h> 355 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \ 356 1.1 mrg do { \ 357 1.3 mrg UWtype __m0 = (m0), __m1 = (m1); \ 358 1.3 mrg ph = _m64_xmahu (__m0, __m1, 0); \ 359 1.3 mrg pl = __m0 * __m1; \ 360 1.1 mrg } while (0) 361 1.1 mrg #endif 362 1.1 mrg #ifndef LONGLONG_STANDALONE 363 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 364 1.1 mrg do { UWtype __di; \ 365 1.1 mrg __di = __MPN(invert_limb) (d); \ 366 1.1 mrg udiv_qrnnd_preinv (q, r, n1, n0, d, __di); \ 367 1.1 mrg } while (0) 368 1.1 mrg #define UDIV_PREINV_ALWAYS 1 369 1.1 mrg #define UDIV_NEEDS_NORMALIZATION 1 370 1.1 mrg #endif 371 1.1 mrg #endif 372 1.1 mrg 373 1.1 mrg 374 1.6 christos #if defined (__GNUC__) || defined(__lint__) 375 1.1 mrg 376 1.1 mrg /* We sometimes need to clobber "cc" with gcc2, but that would not be 377 1.1 mrg understood by gcc1. Use cpp to avoid major code duplication. */ 378 1.1 mrg #if __GNUC__ < 2 379 1.1 mrg #define __CLOBBER_CC 380 1.1 mrg #define __AND_CLOBBER_CC 381 1.1 mrg #else /* __GNUC__ >= 2 */ 382 1.1 mrg #define __CLOBBER_CC : "cc" 383 1.1 mrg #define __AND_CLOBBER_CC , "cc" 384 1.1 mrg #endif /* __GNUC__ < 2 */ 385 1.1 mrg 386 1.1 mrg #if (defined (__a29k__) || defined (_AM29K)) && W_TYPE_SIZE == 32 387 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 388 1.1 mrg __asm__ ("add %1,%4,%5\n\taddc %0,%2,%3" \ 389 1.1 mrg : "=r" (sh), "=&r" (sl) \ 390 1.1 mrg : "r" (ah), "rI" (bh), "%r" (al), "rI" (bl)) 391 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 392 1.1 mrg __asm__ ("sub %1,%4,%5\n\tsubc %0,%2,%3" \ 393 1.1 mrg : "=r" (sh), "=&r" (sl) \ 394 1.1 mrg : "r" (ah), "rI" (bh), "r" (al), "rI" (bl)) 395 1.1 mrg #define umul_ppmm(xh, xl, m0, m1) \ 396 1.1 mrg do { \ 397 1.1 mrg USItype __m0 = (m0), __m1 = (m1); \ 398 1.1 mrg __asm__ ("multiplu %0,%1,%2" \ 399 1.1 mrg : "=r" (xl) \ 400 1.1 mrg : "r" (__m0), "r" (__m1)); \ 401 1.1 mrg __asm__ ("multmu %0,%1,%2" \ 402 1.1 mrg : "=r" (xh) \ 403 1.1 mrg : "r" (__m0), "r" (__m1)); \ 404 1.1 mrg } while (0) 405 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 406 1.1 mrg __asm__ ("dividu %0,%3,%4" \ 407 1.1 mrg : "=r" (q), "=q" (r) \ 408 1.1 mrg : "1" (n1), "r" (n0), "r" (d)) 409 1.1 mrg #define count_leading_zeros(count, x) \ 410 1.1 mrg __asm__ ("clz %0,%1" \ 411 1.1 mrg : "=r" (count) \ 412 1.1 mrg : "r" (x)) 413 1.1 mrg #define COUNT_LEADING_ZEROS_0 32 414 1.1 mrg #endif /* __a29k__ */ 415 1.1 mrg 416 1.1 mrg #if defined (__arc__) 417 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 418 1.1 mrg __asm__ ("add.f\t%1, %4, %5\n\tadc\t%0, %2, %3" \ 419 1.1 mrg : "=r" (sh), \ 420 1.1 mrg "=&r" (sl) \ 421 1.1 mrg : "r" ((USItype) (ah)), \ 422 1.3 mrg "rICal" ((USItype) (bh)), \ 423 1.1 mrg "%r" ((USItype) (al)), \ 424 1.3 mrg "rICal" ((USItype) (bl))) 425 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 426 1.1 mrg __asm__ ("sub.f\t%1, %4, %5\n\tsbc\t%0, %2, %3" \ 427 1.1 mrg : "=r" (sh), \ 428 1.1 mrg "=&r" (sl) \ 429 1.1 mrg : "r" ((USItype) (ah)), \ 430 1.3 mrg "rICal" ((USItype) (bh)), \ 431 1.1 mrg "r" ((USItype) (al)), \ 432 1.3 mrg "rICal" ((USItype) (bl))) 433 1.1 mrg #endif 434 1.1 mrg 435 1.3 mrg #if defined (__arm__) && (defined (__thumb2__) || !defined (__thumb__)) \ 436 1.3 mrg && W_TYPE_SIZE == 32 437 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 438 1.4 mrg do { \ 439 1.5 mrg if (__builtin_constant_p (bl) && -(USItype)(bl) < (USItype)(bl)) \ 440 1.4 mrg __asm__ ("subs\t%1, %4, %5\n\tadc\t%0, %2, %3" \ 441 1.4 mrg : "=r" (sh), "=&r" (sl) \ 442 1.4 mrg : "r" (ah), "rI" (bh), \ 443 1.4 mrg "%r" (al), "rI" (-(USItype)(bl)) __CLOBBER_CC); \ 444 1.4 mrg else \ 445 1.4 mrg __asm__ ("adds\t%1, %4, %5\n\tadc\t%0, %2, %3" \ 446 1.1 mrg : "=r" (sh), "=&r" (sl) \ 447 1.4 mrg : "r" (ah), "rI" (bh), "%r" (al), "rI" (bl) __CLOBBER_CC); \ 448 1.4 mrg } while (0) 449 1.4 mrg /* FIXME: Extend the immediate range for the low word by using both ADDS and 450 1.5 mrg SUBS, since they set carry in the same way. We need separate definitions 451 1.5 mrg for thumb and non-thumb since thumb lacks RSC. */ 452 1.4 mrg #if defined (__thumb__) 453 1.4 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 454 1.4 mrg do { \ 455 1.4 mrg if (__builtin_constant_p (ah) && __builtin_constant_p (bh) \ 456 1.4 mrg && (ah) == (bh)) \ 457 1.4 mrg __asm__ ("subs\t%1, %2, %3\n\tsbc\t%0, %0, %0" \ 458 1.4 mrg : "=r" (sh), "=r" (sl) \ 459 1.4 mrg : "r" (al), "rI" (bl) __CLOBBER_CC); \ 460 1.4 mrg else if (__builtin_constant_p (al)) \ 461 1.4 mrg __asm__ ("rsbs\t%1, %5, %4\n\tsbc\t%0, %2, %3" \ 462 1.4 mrg : "=r" (sh), "=&r" (sl) \ 463 1.4 mrg : "r" (ah), "rI" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \ 464 1.4 mrg else if (__builtin_constant_p (bl)) \ 465 1.4 mrg __asm__ ("subs\t%1, %4, %5\n\tsbc\t%0, %2, %3" \ 466 1.4 mrg : "=r" (sh), "=&r" (sl) \ 467 1.4 mrg : "r" (ah), "rI" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \ 468 1.4 mrg else \ 469 1.4 mrg __asm__ ("subs\t%1, %4, %5\n\tsbc\t%0, %2, %3" \ 470 1.4 mrg : "=r" (sh), "=&r" (sl) \ 471 1.4 mrg : "r" (ah), "rI" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \ 472 1.4 mrg } while (0) 473 1.4 mrg #else 474 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 475 1.1 mrg do { \ 476 1.4 mrg if (__builtin_constant_p (ah) && __builtin_constant_p (bh) \ 477 1.4 mrg && (ah) == (bh)) \ 478 1.4 mrg __asm__ ("subs\t%1, %2, %3\n\tsbc\t%0, %0, %0" \ 479 1.4 mrg : "=r" (sh), "=r" (sl) \ 480 1.4 mrg : "r" (al), "rI" (bl) __CLOBBER_CC); \ 481 1.4 mrg else if (__builtin_constant_p (al)) \ 482 1.1 mrg { \ 483 1.1 mrg if (__builtin_constant_p (ah)) \ 484 1.1 mrg __asm__ ("rsbs\t%1, %5, %4\n\trsc\t%0, %3, %2" \ 485 1.1 mrg : "=r" (sh), "=&r" (sl) \ 486 1.1 mrg : "rI" (ah), "r" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \ 487 1.1 mrg else \ 488 1.1 mrg __asm__ ("rsbs\t%1, %5, %4\n\tsbc\t%0, %2, %3" \ 489 1.1 mrg : "=r" (sh), "=&r" (sl) \ 490 1.1 mrg : "r" (ah), "rI" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \ 491 1.1 mrg } \ 492 1.1 mrg else if (__builtin_constant_p (ah)) \ 493 1.1 mrg { \ 494 1.1 mrg if (__builtin_constant_p (bl)) \ 495 1.1 mrg __asm__ ("subs\t%1, %4, %5\n\trsc\t%0, %3, %2" \ 496 1.1 mrg : "=r" (sh), "=&r" (sl) \ 497 1.1 mrg : "rI" (ah), "r" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \ 498 1.1 mrg else \ 499 1.1 mrg __asm__ ("rsbs\t%1, %5, %4\n\trsc\t%0, %3, %2" \ 500 1.1 mrg : "=r" (sh), "=&r" (sl) \ 501 1.1 mrg : "rI" (ah), "r" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \ 502 1.1 mrg } \ 503 1.1 mrg else if (__builtin_constant_p (bl)) \ 504 1.4 mrg __asm__ ("subs\t%1, %4, %5\n\tsbc\t%0, %2, %3" \ 505 1.4 mrg : "=r" (sh), "=&r" (sl) \ 506 1.4 mrg : "r" (ah), "rI" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \ 507 1.5 mrg else \ 508 1.1 mrg __asm__ ("subs\t%1, %4, %5\n\tsbc\t%0, %2, %3" \ 509 1.1 mrg : "=r" (sh), "=&r" (sl) \ 510 1.4 mrg : "r" (ah), "rI" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \ 511 1.1 mrg } while (0) 512 1.4 mrg #endif 513 1.3 mrg #if defined (__ARM_ARCH_2__) || defined (__ARM_ARCH_2A__) \ 514 1.3 mrg || defined (__ARM_ARCH_3__) 515 1.3 mrg #define umul_ppmm(xh, xl, a, b) \ 516 1.3 mrg do { \ 517 1.3 mrg register USItype __t0, __t1, __t2; \ 518 1.3 mrg __asm__ ("%@ Inlined umul_ppmm\n" \ 519 1.3 mrg " mov %2, %5, lsr #16\n" \ 520 1.3 mrg " mov %0, %6, lsr #16\n" \ 521 1.3 mrg " bic %3, %5, %2, lsl #16\n" \ 522 1.3 mrg " bic %4, %6, %0, lsl #16\n" \ 523 1.3 mrg " mul %1, %3, %4\n" \ 524 1.3 mrg " mul %4, %2, %4\n" \ 525 1.3 mrg " mul %3, %0, %3\n" \ 526 1.3 mrg " mul %0, %2, %0\n" \ 527 1.3 mrg " adds %3, %4, %3\n" \ 528 1.3 mrg " addcs %0, %0, #65536\n" \ 529 1.3 mrg " adds %1, %1, %3, lsl #16\n" \ 530 1.3 mrg " adc %0, %0, %3, lsr #16" \ 531 1.3 mrg : "=&r" ((USItype) (xh)), "=r" ((USItype) (xl)), \ 532 1.3 mrg "=&r" (__t0), "=&r" (__t1), "=r" (__t2) \ 533 1.3 mrg : "r" ((USItype) (a)), "r" ((USItype) (b)) __CLOBBER_CC); \ 534 1.3 mrg } while (0) 535 1.3 mrg #ifndef LONGLONG_STANDALONE 536 1.3 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 537 1.3 mrg do { UWtype __r; \ 538 1.3 mrg (q) = __MPN(udiv_qrnnd) (&__r, (n1), (n0), (d)); \ 539 1.3 mrg (r) = __r; \ 540 1.3 mrg } while (0) 541 1.3 mrg extern UWtype __MPN(udiv_qrnnd) (UWtype *, UWtype, UWtype, UWtype); 542 1.3 mrg #endif /* LONGLONG_STANDALONE */ 543 1.3 mrg #else /* ARMv4 or newer */ 544 1.1 mrg #define umul_ppmm(xh, xl, a, b) \ 545 1.1 mrg __asm__ ("umull %0,%1,%2,%3" : "=&r" (xl), "=&r" (xh) : "r" (a), "r" (b)) 546 1.1 mrg #define smul_ppmm(xh, xl, a, b) \ 547 1.1 mrg __asm__ ("smull %0,%1,%2,%3" : "=&r" (xl), "=&r" (xh) : "r" (a), "r" (b)) 548 1.1 mrg #ifndef LONGLONG_STANDALONE 549 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 550 1.1 mrg do { UWtype __di; \ 551 1.1 mrg __di = __MPN(invert_limb) (d); \ 552 1.1 mrg udiv_qrnnd_preinv (q, r, n1, n0, d, __di); \ 553 1.1 mrg } while (0) 554 1.1 mrg #define UDIV_PREINV_ALWAYS 1 555 1.1 mrg #define UDIV_NEEDS_NORMALIZATION 1 556 1.1 mrg #endif /* LONGLONG_STANDALONE */ 557 1.3 mrg #endif /* defined(__ARM_ARCH_2__) ... */ 558 1.3 mrg #define count_leading_zeros(count, x) count_leading_zeros_gcc_clz(count, x) 559 1.3 mrg #define count_trailing_zeros(count, x) count_trailing_zeros_gcc_ctz(count, x) 560 1.1 mrg #endif /* __arm__ */ 561 1.1 mrg 562 1.2 joerg #if defined (__aarch64__) && W_TYPE_SIZE == 64 563 1.2 joerg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 564 1.4 mrg do { \ 565 1.5 mrg if (__builtin_constant_p (bl) && ~(UDItype)(bl) <= (UDItype)(bl)) \ 566 1.4 mrg __asm__ ("subs\t%1, %x4, %5\n\tadc\t%0, %x2, %x3" \ 567 1.4 mrg : "=r" (sh), "=&r" (sl) \ 568 1.4 mrg : "rZ" ((UDItype)(ah)), "rZ" ((UDItype)(bh)), \ 569 1.4 mrg "%r" ((UDItype)(al)), "rI" (-(UDItype)(bl)) __CLOBBER_CC);\ 570 1.4 mrg else \ 571 1.4 mrg __asm__ ("adds\t%1, %x4, %5\n\tadc\t%0, %x2, %x3" \ 572 1.4 mrg : "=r" (sh), "=&r" (sl) \ 573 1.4 mrg : "rZ" ((UDItype)(ah)), "rZ" ((UDItype)(bh)), \ 574 1.4 mrg "%r" ((UDItype)(al)), "rI" ((UDItype)(bl)) __CLOBBER_CC);\ 575 1.4 mrg } while (0) 576 1.2 joerg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 577 1.4 mrg do { \ 578 1.5 mrg if (__builtin_constant_p (bl) && ~(UDItype)(bl) <= (UDItype)(bl)) \ 579 1.4 mrg __asm__ ("adds\t%1, %x4, %5\n\tsbc\t%0, %x2, %x3" \ 580 1.4 mrg : "=r,r" (sh), "=&r,&r" (sl) \ 581 1.4 mrg : "rZ,rZ" ((UDItype)(ah)), "rZ,rZ" ((UDItype)(bh)), \ 582 1.4 mrg "r,Z" ((UDItype)(al)), "rI,r" (-(UDItype)(bl)) __CLOBBER_CC);\ 583 1.4 mrg else \ 584 1.4 mrg __asm__ ("subs\t%1, %x4, %5\n\tsbc\t%0, %x2, %x3" \ 585 1.4 mrg : "=r,r" (sh), "=&r,&r" (sl) \ 586 1.4 mrg : "rZ,rZ" ((UDItype)(ah)), "rZ,rZ" ((UDItype)(bh)), \ 587 1.4 mrg "r,Z" ((UDItype)(al)), "rI,r" ((UDItype)(bl)) __CLOBBER_CC);\ 588 1.4 mrg } while(0); 589 1.4 mrg #if __GMP_GNUC_PREREQ (4,9) 590 1.4 mrg #define umul_ppmm(w1, w0, u, v) \ 591 1.4 mrg do { \ 592 1.4 mrg typedef unsigned int __ll_UTItype __attribute__((mode(TI))); \ 593 1.4 mrg __ll_UTItype __ll = (__ll_UTItype)(u) * (v); \ 594 1.4 mrg w1 = __ll >> 64; \ 595 1.4 mrg w0 = __ll; \ 596 1.4 mrg } while (0) 597 1.4 mrg #endif 598 1.4 mrg #if !defined (umul_ppmm) 599 1.2 joerg #define umul_ppmm(ph, pl, m0, m1) \ 600 1.2 joerg do { \ 601 1.2 joerg UDItype __m0 = (m0), __m1 = (m1); \ 602 1.3 mrg __asm__ ("umulh\t%0, %1, %2" : "=r" (ph) : "r" (__m0), "r" (__m1)); \ 603 1.2 joerg (pl) = __m0 * __m1; \ 604 1.2 joerg } while (0) 605 1.4 mrg #endif 606 1.3 mrg #define count_leading_zeros(count, x) count_leading_zeros_gcc_clz(count, x) 607 1.3 mrg #define count_trailing_zeros(count, x) count_trailing_zeros_gcc_ctz(count, x) 608 1.2 joerg #endif /* __aarch64__ */ 609 1.2 joerg 610 1.1 mrg #if defined (__clipper__) && W_TYPE_SIZE == 32 611 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 612 1.1 mrg ({union {UDItype __ll; \ 613 1.1 mrg struct {USItype __l, __h;} __i; \ 614 1.1 mrg } __x; \ 615 1.1 mrg __asm__ ("mulwux %2,%0" \ 616 1.1 mrg : "=r" (__x.__ll) \ 617 1.1 mrg : "%0" ((USItype)(u)), "r" ((USItype)(v))); \ 618 1.1 mrg (w1) = __x.__i.__h; (w0) = __x.__i.__l;}) 619 1.1 mrg #define smul_ppmm(w1, w0, u, v) \ 620 1.1 mrg ({union {DItype __ll; \ 621 1.1 mrg struct {SItype __l, __h;} __i; \ 622 1.1 mrg } __x; \ 623 1.1 mrg __asm__ ("mulwx %2,%0" \ 624 1.1 mrg : "=r" (__x.__ll) \ 625 1.1 mrg : "%0" ((SItype)(u)), "r" ((SItype)(v))); \ 626 1.1 mrg (w1) = __x.__i.__h; (w0) = __x.__i.__l;}) 627 1.1 mrg #define __umulsidi3(u, v) \ 628 1.1 mrg ({UDItype __w; \ 629 1.1 mrg __asm__ ("mulwux %2,%0" \ 630 1.1 mrg : "=r" (__w) : "%0" ((USItype)(u)), "r" ((USItype)(v))); \ 631 1.1 mrg __w; }) 632 1.1 mrg #endif /* __clipper__ */ 633 1.1 mrg 634 1.1 mrg /* Fujitsu vector computers. */ 635 1.1 mrg #if defined (__uxp__) && W_TYPE_SIZE == 32 636 1.1 mrg #define umul_ppmm(ph, pl, u, v) \ 637 1.1 mrg do { \ 638 1.1 mrg union {UDItype __ll; \ 639 1.1 mrg struct {USItype __h, __l;} __i; \ 640 1.1 mrg } __x; \ 641 1.1 mrg __asm__ ("mult.lu %1,%2,%0" : "=r" (__x.__ll) : "%r" (u), "rK" (v));\ 642 1.1 mrg (ph) = __x.__i.__h; \ 643 1.1 mrg (pl) = __x.__i.__l; \ 644 1.1 mrg } while (0) 645 1.1 mrg #define smul_ppmm(ph, pl, u, v) \ 646 1.1 mrg do { \ 647 1.1 mrg union {UDItype __ll; \ 648 1.1 mrg struct {USItype __h, __l;} __i; \ 649 1.1 mrg } __x; \ 650 1.1 mrg __asm__ ("mult.l %1,%2,%0" : "=r" (__x.__ll) : "%r" (u), "rK" (v)); \ 651 1.1 mrg (ph) = __x.__i.__h; \ 652 1.1 mrg (pl) = __x.__i.__l; \ 653 1.1 mrg } while (0) 654 1.1 mrg #endif 655 1.1 mrg 656 1.1 mrg #if defined (__gmicro__) && W_TYPE_SIZE == 32 657 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 658 1.1 mrg __asm__ ("add.w %5,%1\n\taddx %3,%0" \ 659 1.1 mrg : "=g" (sh), "=&g" (sl) \ 660 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 661 1.1 mrg "%1" ((USItype)(al)), "g" ((USItype)(bl))) 662 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 663 1.1 mrg __asm__ ("sub.w %5,%1\n\tsubx %3,%0" \ 664 1.1 mrg : "=g" (sh), "=&g" (sl) \ 665 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 666 1.1 mrg "1" ((USItype)(al)), "g" ((USItype)(bl))) 667 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \ 668 1.1 mrg __asm__ ("mulx %3,%0,%1" \ 669 1.1 mrg : "=g" (ph), "=r" (pl) \ 670 1.1 mrg : "%0" ((USItype)(m0)), "g" ((USItype)(m1))) 671 1.1 mrg #define udiv_qrnnd(q, r, nh, nl, d) \ 672 1.1 mrg __asm__ ("divx %4,%0,%1" \ 673 1.1 mrg : "=g" (q), "=r" (r) \ 674 1.1 mrg : "1" ((USItype)(nh)), "0" ((USItype)(nl)), "g" ((USItype)(d))) 675 1.1 mrg #define count_leading_zeros(count, x) \ 676 1.1 mrg __asm__ ("bsch/1 %1,%0" \ 677 1.1 mrg : "=g" (count) : "g" ((USItype)(x)), "0" ((USItype)0)) 678 1.1 mrg #endif 679 1.1 mrg 680 1.1 mrg #if defined (__hppa) && W_TYPE_SIZE == 32 681 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 682 1.1 mrg __asm__ ("add%I5 %5,%r4,%1\n\taddc %r2,%r3,%0" \ 683 1.1 mrg : "=r" (sh), "=&r" (sl) \ 684 1.1 mrg : "rM" (ah), "rM" (bh), "%rM" (al), "rI" (bl)) 685 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 686 1.1 mrg __asm__ ("sub%I4 %4,%r5,%1\n\tsubb %r2,%r3,%0" \ 687 1.1 mrg : "=r" (sh), "=&r" (sl) \ 688 1.1 mrg : "rM" (ah), "rM" (bh), "rI" (al), "rM" (bl)) 689 1.1 mrg #if defined (_PA_RISC1_1) 690 1.1 mrg #define umul_ppmm(wh, wl, u, v) \ 691 1.1 mrg do { \ 692 1.1 mrg union {UDItype __ll; \ 693 1.1 mrg struct {USItype __h, __l;} __i; \ 694 1.1 mrg } __x; \ 695 1.1 mrg __asm__ ("xmpyu %1,%2,%0" : "=*f" (__x.__ll) : "*f" (u), "*f" (v)); \ 696 1.1 mrg (wh) = __x.__i.__h; \ 697 1.1 mrg (wl) = __x.__i.__l; \ 698 1.1 mrg } while (0) 699 1.1 mrg #endif 700 1.1 mrg #define count_leading_zeros(count, x) \ 701 1.1 mrg do { \ 702 1.1 mrg USItype __tmp; \ 703 1.1 mrg __asm__ ( \ 704 1.1 mrg "ldi 1,%0\n" \ 705 1.1 mrg " extru,= %1,15,16,%%r0 ; Bits 31..16 zero?\n" \ 706 1.1 mrg " extru,tr %1,15,16,%1 ; No. Shift down, skip add.\n" \ 707 1.1 mrg " ldo 16(%0),%0 ; Yes. Perform add.\n" \ 708 1.1 mrg " extru,= %1,23,8,%%r0 ; Bits 15..8 zero?\n" \ 709 1.1 mrg " extru,tr %1,23,8,%1 ; No. Shift down, skip add.\n" \ 710 1.1 mrg " ldo 8(%0),%0 ; Yes. Perform add.\n" \ 711 1.1 mrg " extru,= %1,27,4,%%r0 ; Bits 7..4 zero?\n" \ 712 1.1 mrg " extru,tr %1,27,4,%1 ; No. Shift down, skip add.\n" \ 713 1.1 mrg " ldo 4(%0),%0 ; Yes. Perform add.\n" \ 714 1.1 mrg " extru,= %1,29,2,%%r0 ; Bits 3..2 zero?\n" \ 715 1.1 mrg " extru,tr %1,29,2,%1 ; No. Shift down, skip add.\n" \ 716 1.1 mrg " ldo 2(%0),%0 ; Yes. Perform add.\n" \ 717 1.1 mrg " extru %1,30,1,%1 ; Extract bit 1.\n" \ 718 1.1 mrg " sub %0,%1,%0 ; Subtract it.\n" \ 719 1.1 mrg : "=r" (count), "=r" (__tmp) : "1" (x)); \ 720 1.1 mrg } while (0) 721 1.1 mrg #endif /* hppa */ 722 1.1 mrg 723 1.1 mrg /* These macros are for ABI=2.0w. In ABI=2.0n they can't be used, since GCC 724 1.1 mrg (3.2) puts longlong into two adjacent 32-bit registers. Presumably this 725 1.1 mrg is just a case of no direct support for 2.0n but treating it like 1.0. */ 726 1.1 mrg #if defined (__hppa) && W_TYPE_SIZE == 64 && ! defined (_LONG_LONG_LIMB) 727 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 728 1.1 mrg __asm__ ("add%I5 %5,%r4,%1\n\tadd,dc %r2,%r3,%0" \ 729 1.1 mrg : "=r" (sh), "=&r" (sl) \ 730 1.1 mrg : "rM" (ah), "rM" (bh), "%rM" (al), "rI" (bl)) 731 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 732 1.1 mrg __asm__ ("sub%I4 %4,%r5,%1\n\tsub,db %r2,%r3,%0" \ 733 1.1 mrg : "=r" (sh), "=&r" (sl) \ 734 1.1 mrg : "rM" (ah), "rM" (bh), "rI" (al), "rM" (bl)) 735 1.1 mrg #endif /* hppa */ 736 1.1 mrg 737 1.1 mrg #if (defined (__i370__) || defined (__s390__) || defined (__mvs__)) && W_TYPE_SIZE == 32 738 1.2 joerg #if defined (__zarch__) || defined (HAVE_HOST_CPU_s390_zarch) 739 1.2 joerg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 740 1.2 joerg do { \ 741 1.2 joerg /* if (__builtin_constant_p (bl)) \ 742 1.2 joerg __asm__ ("alfi\t%1,%o5\n\talcr\t%0,%3" \ 743 1.2 joerg : "=r" (sh), "=&r" (sl) \ 744 1.2 joerg : "0" (ah), "r" (bh), "%1" (al), "n" (bl) __CLOBBER_CC);\ 745 1.2 joerg else \ 746 1.2 joerg */ __asm__ ("alr\t%1,%5\n\talcr\t%0,%3" \ 747 1.2 joerg : "=r" (sh), "=&r" (sl) \ 748 1.2 joerg : "0" (ah), "r" (bh), "%1" (al), "r" (bl)__CLOBBER_CC); \ 749 1.2 joerg } while (0) 750 1.2 joerg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 751 1.2 joerg do { \ 752 1.2 joerg /* if (__builtin_constant_p (bl)) \ 753 1.2 joerg __asm__ ("slfi\t%1,%o5\n\tslbr\t%0,%3" \ 754 1.2 joerg : "=r" (sh), "=&r" (sl) \ 755 1.2 joerg : "0" (ah), "r" (bh), "1" (al), "n" (bl) __CLOBBER_CC); \ 756 1.2 joerg else \ 757 1.2 joerg */ __asm__ ("slr\t%1,%5\n\tslbr\t%0,%3" \ 758 1.2 joerg : "=r" (sh), "=&r" (sl) \ 759 1.2 joerg : "0" (ah), "r" (bh), "1" (al), "r" (bl) __CLOBBER_CC); \ 760 1.2 joerg } while (0) 761 1.2 joerg #if __GMP_GNUC_PREREQ (4,5) 762 1.2 joerg #define umul_ppmm(xh, xl, m0, m1) \ 763 1.2 joerg do { \ 764 1.2 joerg union {UDItype __ll; \ 765 1.2 joerg struct {USItype __h, __l;} __i; \ 766 1.2 joerg } __x; \ 767 1.2 joerg __x.__ll = (UDItype) (m0) * (UDItype) (m1); \ 768 1.2 joerg (xh) = __x.__i.__h; (xl) = __x.__i.__l; \ 769 1.2 joerg } while (0) 770 1.2 joerg #else 771 1.2 joerg #if 0 772 1.2 joerg /* FIXME: this fails if gcc knows about the 64-bit registers. Use only 773 1.2 joerg with a new enough processor pretending we have 32-bit registers. */ 774 1.2 joerg #define umul_ppmm(xh, xl, m0, m1) \ 775 1.2 joerg do { \ 776 1.2 joerg union {UDItype __ll; \ 777 1.2 joerg struct {USItype __h, __l;} __i; \ 778 1.2 joerg } __x; \ 779 1.2 joerg __asm__ ("mlr\t%0,%2" \ 780 1.2 joerg : "=r" (__x.__ll) \ 781 1.2 joerg : "%0" (m0), "r" (m1)); \ 782 1.2 joerg (xh) = __x.__i.__h; (xl) = __x.__i.__l; \ 783 1.2 joerg } while (0) 784 1.2 joerg #else 785 1.2 joerg #define umul_ppmm(xh, xl, m0, m1) \ 786 1.2 joerg do { \ 787 1.2 joerg /* When we have 64-bit regs and gcc is aware of that, we cannot simply use 788 1.2 joerg DImode for the product, since that would be allocated to a single 64-bit 789 1.2 joerg register, whereas mlr uses the low 32-bits of an even-odd register pair. 790 1.2 joerg */ \ 791 1.2 joerg register USItype __r0 __asm__ ("0"); \ 792 1.2 joerg register USItype __r1 __asm__ ("1") = (m0); \ 793 1.2 joerg __asm__ ("mlr\t%0,%3" \ 794 1.2 joerg : "=r" (__r0), "=r" (__r1) \ 795 1.2 joerg : "r" (__r1), "r" (m1)); \ 796 1.2 joerg (xh) = __r0; (xl) = __r1; \ 797 1.2 joerg } while (0) 798 1.2 joerg #endif /* if 0 */ 799 1.2 joerg #endif 800 1.2 joerg #if 0 801 1.2 joerg /* FIXME: this fails if gcc knows about the 64-bit registers. Use only 802 1.2 joerg with a new enough processor pretending we have 32-bit registers. */ 803 1.2 joerg #define udiv_qrnnd(q, r, n1, n0, d) \ 804 1.2 joerg do { \ 805 1.2 joerg union {UDItype __ll; \ 806 1.2 joerg struct {USItype __h, __l;} __i; \ 807 1.2 joerg } __x; \ 808 1.2 joerg __x.__i.__h = n1; __x.__i.__l = n0; \ 809 1.2 joerg __asm__ ("dlr\t%0,%2" \ 810 1.2 joerg : "=r" (__x.__ll) \ 811 1.2 joerg : "0" (__x.__ll), "r" (d)); \ 812 1.2 joerg (q) = __x.__i.__l; (r) = __x.__i.__h; \ 813 1.2 joerg } while (0) 814 1.2 joerg #else 815 1.2 joerg #define udiv_qrnnd(q, r, n1, n0, d) \ 816 1.2 joerg do { \ 817 1.2 joerg register USItype __r0 __asm__ ("0") = (n1); \ 818 1.2 joerg register USItype __r1 __asm__ ("1") = (n0); \ 819 1.2 joerg __asm__ ("dlr\t%0,%4" \ 820 1.2 joerg : "=r" (__r0), "=r" (__r1) \ 821 1.2 joerg : "r" (__r0), "r" (__r1), "r" (d)); \ 822 1.2 joerg (q) = __r1; (r) = __r0; \ 823 1.2 joerg } while (0) 824 1.2 joerg #endif /* if 0 */ 825 1.2 joerg #else /* if __zarch__ */ 826 1.2 joerg /* FIXME: this fails if gcc knows about the 64-bit registers. */ 827 1.2 joerg #define smul_ppmm(xh, xl, m0, m1) \ 828 1.1 mrg do { \ 829 1.1 mrg union {DItype __ll; \ 830 1.1 mrg struct {USItype __h, __l;} __i; \ 831 1.1 mrg } __x; \ 832 1.2 joerg __asm__ ("mr\t%0,%2" \ 833 1.2 joerg : "=r" (__x.__ll) \ 834 1.2 joerg : "%0" (m0), "r" (m1)); \ 835 1.1 mrg (xh) = __x.__i.__h; (xl) = __x.__i.__l; \ 836 1.1 mrg } while (0) 837 1.2 joerg /* FIXME: this fails if gcc knows about the 64-bit registers. */ 838 1.2 joerg #define sdiv_qrnnd(q, r, n1, n0, d) \ 839 1.1 mrg do { \ 840 1.1 mrg union {DItype __ll; \ 841 1.1 mrg struct {USItype __h, __l;} __i; \ 842 1.1 mrg } __x; \ 843 1.1 mrg __x.__i.__h = n1; __x.__i.__l = n0; \ 844 1.2 joerg __asm__ ("dr\t%0,%2" \ 845 1.1 mrg : "=r" (__x.__ll) \ 846 1.1 mrg : "0" (__x.__ll), "r" (d)); \ 847 1.1 mrg (q) = __x.__i.__l; (r) = __x.__i.__h; \ 848 1.1 mrg } while (0) 849 1.2 joerg #endif /* if __zarch__ */ 850 1.2 joerg #endif 851 1.2 joerg 852 1.2 joerg #if defined (__s390x__) && W_TYPE_SIZE == 64 853 1.2 joerg /* We need to cast operands with register constraints, otherwise their types 854 1.2 joerg will be assumed to be SImode by gcc. For these machines, such operations 855 1.2 joerg will insert a value into the low 32 bits, and leave the high 32 bits with 856 1.2 joerg garbage. */ 857 1.2 joerg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 858 1.2 joerg do { \ 859 1.2 joerg __asm__ ("algr\t%1,%5\n\talcgr\t%0,%3" \ 860 1.2 joerg : "=r" (sh), "=&r" (sl) \ 861 1.2 joerg : "0" ((UDItype)(ah)), "r" ((UDItype)(bh)), \ 862 1.2 joerg "%1" ((UDItype)(al)), "r" ((UDItype)(bl)) __CLOBBER_CC); \ 863 1.2 joerg } while (0) 864 1.2 joerg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 865 1.2 joerg do { \ 866 1.2 joerg __asm__ ("slgr\t%1,%5\n\tslbgr\t%0,%3" \ 867 1.2 joerg : "=r" (sh), "=&r" (sl) \ 868 1.2 joerg : "0" ((UDItype)(ah)), "r" ((UDItype)(bh)), \ 869 1.2 joerg "1" ((UDItype)(al)), "r" ((UDItype)(bl)) __CLOBBER_CC); \ 870 1.2 joerg } while (0) 871 1.2 joerg #define umul_ppmm(xh, xl, m0, m1) \ 872 1.2 joerg do { \ 873 1.2 joerg union {unsigned int __attribute__ ((mode(TI))) __ll; \ 874 1.2 joerg struct {UDItype __h, __l;} __i; \ 875 1.2 joerg } __x; \ 876 1.2 joerg __asm__ ("mlgr\t%0,%2" \ 877 1.2 joerg : "=r" (__x.__ll) \ 878 1.2 joerg : "%0" ((UDItype)(m0)), "r" ((UDItype)(m1))); \ 879 1.2 joerg (xh) = __x.__i.__h; (xl) = __x.__i.__l; \ 880 1.2 joerg } while (0) 881 1.2 joerg #define udiv_qrnnd(q, r, n1, n0, d) \ 882 1.2 joerg do { \ 883 1.2 joerg union {unsigned int __attribute__ ((mode(TI))) __ll; \ 884 1.2 joerg struct {UDItype __h, __l;} __i; \ 885 1.2 joerg } __x; \ 886 1.2 joerg __x.__i.__h = n1; __x.__i.__l = n0; \ 887 1.2 joerg __asm__ ("dlgr\t%0,%2" \ 888 1.2 joerg : "=r" (__x.__ll) \ 889 1.2 joerg : "0" (__x.__ll), "r" ((UDItype)(d))); \ 890 1.2 joerg (q) = __x.__i.__l; (r) = __x.__i.__h; \ 891 1.2 joerg } while (0) 892 1.2 joerg #if 0 /* FIXME: Enable for z10 (?) */ 893 1.2 joerg #define count_leading_zeros(cnt, x) \ 894 1.2 joerg do { \ 895 1.2 joerg union {unsigned int __attribute__ ((mode(TI))) __ll; \ 896 1.2 joerg struct {UDItype __h, __l;} __i; \ 897 1.2 joerg } __clr_cnt; \ 898 1.2 joerg __asm__ ("flogr\t%0,%1" \ 899 1.2 joerg : "=r" (__clr_cnt.__ll) \ 900 1.2 joerg : "r" (x) __CLOBBER_CC); \ 901 1.2 joerg (cnt) = __clr_cnt.__i.__h; \ 902 1.2 joerg } while (0) 903 1.2 joerg #endif 904 1.1 mrg #endif 905 1.1 mrg 906 1.3 mrg /* On x86 and x86_64, every asm implicitly clobbers "flags" and "fpsr", 907 1.3 mrg so we don't need __CLOBBER_CC. */ 908 1.1 mrg #if (defined (__i386__) || defined (__i486__)) && W_TYPE_SIZE == 32 909 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 910 1.1 mrg __asm__ ("addl %5,%k1\n\tadcl %3,%k0" \ 911 1.1 mrg : "=r" (sh), "=&r" (sl) \ 912 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 913 1.1 mrg "%1" ((USItype)(al)), "g" ((USItype)(bl))) 914 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 915 1.1 mrg __asm__ ("subl %5,%k1\n\tsbbl %3,%k0" \ 916 1.1 mrg : "=r" (sh), "=&r" (sl) \ 917 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 918 1.1 mrg "1" ((USItype)(al)), "g" ((USItype)(bl))) 919 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 920 1.1 mrg __asm__ ("mull %3" \ 921 1.1 mrg : "=a" (w0), "=d" (w1) \ 922 1.1 mrg : "%0" ((USItype)(u)), "rm" ((USItype)(v))) 923 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, dx) /* d renamed to dx avoiding "=d" */\ 924 1.1 mrg __asm__ ("divl %4" /* stringification in K&R C */ \ 925 1.1 mrg : "=a" (q), "=d" (r) \ 926 1.1 mrg : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "rm" ((USItype)(dx))) 927 1.1 mrg 928 1.1 mrg #if HAVE_HOST_CPU_i586 || HAVE_HOST_CPU_pentium || HAVE_HOST_CPU_pentiummmx 929 1.1 mrg /* Pentium bsrl takes between 10 and 72 cycles depending where the most 930 1.1 mrg significant 1 bit is, hence the use of the following alternatives. bsfl 931 1.1 mrg is slow too, between 18 and 42 depending where the least significant 1 932 1.1 mrg bit is, so let the generic count_trailing_zeros below make use of the 933 1.1 mrg count_leading_zeros here too. */ 934 1.1 mrg 935 1.1 mrg #if HAVE_HOST_CPU_pentiummmx && ! defined (LONGLONG_STANDALONE) 936 1.1 mrg /* The following should be a fixed 14 or 15 cycles, but possibly plus an L1 937 1.1 mrg cache miss reading from __clz_tab. For P55 it's favoured over the float 938 1.1 mrg below so as to avoid mixing MMX and x87, since the penalty for switching 939 1.1 mrg between the two is about 100 cycles. 940 1.1 mrg 941 1.1 mrg The asm block sets __shift to -3 if the high 24 bits are clear, -2 for 942 1.1 mrg 16, -1 for 8, or 0 otherwise. This could be written equivalently as 943 1.1 mrg follows, but as of gcc 2.95.2 it results in conditional jumps. 944 1.1 mrg 945 1.1 mrg __shift = -(__n < 0x1000000); 946 1.1 mrg __shift -= (__n < 0x10000); 947 1.1 mrg __shift -= (__n < 0x100); 948 1.1 mrg 949 1.1 mrg The middle two sbbl and cmpl's pair, and with luck something gcc 950 1.1 mrg generates might pair with the first cmpl and the last sbbl. The "32+1" 951 1.1 mrg constant could be folded into __clz_tab[], but it doesn't seem worth 952 1.1 mrg making a different table just for that. */ 953 1.1 mrg 954 1.1 mrg #define count_leading_zeros(c,n) \ 955 1.1 mrg do { \ 956 1.1 mrg USItype __n = (n); \ 957 1.1 mrg USItype __shift; \ 958 1.1 mrg __asm__ ("cmpl $0x1000000, %1\n" \ 959 1.1 mrg "sbbl %0, %0\n" \ 960 1.1 mrg "cmpl $0x10000, %1\n" \ 961 1.1 mrg "sbbl $0, %0\n" \ 962 1.1 mrg "cmpl $0x100, %1\n" \ 963 1.1 mrg "sbbl $0, %0\n" \ 964 1.1 mrg : "=&r" (__shift) : "r" (__n)); \ 965 1.1 mrg __shift = __shift*8 + 24 + 1; \ 966 1.1 mrg (c) = 32 + 1 - __shift - __clz_tab[__n >> __shift]; \ 967 1.1 mrg } while (0) 968 1.1 mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB 969 1.1 mrg #define COUNT_LEADING_ZEROS_0 31 /* n==0 indistinguishable from n==1 */ 970 1.1 mrg 971 1.1 mrg #else /* ! pentiummmx || LONGLONG_STANDALONE */ 972 1.1 mrg /* The following should be a fixed 14 cycles or so. Some scheduling 973 1.1 mrg opportunities should be available between the float load/store too. This 974 1.1 mrg sort of code is used in gcc 3 for __builtin_ffs (with "n&-n") and is 975 1.1 mrg apparently suggested by the Intel optimizing manual (don't know exactly 976 1.1 mrg where). gcc 2.95 or up will be best for this, so the "double" is 977 1.1 mrg correctly aligned on the stack. */ 978 1.1 mrg #define count_leading_zeros(c,n) \ 979 1.1 mrg do { \ 980 1.1 mrg union { \ 981 1.1 mrg double d; \ 982 1.1 mrg unsigned a[2]; \ 983 1.1 mrg } __u; \ 984 1.1 mrg __u.d = (UWtype) (n); \ 985 1.1 mrg (c) = 0x3FF + 31 - (__u.a[1] >> 20); \ 986 1.1 mrg } while (0) 987 1.1 mrg #define COUNT_LEADING_ZEROS_0 (0x3FF + 31) 988 1.1 mrg #endif /* pentiummx */ 989 1.1 mrg 990 1.1 mrg #else /* ! pentium */ 991 1.1 mrg 992 1.1 mrg #if __GMP_GNUC_PREREQ (3,4) /* using bsrl */ 993 1.1 mrg #define count_leading_zeros(count,x) count_leading_zeros_gcc_clz(count,x) 994 1.1 mrg #endif /* gcc clz */ 995 1.1 mrg 996 1.1 mrg /* On P6, gcc prior to 3.0 generates a partial register stall for 997 1.1 mrg __cbtmp^31, due to using "xorb $31" instead of "xorl $31", the former 998 1.1 mrg being 1 code byte smaller. "31-__cbtmp" is a workaround, probably at the 999 1.1 mrg cost of one extra instruction. Do this for "i386" too, since that means 1000 1.1 mrg generic x86. */ 1001 1.3 mrg #if ! defined (count_leading_zeros) && __GNUC__ < 3 \ 1002 1.1 mrg && (HAVE_HOST_CPU_i386 \ 1003 1.1 mrg || HAVE_HOST_CPU_i686 \ 1004 1.1 mrg || HAVE_HOST_CPU_pentiumpro \ 1005 1.1 mrg || HAVE_HOST_CPU_pentium2 \ 1006 1.1 mrg || HAVE_HOST_CPU_pentium3) 1007 1.1 mrg #define count_leading_zeros(count, x) \ 1008 1.1 mrg do { \ 1009 1.1 mrg USItype __cbtmp; \ 1010 1.1 mrg ASSERT ((x) != 0); \ 1011 1.1 mrg __asm__ ("bsrl %1,%0" : "=r" (__cbtmp) : "rm" ((USItype)(x))); \ 1012 1.1 mrg (count) = 31 - __cbtmp; \ 1013 1.1 mrg } while (0) 1014 1.1 mrg #endif /* gcc<3 asm bsrl */ 1015 1.1 mrg 1016 1.1 mrg #ifndef count_leading_zeros 1017 1.1 mrg #define count_leading_zeros(count, x) \ 1018 1.1 mrg do { \ 1019 1.1 mrg USItype __cbtmp; \ 1020 1.1 mrg ASSERT ((x) != 0); \ 1021 1.1 mrg __asm__ ("bsrl %1,%0" : "=r" (__cbtmp) : "rm" ((USItype)(x))); \ 1022 1.1 mrg (count) = __cbtmp ^ 31; \ 1023 1.1 mrg } while (0) 1024 1.1 mrg #endif /* asm bsrl */ 1025 1.1 mrg 1026 1.1 mrg #if __GMP_GNUC_PREREQ (3,4) /* using bsfl */ 1027 1.1 mrg #define count_trailing_zeros(count,x) count_trailing_zeros_gcc_ctz(count,x) 1028 1.1 mrg #endif /* gcc ctz */ 1029 1.1 mrg 1030 1.1 mrg #ifndef count_trailing_zeros 1031 1.1 mrg #define count_trailing_zeros(count, x) \ 1032 1.1 mrg do { \ 1033 1.1 mrg ASSERT ((x) != 0); \ 1034 1.1 mrg __asm__ ("bsfl %1,%k0" : "=r" (count) : "rm" ((USItype)(x))); \ 1035 1.1 mrg } while (0) 1036 1.1 mrg #endif /* asm bsfl */ 1037 1.1 mrg 1038 1.1 mrg #endif /* ! pentium */ 1039 1.1 mrg 1040 1.1 mrg #endif /* 80x86 */ 1041 1.1 mrg 1042 1.1 mrg #if defined (__amd64__) && W_TYPE_SIZE == 64 1043 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1044 1.1 mrg __asm__ ("addq %5,%q1\n\tadcq %3,%q0" \ 1045 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1046 1.1 mrg : "0" ((UDItype)(ah)), "rme" ((UDItype)(bh)), \ 1047 1.1 mrg "%1" ((UDItype)(al)), "rme" ((UDItype)(bl))) 1048 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1049 1.1 mrg __asm__ ("subq %5,%q1\n\tsbbq %3,%q0" \ 1050 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1051 1.1 mrg : "0" ((UDItype)(ah)), "rme" ((UDItype)(bh)), \ 1052 1.1 mrg "1" ((UDItype)(al)), "rme" ((UDItype)(bl))) 1053 1.4 mrg #if X86_ASM_MULX \ 1054 1.4 mrg && (HAVE_HOST_CPU_haswell || HAVE_HOST_CPU_broadwell \ 1055 1.4 mrg || HAVE_HOST_CPU_skylake || HAVE_HOST_CPU_bd4 || HAVE_HOST_CPU_zen) 1056 1.4 mrg #define umul_ppmm(w1, w0, u, v) \ 1057 1.5 mrg __asm__ ("mulx\t%3, %q0, %q1" \ 1058 1.4 mrg : "=r" (w0), "=r" (w1) \ 1059 1.4 mrg : "%d" ((UDItype)(u)), "rm" ((UDItype)(v))) 1060 1.4 mrg #else 1061 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1062 1.4 mrg __asm__ ("mulq\t%3" \ 1063 1.1 mrg : "=a" (w0), "=d" (w1) \ 1064 1.1 mrg : "%0" ((UDItype)(u)), "rm" ((UDItype)(v))) 1065 1.4 mrg #endif 1066 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, dx) /* d renamed to dx avoiding "=d" */\ 1067 1.1 mrg __asm__ ("divq %4" /* stringification in K&R C */ \ 1068 1.1 mrg : "=a" (q), "=d" (r) \ 1069 1.1 mrg : "0" ((UDItype)(n0)), "1" ((UDItype)(n1)), "rm" ((UDItype)(dx))) 1070 1.4 mrg 1071 1.4 mrg #if HAVE_HOST_CPU_haswell || HAVE_HOST_CPU_broadwell || HAVE_HOST_CPU_skylake \ 1072 1.4 mrg || HAVE_HOST_CPU_k10 || HAVE_HOST_CPU_bd1 || HAVE_HOST_CPU_bd2 \ 1073 1.4 mrg || HAVE_HOST_CPU_bd3 || HAVE_HOST_CPU_bd4 || HAVE_HOST_CPU_zen \ 1074 1.4 mrg || HAVE_HOST_CPU_bobcat || HAVE_HOST_CPU_jaguar 1075 1.4 mrg #define count_leading_zeros(count, x) \ 1076 1.4 mrg do { \ 1077 1.4 mrg /* This is lzcnt, spelled for older assemblers. Destination and */ \ 1078 1.4 mrg /* source must be a 64-bit registers, hence cast and %q. */ \ 1079 1.4 mrg __asm__ ("rep;bsr\t%1, %q0" : "=r" (count) : "rm" ((UDItype)(x))); \ 1080 1.4 mrg } while (0) 1081 1.4 mrg #define COUNT_LEADING_ZEROS_0 64 1082 1.4 mrg #else 1083 1.1 mrg #define count_leading_zeros(count, x) \ 1084 1.1 mrg do { \ 1085 1.1 mrg UDItype __cbtmp; \ 1086 1.1 mrg ASSERT ((x) != 0); \ 1087 1.4 mrg __asm__ ("bsr\t%1,%0" : "=r" (__cbtmp) : "rm" ((UDItype)(x))); \ 1088 1.1 mrg (count) = __cbtmp ^ 63; \ 1089 1.1 mrg } while (0) 1090 1.4 mrg #endif 1091 1.4 mrg 1092 1.4 mrg #if HAVE_HOST_CPU_bd2 || HAVE_HOST_CPU_bd3 || HAVE_HOST_CPU_bd4 \ 1093 1.4 mrg || HAVE_HOST_CPU_zen || HAVE_HOST_CPU_jaguar 1094 1.4 mrg #define count_trailing_zeros(count, x) \ 1095 1.4 mrg do { \ 1096 1.4 mrg /* This is tzcnt, spelled for older assemblers. Destination and */ \ 1097 1.4 mrg /* source must be a 64-bit registers, hence cast and %q. */ \ 1098 1.4 mrg __asm__ ("rep;bsf\t%1, %q0" : "=r" (count) : "rm" ((UDItype)(x))); \ 1099 1.4 mrg } while (0) 1100 1.4 mrg #define COUNT_TRAILING_ZEROS_0 64 1101 1.4 mrg #else 1102 1.1 mrg #define count_trailing_zeros(count, x) \ 1103 1.1 mrg do { \ 1104 1.1 mrg ASSERT ((x) != 0); \ 1105 1.4 mrg __asm__ ("bsf\t%1, %q0" : "=r" (count) : "rm" ((UDItype)(x))); \ 1106 1.1 mrg } while (0) 1107 1.4 mrg #endif 1108 1.3 mrg #endif /* __amd64__ */ 1109 1.1 mrg 1110 1.1 mrg #if defined (__i860__) && W_TYPE_SIZE == 32 1111 1.1 mrg #define rshift_rhlc(r,h,l,c) \ 1112 1.1 mrg __asm__ ("shr %3,r0,r0\;shrd %1,%2,%0" \ 1113 1.1 mrg "=r" (r) : "r" (h), "r" (l), "rn" (c)) 1114 1.1 mrg #endif /* i860 */ 1115 1.1 mrg 1116 1.1 mrg #if defined (__i960__) && W_TYPE_SIZE == 32 1117 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1118 1.1 mrg __asm__ ("cmpo 1,0\;addc %5,%4,%1\;addc %3,%2,%0" \ 1119 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1120 1.1 mrg : "dI" (ah), "dI" (bh), "%dI" (al), "dI" (bl)) 1121 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1122 1.1 mrg __asm__ ("cmpo 0,0\;subc %5,%4,%1\;subc %3,%2,%0" \ 1123 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1124 1.1 mrg : "dI" (ah), "dI" (bh), "dI" (al), "dI" (bl)) 1125 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1126 1.1 mrg ({union {UDItype __ll; \ 1127 1.1 mrg struct {USItype __l, __h;} __i; \ 1128 1.1 mrg } __x; \ 1129 1.1 mrg __asm__ ("emul %2,%1,%0" \ 1130 1.1 mrg : "=d" (__x.__ll) : "%dI" (u), "dI" (v)); \ 1131 1.1 mrg (w1) = __x.__i.__h; (w0) = __x.__i.__l;}) 1132 1.1 mrg #define __umulsidi3(u, v) \ 1133 1.1 mrg ({UDItype __w; \ 1134 1.1 mrg __asm__ ("emul %2,%1,%0" : "=d" (__w) : "%dI" (u), "dI" (v)); \ 1135 1.1 mrg __w; }) 1136 1.1 mrg #define udiv_qrnnd(q, r, nh, nl, d) \ 1137 1.1 mrg do { \ 1138 1.1 mrg union {UDItype __ll; \ 1139 1.1 mrg struct {USItype __l, __h;} __i; \ 1140 1.1 mrg } __nn; \ 1141 1.1 mrg __nn.__i.__h = (nh); __nn.__i.__l = (nl); \ 1142 1.1 mrg __asm__ ("ediv %d,%n,%0" \ 1143 1.1 mrg : "=d" (__rq.__ll) : "dI" (__nn.__ll), "dI" (d)); \ 1144 1.1 mrg (r) = __rq.__i.__l; (q) = __rq.__i.__h; \ 1145 1.1 mrg } while (0) 1146 1.1 mrg #define count_leading_zeros(count, x) \ 1147 1.1 mrg do { \ 1148 1.1 mrg USItype __cbtmp; \ 1149 1.1 mrg __asm__ ("scanbit %1,%0" : "=r" (__cbtmp) : "r" (x)); \ 1150 1.1 mrg (count) = __cbtmp ^ 31; \ 1151 1.1 mrg } while (0) 1152 1.1 mrg #define COUNT_LEADING_ZEROS_0 (-32) /* sic */ 1153 1.1 mrg #if defined (__i960mx) /* what is the proper symbol to test??? */ 1154 1.1 mrg #define rshift_rhlc(r,h,l,c) \ 1155 1.1 mrg do { \ 1156 1.1 mrg union {UDItype __ll; \ 1157 1.1 mrg struct {USItype __l, __h;} __i; \ 1158 1.1 mrg } __nn; \ 1159 1.1 mrg __nn.__i.__h = (h); __nn.__i.__l = (l); \ 1160 1.1 mrg __asm__ ("shre %2,%1,%0" : "=d" (r) : "dI" (__nn.__ll), "dI" (c)); \ 1161 1.1 mrg } 1162 1.1 mrg #endif /* i960mx */ 1163 1.1 mrg #endif /* i960 */ 1164 1.1 mrg 1165 1.1 mrg #if (defined (__mc68000__) || defined (__mc68020__) || defined(mc68020) \ 1166 1.1 mrg || defined (__m68k__) || defined (__mc5200__) || defined (__mc5206e__) \ 1167 1.1 mrg || defined (__mc5307__)) && W_TYPE_SIZE == 32 1168 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1169 1.1 mrg __asm__ ("add%.l %5,%1\n\taddx%.l %3,%0" \ 1170 1.1 mrg : "=d" (sh), "=&d" (sl) \ 1171 1.1 mrg : "0" ((USItype)(ah)), "d" ((USItype)(bh)), \ 1172 1.1 mrg "%1" ((USItype)(al)), "g" ((USItype)(bl))) 1173 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1174 1.1 mrg __asm__ ("sub%.l %5,%1\n\tsubx%.l %3,%0" \ 1175 1.1 mrg : "=d" (sh), "=&d" (sl) \ 1176 1.1 mrg : "0" ((USItype)(ah)), "d" ((USItype)(bh)), \ 1177 1.1 mrg "1" ((USItype)(al)), "g" ((USItype)(bl))) 1178 1.1 mrg /* The '020, '030, '040 and CPU32 have 32x32->64 and 64/32->32q-32r. */ 1179 1.1 mrg #if defined (__mc68020__) || defined(mc68020) \ 1180 1.1 mrg || defined (__mc68030__) || defined (mc68030) \ 1181 1.1 mrg || defined (__mc68040__) || defined (mc68040) \ 1182 1.1 mrg || defined (__mcpu32__) || defined (mcpu32) \ 1183 1.1 mrg || defined (__NeXT__) 1184 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1185 1.1 mrg __asm__ ("mulu%.l %3,%1:%0" \ 1186 1.1 mrg : "=d" (w0), "=d" (w1) \ 1187 1.1 mrg : "%0" ((USItype)(u)), "dmi" ((USItype)(v))) 1188 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 1189 1.1 mrg __asm__ ("divu%.l %4,%1:%0" \ 1190 1.1 mrg : "=d" (q), "=d" (r) \ 1191 1.1 mrg : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "dmi" ((USItype)(d))) 1192 1.1 mrg #define sdiv_qrnnd(q, r, n1, n0, d) \ 1193 1.1 mrg __asm__ ("divs%.l %4,%1:%0" \ 1194 1.1 mrg : "=d" (q), "=d" (r) \ 1195 1.1 mrg : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "dmi" ((USItype)(d))) 1196 1.1 mrg #else /* for other 68k family members use 16x16->32 multiplication */ 1197 1.1 mrg #define umul_ppmm(xh, xl, a, b) \ 1198 1.1 mrg do { USItype __umul_tmp1, __umul_tmp2; \ 1199 1.1 mrg __asm__ ("| Inlined umul_ppmm\n" \ 1200 1.1 mrg " move%.l %5,%3\n" \ 1201 1.1 mrg " move%.l %2,%0\n" \ 1202 1.1 mrg " move%.w %3,%1\n" \ 1203 1.1 mrg " swap %3\n" \ 1204 1.1 mrg " swap %0\n" \ 1205 1.1 mrg " mulu%.w %2,%1\n" \ 1206 1.1 mrg " mulu%.w %3,%0\n" \ 1207 1.1 mrg " mulu%.w %2,%3\n" \ 1208 1.1 mrg " swap %2\n" \ 1209 1.1 mrg " mulu%.w %5,%2\n" \ 1210 1.1 mrg " add%.l %3,%2\n" \ 1211 1.1 mrg " jcc 1f\n" \ 1212 1.1 mrg " add%.l %#0x10000,%0\n" \ 1213 1.1 mrg "1: move%.l %2,%3\n" \ 1214 1.1 mrg " clr%.w %2\n" \ 1215 1.1 mrg " swap %2\n" \ 1216 1.1 mrg " swap %3\n" \ 1217 1.1 mrg " clr%.w %3\n" \ 1218 1.1 mrg " add%.l %3,%1\n" \ 1219 1.1 mrg " addx%.l %2,%0\n" \ 1220 1.1 mrg " | End inlined umul_ppmm" \ 1221 1.1 mrg : "=&d" (xh), "=&d" (xl), \ 1222 1.1 mrg "=d" (__umul_tmp1), "=&d" (__umul_tmp2) \ 1223 1.1 mrg : "%2" ((USItype)(a)), "d" ((USItype)(b))); \ 1224 1.1 mrg } while (0) 1225 1.1 mrg #endif /* not mc68020 */ 1226 1.1 mrg /* The '020, '030, '040 and '060 have bitfield insns. 1227 1.1 mrg GCC 3.4 defines __mc68020__ when in CPU32 mode, check for __mcpu32__ to 1228 1.1 mrg exclude bfffo on that chip (bitfield insns not available). */ 1229 1.1 mrg #if (defined (__mc68020__) || defined (mc68020) \ 1230 1.1 mrg || defined (__mc68030__) || defined (mc68030) \ 1231 1.1 mrg || defined (__mc68040__) || defined (mc68040) \ 1232 1.1 mrg || defined (__mc68060__) || defined (mc68060) \ 1233 1.3 mrg || defined (__NeXT__)) \ 1234 1.1 mrg && ! defined (__mcpu32__) 1235 1.1 mrg #define count_leading_zeros(count, x) \ 1236 1.1 mrg __asm__ ("bfffo %1{%b2:%b2},%0" \ 1237 1.1 mrg : "=d" (count) \ 1238 1.1 mrg : "od" ((USItype) (x)), "n" (0)) 1239 1.1 mrg #define COUNT_LEADING_ZEROS_0 32 1240 1.1 mrg #endif 1241 1.1 mrg #endif /* mc68000 */ 1242 1.1 mrg 1243 1.1 mrg #if defined (__m88000__) && W_TYPE_SIZE == 32 1244 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1245 1.1 mrg __asm__ ("addu.co %1,%r4,%r5\n\taddu.ci %0,%r2,%r3" \ 1246 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1247 1.1 mrg : "rJ" (ah), "rJ" (bh), "%rJ" (al), "rJ" (bl)) 1248 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1249 1.1 mrg __asm__ ("subu.co %1,%r4,%r5\n\tsubu.ci %0,%r2,%r3" \ 1250 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1251 1.1 mrg : "rJ" (ah), "rJ" (bh), "rJ" (al), "rJ" (bl)) 1252 1.1 mrg #define count_leading_zeros(count, x) \ 1253 1.1 mrg do { \ 1254 1.1 mrg USItype __cbtmp; \ 1255 1.1 mrg __asm__ ("ff1 %0,%1" : "=r" (__cbtmp) : "r" (x)); \ 1256 1.1 mrg (count) = __cbtmp ^ 31; \ 1257 1.1 mrg } while (0) 1258 1.1 mrg #define COUNT_LEADING_ZEROS_0 63 /* sic */ 1259 1.1 mrg #if defined (__m88110__) 1260 1.1 mrg #define umul_ppmm(wh, wl, u, v) \ 1261 1.1 mrg do { \ 1262 1.1 mrg union {UDItype __ll; \ 1263 1.1 mrg struct {USItype __h, __l;} __i; \ 1264 1.1 mrg } __x; \ 1265 1.1 mrg __asm__ ("mulu.d %0,%1,%2" : "=r" (__x.__ll) : "r" (u), "r" (v)); \ 1266 1.1 mrg (wh) = __x.__i.__h; \ 1267 1.1 mrg (wl) = __x.__i.__l; \ 1268 1.1 mrg } while (0) 1269 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 1270 1.1 mrg ({union {UDItype __ll; \ 1271 1.1 mrg struct {USItype __h, __l;} __i; \ 1272 1.1 mrg } __x, __q; \ 1273 1.1 mrg __x.__i.__h = (n1); __x.__i.__l = (n0); \ 1274 1.1 mrg __asm__ ("divu.d %0,%1,%2" \ 1275 1.1 mrg : "=r" (__q.__ll) : "r" (__x.__ll), "r" (d)); \ 1276 1.1 mrg (r) = (n0) - __q.__l * (d); (q) = __q.__l; }) 1277 1.1 mrg #endif /* __m88110__ */ 1278 1.1 mrg #endif /* __m88000__ */ 1279 1.1 mrg 1280 1.1 mrg #if defined (__mips) && W_TYPE_SIZE == 32 1281 1.2 joerg #if __GMP_GNUC_PREREQ (4,4) || defined(__clang__) 1282 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1283 1.1 mrg do { \ 1284 1.1 mrg UDItype __ll = (UDItype)(u) * (v); \ 1285 1.1 mrg w1 = __ll >> 32; \ 1286 1.1 mrg w0 = __ll; \ 1287 1.1 mrg } while (0) 1288 1.1 mrg #endif 1289 1.3 mrg #if !defined (umul_ppmm) && __GMP_GNUC_PREREQ (2,7) && !defined (__clang__) 1290 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1291 1.1 mrg __asm__ ("multu %2,%3" : "=l" (w0), "=h" (w1) : "d" (u), "d" (v)) 1292 1.1 mrg #endif 1293 1.1 mrg #if !defined (umul_ppmm) 1294 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1295 1.1 mrg __asm__ ("multu %2,%3\n\tmflo %0\n\tmfhi %1" \ 1296 1.1 mrg : "=d" (w0), "=d" (w1) : "d" (u), "d" (v)) 1297 1.1 mrg #endif 1298 1.1 mrg #endif /* __mips */ 1299 1.1 mrg 1300 1.1 mrg #if (defined (__mips) && __mips >= 3) && W_TYPE_SIZE == 64 1301 1.4 mrg #if defined (_MIPS_ARCH_MIPS64R6) 1302 1.4 mrg #define umul_ppmm(w1, w0, u, v) \ 1303 1.4 mrg do { \ 1304 1.4 mrg UDItype __m0 = (u), __m1 = (v); \ 1305 1.4 mrg (w0) = __m0 * __m1; \ 1306 1.4 mrg __asm__ ("dmuhu\t%0, %1, %2" : "=d" (w1) : "d" (__m0), "d" (__m1)); \ 1307 1.4 mrg } while (0) 1308 1.4 mrg #endif 1309 1.4 mrg #if !defined (umul_ppmm) && (__GMP_GNUC_PREREQ (4,4) || defined(__clang__)) 1310 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1311 1.1 mrg do { \ 1312 1.1 mrg typedef unsigned int __ll_UTItype __attribute__((mode(TI))); \ 1313 1.1 mrg __ll_UTItype __ll = (__ll_UTItype)(u) * (v); \ 1314 1.1 mrg w1 = __ll >> 64; \ 1315 1.1 mrg w0 = __ll; \ 1316 1.1 mrg } while (0) 1317 1.1 mrg #endif 1318 1.3 mrg #if !defined (umul_ppmm) && __GMP_GNUC_PREREQ (2,7) && !defined (__clang__) 1319 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1320 1.3 mrg __asm__ ("dmultu %2,%3" \ 1321 1.3 mrg : "=l" (w0), "=h" (w1) \ 1322 1.3 mrg : "d" ((UDItype)(u)), "d" ((UDItype)(v))) 1323 1.1 mrg #endif 1324 1.1 mrg #if !defined (umul_ppmm) 1325 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1326 1.1 mrg __asm__ ("dmultu %2,%3\n\tmflo %0\n\tmfhi %1" \ 1327 1.3 mrg : "=d" (w0), "=d" (w1) \ 1328 1.3 mrg : "d" ((UDItype)(u)), "d" ((UDItype)(v))) 1329 1.1 mrg #endif 1330 1.1 mrg #endif /* __mips */ 1331 1.1 mrg 1332 1.1 mrg #if defined (__mmix__) && W_TYPE_SIZE == 64 1333 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1334 1.1 mrg __asm__ ("MULU %0,%2,%3" : "=r" (w0), "=z" (w1) : "r" (u), "r" (v)) 1335 1.1 mrg #endif 1336 1.1 mrg 1337 1.1 mrg #if defined (__ns32000__) && W_TYPE_SIZE == 32 1338 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1339 1.1 mrg ({union {UDItype __ll; \ 1340 1.1 mrg struct {USItype __l, __h;} __i; \ 1341 1.1 mrg } __x; \ 1342 1.1 mrg __asm__ ("meid %2,%0" \ 1343 1.1 mrg : "=g" (__x.__ll) \ 1344 1.1 mrg : "%0" ((USItype)(u)), "g" ((USItype)(v))); \ 1345 1.1 mrg (w1) = __x.__i.__h; (w0) = __x.__i.__l;}) 1346 1.1 mrg #define __umulsidi3(u, v) \ 1347 1.1 mrg ({UDItype __w; \ 1348 1.1 mrg __asm__ ("meid %2,%0" \ 1349 1.1 mrg : "=g" (__w) \ 1350 1.1 mrg : "%0" ((USItype)(u)), "g" ((USItype)(v))); \ 1351 1.1 mrg __w; }) 1352 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 1353 1.1 mrg ({union {UDItype __ll; \ 1354 1.1 mrg struct {USItype __l, __h;} __i; \ 1355 1.1 mrg } __x; \ 1356 1.1 mrg __x.__i.__h = (n1); __x.__i.__l = (n0); \ 1357 1.1 mrg __asm__ ("deid %2,%0" \ 1358 1.1 mrg : "=g" (__x.__ll) \ 1359 1.1 mrg : "0" (__x.__ll), "g" ((USItype)(d))); \ 1360 1.1 mrg (r) = __x.__i.__l; (q) = __x.__i.__h; }) 1361 1.1 mrg #define count_trailing_zeros(count,x) \ 1362 1.1 mrg do { \ 1363 1.1 mrg __asm__ ("ffsd %2,%0" \ 1364 1.1 mrg : "=r" (count) \ 1365 1.1 mrg : "0" ((USItype) 0), "r" ((USItype) (x))); \ 1366 1.1 mrg } while (0) 1367 1.1 mrg #endif /* __ns32000__ */ 1368 1.1 mrg 1369 1.1 mrg /* In the past we had a block of various #defines tested 1370 1.1 mrg _ARCH_PPC - AIX 1371 1.1 mrg _ARCH_PWR - AIX 1372 1.1 mrg __powerpc__ - gcc 1373 1.1 mrg __POWERPC__ - BEOS 1374 1.1 mrg __ppc__ - Darwin 1375 1.1 mrg PPC - old gcc, GNU/Linux, SysV 1376 1.1 mrg The plain PPC test was not good for vxWorks, since PPC is defined on all 1377 1.1 mrg CPUs there (eg. m68k too), as a constant one is expected to compare 1378 1.1 mrg CPU_FAMILY against. 1379 1.1 mrg 1380 1.1 mrg At any rate, this was pretty unattractive and a bit fragile. The use of 1381 1.1 mrg HAVE_HOST_CPU_FAMILY is designed to cut through it all and be sure of 1382 1.1 mrg getting the desired effect. 1383 1.1 mrg 1384 1.1 mrg ENHANCE-ME: We should test _IBMR2 here when we add assembly support for 1385 1.1 mrg the system vendor compilers. (Is that vendor compilers with inline asm, 1386 1.1 mrg or what?) */ 1387 1.1 mrg 1388 1.3 mrg #if (HAVE_HOST_CPU_FAMILY_power || HAVE_HOST_CPU_FAMILY_powerpc) \ 1389 1.1 mrg && W_TYPE_SIZE == 32 1390 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1391 1.1 mrg do { \ 1392 1.1 mrg if (__builtin_constant_p (bh) && (bh) == 0) \ 1393 1.3 mrg __asm__ ("add%I4c %1,%3,%4\n\taddze %0,%2" \ 1394 1.4 mrg : "=r" (sh), "=&r" (sl) : "r" (ah), "%r" (al), "rI" (bl) \ 1395 1.4 mrg __CLOBBER_CC); \ 1396 1.1 mrg else if (__builtin_constant_p (bh) && (bh) == ~(USItype) 0) \ 1397 1.3 mrg __asm__ ("add%I4c %1,%3,%4\n\taddme %0,%2" \ 1398 1.4 mrg : "=r" (sh), "=&r" (sl) : "r" (ah), "%r" (al), "rI" (bl) \ 1399 1.4 mrg __CLOBBER_CC); \ 1400 1.1 mrg else \ 1401 1.3 mrg __asm__ ("add%I5c %1,%4,%5\n\tadde %0,%2,%3" \ 1402 1.4 mrg : "=r" (sh), "=&r" (sl) \ 1403 1.4 mrg : "r" (ah), "r" (bh), "%r" (al), "rI" (bl) \ 1404 1.4 mrg __CLOBBER_CC); \ 1405 1.1 mrg } while (0) 1406 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1407 1.1 mrg do { \ 1408 1.1 mrg if (__builtin_constant_p (ah) && (ah) == 0) \ 1409 1.3 mrg __asm__ ("subf%I3c %1,%4,%3\n\tsubfze %0,%2" \ 1410 1.4 mrg : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl) \ 1411 1.4 mrg __CLOBBER_CC); \ 1412 1.1 mrg else if (__builtin_constant_p (ah) && (ah) == ~(USItype) 0) \ 1413 1.3 mrg __asm__ ("subf%I3c %1,%4,%3\n\tsubfme %0,%2" \ 1414 1.4 mrg : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl) \ 1415 1.4 mrg __CLOBBER_CC); \ 1416 1.1 mrg else if (__builtin_constant_p (bh) && (bh) == 0) \ 1417 1.3 mrg __asm__ ("subf%I3c %1,%4,%3\n\taddme %0,%2" \ 1418 1.4 mrg : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "r" (bl) \ 1419 1.4 mrg __CLOBBER_CC); \ 1420 1.1 mrg else if (__builtin_constant_p (bh) && (bh) == ~(USItype) 0) \ 1421 1.3 mrg __asm__ ("subf%I3c %1,%4,%3\n\taddze %0,%2" \ 1422 1.4 mrg : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "r" (bl) \ 1423 1.4 mrg __CLOBBER_CC); \ 1424 1.1 mrg else \ 1425 1.3 mrg __asm__ ("subf%I4c %1,%5,%4\n\tsubfe %0,%3,%2" \ 1426 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1427 1.4 mrg : "r" (ah), "r" (bh), "rI" (al), "r" (bl) \ 1428 1.4 mrg __CLOBBER_CC); \ 1429 1.1 mrg } while (0) 1430 1.1 mrg #define count_leading_zeros(count, x) \ 1431 1.2 joerg __asm__ ("cntlzw %0,%1" : "=r" (count) : "r" (x)) 1432 1.1 mrg #define COUNT_LEADING_ZEROS_0 32 1433 1.1 mrg #if HAVE_HOST_CPU_FAMILY_powerpc 1434 1.2 joerg #if __GMP_GNUC_PREREQ (4,4) || defined(__clang__) 1435 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1436 1.1 mrg do { \ 1437 1.1 mrg UDItype __ll = (UDItype)(u) * (v); \ 1438 1.1 mrg w1 = __ll >> 32; \ 1439 1.1 mrg w0 = __ll; \ 1440 1.1 mrg } while (0) 1441 1.1 mrg #endif 1442 1.1 mrg #if !defined (umul_ppmm) 1443 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \ 1444 1.1 mrg do { \ 1445 1.1 mrg USItype __m0 = (m0), __m1 = (m1); \ 1446 1.1 mrg __asm__ ("mulhwu %0,%1,%2" : "=r" (ph) : "%r" (m0), "r" (m1)); \ 1447 1.1 mrg (pl) = __m0 * __m1; \ 1448 1.1 mrg } while (0) 1449 1.1 mrg #endif 1450 1.1 mrg #define smul_ppmm(ph, pl, m0, m1) \ 1451 1.1 mrg do { \ 1452 1.1 mrg SItype __m0 = (m0), __m1 = (m1); \ 1453 1.1 mrg __asm__ ("mulhw %0,%1,%2" : "=r" (ph) : "%r" (m0), "r" (m1)); \ 1454 1.1 mrg (pl) = __m0 * __m1; \ 1455 1.1 mrg } while (0) 1456 1.1 mrg #else 1457 1.1 mrg #define smul_ppmm(xh, xl, m0, m1) \ 1458 1.1 mrg __asm__ ("mul %0,%2,%3" : "=r" (xh), "=q" (xl) : "r" (m0), "r" (m1)) 1459 1.1 mrg #define sdiv_qrnnd(q, r, nh, nl, d) \ 1460 1.1 mrg __asm__ ("div %0,%2,%4" : "=r" (q), "=q" (r) : "r" (nh), "1" (nl), "r" (d)) 1461 1.1 mrg #endif 1462 1.1 mrg #endif /* 32-bit POWER architecture variants. */ 1463 1.1 mrg 1464 1.1 mrg /* We should test _IBMR2 here when we add assembly support for the system 1465 1.1 mrg vendor compilers. */ 1466 1.1 mrg #if HAVE_HOST_CPU_FAMILY_powerpc && W_TYPE_SIZE == 64 1467 1.1 mrg #if !defined (_LONG_LONG_LIMB) 1468 1.1 mrg /* _LONG_LONG_LIMB is ABI=mode32 where adde operates on 32-bit values. So 1469 1.1 mrg use adde etc only when not _LONG_LONG_LIMB. */ 1470 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1471 1.1 mrg do { \ 1472 1.1 mrg if (__builtin_constant_p (bh) && (bh) == 0) \ 1473 1.3 mrg __asm__ ("add%I4c %1,%3,%4\n\taddze %0,%2" \ 1474 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1475 1.3 mrg : "r" ((UDItype)(ah)), \ 1476 1.4 mrg "%r" ((UDItype)(al)), "rI" ((UDItype)(bl)) \ 1477 1.4 mrg __CLOBBER_CC); \ 1478 1.1 mrg else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0) \ 1479 1.3 mrg __asm__ ("add%I4c %1,%3,%4\n\taddme %0,%2" \ 1480 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1481 1.3 mrg : "r" ((UDItype)(ah)), \ 1482 1.4 mrg "%r" ((UDItype)(al)), "rI" ((UDItype)(bl)) \ 1483 1.4 mrg __CLOBBER_CC); \ 1484 1.1 mrg else \ 1485 1.3 mrg __asm__ ("add%I5c %1,%4,%5\n\tadde %0,%2,%3" \ 1486 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1487 1.3 mrg : "r" ((UDItype)(ah)), "r" ((UDItype)(bh)), \ 1488 1.4 mrg "%r" ((UDItype)(al)), "rI" ((UDItype)(bl)) \ 1489 1.4 mrg __CLOBBER_CC); \ 1490 1.1 mrg } while (0) 1491 1.1 mrg /* We use "*rI" for the constant operand here, since with just "I", gcc barfs. 1492 1.1 mrg This might seem strange, but gcc folds away the dead code late. */ 1493 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1494 1.3 mrg do { \ 1495 1.5 mrg if (__builtin_constant_p (bl) \ 1496 1.5 mrg && (bl) > -0x8000 && (bl) <= 0x8000 && (bl) != 0) { \ 1497 1.3 mrg if (__builtin_constant_p (ah) && (ah) == 0) \ 1498 1.3 mrg __asm__ ("addic %1,%3,%4\n\tsubfze %0,%2" \ 1499 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1500 1.3 mrg : "r" ((UDItype)(bh)), \ 1501 1.5 mrg "r" ((UDItype)(al)), "*rI" (-((UDItype)(bl))) \ 1502 1.4 mrg __CLOBBER_CC); \ 1503 1.3 mrg else if (__builtin_constant_p (ah) && (ah) == ~(UDItype) 0) \ 1504 1.3 mrg __asm__ ("addic %1,%3,%4\n\tsubfme %0,%2" \ 1505 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1506 1.3 mrg : "r" ((UDItype)(bh)), \ 1507 1.5 mrg "r" ((UDItype)(al)), "*rI" (-((UDItype)(bl))) \ 1508 1.4 mrg __CLOBBER_CC); \ 1509 1.3 mrg else if (__builtin_constant_p (bh) && (bh) == 0) \ 1510 1.3 mrg __asm__ ("addic %1,%3,%4\n\taddme %0,%2" \ 1511 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1512 1.5 mrg : "r" ((UDItype)(ah)), \ 1513 1.5 mrg "r" ((UDItype)(al)), "*rI" (-((UDItype)(bl))) \ 1514 1.4 mrg __CLOBBER_CC); \ 1515 1.3 mrg else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0) \ 1516 1.3 mrg __asm__ ("addic %1,%3,%4\n\taddze %0,%2" \ 1517 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1518 1.5 mrg : "r" ((UDItype)(ah)), \ 1519 1.5 mrg "r" ((UDItype)(al)), "*rI" (-((UDItype)(bl))) \ 1520 1.4 mrg __CLOBBER_CC); \ 1521 1.3 mrg else \ 1522 1.3 mrg __asm__ ("addic %1,%4,%5\n\tsubfe %0,%3,%2" \ 1523 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1524 1.5 mrg : "r" ((UDItype)(ah)), "r" ((UDItype)(bh)), \ 1525 1.5 mrg "r" ((UDItype)(al)), "*rI" (-((UDItype)(bl))) \ 1526 1.4 mrg __CLOBBER_CC); \ 1527 1.3 mrg } else { \ 1528 1.3 mrg if (__builtin_constant_p (ah) && (ah) == 0) \ 1529 1.3 mrg __asm__ ("subf%I3c %1,%4,%3\n\tsubfze %0,%2" \ 1530 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1531 1.3 mrg : "r" ((UDItype)(bh)), \ 1532 1.4 mrg "rI" ((UDItype)(al)), "r" ((UDItype)(bl)) \ 1533 1.4 mrg __CLOBBER_CC); \ 1534 1.3 mrg else if (__builtin_constant_p (ah) && (ah) == ~(UDItype) 0) \ 1535 1.3 mrg __asm__ ("subf%I3c %1,%4,%3\n\tsubfme %0,%2" \ 1536 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1537 1.3 mrg : "r" ((UDItype)(bh)), \ 1538 1.4 mrg "rI" ((UDItype)(al)), "r" ((UDItype)(bl)) \ 1539 1.4 mrg __CLOBBER_CC); \ 1540 1.3 mrg else if (__builtin_constant_p (bh) && (bh) == 0) \ 1541 1.3 mrg __asm__ ("subf%I3c %1,%4,%3\n\taddme %0,%2" \ 1542 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1543 1.3 mrg : "r" ((UDItype)(ah)), \ 1544 1.4 mrg "rI" ((UDItype)(al)), "r" ((UDItype)(bl)) \ 1545 1.4 mrg __CLOBBER_CC); \ 1546 1.3 mrg else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0) \ 1547 1.3 mrg __asm__ ("subf%I3c %1,%4,%3\n\taddze %0,%2" \ 1548 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1549 1.3 mrg : "r" ((UDItype)(ah)), \ 1550 1.4 mrg "rI" ((UDItype)(al)), "r" ((UDItype)(bl)) \ 1551 1.4 mrg __CLOBBER_CC); \ 1552 1.3 mrg else \ 1553 1.3 mrg __asm__ ("subf%I4c %1,%5,%4\n\tsubfe %0,%3,%2" \ 1554 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1555 1.3 mrg : "r" ((UDItype)(ah)), "r" ((UDItype)(bh)), \ 1556 1.4 mrg "rI" ((UDItype)(al)), "r" ((UDItype)(bl)) \ 1557 1.4 mrg __CLOBBER_CC); \ 1558 1.3 mrg } \ 1559 1.1 mrg } while (0) 1560 1.1 mrg #endif /* ! _LONG_LONG_LIMB */ 1561 1.1 mrg #define count_leading_zeros(count, x) \ 1562 1.1 mrg __asm__ ("cntlzd %0,%1" : "=r" (count) : "r" (x)) 1563 1.1 mrg #define COUNT_LEADING_ZEROS_0 64 1564 1.4 mrg #if __GMP_GNUC_PREREQ (4,8) 1565 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1566 1.1 mrg do { \ 1567 1.1 mrg typedef unsigned int __ll_UTItype __attribute__((mode(TI))); \ 1568 1.1 mrg __ll_UTItype __ll = (__ll_UTItype)(u) * (v); \ 1569 1.1 mrg w1 = __ll >> 64; \ 1570 1.1 mrg w0 = __ll; \ 1571 1.1 mrg } while (0) 1572 1.1 mrg #endif 1573 1.1 mrg #if !defined (umul_ppmm) 1574 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \ 1575 1.1 mrg do { \ 1576 1.1 mrg UDItype __m0 = (m0), __m1 = (m1); \ 1577 1.3 mrg __asm__ ("mulhdu %0,%1,%2" : "=r" (ph) : "%r" (__m0), "r" (__m1)); \ 1578 1.1 mrg (pl) = __m0 * __m1; \ 1579 1.1 mrg } while (0) 1580 1.1 mrg #endif 1581 1.1 mrg #define smul_ppmm(ph, pl, m0, m1) \ 1582 1.1 mrg do { \ 1583 1.1 mrg DItype __m0 = (m0), __m1 = (m1); \ 1584 1.3 mrg __asm__ ("mulhd %0,%1,%2" : "=r" (ph) : "%r" (__m0), "r" (__m1)); \ 1585 1.1 mrg (pl) = __m0 * __m1; \ 1586 1.1 mrg } while (0) 1587 1.1 mrg #endif /* 64-bit PowerPC. */ 1588 1.1 mrg 1589 1.1 mrg #if defined (__pyr__) && W_TYPE_SIZE == 32 1590 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1591 1.1 mrg __asm__ ("addw %5,%1\n\taddwc %3,%0" \ 1592 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1593 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 1594 1.1 mrg "%1" ((USItype)(al)), "g" ((USItype)(bl))) 1595 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1596 1.1 mrg __asm__ ("subw %5,%1\n\tsubwb %3,%0" \ 1597 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1598 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 1599 1.1 mrg "1" ((USItype)(al)), "g" ((USItype)(bl))) 1600 1.1 mrg /* This insn works on Pyramids with AP, XP, or MI CPUs, but not with SP. */ 1601 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1602 1.1 mrg ({union {UDItype __ll; \ 1603 1.1 mrg struct {USItype __h, __l;} __i; \ 1604 1.1 mrg } __x; \ 1605 1.1 mrg __asm__ ("movw %1,%R0\n\tuemul %2,%0" \ 1606 1.1 mrg : "=&r" (__x.__ll) \ 1607 1.1 mrg : "g" ((USItype) (u)), "g" ((USItype)(v))); \ 1608 1.1 mrg (w1) = __x.__i.__h; (w0) = __x.__i.__l;}) 1609 1.1 mrg #endif /* __pyr__ */ 1610 1.1 mrg 1611 1.1 mrg #if defined (__ibm032__) /* RT/ROMP */ && W_TYPE_SIZE == 32 1612 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1613 1.1 mrg __asm__ ("a %1,%5\n\tae %0,%3" \ 1614 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1615 1.1 mrg : "0" ((USItype)(ah)), "r" ((USItype)(bh)), \ 1616 1.1 mrg "%1" ((USItype)(al)), "r" ((USItype)(bl))) 1617 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1618 1.1 mrg __asm__ ("s %1,%5\n\tse %0,%3" \ 1619 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1620 1.1 mrg : "0" ((USItype)(ah)), "r" ((USItype)(bh)), \ 1621 1.1 mrg "1" ((USItype)(al)), "r" ((USItype)(bl))) 1622 1.1 mrg #define smul_ppmm(ph, pl, m0, m1) \ 1623 1.1 mrg __asm__ ( \ 1624 1.1 mrg "s r2,r2\n" \ 1625 1.1 mrg " mts r10,%2\n" \ 1626 1.1 mrg " m r2,%3\n" \ 1627 1.1 mrg " m r2,%3\n" \ 1628 1.1 mrg " m r2,%3\n" \ 1629 1.1 mrg " m r2,%3\n" \ 1630 1.1 mrg " m r2,%3\n" \ 1631 1.1 mrg " m r2,%3\n" \ 1632 1.1 mrg " m r2,%3\n" \ 1633 1.1 mrg " m r2,%3\n" \ 1634 1.1 mrg " m r2,%3\n" \ 1635 1.1 mrg " m r2,%3\n" \ 1636 1.1 mrg " m r2,%3\n" \ 1637 1.1 mrg " m r2,%3\n" \ 1638 1.1 mrg " m r2,%3\n" \ 1639 1.1 mrg " m r2,%3\n" \ 1640 1.1 mrg " m r2,%3\n" \ 1641 1.1 mrg " m r2,%3\n" \ 1642 1.1 mrg " cas %0,r2,r0\n" \ 1643 1.1 mrg " mfs r10,%1" \ 1644 1.1 mrg : "=r" (ph), "=r" (pl) \ 1645 1.1 mrg : "%r" ((USItype)(m0)), "r" ((USItype)(m1)) \ 1646 1.1 mrg : "r2") 1647 1.1 mrg #define count_leading_zeros(count, x) \ 1648 1.1 mrg do { \ 1649 1.1 mrg if ((x) >= 0x10000) \ 1650 1.1 mrg __asm__ ("clz %0,%1" \ 1651 1.1 mrg : "=r" (count) : "r" ((USItype)(x) >> 16)); \ 1652 1.1 mrg else \ 1653 1.1 mrg { \ 1654 1.1 mrg __asm__ ("clz %0,%1" \ 1655 1.1 mrg : "=r" (count) : "r" ((USItype)(x))); \ 1656 1.1 mrg (count) += 16; \ 1657 1.1 mrg } \ 1658 1.1 mrg } while (0) 1659 1.1 mrg #endif /* RT/ROMP */ 1660 1.1 mrg 1661 1.4 mrg #if defined (__riscv64) && W_TYPE_SIZE == 64 1662 1.4 mrg #define umul_ppmm(ph, pl, u, v) \ 1663 1.4 mrg do { \ 1664 1.4 mrg UDItype __u = (u), __v = (v); \ 1665 1.4 mrg (pl) = __u * __v; \ 1666 1.4 mrg __asm__ ("mulhu\t%2, %1, %0" : "=r" (ph) : "%r" (__u), "r" (__v)); \ 1667 1.4 mrg } while (0) 1668 1.4 mrg #endif 1669 1.4 mrg 1670 1.2 joerg #if (defined (__SH2__) || defined (__SH3__) || defined (__SH4__)) && W_TYPE_SIZE == 32 1671 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1672 1.1 mrg __asm__ ("dmulu.l %2,%3\n\tsts macl,%1\n\tsts mach,%0" \ 1673 1.1 mrg : "=r" (w1), "=r" (w0) : "r" (u), "r" (v) : "macl", "mach") 1674 1.1 mrg #endif 1675 1.1 mrg 1676 1.1 mrg #if defined (__sparc__) && W_TYPE_SIZE == 32 1677 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1678 1.1 mrg __asm__ ("addcc %r4,%5,%1\n\taddx %r2,%3,%0" \ 1679 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1680 1.1 mrg : "rJ" (ah), "rI" (bh),"%rJ" (al), "rI" (bl) \ 1681 1.1 mrg __CLOBBER_CC) 1682 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1683 1.1 mrg __asm__ ("subcc %r4,%5,%1\n\tsubx %r2,%3,%0" \ 1684 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1685 1.1 mrg : "rJ" (ah), "rI" (bh), "rJ" (al), "rI" (bl) \ 1686 1.1 mrg __CLOBBER_CC) 1687 1.1 mrg /* FIXME: When gcc -mcpu=v9 is used on solaris, gcc/config/sol2-sld-64.h 1688 1.1 mrg doesn't define anything to indicate that to us, it only sets __sparcv8. */ 1689 1.1 mrg #if defined (__sparc_v9__) || defined (__sparcv9) 1690 1.1 mrg /* Perhaps we should use floating-point operations here? */ 1691 1.1 mrg #if 0 1692 1.1 mrg /* Triggers a bug making mpz/tests/t-gcd.c fail. 1693 1.1 mrg Perhaps we simply need explicitly zero-extend the inputs? */ 1694 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1695 1.1 mrg __asm__ ("mulx %2,%3,%%g1; srl %%g1,0,%1; srlx %%g1,32,%0" : \ 1696 1.1 mrg "=r" (w1), "=r" (w0) : "r" (u), "r" (v) : "g1") 1697 1.1 mrg #else 1698 1.1 mrg /* Use v8 umul until above bug is fixed. */ 1699 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1700 1.1 mrg __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v)) 1701 1.1 mrg #endif 1702 1.1 mrg /* Use a plain v8 divide for v9. */ 1703 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 1704 1.1 mrg do { \ 1705 1.1 mrg USItype __q; \ 1706 1.1 mrg __asm__ ("mov %1,%%y;nop;nop;nop;udiv %2,%3,%0" \ 1707 1.1 mrg : "=r" (__q) : "r" (n1), "r" (n0), "r" (d)); \ 1708 1.1 mrg (r) = (n0) - __q * (d); \ 1709 1.1 mrg (q) = __q; \ 1710 1.1 mrg } while (0) 1711 1.1 mrg #else 1712 1.1 mrg #if defined (__sparc_v8__) /* gcc normal */ \ 1713 1.1 mrg || defined (__sparcv8) /* gcc solaris */ \ 1714 1.1 mrg || HAVE_HOST_CPU_supersparc 1715 1.1 mrg /* Don't match immediate range because, 1) it is not often useful, 1716 1.1 mrg 2) the 'I' flag thinks of the range as a 13 bit signed interval, 1717 1.1 mrg while we want to match a 13 bit interval, sign extended to 32 bits, 1718 1.1 mrg but INTERPRETED AS UNSIGNED. */ 1719 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1720 1.1 mrg __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v)) 1721 1.1 mrg 1722 1.1 mrg #if HAVE_HOST_CPU_supersparc 1723 1.1 mrg #else 1724 1.1 mrg /* Don't use this on SuperSPARC because its udiv only handles 53 bit 1725 1.1 mrg dividends and will trap to the kernel for the rest. */ 1726 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 1727 1.1 mrg do { \ 1728 1.1 mrg USItype __q; \ 1729 1.1 mrg __asm__ ("mov %1,%%y;nop;nop;nop;udiv %2,%3,%0" \ 1730 1.1 mrg : "=r" (__q) : "r" (n1), "r" (n0), "r" (d)); \ 1731 1.1 mrg (r) = (n0) - __q * (d); \ 1732 1.1 mrg (q) = __q; \ 1733 1.1 mrg } while (0) 1734 1.1 mrg #endif /* HAVE_HOST_CPU_supersparc */ 1735 1.1 mrg 1736 1.1 mrg #else /* ! __sparc_v8__ */ 1737 1.1 mrg #if defined (__sparclite__) 1738 1.1 mrg /* This has hardware multiply but not divide. It also has two additional 1739 1.1 mrg instructions scan (ffs from high bit) and divscc. */ 1740 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1741 1.1 mrg __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v)) 1742 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 1743 1.1 mrg __asm__ ("! Inlined udiv_qrnnd\n" \ 1744 1.1 mrg " wr %%g0,%2,%%y ! Not a delayed write for sparclite\n" \ 1745 1.1 mrg " tst %%g0\n" \ 1746 1.1 mrg " divscc %3,%4,%%g1\n" \ 1747 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1748 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1749 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1750 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1751 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1752 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1753 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1754 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1755 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1756 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1757 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1758 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1759 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1760 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1761 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1762 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1763 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1764 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1765 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1766 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1767 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1768 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1769 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1770 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1771 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1772 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1773 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1774 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1775 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1776 1.1 mrg " divscc %%g1,%4,%%g1\n" \ 1777 1.1 mrg " divscc %%g1,%4,%0\n" \ 1778 1.1 mrg " rd %%y,%1\n" \ 1779 1.1 mrg " bl,a 1f\n" \ 1780 1.1 mrg " add %1,%4,%1\n" \ 1781 1.1 mrg "1: ! End of inline udiv_qrnnd" \ 1782 1.1 mrg : "=r" (q), "=r" (r) : "r" (n1), "r" (n0), "rI" (d) \ 1783 1.1 mrg : "%g1" __AND_CLOBBER_CC) 1784 1.1 mrg #define count_leading_zeros(count, x) \ 1785 1.1 mrg __asm__ ("scan %1,1,%0" : "=r" (count) : "r" (x)) 1786 1.1 mrg /* Early sparclites return 63 for an argument of 0, but they warn that future 1787 1.1 mrg implementations might change this. Therefore, leave COUNT_LEADING_ZEROS_0 1788 1.1 mrg undefined. */ 1789 1.1 mrg #endif /* __sparclite__ */ 1790 1.1 mrg #endif /* __sparc_v8__ */ 1791 1.1 mrg #endif /* __sparc_v9__ */ 1792 1.1 mrg /* Default to sparc v7 versions of umul_ppmm and udiv_qrnnd. */ 1793 1.1 mrg #ifndef umul_ppmm 1794 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 1795 1.1 mrg __asm__ ("! Inlined umul_ppmm\n" \ 1796 1.1 mrg " wr %%g0,%2,%%y ! SPARC has 0-3 delay insn after a wr\n" \ 1797 1.1 mrg " sra %3,31,%%g2 ! Don't move this insn\n" \ 1798 1.1 mrg " and %2,%%g2,%%g2 ! Don't move this insn\n" \ 1799 1.1 mrg " andcc %%g0,0,%%g1 ! Don't move this insn\n" \ 1800 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1801 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1802 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1803 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1804 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1805 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1806 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1807 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1808 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1809 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1810 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1811 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1812 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1813 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1814 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1815 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1816 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1817 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1818 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1819 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1820 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1821 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1822 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1823 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1824 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1825 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1826 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1827 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1828 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1829 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1830 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1831 1.1 mrg " mulscc %%g1,%3,%%g1\n" \ 1832 1.1 mrg " mulscc %%g1,0,%%g1\n" \ 1833 1.1 mrg " add %%g1,%%g2,%0\n" \ 1834 1.1 mrg " rd %%y,%1" \ 1835 1.1 mrg : "=r" (w1), "=r" (w0) : "%rI" (u), "r" (v) \ 1836 1.1 mrg : "%g1", "%g2" __AND_CLOBBER_CC) 1837 1.1 mrg #endif 1838 1.1 mrg #ifndef udiv_qrnnd 1839 1.1 mrg #ifndef LONGLONG_STANDALONE 1840 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 1841 1.1 mrg do { UWtype __r; \ 1842 1.1 mrg (q) = __MPN(udiv_qrnnd) (&__r, (n1), (n0), (d)); \ 1843 1.1 mrg (r) = __r; \ 1844 1.1 mrg } while (0) 1845 1.2 joerg extern UWtype __MPN(udiv_qrnnd) (UWtype *, UWtype, UWtype, UWtype); 1846 1.1 mrg #endif /* LONGLONG_STANDALONE */ 1847 1.1 mrg #endif /* udiv_qrnnd */ 1848 1.1 mrg #endif /* __sparc__ */ 1849 1.1 mrg 1850 1.1 mrg #if defined (__sparc__) && W_TYPE_SIZE == 64 1851 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1852 1.1 mrg __asm__ ( \ 1853 1.1 mrg "addcc %r4,%5,%1\n" \ 1854 1.1 mrg " addccc %r6,%7,%%g0\n" \ 1855 1.1 mrg " addc %r2,%3,%0" \ 1856 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1857 1.3 mrg : "rJ" ((UDItype)(ah)), "rI" ((UDItype)(bh)), \ 1858 1.3 mrg "%rJ" ((UDItype)(al)), "rI" ((UDItype)(bl)), \ 1859 1.3 mrg "%rJ" ((UDItype)(al) >> 32), "rI" ((UDItype)(bl) >> 32) \ 1860 1.1 mrg __CLOBBER_CC) 1861 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1862 1.1 mrg __asm__ ( \ 1863 1.1 mrg "subcc %r4,%5,%1\n" \ 1864 1.1 mrg " subccc %r6,%7,%%g0\n" \ 1865 1.1 mrg " subc %r2,%3,%0" \ 1866 1.3 mrg : "=r" (sh), "=&r" (sl) \ 1867 1.3 mrg : "rJ" ((UDItype)(ah)), "rI" ((UDItype)(bh)), \ 1868 1.3 mrg "rJ" ((UDItype)(al)), "rI" ((UDItype)(bl)), \ 1869 1.3 mrg "rJ" ((UDItype)(al) >> 32), "rI" ((UDItype)(bl) >> 32) \ 1870 1.3 mrg __CLOBBER_CC) 1871 1.3 mrg #if __VIS__ >= 0x300 1872 1.3 mrg #undef add_ssaaaa 1873 1.3 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1874 1.3 mrg __asm__ ( \ 1875 1.3 mrg "addcc %r4, %5, %1\n" \ 1876 1.3 mrg " addxc %r2, %r3, %0" \ 1877 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1878 1.3 mrg : "rJ" ((UDItype)(ah)), "rJ" ((UDItype)(bh)), \ 1879 1.3 mrg "%rJ" ((UDItype)(al)), "rI" ((UDItype)(bl)) __CLOBBER_CC) 1880 1.3 mrg #define umul_ppmm(ph, pl, m0, m1) \ 1881 1.3 mrg do { \ 1882 1.3 mrg UDItype __m0 = (m0), __m1 = (m1); \ 1883 1.3 mrg (pl) = __m0 * __m1; \ 1884 1.3 mrg __asm__ ("umulxhi\t%2, %1, %0" \ 1885 1.3 mrg : "=r" (ph) \ 1886 1.3 mrg : "%r" (__m0), "r" (__m1)); \ 1887 1.3 mrg } while (0) 1888 1.3 mrg #define count_leading_zeros(count, x) \ 1889 1.3 mrg __asm__ ("lzd\t%1,%0" : "=r" (count) : "r" (x)) 1890 1.3 mrg /* Needed by count_leading_zeros_32 in sparc64.h. */ 1891 1.3 mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB 1892 1.3 mrg #endif 1893 1.1 mrg #endif 1894 1.1 mrg 1895 1.2 joerg #if (defined (__vax) || defined (__vax__)) && W_TYPE_SIZE == 32 1896 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1897 1.1 mrg __asm__ ("addl2 %5,%1\n\tadwc %3,%0" \ 1898 1.1 mrg : "=g" (sh), "=&g" (sl) \ 1899 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 1900 1.1 mrg "%1" ((USItype)(al)), "g" ((USItype)(bl))) 1901 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1902 1.1 mrg __asm__ ("subl2 %5,%1\n\tsbwc %3,%0" \ 1903 1.1 mrg : "=g" (sh), "=&g" (sl) \ 1904 1.1 mrg : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 1905 1.1 mrg "1" ((USItype)(al)), "g" ((USItype)(bl))) 1906 1.1 mrg #define smul_ppmm(xh, xl, m0, m1) \ 1907 1.1 mrg do { \ 1908 1.1 mrg union {UDItype __ll; \ 1909 1.1 mrg struct {USItype __l, __h;} __i; \ 1910 1.1 mrg } __x; \ 1911 1.1 mrg USItype __m0 = (m0), __m1 = (m1); \ 1912 1.1 mrg __asm__ ("emul %1,%2,$0,%0" \ 1913 1.1 mrg : "=g" (__x.__ll) : "g" (__m0), "g" (__m1)); \ 1914 1.1 mrg (xh) = __x.__i.__h; (xl) = __x.__i.__l; \ 1915 1.1 mrg } while (0) 1916 1.1 mrg #define sdiv_qrnnd(q, r, n1, n0, d) \ 1917 1.1 mrg do { \ 1918 1.1 mrg union {DItype __ll; \ 1919 1.1 mrg struct {SItype __l, __h;} __i; \ 1920 1.1 mrg } __x; \ 1921 1.1 mrg __x.__i.__h = n1; __x.__i.__l = n0; \ 1922 1.1 mrg __asm__ ("ediv %3,%2,%0,%1" \ 1923 1.1 mrg : "=g" (q), "=g" (r) : "g" (__x.__ll), "g" (d)); \ 1924 1.1 mrg } while (0) 1925 1.1 mrg #if 0 1926 1.1 mrg /* FIXME: This instruction appears to be unimplemented on some systems (vax 1927 1.1 mrg 8800 maybe). */ 1928 1.1 mrg #define count_trailing_zeros(count,x) \ 1929 1.1 mrg do { \ 1930 1.1 mrg __asm__ ("ffs 0, 31, %1, %0" \ 1931 1.1 mrg : "=g" (count) \ 1932 1.1 mrg : "g" ((USItype) (x))); \ 1933 1.1 mrg } while (0) 1934 1.1 mrg #endif 1935 1.2 joerg #endif /* vax */ 1936 1.1 mrg 1937 1.1 mrg #if defined (__z8000__) && W_TYPE_SIZE == 16 1938 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1939 1.1 mrg __asm__ ("add %H1,%H5\n\tadc %H0,%H3" \ 1940 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1941 1.1 mrg : "0" ((unsigned int)(ah)), "r" ((unsigned int)(bh)), \ 1942 1.1 mrg "%1" ((unsigned int)(al)), "rQR" ((unsigned int)(bl))) 1943 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1944 1.1 mrg __asm__ ("sub %H1,%H5\n\tsbc %H0,%H3" \ 1945 1.1 mrg : "=r" (sh), "=&r" (sl) \ 1946 1.1 mrg : "0" ((unsigned int)(ah)), "r" ((unsigned int)(bh)), \ 1947 1.1 mrg "1" ((unsigned int)(al)), "rQR" ((unsigned int)(bl))) 1948 1.1 mrg #define umul_ppmm(xh, xl, m0, m1) \ 1949 1.1 mrg do { \ 1950 1.1 mrg union {long int __ll; \ 1951 1.1 mrg struct {unsigned int __h, __l;} __i; \ 1952 1.1 mrg } __x; \ 1953 1.1 mrg unsigned int __m0 = (m0), __m1 = (m1); \ 1954 1.1 mrg __asm__ ("mult %S0,%H3" \ 1955 1.1 mrg : "=r" (__x.__i.__h), "=r" (__x.__i.__l) \ 1956 1.1 mrg : "%1" (m0), "rQR" (m1)); \ 1957 1.1 mrg (xh) = __x.__i.__h; (xl) = __x.__i.__l; \ 1958 1.1 mrg (xh) += ((((signed int) __m0 >> 15) & __m1) \ 1959 1.1 mrg + (((signed int) __m1 >> 15) & __m0)); \ 1960 1.1 mrg } while (0) 1961 1.1 mrg #endif /* __z8000__ */ 1962 1.1 mrg 1963 1.1 mrg #endif /* __GNUC__ */ 1964 1.1 mrg 1965 1.1 mrg #endif /* NO_ASM */ 1966 1.1 mrg 1967 1.1 mrg 1968 1.2 joerg /* FIXME: "sidi" here is highly doubtful, should sometimes be "diti". */ 1969 1.1 mrg #if !defined (umul_ppmm) && defined (__umulsidi3) 1970 1.1 mrg #define umul_ppmm(ph, pl, m0, m1) \ 1971 1.3 mrg do { \ 1972 1.1 mrg UDWtype __ll = __umulsidi3 (m0, m1); \ 1973 1.1 mrg ph = (UWtype) (__ll >> W_TYPE_SIZE); \ 1974 1.1 mrg pl = (UWtype) __ll; \ 1975 1.3 mrg } while (0) 1976 1.1 mrg #endif 1977 1.1 mrg 1978 1.1 mrg #if !defined (__umulsidi3) 1979 1.1 mrg #define __umulsidi3(u, v) \ 1980 1.1 mrg ({UWtype __hi, __lo; \ 1981 1.1 mrg umul_ppmm (__hi, __lo, u, v); \ 1982 1.1 mrg ((UDWtype) __hi << W_TYPE_SIZE) | __lo; }) 1983 1.1 mrg #endif 1984 1.1 mrg 1985 1.1 mrg 1986 1.3 mrg #if defined (__cplusplus) 1987 1.3 mrg #define __longlong_h_C "C" 1988 1.3 mrg #else 1989 1.3 mrg #define __longlong_h_C 1990 1.3 mrg #endif 1991 1.3 mrg 1992 1.1 mrg /* Use mpn_umul_ppmm or mpn_udiv_qrnnd functions, if they exist. The "_r" 1993 1.1 mrg forms have "reversed" arguments, meaning the pointer is last, which 1994 1.1 mrg sometimes allows better parameter passing, in particular on 64-bit 1995 1.1 mrg hppa. */ 1996 1.1 mrg 1997 1.1 mrg #define mpn_umul_ppmm __MPN(umul_ppmm) 1998 1.3 mrg extern __longlong_h_C UWtype mpn_umul_ppmm (UWtype *, UWtype, UWtype); 1999 1.1 mrg 2000 1.1 mrg #if ! defined (umul_ppmm) && HAVE_NATIVE_mpn_umul_ppmm \ 2001 1.1 mrg && ! defined (LONGLONG_STANDALONE) 2002 1.3 mrg #define umul_ppmm(wh, wl, u, v) \ 2003 1.3 mrg do { \ 2004 1.3 mrg UWtype __umul_ppmm__p0; \ 2005 1.3 mrg (wh) = mpn_umul_ppmm (&__umul_ppmm__p0, (UWtype) (u), (UWtype) (v));\ 2006 1.3 mrg (wl) = __umul_ppmm__p0; \ 2007 1.1 mrg } while (0) 2008 1.1 mrg #endif 2009 1.1 mrg 2010 1.1 mrg #define mpn_umul_ppmm_r __MPN(umul_ppmm_r) 2011 1.3 mrg extern __longlong_h_C UWtype mpn_umul_ppmm_r (UWtype, UWtype, UWtype *); 2012 1.1 mrg 2013 1.1 mrg #if ! defined (umul_ppmm) && HAVE_NATIVE_mpn_umul_ppmm_r \ 2014 1.1 mrg && ! defined (LONGLONG_STANDALONE) 2015 1.3 mrg #define umul_ppmm(wh, wl, u, v) \ 2016 1.3 mrg do { \ 2017 1.3 mrg UWtype __umul_p0; \ 2018 1.3 mrg (wh) = mpn_umul_ppmm_r ((UWtype) (u), (UWtype) (v), &__umul_p0); \ 2019 1.3 mrg (wl) = __umul_p0; \ 2020 1.1 mrg } while (0) 2021 1.1 mrg #endif 2022 1.1 mrg 2023 1.1 mrg #define mpn_udiv_qrnnd __MPN(udiv_qrnnd) 2024 1.3 mrg extern __longlong_h_C UWtype mpn_udiv_qrnnd (UWtype *, UWtype, UWtype, UWtype); 2025 1.1 mrg 2026 1.1 mrg #if ! defined (udiv_qrnnd) && HAVE_NATIVE_mpn_udiv_qrnnd \ 2027 1.1 mrg && ! defined (LONGLONG_STANDALONE) 2028 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 2029 1.1 mrg do { \ 2030 1.3 mrg UWtype __udiv_qrnnd_r; \ 2031 1.3 mrg (q) = mpn_udiv_qrnnd (&__udiv_qrnnd_r, \ 2032 1.1 mrg (UWtype) (n1), (UWtype) (n0), (UWtype) d); \ 2033 1.3 mrg (r) = __udiv_qrnnd_r; \ 2034 1.1 mrg } while (0) 2035 1.1 mrg #endif 2036 1.1 mrg 2037 1.1 mrg #define mpn_udiv_qrnnd_r __MPN(udiv_qrnnd_r) 2038 1.3 mrg extern __longlong_h_C UWtype mpn_udiv_qrnnd_r (UWtype, UWtype, UWtype, UWtype *); 2039 1.1 mrg 2040 1.1 mrg #if ! defined (udiv_qrnnd) && HAVE_NATIVE_mpn_udiv_qrnnd_r \ 2041 1.1 mrg && ! defined (LONGLONG_STANDALONE) 2042 1.1 mrg #define udiv_qrnnd(q, r, n1, n0, d) \ 2043 1.1 mrg do { \ 2044 1.3 mrg UWtype __udiv_qrnnd_r; \ 2045 1.1 mrg (q) = mpn_udiv_qrnnd_r ((UWtype) (n1), (UWtype) (n0), (UWtype) d, \ 2046 1.3 mrg &__udiv_qrnnd_r); \ 2047 1.3 mrg (r) = __udiv_qrnnd_r; \ 2048 1.1 mrg } while (0) 2049 1.1 mrg #endif 2050 1.1 mrg 2051 1.1 mrg 2052 1.1 mrg /* If this machine has no inline assembler, use C macros. */ 2053 1.1 mrg 2054 1.1 mrg #if !defined (add_ssaaaa) 2055 1.1 mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 2056 1.1 mrg do { \ 2057 1.1 mrg UWtype __x; \ 2058 1.5 mrg UWtype __al = (al); \ 2059 1.5 mrg UWtype __bl = (bl); \ 2060 1.5 mrg __x = __al + __bl; \ 2061 1.5 mrg (sh) = (ah) + (bh) + (__x < __al); \ 2062 1.1 mrg (sl) = __x; \ 2063 1.1 mrg } while (0) 2064 1.1 mrg #endif 2065 1.1 mrg 2066 1.1 mrg #if !defined (sub_ddmmss) 2067 1.1 mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 2068 1.1 mrg do { \ 2069 1.1 mrg UWtype __x; \ 2070 1.5 mrg UWtype __al = (al); \ 2071 1.5 mrg UWtype __bl = (bl); \ 2072 1.5 mrg __x = __al - __bl; \ 2073 1.5 mrg (sh) = (ah) - (bh) - (__al < __bl); \ 2074 1.1 mrg (sl) = __x; \ 2075 1.1 mrg } while (0) 2076 1.1 mrg #endif 2077 1.1 mrg 2078 1.1 mrg /* If we lack umul_ppmm but have smul_ppmm, define umul_ppmm in terms of 2079 1.1 mrg smul_ppmm. */ 2080 1.1 mrg #if !defined (umul_ppmm) && defined (smul_ppmm) 2081 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 2082 1.1 mrg do { \ 2083 1.1 mrg UWtype __w1; \ 2084 1.1 mrg UWtype __xm0 = (u), __xm1 = (v); \ 2085 1.1 mrg smul_ppmm (__w1, w0, __xm0, __xm1); \ 2086 1.1 mrg (w1) = __w1 + (-(__xm0 >> (W_TYPE_SIZE - 1)) & __xm1) \ 2087 1.1 mrg + (-(__xm1 >> (W_TYPE_SIZE - 1)) & __xm0); \ 2088 1.1 mrg } while (0) 2089 1.1 mrg #endif 2090 1.1 mrg 2091 1.1 mrg /* If we still don't have umul_ppmm, define it using plain C. 2092 1.1 mrg 2093 1.1 mrg For reference, when this code is used for squaring (ie. u and v identical 2094 1.1 mrg expressions), gcc recognises __x1 and __x2 are the same and generates 3 2095 1.1 mrg multiplies, not 4. The subsequent additions could be optimized a bit, 2096 1.1 mrg but the only place GMP currently uses such a square is mpn_sqr_basecase, 2097 1.1 mrg and chips obliged to use this generic C umul will have plenty of worse 2098 1.1 mrg performance problems than a couple of extra instructions on the diagonal 2099 1.1 mrg of sqr_basecase. */ 2100 1.1 mrg 2101 1.1 mrg #if !defined (umul_ppmm) 2102 1.1 mrg #define umul_ppmm(w1, w0, u, v) \ 2103 1.1 mrg do { \ 2104 1.1 mrg UWtype __x0, __x1, __x2, __x3; \ 2105 1.1 mrg UHWtype __ul, __vl, __uh, __vh; \ 2106 1.1 mrg UWtype __u = (u), __v = (v); \ 2107 1.1 mrg \ 2108 1.1 mrg __ul = __ll_lowpart (__u); \ 2109 1.1 mrg __uh = __ll_highpart (__u); \ 2110 1.1 mrg __vl = __ll_lowpart (__v); \ 2111 1.1 mrg __vh = __ll_highpart (__v); \ 2112 1.1 mrg \ 2113 1.1 mrg __x0 = (UWtype) __ul * __vl; \ 2114 1.1 mrg __x1 = (UWtype) __ul * __vh; \ 2115 1.1 mrg __x2 = (UWtype) __uh * __vl; \ 2116 1.1 mrg __x3 = (UWtype) __uh * __vh; \ 2117 1.1 mrg \ 2118 1.1 mrg __x1 += __ll_highpart (__x0);/* this can't give carry */ \ 2119 1.1 mrg __x1 += __x2; /* but this indeed can */ \ 2120 1.1 mrg if (__x1 < __x2) /* did we get it? */ \ 2121 1.1 mrg __x3 += __ll_B; /* yes, add it in the proper pos. */ \ 2122 1.1 mrg \ 2123 1.1 mrg (w1) = __x3 + __ll_highpart (__x1); \ 2124 1.1 mrg (w0) = (__x1 << W_TYPE_SIZE/2) + __ll_lowpart (__x0); \ 2125 1.1 mrg } while (0) 2126 1.1 mrg #endif 2127 1.1 mrg 2128 1.1 mrg /* If we don't have smul_ppmm, define it using umul_ppmm (which surely will 2129 1.1 mrg exist in one form or another. */ 2130 1.1 mrg #if !defined (smul_ppmm) 2131 1.1 mrg #define smul_ppmm(w1, w0, u, v) \ 2132 1.1 mrg do { \ 2133 1.1 mrg UWtype __w1; \ 2134 1.1 mrg UWtype __xm0 = (u), __xm1 = (v); \ 2135 1.1 mrg umul_ppmm (__w1, w0, __xm0, __xm1); \ 2136 1.1 mrg (w1) = __w1 - (-(__xm0 >> (W_TYPE_SIZE - 1)) & __xm1) \ 2137 1.1 mrg - (-(__xm1 >> (W_TYPE_SIZE - 1)) & __xm0); \ 2138 1.1 mrg } while (0) 2139 1.1 mrg #endif 2140 1.1 mrg 2141 1.1 mrg /* Define this unconditionally, so it can be used for debugging. */ 2142 1.1 mrg #define __udiv_qrnnd_c(q, r, n1, n0, d) \ 2143 1.1 mrg do { \ 2144 1.1 mrg UWtype __d1, __d0, __q1, __q0, __r1, __r0, __m; \ 2145 1.1 mrg \ 2146 1.1 mrg ASSERT ((d) != 0); \ 2147 1.1 mrg ASSERT ((n1) < (d)); \ 2148 1.1 mrg \ 2149 1.1 mrg __d1 = __ll_highpart (d); \ 2150 1.1 mrg __d0 = __ll_lowpart (d); \ 2151 1.1 mrg \ 2152 1.1 mrg __q1 = (n1) / __d1; \ 2153 1.1 mrg __r1 = (n1) - __q1 * __d1; \ 2154 1.1 mrg __m = __q1 * __d0; \ 2155 1.1 mrg __r1 = __r1 * __ll_B | __ll_highpart (n0); \ 2156 1.1 mrg if (__r1 < __m) \ 2157 1.1 mrg { \ 2158 1.1 mrg __q1--, __r1 += (d); \ 2159 1.1 mrg if (__r1 >= (d)) /* i.e. we didn't get carry when adding to __r1 */\ 2160 1.1 mrg if (__r1 < __m) \ 2161 1.1 mrg __q1--, __r1 += (d); \ 2162 1.1 mrg } \ 2163 1.1 mrg __r1 -= __m; \ 2164 1.1 mrg \ 2165 1.1 mrg __q0 = __r1 / __d1; \ 2166 1.1 mrg __r0 = __r1 - __q0 * __d1; \ 2167 1.1 mrg __m = __q0 * __d0; \ 2168 1.1 mrg __r0 = __r0 * __ll_B | __ll_lowpart (n0); \ 2169 1.1 mrg if (__r0 < __m) \ 2170 1.1 mrg { \ 2171 1.1 mrg __q0--, __r0 += (d); \ 2172 1.1 mrg if (__r0 >= (d)) \ 2173 1.1 mrg if (__r0 < __m) \ 2174 1.1 mrg __q0--, __r0 += (d); \ 2175 1.1 mrg } \ 2176 1.1 mrg __r0 -= __m; \ 2177 1.1 mrg \ 2178 1.1 mrg (q) = __q1 * __ll_B | __q0; \ 2179 1.1 mrg (r) = __r0; \ 2180 1.1 mrg } while (0) 2181 1.1 mrg 2182 1.1 mrg /* If the processor has no udiv_qrnnd but sdiv_qrnnd, go through 2183 1.1 mrg __udiv_w_sdiv (defined in libgcc or elsewhere). */ 2184 1.3 mrg #if !defined (udiv_qrnnd) && defined (sdiv_qrnnd) \ 2185 1.3 mrg && ! defined (LONGLONG_STANDALONE) 2186 1.1 mrg #define udiv_qrnnd(q, r, nh, nl, d) \ 2187 1.1 mrg do { \ 2188 1.1 mrg UWtype __r; \ 2189 1.1 mrg (q) = __MPN(udiv_w_sdiv) (&__r, nh, nl, d); \ 2190 1.1 mrg (r) = __r; \ 2191 1.1 mrg } while (0) 2192 1.2 joerg __GMP_DECLSPEC UWtype __MPN(udiv_w_sdiv) (UWtype *, UWtype, UWtype, UWtype); 2193 1.1 mrg #endif 2194 1.1 mrg 2195 1.1 mrg /* If udiv_qrnnd was not defined for this processor, use __udiv_qrnnd_c. */ 2196 1.1 mrg #if !defined (udiv_qrnnd) 2197 1.1 mrg #define UDIV_NEEDS_NORMALIZATION 1 2198 1.1 mrg #define udiv_qrnnd __udiv_qrnnd_c 2199 1.1 mrg #endif 2200 1.1 mrg 2201 1.1 mrg #if !defined (count_leading_zeros) 2202 1.1 mrg #define count_leading_zeros(count, x) \ 2203 1.1 mrg do { \ 2204 1.1 mrg UWtype __xr = (x); \ 2205 1.1 mrg UWtype __a; \ 2206 1.1 mrg \ 2207 1.1 mrg if (W_TYPE_SIZE == 32) \ 2208 1.1 mrg { \ 2209 1.1 mrg __a = __xr < ((UWtype) 1 << 2*__BITS4) \ 2210 1.1 mrg ? (__xr < ((UWtype) 1 << __BITS4) ? 1 : __BITS4 + 1) \ 2211 1.1 mrg : (__xr < ((UWtype) 1 << 3*__BITS4) ? 2*__BITS4 + 1 \ 2212 1.1 mrg : 3*__BITS4 + 1); \ 2213 1.1 mrg } \ 2214 1.1 mrg else \ 2215 1.1 mrg { \ 2216 1.1 mrg for (__a = W_TYPE_SIZE - 8; __a > 0; __a -= 8) \ 2217 1.1 mrg if (((__xr >> __a) & 0xff) != 0) \ 2218 1.1 mrg break; \ 2219 1.1 mrg ++__a; \ 2220 1.1 mrg } \ 2221 1.1 mrg \ 2222 1.1 mrg (count) = W_TYPE_SIZE + 1 - __a - __clz_tab[__xr >> __a]; \ 2223 1.1 mrg } while (0) 2224 1.1 mrg /* This version gives a well-defined value for zero. */ 2225 1.1 mrg #define COUNT_LEADING_ZEROS_0 (W_TYPE_SIZE - 1) 2226 1.1 mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB 2227 1.2 joerg #define COUNT_LEADING_ZEROS_SLOW 2228 1.1 mrg #endif 2229 1.1 mrg 2230 1.1 mrg /* clz_tab needed by mpn/x86/pentium/mod_1.asm in a fat binary */ 2231 1.1 mrg #if HAVE_HOST_CPU_FAMILY_x86 && WANT_FAT_BINARY 2232 1.1 mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB 2233 1.1 mrg #endif 2234 1.1 mrg 2235 1.1 mrg #ifdef COUNT_LEADING_ZEROS_NEED_CLZ_TAB 2236 1.2 joerg extern const unsigned char __GMP_DECLSPEC __clz_tab[129]; 2237 1.1 mrg #endif 2238 1.1 mrg 2239 1.1 mrg #if !defined (count_trailing_zeros) 2240 1.2 joerg #if !defined (COUNT_LEADING_ZEROS_SLOW) 2241 1.2 joerg /* Define count_trailing_zeros using an asm count_leading_zeros. */ 2242 1.2 joerg #define count_trailing_zeros(count, x) \ 2243 1.1 mrg do { \ 2244 1.1 mrg UWtype __ctz_x = (x); \ 2245 1.1 mrg UWtype __ctz_c; \ 2246 1.1 mrg ASSERT (__ctz_x != 0); \ 2247 1.1 mrg count_leading_zeros (__ctz_c, __ctz_x & -__ctz_x); \ 2248 1.1 mrg (count) = W_TYPE_SIZE - 1 - __ctz_c; \ 2249 1.1 mrg } while (0) 2250 1.2 joerg #else 2251 1.2 joerg /* Define count_trailing_zeros in plain C, assuming small counts are common. 2252 1.2 joerg We use clz_tab without ado, since the C count_leading_zeros above will have 2253 1.2 joerg pulled it in. */ 2254 1.2 joerg #define count_trailing_zeros(count, x) \ 2255 1.2 joerg do { \ 2256 1.2 joerg UWtype __ctz_x = (x); \ 2257 1.2 joerg int __ctz_c; \ 2258 1.2 joerg \ 2259 1.2 joerg if (LIKELY ((__ctz_x & 0xff) != 0)) \ 2260 1.2 joerg (count) = __clz_tab[__ctz_x & -__ctz_x] - 2; \ 2261 1.2 joerg else \ 2262 1.2 joerg { \ 2263 1.2 joerg for (__ctz_c = 8 - 2; __ctz_c < W_TYPE_SIZE - 2; __ctz_c += 8) \ 2264 1.2 joerg { \ 2265 1.2 joerg __ctz_x >>= 8; \ 2266 1.2 joerg if (LIKELY ((__ctz_x & 0xff) != 0)) \ 2267 1.2 joerg break; \ 2268 1.2 joerg } \ 2269 1.2 joerg \ 2270 1.2 joerg (count) = __ctz_c + __clz_tab[__ctz_x & -__ctz_x]; \ 2271 1.2 joerg } \ 2272 1.2 joerg } while (0) 2273 1.2 joerg #endif 2274 1.1 mrg #endif 2275 1.1 mrg 2276 1.1 mrg #ifndef UDIV_NEEDS_NORMALIZATION 2277 1.1 mrg #define UDIV_NEEDS_NORMALIZATION 0 2278 1.1 mrg #endif 2279 1.1 mrg 2280 1.1 mrg /* Whether udiv_qrnnd is actually implemented with udiv_qrnnd_preinv, and 2281 1.1 mrg that hence the latter should always be used. */ 2282 1.1 mrg #ifndef UDIV_PREINV_ALWAYS 2283 1.1 mrg #define UDIV_PREINV_ALWAYS 0 2284 1.1 mrg #endif 2285