Home | History | Annotate | Line # | Download | only in dist
      1  1.1       mrg /* longlong.h -- definitions for mixed size 32/64 bit arithmetic.
      2  1.1       mrg 
      3  1.4       mrg Copyright 1991-1994, 1996, 1997, 1999-2005, 2007-2009, 2011-2020 Free Software
      4  1.3       mrg Foundation, Inc.
      5  1.1       mrg 
      6  1.3       mrg This file is part of the GNU MP Library.
      7  1.1       mrg 
      8  1.3       mrg The GNU MP Library is free software; you can redistribute it and/or modify
      9  1.3       mrg it under the terms of either:
     10  1.3       mrg 
     11  1.3       mrg   * the GNU Lesser General Public License as published by the Free
     12  1.3       mrg     Software Foundation; either version 3 of the License, or (at your
     13  1.3       mrg     option) any later version.
     14  1.3       mrg 
     15  1.3       mrg or
     16  1.3       mrg 
     17  1.3       mrg   * the GNU General Public License as published by the Free Software
     18  1.3       mrg     Foundation; either version 2 of the License, or (at your option) any
     19  1.3       mrg     later version.
     20  1.3       mrg 
     21  1.3       mrg or both in parallel, as here.
     22  1.3       mrg 
     23  1.3       mrg The GNU MP Library is distributed in the hope that it will be useful, but
     24  1.3       mrg WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
     25  1.3       mrg or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
     26  1.3       mrg for more details.
     27  1.3       mrg 
     28  1.3       mrg You should have received copies of the GNU General Public License and the
     29  1.3       mrg GNU Lesser General Public License along with the GNU MP Library.  If not,
     30  1.3       mrg see https://www.gnu.org/licenses/.  */
     31  1.1       mrg 
     32  1.1       mrg /* You have to define the following before including this file:
     33  1.1       mrg 
     34  1.1       mrg    UWtype -- An unsigned type, default type for operations (typically a "word")
     35  1.1       mrg    UHWtype -- An unsigned type, at least half the size of UWtype
     36  1.1       mrg    UDWtype -- An unsigned type, at least twice as large a UWtype
     37  1.1       mrg    W_TYPE_SIZE -- size in bits of UWtype
     38  1.1       mrg 
     39  1.1       mrg    SItype, USItype -- Signed and unsigned 32 bit types
     40  1.1       mrg    DItype, UDItype -- Signed and unsigned 64 bit types
     41  1.1       mrg 
     42  1.1       mrg    On a 32 bit machine UWtype should typically be USItype;
     43  1.1       mrg    on a 64 bit machine, UWtype should typically be UDItype.
     44  1.1       mrg 
     45  1.1       mrg    Optionally, define:
     46  1.1       mrg 
     47  1.1       mrg    LONGLONG_STANDALONE -- Avoid code that needs machine-dependent support files
     48  1.1       mrg    NO_ASM -- Disable inline asm
     49  1.1       mrg 
     50  1.1       mrg 
     51  1.1       mrg    CAUTION!  Using this version of longlong.h outside of GMP is not safe.  You
     52  1.1       mrg    need to include gmp.h and gmp-impl.h, or certain things might not work as
     53  1.1       mrg    expected.
     54  1.1       mrg */
     55  1.1       mrg 
     56  1.1       mrg #define __BITS4 (W_TYPE_SIZE / 4)
     57  1.1       mrg #define __ll_B ((UWtype) 1 << (W_TYPE_SIZE / 2))
     58  1.1       mrg #define __ll_lowpart(t) ((UWtype) (t) & (__ll_B - 1))
     59  1.1       mrg #define __ll_highpart(t) ((UWtype) (t) >> (W_TYPE_SIZE / 2))
     60  1.1       mrg 
     61  1.1       mrg /* This is used to make sure no undesirable sharing between different libraries
     62  1.1       mrg    that use this file takes place.  */
     63  1.1       mrg #ifndef __MPN
     64  1.1       mrg #define __MPN(x) __##x
     65  1.1       mrg #endif
     66  1.1       mrg 
     67  1.1       mrg /* Define auxiliary asm macros.
     68  1.1       mrg 
     69  1.1       mrg    1) umul_ppmm(high_prod, low_prod, multiplier, multiplicand) multiplies two
     70  1.1       mrg    UWtype integers MULTIPLIER and MULTIPLICAND, and generates a two UWtype
     71  1.1       mrg    word product in HIGH_PROD and LOW_PROD.
     72  1.1       mrg 
     73  1.1       mrg    2) __umulsidi3(a,b) multiplies two UWtype integers A and B, and returns a
     74  1.1       mrg    UDWtype product.  This is just a variant of umul_ppmm.
     75  1.1       mrg 
     76  1.1       mrg    3) udiv_qrnnd(quotient, remainder, high_numerator, low_numerator,
     77  1.1       mrg    denominator) divides a UDWtype, composed by the UWtype integers
     78  1.1       mrg    HIGH_NUMERATOR and LOW_NUMERATOR, by DENOMINATOR and places the quotient
     79  1.1       mrg    in QUOTIENT and the remainder in REMAINDER.  HIGH_NUMERATOR must be less
     80  1.1       mrg    than DENOMINATOR for correct operation.  If, in addition, the most
     81  1.1       mrg    significant bit of DENOMINATOR must be 1, then the pre-processor symbol
     82  1.1       mrg    UDIV_NEEDS_NORMALIZATION is defined to 1.
     83  1.1       mrg 
     84  1.1       mrg    4) sdiv_qrnnd(quotient, remainder, high_numerator, low_numerator,
     85  1.1       mrg    denominator).  Like udiv_qrnnd but the numbers are signed.  The quotient
     86  1.1       mrg    is rounded towards 0.
     87  1.1       mrg 
     88  1.1       mrg    5) count_leading_zeros(count, x) counts the number of zero-bits from the
     89  1.1       mrg    msb to the first non-zero bit in the UWtype X.  This is the number of
     90  1.1       mrg    steps X needs to be shifted left to set the msb.  Undefined for X == 0,
     91  1.1       mrg    unless the symbol COUNT_LEADING_ZEROS_0 is defined to some value.
     92  1.1       mrg 
     93  1.1       mrg    6) count_trailing_zeros(count, x) like count_leading_zeros, but counts
     94  1.1       mrg    from the least significant end.
     95  1.1       mrg 
     96  1.1       mrg    7) add_ssaaaa(high_sum, low_sum, high_addend_1, low_addend_1,
     97  1.1       mrg    high_addend_2, low_addend_2) adds two UWtype integers, composed by
     98  1.1       mrg    HIGH_ADDEND_1 and LOW_ADDEND_1, and HIGH_ADDEND_2 and LOW_ADDEND_2
     99  1.1       mrg    respectively.  The result is placed in HIGH_SUM and LOW_SUM.  Overflow
    100  1.1       mrg    (i.e. carry out) is not stored anywhere, and is lost.
    101  1.1       mrg 
    102  1.1       mrg    8) sub_ddmmss(high_difference, low_difference, high_minuend, low_minuend,
    103  1.1       mrg    high_subtrahend, low_subtrahend) subtracts two two-word UWtype integers,
    104  1.1       mrg    composed by HIGH_MINUEND_1 and LOW_MINUEND_1, and HIGH_SUBTRAHEND_2 and
    105  1.1       mrg    LOW_SUBTRAHEND_2 respectively.  The result is placed in HIGH_DIFFERENCE
    106  1.1       mrg    and LOW_DIFFERENCE.  Overflow (i.e. carry out) is not stored anywhere,
    107  1.1       mrg    and is lost.
    108  1.1       mrg 
    109  1.1       mrg    If any of these macros are left undefined for a particular CPU,
    110  1.1       mrg    C macros are used.
    111  1.1       mrg 
    112  1.1       mrg 
    113  1.1       mrg    Notes:
    114  1.1       mrg 
    115  1.1       mrg    For add_ssaaaa the two high and two low addends can both commute, but
    116  1.1       mrg    unfortunately gcc only supports one "%" commutative in each asm block.
    117  1.1       mrg    This has always been so but is only documented in recent versions
    118  1.1       mrg    (eg. pre-release 3.3).  Having two or more "%"s can cause an internal
    119  1.1       mrg    compiler error in certain rare circumstances.
    120  1.1       mrg 
    121  1.1       mrg    Apparently it was only the last "%" that was ever actually respected, so
    122  1.1       mrg    the code has been updated to leave just that.  Clearly there's a free
    123  1.1       mrg    choice whether high or low should get it, if there's a reason to favour
    124  1.1       mrg    one over the other.  Also obviously when the constraints on the two
    125  1.1       mrg    operands are identical there's no benefit to the reloader in any "%" at
    126  1.1       mrg    all.
    127  1.1       mrg 
    128  1.1       mrg    */
    129  1.1       mrg 
    130  1.1       mrg /* The CPUs come in alphabetical order below.
    131  1.1       mrg 
    132  1.1       mrg    Please add support for more CPUs here, or improve the current support
    133  1.1       mrg    for the CPUs below!  */
    134  1.1       mrg 
    135  1.1       mrg 
    136  1.1       mrg /* count_leading_zeros_gcc_clz is count_leading_zeros implemented with gcc
    137  1.1       mrg    3.4 __builtin_clzl or __builtin_clzll, according to our limb size.
    138  1.1       mrg    Similarly count_trailing_zeros_gcc_ctz using __builtin_ctzl or
    139  1.1       mrg    __builtin_ctzll.
    140  1.1       mrg 
    141  1.1       mrg    These builtins are only used when we check what code comes out, on some
    142  1.1       mrg    chips they're merely libgcc calls, where we will instead want an inline
    143  1.1       mrg    in that case (either asm or generic C).
    144  1.1       mrg 
    145  1.1       mrg    These builtins are better than an asm block of the same insn, since an
    146  1.1       mrg    asm block doesn't give gcc any information about scheduling or resource
    147  1.1       mrg    usage.  We keep an asm block for use on prior versions of gcc though.
    148  1.1       mrg 
    149  1.1       mrg    For reference, __builtin_ffs existed in gcc prior to __builtin_clz, but
    150  1.1       mrg    it's not used (for count_leading_zeros) because it generally gives extra
    151  1.1       mrg    code to ensure the result is 0 when the input is 0, which we don't need
    152  1.1       mrg    or want.  */
    153  1.1       mrg 
    154  1.1       mrg #ifdef _LONG_LONG_LIMB
    155  1.3       mrg #define count_leading_zeros_gcc_clz(count,x)	\
    156  1.3       mrg   do {						\
    157  1.3       mrg     ASSERT ((x) != 0);				\
    158  1.3       mrg     (count) = __builtin_clzll (x);		\
    159  1.1       mrg   } while (0)
    160  1.1       mrg #else
    161  1.3       mrg #define count_leading_zeros_gcc_clz(count,x)	\
    162  1.3       mrg   do {						\
    163  1.3       mrg     ASSERT ((x) != 0);				\
    164  1.3       mrg     (count) = __builtin_clzl (x);		\
    165  1.1       mrg   } while (0)
    166  1.1       mrg #endif
    167  1.1       mrg 
    168  1.1       mrg #ifdef _LONG_LONG_LIMB
    169  1.3       mrg #define count_trailing_zeros_gcc_ctz(count,x)	\
    170  1.3       mrg   do {						\
    171  1.3       mrg     ASSERT ((x) != 0);				\
    172  1.3       mrg     (count) = __builtin_ctzll (x);		\
    173  1.1       mrg   } while (0)
    174  1.1       mrg #else
    175  1.3       mrg #define count_trailing_zeros_gcc_ctz(count,x)	\
    176  1.3       mrg   do {						\
    177  1.3       mrg     ASSERT ((x) != 0);				\
    178  1.3       mrg     (count) = __builtin_ctzl (x);		\
    179  1.1       mrg   } while (0)
    180  1.1       mrg #endif
    181  1.1       mrg 
    182  1.1       mrg 
    183  1.1       mrg /* FIXME: The macros using external routines like __MPN(count_leading_zeros)
    184  1.1       mrg    don't need to be under !NO_ASM */
    185  1.1       mrg #if ! defined (NO_ASM)
    186  1.1       mrg 
    187  1.1       mrg #if defined (__alpha) && W_TYPE_SIZE == 64
    188  1.1       mrg /* Most alpha-based machines, except Cray systems. */
    189  1.1       mrg #if defined (__GNUC__)
    190  1.1       mrg #if __GMP_GNUC_PREREQ (3,3)
    191  1.1       mrg #define umul_ppmm(ph, pl, m0, m1) \
    192  1.1       mrg   do {									\
    193  1.1       mrg     UDItype __m0 = (m0), __m1 = (m1);					\
    194  1.1       mrg     (ph) = __builtin_alpha_umulh (__m0, __m1);				\
    195  1.1       mrg     (pl) = __m0 * __m1;							\
    196  1.1       mrg   } while (0)
    197  1.1       mrg #else
    198  1.1       mrg #define umul_ppmm(ph, pl, m0, m1) \
    199  1.1       mrg   do {									\
    200  1.1       mrg     UDItype __m0 = (m0), __m1 = (m1);					\
    201  1.1       mrg     __asm__ ("umulh %r1,%2,%0"						\
    202  1.1       mrg 	     : "=r" (ph)						\
    203  1.3       mrg 	     : "%rJ" (__m0), "rI" (__m1));				\
    204  1.1       mrg     (pl) = __m0 * __m1;							\
    205  1.1       mrg   } while (0)
    206  1.1       mrg #endif
    207  1.1       mrg #else /* ! __GNUC__ */
    208  1.1       mrg #include <machine/builtins.h>
    209  1.1       mrg #define umul_ppmm(ph, pl, m0, m1) \
    210  1.1       mrg   do {									\
    211  1.1       mrg     UDItype __m0 = (m0), __m1 = (m1);					\
    212  1.3       mrg     (ph) = __UMULH (__m0, __m1);					\
    213  1.1       mrg     (pl) = __m0 * __m1;							\
    214  1.1       mrg   } while (0)
    215  1.1       mrg #endif
    216  1.1       mrg #ifndef LONGLONG_STANDALONE
    217  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, d) \
    218  1.1       mrg   do { UWtype __di;							\
    219  1.1       mrg     __di = __MPN(invert_limb) (d);					\
    220  1.1       mrg     udiv_qrnnd_preinv (q, r, n1, n0, d, __di);				\
    221  1.1       mrg   } while (0)
    222  1.1       mrg #define UDIV_PREINV_ALWAYS  1
    223  1.1       mrg #define UDIV_NEEDS_NORMALIZATION 1
    224  1.1       mrg #endif /* LONGLONG_STANDALONE */
    225  1.1       mrg 
    226  1.1       mrg /* clz_tab is required in all configurations, since mpn/alpha/cntlz.asm
    227  1.1       mrg    always goes into libgmp.so, even when not actually used.  */
    228  1.1       mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
    229  1.1       mrg 
    230  1.1       mrg #if defined (__GNUC__) && HAVE_HOST_CPU_alpha_CIX
    231  1.1       mrg #define count_leading_zeros(COUNT,X) \
    232  1.1       mrg   __asm__("ctlz %1,%0" : "=r"(COUNT) : "r"(X))
    233  1.1       mrg #define count_trailing_zeros(COUNT,X) \
    234  1.1       mrg   __asm__("cttz %1,%0" : "=r"(COUNT) : "r"(X))
    235  1.1       mrg #endif /* clz/ctz using cix */
    236  1.1       mrg 
    237  1.3       mrg #if ! defined (count_leading_zeros)				\
    238  1.1       mrg   && defined (__GNUC__) && ! defined (LONGLONG_STANDALONE)
    239  1.1       mrg /* ALPHA_CMPBGE_0 gives "cmpbge $31,src,dst", ie. test src bytes == 0.
    240  1.1       mrg    "$31" is written explicitly in the asm, since an "r" constraint won't
    241  1.1       mrg    select reg 31.  There seems no need to worry about "r31" syntax for cray,
    242  1.3       mrg    since gcc itself (pre-release 3.4) emits just $31 in various places.	 */
    243  1.3       mrg #define ALPHA_CMPBGE_0(dst, src)					\
    244  1.1       mrg   do { asm ("cmpbge $31, %1, %0" : "=r" (dst) : "r" (src)); } while (0)
    245  1.1       mrg /* Zero bytes are turned into bits with cmpbge, a __clz_tab lookup counts
    246  1.1       mrg    them, locating the highest non-zero byte.  A second __clz_tab lookup
    247  1.1       mrg    counts the leading zero bits in that byte, giving the result.  */
    248  1.3       mrg #define count_leading_zeros(count, x)					\
    249  1.3       mrg   do {									\
    250  1.3       mrg     UWtype  __clz__b, __clz__c, __clz__x = (x);				\
    251  1.3       mrg     ALPHA_CMPBGE_0 (__clz__b,  __clz__x);	    /* zero bytes */	\
    252  1.3       mrg     __clz__b = __clz_tab [(__clz__b >> 1) ^ 0x7F];  /* 8 to 1 byte */	\
    253  1.3       mrg     __clz__b = __clz__b * 8 - 7;		    /* 57 to 1 shift */ \
    254  1.3       mrg     __clz__x >>= __clz__b;						\
    255  1.3       mrg     __clz__c = __clz_tab [__clz__x];		    /* 8 to 1 bit */	\
    256  1.3       mrg     __clz__b = 65 - __clz__b;						\
    257  1.3       mrg     (count) = __clz__b - __clz__c;					\
    258  1.1       mrg   } while (0)
    259  1.1       mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
    260  1.1       mrg #endif /* clz using cmpbge */
    261  1.1       mrg 
    262  1.1       mrg #if ! defined (count_leading_zeros) && ! defined (LONGLONG_STANDALONE)
    263  1.1       mrg #if HAVE_ATTRIBUTE_CONST
    264  1.2     joerg long __MPN(count_leading_zeros) (UDItype) __attribute__ ((const));
    265  1.1       mrg #else
    266  1.2     joerg long __MPN(count_leading_zeros) (UDItype);
    267  1.1       mrg #endif
    268  1.1       mrg #define count_leading_zeros(count, x) \
    269  1.1       mrg   ((count) = __MPN(count_leading_zeros) (x))
    270  1.1       mrg #endif /* clz using mpn */
    271  1.1       mrg #endif /* __alpha */
    272  1.1       mrg 
    273  1.2     joerg #if defined (__AVR) && W_TYPE_SIZE == 8
    274  1.2     joerg #define umul_ppmm(ph, pl, m0, m1) \
    275  1.2     joerg   do {									\
    276  1.2     joerg     unsigned short __p = (unsigned short) (m0) * (m1);			\
    277  1.2     joerg     (ph) = __p >> 8;							\
    278  1.2     joerg     (pl) = __p;								\
    279  1.2     joerg   } while (0)
    280  1.2     joerg #endif /* AVR */
    281  1.2     joerg 
    282  1.1       mrg #if defined (_CRAY) && W_TYPE_SIZE == 64
    283  1.1       mrg #include <intrinsics.h>
    284  1.1       mrg #define UDIV_PREINV_ALWAYS  1
    285  1.1       mrg #define UDIV_NEEDS_NORMALIZATION 1
    286  1.2     joerg long __MPN(count_leading_zeros) (UDItype);
    287  1.1       mrg #define count_leading_zeros(count, x) \
    288  1.1       mrg   ((count) = _leadz ((UWtype) (x)))
    289  1.1       mrg #if defined (_CRAYIEEE)		/* I.e., Cray T90/ieee, T3D, and T3E */
    290  1.1       mrg #define umul_ppmm(ph, pl, m0, m1) \
    291  1.1       mrg   do {									\
    292  1.1       mrg     UDItype __m0 = (m0), __m1 = (m1);					\
    293  1.3       mrg     (ph) = _int_mult_upper (__m0, __m1);				\
    294  1.1       mrg     (pl) = __m0 * __m1;							\
    295  1.1       mrg   } while (0)
    296  1.1       mrg #ifndef LONGLONG_STANDALONE
    297  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, d) \
    298  1.1       mrg   do { UWtype __di;							\
    299  1.1       mrg     __di = __MPN(invert_limb) (d);					\
    300  1.1       mrg     udiv_qrnnd_preinv (q, r, n1, n0, d, __di);				\
    301  1.1       mrg   } while (0)
    302  1.1       mrg #endif /* LONGLONG_STANDALONE */
    303  1.1       mrg #endif /* _CRAYIEEE */
    304  1.1       mrg #endif /* _CRAY */
    305  1.1       mrg 
    306  1.1       mrg #if defined (__ia64) && W_TYPE_SIZE == 64
    307  1.1       mrg /* This form encourages gcc (pre-release 3.4 at least) to emit predicated
    308  1.1       mrg    "sub r=r,r" and "sub r=r,r,1", giving a 2 cycle latency.  The generic
    309  1.1       mrg    code using "al<bl" arithmetically comes out making an actual 0 or 1 in a
    310  1.1       mrg    register, which takes an extra cycle.  */
    311  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl)      \
    312  1.3       mrg   do {						\
    313  1.3       mrg     UWtype __x;					\
    314  1.3       mrg     __x = (al) - (bl);				\
    315  1.3       mrg     if ((al) < (bl))				\
    316  1.3       mrg       (sh) = (ah) - (bh) - 1;			\
    317  1.3       mrg     else					\
    318  1.3       mrg       (sh) = (ah) - (bh);			\
    319  1.3       mrg     (sl) = __x;					\
    320  1.1       mrg   } while (0)
    321  1.1       mrg #if defined (__GNUC__) && ! defined (__INTEL_COMPILER)
    322  1.1       mrg /* Do both product parts in assembly, since that gives better code with
    323  1.1       mrg    all gcc versions.  Some callers will just use the upper part, and in
    324  1.1       mrg    that situation we waste an instruction, but not any cycles.  */
    325  1.1       mrg #define umul_ppmm(ph, pl, m0, m1) \
    326  1.1       mrg     __asm__ ("xma.hu %0 = %2, %3, f0\n\txma.l %1 = %2, %3, f0"		\
    327  1.1       mrg 	     : "=&f" (ph), "=f" (pl)					\
    328  1.1       mrg 	     : "f" (m0), "f" (m1))
    329  1.1       mrg #define count_leading_zeros(count, x) \
    330  1.1       mrg   do {									\
    331  1.1       mrg     UWtype _x = (x), _y, _a, _c;					\
    332  1.1       mrg     __asm__ ("mux1 %0 = %1, @rev" : "=r" (_y) : "r" (_x));		\
    333  1.1       mrg     __asm__ ("czx1.l %0 = %1" : "=r" (_a) : "r" (-_y | _y));		\
    334  1.1       mrg     _c = (_a - 1) << 3;							\
    335  1.1       mrg     _x >>= _c;								\
    336  1.1       mrg     if (_x >= 1 << 4)							\
    337  1.1       mrg       _x >>= 4, _c += 4;						\
    338  1.1       mrg     if (_x >= 1 << 2)							\
    339  1.1       mrg       _x >>= 2, _c += 2;						\
    340  1.1       mrg     _c += _x >> 1;							\
    341  1.1       mrg     (count) =  W_TYPE_SIZE - 1 - _c;					\
    342  1.1       mrg   } while (0)
    343  1.1       mrg /* similar to what gcc does for __builtin_ffs, but 0 based rather than 1
    344  1.1       mrg    based, and we don't need a special case for x==0 here */
    345  1.1       mrg #define count_trailing_zeros(count, x)					\
    346  1.1       mrg   do {									\
    347  1.1       mrg     UWtype __ctz_x = (x);						\
    348  1.1       mrg     __asm__ ("popcnt %0 = %1"						\
    349  1.1       mrg 	     : "=r" (count)						\
    350  1.1       mrg 	     : "r" ((__ctz_x-1) & ~__ctz_x));				\
    351  1.1       mrg   } while (0)
    352  1.1       mrg #endif
    353  1.1       mrg #if defined (__INTEL_COMPILER)
    354  1.1       mrg #include <ia64intrin.h>
    355  1.1       mrg #define umul_ppmm(ph, pl, m0, m1)					\
    356  1.1       mrg   do {									\
    357  1.3       mrg     UWtype __m0 = (m0), __m1 = (m1);					\
    358  1.3       mrg     ph = _m64_xmahu (__m0, __m1, 0);					\
    359  1.3       mrg     pl = __m0 * __m1;							\
    360  1.1       mrg   } while (0)
    361  1.1       mrg #endif
    362  1.1       mrg #ifndef LONGLONG_STANDALONE
    363  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, d) \
    364  1.1       mrg   do { UWtype __di;							\
    365  1.1       mrg     __di = __MPN(invert_limb) (d);					\
    366  1.1       mrg     udiv_qrnnd_preinv (q, r, n1, n0, d, __di);				\
    367  1.1       mrg   } while (0)
    368  1.1       mrg #define UDIV_PREINV_ALWAYS  1
    369  1.1       mrg #define UDIV_NEEDS_NORMALIZATION 1
    370  1.1       mrg #endif
    371  1.1       mrg #endif
    372  1.1       mrg 
    373  1.1       mrg 
    374  1.6  christos #if defined (__GNUC__) || defined(__lint__)
    375  1.1       mrg 
    376  1.1       mrg /* We sometimes need to clobber "cc" with gcc2, but that would not be
    377  1.1       mrg    understood by gcc1.  Use cpp to avoid major code duplication.  */
    378  1.1       mrg #if __GNUC__ < 2
    379  1.1       mrg #define __CLOBBER_CC
    380  1.1       mrg #define __AND_CLOBBER_CC
    381  1.1       mrg #else /* __GNUC__ >= 2 */
    382  1.1       mrg #define __CLOBBER_CC : "cc"
    383  1.1       mrg #define __AND_CLOBBER_CC , "cc"
    384  1.1       mrg #endif /* __GNUC__ < 2 */
    385  1.1       mrg 
    386  1.1       mrg #if (defined (__a29k__) || defined (_AM29K)) && W_TYPE_SIZE == 32
    387  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
    388  1.1       mrg   __asm__ ("add %1,%4,%5\n\taddc %0,%2,%3"				\
    389  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
    390  1.1       mrg 	   : "r" (ah), "rI" (bh), "%r" (al), "rI" (bl))
    391  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
    392  1.1       mrg   __asm__ ("sub %1,%4,%5\n\tsubc %0,%2,%3"				\
    393  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
    394  1.1       mrg 	   : "r" (ah), "rI" (bh), "r" (al), "rI" (bl))
    395  1.1       mrg #define umul_ppmm(xh, xl, m0, m1) \
    396  1.1       mrg   do {									\
    397  1.1       mrg     USItype __m0 = (m0), __m1 = (m1);					\
    398  1.1       mrg     __asm__ ("multiplu %0,%1,%2"					\
    399  1.1       mrg 	     : "=r" (xl)						\
    400  1.1       mrg 	     : "r" (__m0), "r" (__m1));					\
    401  1.1       mrg     __asm__ ("multmu %0,%1,%2"						\
    402  1.1       mrg 	     : "=r" (xh)						\
    403  1.1       mrg 	     : "r" (__m0), "r" (__m1));					\
    404  1.1       mrg   } while (0)
    405  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, d) \
    406  1.1       mrg   __asm__ ("dividu %0,%3,%4"						\
    407  1.1       mrg 	   : "=r" (q), "=q" (r)						\
    408  1.1       mrg 	   : "1" (n1), "r" (n0), "r" (d))
    409  1.1       mrg #define count_leading_zeros(count, x) \
    410  1.1       mrg     __asm__ ("clz %0,%1"						\
    411  1.1       mrg 	     : "=r" (count)						\
    412  1.1       mrg 	     : "r" (x))
    413  1.1       mrg #define COUNT_LEADING_ZEROS_0 32
    414  1.1       mrg #endif /* __a29k__ */
    415  1.1       mrg 
    416  1.1       mrg #if defined (__arc__)
    417  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
    418  1.1       mrg   __asm__ ("add.f\t%1, %4, %5\n\tadc\t%0, %2, %3"			\
    419  1.1       mrg 	   : "=r" (sh),							\
    420  1.1       mrg 	     "=&r" (sl)							\
    421  1.1       mrg 	   : "r"  ((USItype) (ah)),					\
    422  1.3       mrg 	     "rICal" ((USItype) (bh)),					\
    423  1.1       mrg 	     "%r" ((USItype) (al)),					\
    424  1.3       mrg 	     "rICal" ((USItype) (bl)))
    425  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
    426  1.1       mrg   __asm__ ("sub.f\t%1, %4, %5\n\tsbc\t%0, %2, %3"			\
    427  1.1       mrg 	   : "=r" (sh),							\
    428  1.1       mrg 	     "=&r" (sl)							\
    429  1.1       mrg 	   : "r" ((USItype) (ah)),					\
    430  1.3       mrg 	     "rICal" ((USItype) (bh)),					\
    431  1.1       mrg 	     "r" ((USItype) (al)),					\
    432  1.3       mrg 	     "rICal" ((USItype) (bl)))
    433  1.1       mrg #endif
    434  1.1       mrg 
    435  1.3       mrg #if defined (__arm__) && (defined (__thumb2__) || !defined (__thumb__)) \
    436  1.3       mrg     && W_TYPE_SIZE == 32
    437  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
    438  1.4       mrg   do {									\
    439  1.5       mrg     if (__builtin_constant_p (bl) && -(USItype)(bl) < (USItype)(bl))	\
    440  1.4       mrg       __asm__ ("subs\t%1, %4, %5\n\tadc\t%0, %2, %3"			\
    441  1.4       mrg 	   : "=r" (sh), "=&r" (sl)					\
    442  1.4       mrg 	       : "r" (ah), "rI" (bh),					\
    443  1.4       mrg 		 "%r" (al), "rI" (-(USItype)(bl)) __CLOBBER_CC);	\
    444  1.4       mrg     else								\
    445  1.4       mrg       __asm__ ("adds\t%1, %4, %5\n\tadc\t%0, %2, %3"			\
    446  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
    447  1.4       mrg 	   : "r" (ah), "rI" (bh), "%r" (al), "rI" (bl) __CLOBBER_CC);	\
    448  1.4       mrg   } while (0)
    449  1.4       mrg /* FIXME: Extend the immediate range for the low word by using both ADDS and
    450  1.5       mrg    SUBS, since they set carry in the same way.  We need separate definitions
    451  1.5       mrg    for thumb and non-thumb since thumb lacks RSC.  */
    452  1.4       mrg #if defined (__thumb__)
    453  1.4       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
    454  1.4       mrg   do {									\
    455  1.4       mrg     if (__builtin_constant_p (ah) && __builtin_constant_p (bh)		\
    456  1.4       mrg 	&& (ah) == (bh))						\
    457  1.4       mrg       __asm__ ("subs\t%1, %2, %3\n\tsbc\t%0, %0, %0"			\
    458  1.4       mrg 	       : "=r" (sh), "=r" (sl)					\
    459  1.4       mrg 	       : "r" (al), "rI" (bl) __CLOBBER_CC);			\
    460  1.4       mrg     else if (__builtin_constant_p (al))					\
    461  1.4       mrg       __asm__ ("rsbs\t%1, %5, %4\n\tsbc\t%0, %2, %3"			\
    462  1.4       mrg 	       : "=r" (sh), "=&r" (sl)					\
    463  1.4       mrg 	       : "r" (ah), "rI" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \
    464  1.4       mrg     else if (__builtin_constant_p (bl))					\
    465  1.4       mrg       __asm__ ("subs\t%1, %4, %5\n\tsbc\t%0, %2, %3"			\
    466  1.4       mrg 	       : "=r" (sh), "=&r" (sl)					\
    467  1.4       mrg 	       : "r" (ah), "rI" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \
    468  1.4       mrg     else								\
    469  1.4       mrg       __asm__ ("subs\t%1, %4, %5\n\tsbc\t%0, %2, %3"			\
    470  1.4       mrg 	       : "=r" (sh), "=&r" (sl)					\
    471  1.4       mrg 	       : "r" (ah), "rI" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \
    472  1.4       mrg     } while (0)
    473  1.4       mrg #else
    474  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
    475  1.1       mrg   do {									\
    476  1.4       mrg     if (__builtin_constant_p (ah) && __builtin_constant_p (bh)		\
    477  1.4       mrg 	&& (ah) == (bh))						\
    478  1.4       mrg       __asm__ ("subs\t%1, %2, %3\n\tsbc\t%0, %0, %0"			\
    479  1.4       mrg 	       : "=r" (sh), "=r" (sl)					\
    480  1.4       mrg 	       : "r" (al), "rI" (bl) __CLOBBER_CC);			\
    481  1.4       mrg     else if (__builtin_constant_p (al))					\
    482  1.1       mrg       {									\
    483  1.1       mrg 	if (__builtin_constant_p (ah))					\
    484  1.1       mrg 	  __asm__ ("rsbs\t%1, %5, %4\n\trsc\t%0, %3, %2"		\
    485  1.1       mrg 		   : "=r" (sh), "=&r" (sl)				\
    486  1.1       mrg 		   : "rI" (ah), "r" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \
    487  1.1       mrg 	else								\
    488  1.1       mrg 	  __asm__ ("rsbs\t%1, %5, %4\n\tsbc\t%0, %2, %3"		\
    489  1.1       mrg 		   : "=r" (sh), "=&r" (sl)				\
    490  1.1       mrg 		   : "r" (ah), "rI" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \
    491  1.1       mrg       }									\
    492  1.1       mrg     else if (__builtin_constant_p (ah))					\
    493  1.1       mrg       {									\
    494  1.1       mrg 	if (__builtin_constant_p (bl))					\
    495  1.1       mrg 	  __asm__ ("subs\t%1, %4, %5\n\trsc\t%0, %3, %2"		\
    496  1.1       mrg 		   : "=r" (sh), "=&r" (sl)				\
    497  1.1       mrg 		   : "rI" (ah), "r" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \
    498  1.1       mrg 	else								\
    499  1.1       mrg 	  __asm__ ("rsbs\t%1, %5, %4\n\trsc\t%0, %3, %2"		\
    500  1.1       mrg 		   : "=r" (sh), "=&r" (sl)				\
    501  1.1       mrg 		   : "rI" (ah), "r" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \
    502  1.1       mrg       }									\
    503  1.1       mrg     else if (__builtin_constant_p (bl))					\
    504  1.4       mrg       __asm__ ("subs\t%1, %4, %5\n\tsbc\t%0, %2, %3"			\
    505  1.4       mrg 	       : "=r" (sh), "=&r" (sl)					\
    506  1.4       mrg 	       : "r" (ah), "rI" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \
    507  1.5       mrg     else								\
    508  1.1       mrg       __asm__ ("subs\t%1, %4, %5\n\tsbc\t%0, %2, %3"			\
    509  1.1       mrg 	       : "=r" (sh), "=&r" (sl)					\
    510  1.4       mrg 	       : "r" (ah), "rI" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \
    511  1.1       mrg     } while (0)
    512  1.4       mrg #endif
    513  1.3       mrg #if defined (__ARM_ARCH_2__) || defined (__ARM_ARCH_2A__) \
    514  1.3       mrg     || defined (__ARM_ARCH_3__)
    515  1.3       mrg #define umul_ppmm(xh, xl, a, b)						\
    516  1.3       mrg   do {									\
    517  1.3       mrg     register USItype __t0, __t1, __t2;					\
    518  1.3       mrg     __asm__ ("%@ Inlined umul_ppmm\n"					\
    519  1.3       mrg 	   "	mov	%2, %5, lsr #16\n"				\
    520  1.3       mrg 	   "	mov	%0, %6, lsr #16\n"				\
    521  1.3       mrg 	   "	bic	%3, %5, %2, lsl #16\n"				\
    522  1.3       mrg 	   "	bic	%4, %6, %0, lsl #16\n"				\
    523  1.3       mrg 	   "	mul	%1, %3, %4\n"					\
    524  1.3       mrg 	   "	mul	%4, %2, %4\n"					\
    525  1.3       mrg 	   "	mul	%3, %0, %3\n"					\
    526  1.3       mrg 	   "	mul	%0, %2, %0\n"					\
    527  1.3       mrg 	   "	adds	%3, %4, %3\n"					\
    528  1.3       mrg 	   "	addcs	%0, %0, #65536\n"				\
    529  1.3       mrg 	   "	adds	%1, %1, %3, lsl #16\n"				\
    530  1.3       mrg 	   "	adc	%0, %0, %3, lsr #16"				\
    531  1.3       mrg 	   : "=&r" ((USItype) (xh)), "=r" ((USItype) (xl)),		\
    532  1.3       mrg 	     "=&r" (__t0), "=&r" (__t1), "=r" (__t2)			\
    533  1.3       mrg 	   : "r" ((USItype) (a)), "r" ((USItype) (b)) __CLOBBER_CC);	\
    534  1.3       mrg   } while (0)
    535  1.3       mrg #ifndef LONGLONG_STANDALONE
    536  1.3       mrg #define udiv_qrnnd(q, r, n1, n0, d) \
    537  1.3       mrg   do { UWtype __r;							\
    538  1.3       mrg     (q) = __MPN(udiv_qrnnd) (&__r, (n1), (n0), (d));			\
    539  1.3       mrg     (r) = __r;								\
    540  1.3       mrg   } while (0)
    541  1.3       mrg extern UWtype __MPN(udiv_qrnnd) (UWtype *, UWtype, UWtype, UWtype);
    542  1.3       mrg #endif /* LONGLONG_STANDALONE */
    543  1.3       mrg #else /* ARMv4 or newer */
    544  1.1       mrg #define umul_ppmm(xh, xl, a, b) \
    545  1.1       mrg   __asm__ ("umull %0,%1,%2,%3" : "=&r" (xl), "=&r" (xh) : "r" (a), "r" (b))
    546  1.1       mrg #define smul_ppmm(xh, xl, a, b) \
    547  1.1       mrg   __asm__ ("smull %0,%1,%2,%3" : "=&r" (xl), "=&r" (xh) : "r" (a), "r" (b))
    548  1.1       mrg #ifndef LONGLONG_STANDALONE
    549  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, d) \
    550  1.1       mrg   do { UWtype __di;							\
    551  1.1       mrg     __di = __MPN(invert_limb) (d);					\
    552  1.1       mrg     udiv_qrnnd_preinv (q, r, n1, n0, d, __di);				\
    553  1.1       mrg   } while (0)
    554  1.1       mrg #define UDIV_PREINV_ALWAYS  1
    555  1.1       mrg #define UDIV_NEEDS_NORMALIZATION 1
    556  1.1       mrg #endif /* LONGLONG_STANDALONE */
    557  1.3       mrg #endif /* defined(__ARM_ARCH_2__) ... */
    558  1.3       mrg #define count_leading_zeros(count, x)  count_leading_zeros_gcc_clz(count, x)
    559  1.3       mrg #define count_trailing_zeros(count, x)  count_trailing_zeros_gcc_ctz(count, x)
    560  1.1       mrg #endif /* __arm__ */
    561  1.1       mrg 
    562  1.2     joerg #if defined (__aarch64__) && W_TYPE_SIZE == 64
    563  1.2     joerg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
    564  1.4       mrg   do {									\
    565  1.5       mrg     if (__builtin_constant_p (bl) && ~(UDItype)(bl) <= (UDItype)(bl))	\
    566  1.4       mrg       __asm__ ("subs\t%1, %x4, %5\n\tadc\t%0, %x2, %x3"			\
    567  1.4       mrg 	       : "=r" (sh), "=&r" (sl)					\
    568  1.4       mrg 	       : "rZ" ((UDItype)(ah)), "rZ" ((UDItype)(bh)),		\
    569  1.4       mrg 		 "%r" ((UDItype)(al)), "rI" (-(UDItype)(bl)) __CLOBBER_CC);\
    570  1.4       mrg     else								\
    571  1.4       mrg       __asm__ ("adds\t%1, %x4, %5\n\tadc\t%0, %x2, %x3"			\
    572  1.4       mrg 	       : "=r" (sh), "=&r" (sl)					\
    573  1.4       mrg 	       : "rZ" ((UDItype)(ah)), "rZ" ((UDItype)(bh)),		\
    574  1.4       mrg 		 "%r" ((UDItype)(al)), "rI" ((UDItype)(bl)) __CLOBBER_CC);\
    575  1.4       mrg   } while (0)
    576  1.2     joerg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
    577  1.4       mrg   do {									\
    578  1.5       mrg     if (__builtin_constant_p (bl) && ~(UDItype)(bl) <= (UDItype)(bl))	\
    579  1.4       mrg       __asm__ ("adds\t%1, %x4, %5\n\tsbc\t%0, %x2, %x3"			\
    580  1.4       mrg 	       : "=r,r" (sh), "=&r,&r" (sl)				\
    581  1.4       mrg 	       : "rZ,rZ" ((UDItype)(ah)), "rZ,rZ" ((UDItype)(bh)),	\
    582  1.4       mrg 		 "r,Z"   ((UDItype)(al)), "rI,r" (-(UDItype)(bl)) __CLOBBER_CC);\
    583  1.4       mrg     else								\
    584  1.4       mrg       __asm__ ("subs\t%1, %x4, %5\n\tsbc\t%0, %x2, %x3"			\
    585  1.4       mrg 	       : "=r,r" (sh), "=&r,&r" (sl)				\
    586  1.4       mrg 	       : "rZ,rZ" ((UDItype)(ah)), "rZ,rZ" ((UDItype)(bh)),	\
    587  1.4       mrg 		 "r,Z"   ((UDItype)(al)), "rI,r"  ((UDItype)(bl)) __CLOBBER_CC);\
    588  1.4       mrg   } while(0);
    589  1.4       mrg #if __GMP_GNUC_PREREQ (4,9)
    590  1.4       mrg #define umul_ppmm(w1, w0, u, v) \
    591  1.4       mrg   do {									\
    592  1.4       mrg     typedef unsigned int __ll_UTItype __attribute__((mode(TI)));	\
    593  1.4       mrg     __ll_UTItype __ll = (__ll_UTItype)(u) * (v);			\
    594  1.4       mrg     w1 = __ll >> 64;							\
    595  1.4       mrg     w0 = __ll;								\
    596  1.4       mrg   } while (0)
    597  1.4       mrg #endif
    598  1.4       mrg #if !defined (umul_ppmm)
    599  1.2     joerg #define umul_ppmm(ph, pl, m0, m1) \
    600  1.2     joerg   do {									\
    601  1.2     joerg     UDItype __m0 = (m0), __m1 = (m1);					\
    602  1.3       mrg     __asm__ ("umulh\t%0, %1, %2" : "=r" (ph) : "r" (__m0), "r" (__m1));	\
    603  1.2     joerg     (pl) = __m0 * __m1;							\
    604  1.2     joerg   } while (0)
    605  1.4       mrg #endif
    606  1.3       mrg #define count_leading_zeros(count, x)  count_leading_zeros_gcc_clz(count, x)
    607  1.3       mrg #define count_trailing_zeros(count, x)  count_trailing_zeros_gcc_ctz(count, x)
    608  1.2     joerg #endif /* __aarch64__ */
    609  1.2     joerg 
    610  1.1       mrg #if defined (__clipper__) && W_TYPE_SIZE == 32
    611  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
    612  1.1       mrg   ({union {UDItype __ll;						\
    613  1.1       mrg 	   struct {USItype __l, __h;} __i;				\
    614  1.1       mrg 	  } __x;							\
    615  1.1       mrg   __asm__ ("mulwux %2,%0"						\
    616  1.1       mrg 	   : "=r" (__x.__ll)						\
    617  1.1       mrg 	   : "%0" ((USItype)(u)), "r" ((USItype)(v)));			\
    618  1.1       mrg   (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
    619  1.1       mrg #define smul_ppmm(w1, w0, u, v) \
    620  1.1       mrg   ({union {DItype __ll;							\
    621  1.1       mrg 	   struct {SItype __l, __h;} __i;				\
    622  1.1       mrg 	  } __x;							\
    623  1.1       mrg   __asm__ ("mulwx %2,%0"						\
    624  1.1       mrg 	   : "=r" (__x.__ll)						\
    625  1.1       mrg 	   : "%0" ((SItype)(u)), "r" ((SItype)(v)));			\
    626  1.1       mrg   (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
    627  1.1       mrg #define __umulsidi3(u, v) \
    628  1.1       mrg   ({UDItype __w;							\
    629  1.1       mrg     __asm__ ("mulwux %2,%0"						\
    630  1.1       mrg 	     : "=r" (__w) : "%0" ((USItype)(u)), "r" ((USItype)(v)));	\
    631  1.1       mrg     __w; })
    632  1.1       mrg #endif /* __clipper__ */
    633  1.1       mrg 
    634  1.1       mrg /* Fujitsu vector computers.  */
    635  1.1       mrg #if defined (__uxp__) && W_TYPE_SIZE == 32
    636  1.1       mrg #define umul_ppmm(ph, pl, u, v) \
    637  1.1       mrg   do {									\
    638  1.1       mrg     union {UDItype __ll;						\
    639  1.1       mrg 	   struct {USItype __h, __l;} __i;				\
    640  1.1       mrg 	  } __x;							\
    641  1.1       mrg     __asm__ ("mult.lu %1,%2,%0"	: "=r" (__x.__ll) : "%r" (u), "rK" (v));\
    642  1.1       mrg     (ph) = __x.__i.__h;							\
    643  1.1       mrg     (pl) = __x.__i.__l;							\
    644  1.1       mrg   } while (0)
    645  1.1       mrg #define smul_ppmm(ph, pl, u, v) \
    646  1.1       mrg   do {									\
    647  1.1       mrg     union {UDItype __ll;						\
    648  1.1       mrg 	   struct {USItype __h, __l;} __i;				\
    649  1.1       mrg 	  } __x;							\
    650  1.1       mrg     __asm__ ("mult.l %1,%2,%0" : "=r" (__x.__ll) : "%r" (u), "rK" (v));	\
    651  1.1       mrg     (ph) = __x.__i.__h;							\
    652  1.1       mrg     (pl) = __x.__i.__l;							\
    653  1.1       mrg   } while (0)
    654  1.1       mrg #endif
    655  1.1       mrg 
    656  1.1       mrg #if defined (__gmicro__) && W_TYPE_SIZE == 32
    657  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
    658  1.1       mrg   __asm__ ("add.w %5,%1\n\taddx %3,%0"					\
    659  1.1       mrg 	   : "=g" (sh), "=&g" (sl)					\
    660  1.1       mrg 	   : "0"  ((USItype)(ah)), "g" ((USItype)(bh)),			\
    661  1.1       mrg 	     "%1" ((USItype)(al)), "g" ((USItype)(bl)))
    662  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
    663  1.1       mrg   __asm__ ("sub.w %5,%1\n\tsubx %3,%0"					\
    664  1.1       mrg 	   : "=g" (sh), "=&g" (sl)					\
    665  1.1       mrg 	   : "0" ((USItype)(ah)), "g" ((USItype)(bh)),			\
    666  1.1       mrg 	     "1" ((USItype)(al)), "g" ((USItype)(bl)))
    667  1.1       mrg #define umul_ppmm(ph, pl, m0, m1) \
    668  1.1       mrg   __asm__ ("mulx %3,%0,%1"						\
    669  1.1       mrg 	   : "=g" (ph), "=r" (pl)					\
    670  1.1       mrg 	   : "%0" ((USItype)(m0)), "g" ((USItype)(m1)))
    671  1.1       mrg #define udiv_qrnnd(q, r, nh, nl, d) \
    672  1.1       mrg   __asm__ ("divx %4,%0,%1"						\
    673  1.1       mrg 	   : "=g" (q), "=r" (r)						\
    674  1.1       mrg 	   : "1" ((USItype)(nh)), "0" ((USItype)(nl)), "g" ((USItype)(d)))
    675  1.1       mrg #define count_leading_zeros(count, x) \
    676  1.1       mrg   __asm__ ("bsch/1 %1,%0"						\
    677  1.1       mrg 	   : "=g" (count) : "g" ((USItype)(x)), "0" ((USItype)0))
    678  1.1       mrg #endif
    679  1.1       mrg 
    680  1.1       mrg #if defined (__hppa) && W_TYPE_SIZE == 32
    681  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
    682  1.1       mrg   __asm__ ("add%I5 %5,%r4,%1\n\taddc %r2,%r3,%0"			\
    683  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
    684  1.1       mrg 	   : "rM" (ah), "rM" (bh), "%rM" (al), "rI" (bl))
    685  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
    686  1.1       mrg   __asm__ ("sub%I4 %4,%r5,%1\n\tsubb %r2,%r3,%0"			\
    687  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
    688  1.1       mrg 	   : "rM" (ah), "rM" (bh), "rI" (al), "rM" (bl))
    689  1.1       mrg #if defined (_PA_RISC1_1)
    690  1.1       mrg #define umul_ppmm(wh, wl, u, v) \
    691  1.1       mrg   do {									\
    692  1.1       mrg     union {UDItype __ll;						\
    693  1.1       mrg 	   struct {USItype __h, __l;} __i;				\
    694  1.1       mrg 	  } __x;							\
    695  1.1       mrg     __asm__ ("xmpyu %1,%2,%0" : "=*f" (__x.__ll) : "*f" (u), "*f" (v));	\
    696  1.1       mrg     (wh) = __x.__i.__h;							\
    697  1.1       mrg     (wl) = __x.__i.__l;							\
    698  1.1       mrg   } while (0)
    699  1.1       mrg #endif
    700  1.1       mrg #define count_leading_zeros(count, x) \
    701  1.1       mrg   do {									\
    702  1.1       mrg     USItype __tmp;							\
    703  1.1       mrg     __asm__ (								\
    704  1.1       mrg        "ldi		1,%0\n"						\
    705  1.1       mrg "	extru,=		%1,15,16,%%r0	; Bits 31..16 zero?\n"		\
    706  1.1       mrg "	extru,tr	%1,15,16,%1	; No.  Shift down, skip add.\n"	\
    707  1.1       mrg "	ldo		16(%0),%0	; Yes.  Perform add.\n"		\
    708  1.1       mrg "	extru,=		%1,23,8,%%r0	; Bits 15..8 zero?\n"		\
    709  1.1       mrg "	extru,tr	%1,23,8,%1	; No.  Shift down, skip add.\n"	\
    710  1.1       mrg "	ldo		8(%0),%0	; Yes.  Perform add.\n"		\
    711  1.1       mrg "	extru,=		%1,27,4,%%r0	; Bits 7..4 zero?\n"		\
    712  1.1       mrg "	extru,tr	%1,27,4,%1	; No.  Shift down, skip add.\n"	\
    713  1.1       mrg "	ldo		4(%0),%0	; Yes.  Perform add.\n"		\
    714  1.1       mrg "	extru,=		%1,29,2,%%r0	; Bits 3..2 zero?\n"		\
    715  1.1       mrg "	extru,tr	%1,29,2,%1	; No.  Shift down, skip add.\n"	\
    716  1.1       mrg "	ldo		2(%0),%0	; Yes.  Perform add.\n"		\
    717  1.1       mrg "	extru		%1,30,1,%1	; Extract bit 1.\n"		\
    718  1.1       mrg "	sub		%0,%1,%0	; Subtract it.\n"		\
    719  1.1       mrg 	: "=r" (count), "=r" (__tmp) : "1" (x));			\
    720  1.1       mrg   } while (0)
    721  1.1       mrg #endif /* hppa */
    722  1.1       mrg 
    723  1.1       mrg /* These macros are for ABI=2.0w.  In ABI=2.0n they can't be used, since GCC
    724  1.1       mrg    (3.2) puts longlong into two adjacent 32-bit registers.  Presumably this
    725  1.1       mrg    is just a case of no direct support for 2.0n but treating it like 1.0. */
    726  1.1       mrg #if defined (__hppa) && W_TYPE_SIZE == 64 && ! defined (_LONG_LONG_LIMB)
    727  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
    728  1.1       mrg   __asm__ ("add%I5 %5,%r4,%1\n\tadd,dc %r2,%r3,%0"			\
    729  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
    730  1.1       mrg 	   : "rM" (ah), "rM" (bh), "%rM" (al), "rI" (bl))
    731  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
    732  1.1       mrg   __asm__ ("sub%I4 %4,%r5,%1\n\tsub,db %r2,%r3,%0"			\
    733  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
    734  1.1       mrg 	   : "rM" (ah), "rM" (bh), "rI" (al), "rM" (bl))
    735  1.1       mrg #endif /* hppa */
    736  1.1       mrg 
    737  1.1       mrg #if (defined (__i370__) || defined (__s390__) || defined (__mvs__)) && W_TYPE_SIZE == 32
    738  1.2     joerg #if defined (__zarch__) || defined (HAVE_HOST_CPU_s390_zarch)
    739  1.2     joerg #define add_ssaaaa(sh, sl, ah, al, bh, bl)				\
    740  1.2     joerg   do {									\
    741  1.2     joerg /*  if (__builtin_constant_p (bl))					\
    742  1.2     joerg       __asm__ ("alfi\t%1,%o5\n\talcr\t%0,%3"				\
    743  1.2     joerg 	       : "=r" (sh), "=&r" (sl)					\
    744  1.2     joerg 	       : "0"  (ah), "r" (bh), "%1" (al), "n" (bl) __CLOBBER_CC);\
    745  1.2     joerg     else								\
    746  1.2     joerg */    __asm__ ("alr\t%1,%5\n\talcr\t%0,%3"				\
    747  1.2     joerg 	       : "=r" (sh), "=&r" (sl)					\
    748  1.2     joerg 	       : "0"  (ah), "r" (bh), "%1" (al), "r" (bl)__CLOBBER_CC);	\
    749  1.2     joerg   } while (0)
    750  1.2     joerg #define sub_ddmmss(sh, sl, ah, al, bh, bl)				\
    751  1.2     joerg   do {									\
    752  1.2     joerg /*  if (__builtin_constant_p (bl))					\
    753  1.2     joerg       __asm__ ("slfi\t%1,%o5\n\tslbr\t%0,%3"				\
    754  1.2     joerg 	       : "=r" (sh), "=&r" (sl)					\
    755  1.2     joerg 	       : "0" (ah), "r" (bh), "1" (al), "n" (bl) __CLOBBER_CC);	\
    756  1.2     joerg     else								\
    757  1.2     joerg */    __asm__ ("slr\t%1,%5\n\tslbr\t%0,%3"				\
    758  1.2     joerg 	       : "=r" (sh), "=&r" (sl)					\
    759  1.2     joerg 	       : "0" (ah), "r" (bh), "1" (al), "r" (bl) __CLOBBER_CC);	\
    760  1.2     joerg   } while (0)
    761  1.2     joerg #if __GMP_GNUC_PREREQ (4,5)
    762  1.2     joerg #define umul_ppmm(xh, xl, m0, m1)					\
    763  1.2     joerg   do {									\
    764  1.2     joerg     union {UDItype __ll;						\
    765  1.2     joerg 	   struct {USItype __h, __l;} __i;				\
    766  1.2     joerg 	  } __x;							\
    767  1.2     joerg     __x.__ll = (UDItype) (m0) * (UDItype) (m1);				\
    768  1.2     joerg     (xh) = __x.__i.__h; (xl) = __x.__i.__l;				\
    769  1.2     joerg   } while (0)
    770  1.2     joerg #else
    771  1.2     joerg #if 0
    772  1.2     joerg /* FIXME: this fails if gcc knows about the 64-bit registers.  Use only
    773  1.2     joerg    with a new enough processor pretending we have 32-bit registers.  */
    774  1.2     joerg #define umul_ppmm(xh, xl, m0, m1)					\
    775  1.2     joerg   do {									\
    776  1.2     joerg     union {UDItype __ll;						\
    777  1.2     joerg 	   struct {USItype __h, __l;} __i;				\
    778  1.2     joerg 	  } __x;							\
    779  1.2     joerg     __asm__ ("mlr\t%0,%2"						\
    780  1.2     joerg 	     : "=r" (__x.__ll)						\
    781  1.2     joerg 	     : "%0" (m0), "r" (m1));					\
    782  1.2     joerg     (xh) = __x.__i.__h; (xl) = __x.__i.__l;				\
    783  1.2     joerg   } while (0)
    784  1.2     joerg #else
    785  1.2     joerg #define umul_ppmm(xh, xl, m0, m1)					\
    786  1.2     joerg   do {									\
    787  1.2     joerg   /* When we have 64-bit regs and gcc is aware of that, we cannot simply use
    788  1.2     joerg      DImode for the product, since that would be allocated to a single 64-bit
    789  1.2     joerg      register, whereas mlr uses the low 32-bits of an even-odd register pair.
    790  1.2     joerg   */									\
    791  1.2     joerg     register USItype __r0 __asm__ ("0");				\
    792  1.2     joerg     register USItype __r1 __asm__ ("1") = (m0);				\
    793  1.2     joerg     __asm__ ("mlr\t%0,%3"						\
    794  1.2     joerg 	     : "=r" (__r0), "=r" (__r1)					\
    795  1.2     joerg 	     : "r" (__r1), "r" (m1));					\
    796  1.2     joerg     (xh) = __r0; (xl) = __r1;						\
    797  1.2     joerg   } while (0)
    798  1.2     joerg #endif /* if 0 */
    799  1.2     joerg #endif
    800  1.2     joerg #if 0
    801  1.2     joerg /* FIXME: this fails if gcc knows about the 64-bit registers.  Use only
    802  1.2     joerg    with a new enough processor pretending we have 32-bit registers.  */
    803  1.2     joerg #define udiv_qrnnd(q, r, n1, n0, d)					\
    804  1.2     joerg   do {									\
    805  1.2     joerg     union {UDItype __ll;						\
    806  1.2     joerg 	   struct {USItype __h, __l;} __i;				\
    807  1.2     joerg 	  } __x;							\
    808  1.2     joerg     __x.__i.__h = n1; __x.__i.__l = n0;					\
    809  1.2     joerg     __asm__ ("dlr\t%0,%2"						\
    810  1.2     joerg 	     : "=r" (__x.__ll)						\
    811  1.2     joerg 	     : "0" (__x.__ll), "r" (d));				\
    812  1.2     joerg     (q) = __x.__i.__l; (r) = __x.__i.__h;				\
    813  1.2     joerg   } while (0)
    814  1.2     joerg #else
    815  1.2     joerg #define udiv_qrnnd(q, r, n1, n0, d)					\
    816  1.2     joerg   do {									\
    817  1.2     joerg     register USItype __r0 __asm__ ("0") = (n1);				\
    818  1.2     joerg     register USItype __r1 __asm__ ("1") = (n0);				\
    819  1.2     joerg     __asm__ ("dlr\t%0,%4"						\
    820  1.2     joerg 	     : "=r" (__r0), "=r" (__r1)					\
    821  1.2     joerg 	     : "r" (__r0), "r" (__r1), "r" (d));			\
    822  1.2     joerg     (q) = __r1; (r) = __r0;						\
    823  1.2     joerg   } while (0)
    824  1.2     joerg #endif /* if 0 */
    825  1.2     joerg #else /* if __zarch__ */
    826  1.2     joerg /* FIXME: this fails if gcc knows about the 64-bit registers.  */
    827  1.2     joerg #define smul_ppmm(xh, xl, m0, m1)					\
    828  1.1       mrg   do {									\
    829  1.1       mrg     union {DItype __ll;							\
    830  1.1       mrg 	   struct {USItype __h, __l;} __i;				\
    831  1.1       mrg 	  } __x;							\
    832  1.2     joerg     __asm__ ("mr\t%0,%2"						\
    833  1.2     joerg 	     : "=r" (__x.__ll)						\
    834  1.2     joerg 	     : "%0" (m0), "r" (m1));					\
    835  1.1       mrg     (xh) = __x.__i.__h; (xl) = __x.__i.__l;				\
    836  1.1       mrg   } while (0)
    837  1.2     joerg /* FIXME: this fails if gcc knows about the 64-bit registers.  */
    838  1.2     joerg #define sdiv_qrnnd(q, r, n1, n0, d)					\
    839  1.1       mrg   do {									\
    840  1.1       mrg     union {DItype __ll;							\
    841  1.1       mrg 	   struct {USItype __h, __l;} __i;				\
    842  1.1       mrg 	  } __x;							\
    843  1.1       mrg     __x.__i.__h = n1; __x.__i.__l = n0;					\
    844  1.2     joerg     __asm__ ("dr\t%0,%2"						\
    845  1.1       mrg 	     : "=r" (__x.__ll)						\
    846  1.1       mrg 	     : "0" (__x.__ll), "r" (d));				\
    847  1.1       mrg     (q) = __x.__i.__l; (r) = __x.__i.__h;				\
    848  1.1       mrg   } while (0)
    849  1.2     joerg #endif /* if __zarch__ */
    850  1.2     joerg #endif
    851  1.2     joerg 
    852  1.2     joerg #if defined (__s390x__) && W_TYPE_SIZE == 64
    853  1.2     joerg /* We need to cast operands with register constraints, otherwise their types
    854  1.2     joerg    will be assumed to be SImode by gcc.  For these machines, such operations
    855  1.2     joerg    will insert a value into the low 32 bits, and leave the high 32 bits with
    856  1.2     joerg    garbage.  */
    857  1.2     joerg #define add_ssaaaa(sh, sl, ah, al, bh, bl)				\
    858  1.2     joerg   do {									\
    859  1.2     joerg     __asm__ ("algr\t%1,%5\n\talcgr\t%0,%3"				\
    860  1.2     joerg 	       : "=r" (sh), "=&r" (sl)					\
    861  1.2     joerg 	       : "0"  ((UDItype)(ah)), "r" ((UDItype)(bh)),		\
    862  1.2     joerg 		 "%1" ((UDItype)(al)), "r" ((UDItype)(bl)) __CLOBBER_CC); \
    863  1.2     joerg   } while (0)
    864  1.2     joerg #define sub_ddmmss(sh, sl, ah, al, bh, bl)				\
    865  1.2     joerg   do {									\
    866  1.2     joerg     __asm__ ("slgr\t%1,%5\n\tslbgr\t%0,%3"				\
    867  1.2     joerg 	     : "=r" (sh), "=&r" (sl)					\
    868  1.2     joerg 	     : "0" ((UDItype)(ah)), "r" ((UDItype)(bh)),		\
    869  1.2     joerg 	       "1" ((UDItype)(al)), "r" ((UDItype)(bl)) __CLOBBER_CC);	\
    870  1.2     joerg   } while (0)
    871  1.2     joerg #define umul_ppmm(xh, xl, m0, m1)					\
    872  1.2     joerg   do {									\
    873  1.2     joerg     union {unsigned int __attribute__ ((mode(TI))) __ll;		\
    874  1.2     joerg 	   struct {UDItype __h, __l;} __i;				\
    875  1.2     joerg 	  } __x;							\
    876  1.2     joerg     __asm__ ("mlgr\t%0,%2"						\
    877  1.2     joerg 	     : "=r" (__x.__ll)						\
    878  1.2     joerg 	     : "%0" ((UDItype)(m0)), "r" ((UDItype)(m1)));		\
    879  1.2     joerg     (xh) = __x.__i.__h; (xl) = __x.__i.__l;				\
    880  1.2     joerg   } while (0)
    881  1.2     joerg #define udiv_qrnnd(q, r, n1, n0, d)					\
    882  1.2     joerg   do {									\
    883  1.2     joerg     union {unsigned int __attribute__ ((mode(TI))) __ll;		\
    884  1.2     joerg 	   struct {UDItype __h, __l;} __i;				\
    885  1.2     joerg 	  } __x;							\
    886  1.2     joerg     __x.__i.__h = n1; __x.__i.__l = n0;					\
    887  1.2     joerg     __asm__ ("dlgr\t%0,%2"						\
    888  1.2     joerg 	     : "=r" (__x.__ll)						\
    889  1.2     joerg 	     : "0" (__x.__ll), "r" ((UDItype)(d)));			\
    890  1.2     joerg     (q) = __x.__i.__l; (r) = __x.__i.__h;				\
    891  1.2     joerg   } while (0)
    892  1.2     joerg #if 0 /* FIXME: Enable for z10 (?) */
    893  1.2     joerg #define count_leading_zeros(cnt, x)					\
    894  1.2     joerg   do {									\
    895  1.2     joerg     union {unsigned int __attribute__ ((mode(TI))) __ll;		\
    896  1.2     joerg 	   struct {UDItype __h, __l;} __i;				\
    897  1.2     joerg 	  } __clr_cnt;							\
    898  1.2     joerg     __asm__ ("flogr\t%0,%1"						\
    899  1.2     joerg 	     : "=r" (__clr_cnt.__ll)					\
    900  1.2     joerg 	     : "r" (x) __CLOBBER_CC);					\
    901  1.2     joerg     (cnt) = __clr_cnt.__i.__h;						\
    902  1.2     joerg   } while (0)
    903  1.2     joerg #endif
    904  1.1       mrg #endif
    905  1.1       mrg 
    906  1.3       mrg /* On x86 and x86_64, every asm implicitly clobbers "flags" and "fpsr",
    907  1.3       mrg    so we don't need __CLOBBER_CC.  */
    908  1.1       mrg #if (defined (__i386__) || defined (__i486__)) && W_TYPE_SIZE == 32
    909  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
    910  1.1       mrg   __asm__ ("addl %5,%k1\n\tadcl %3,%k0"					\
    911  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
    912  1.1       mrg 	   : "0"  ((USItype)(ah)), "g" ((USItype)(bh)),			\
    913  1.1       mrg 	     "%1" ((USItype)(al)), "g" ((USItype)(bl)))
    914  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
    915  1.1       mrg   __asm__ ("subl %5,%k1\n\tsbbl %3,%k0"					\
    916  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
    917  1.1       mrg 	   : "0" ((USItype)(ah)), "g" ((USItype)(bh)),			\
    918  1.1       mrg 	     "1" ((USItype)(al)), "g" ((USItype)(bl)))
    919  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
    920  1.1       mrg   __asm__ ("mull %3"							\
    921  1.1       mrg 	   : "=a" (w0), "=d" (w1)					\
    922  1.1       mrg 	   : "%0" ((USItype)(u)), "rm" ((USItype)(v)))
    923  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, dx) /* d renamed to dx avoiding "=d" */\
    924  1.1       mrg   __asm__ ("divl %4"		     /* stringification in K&R C */	\
    925  1.1       mrg 	   : "=a" (q), "=d" (r)						\
    926  1.1       mrg 	   : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "rm" ((USItype)(dx)))
    927  1.1       mrg 
    928  1.1       mrg #if HAVE_HOST_CPU_i586 || HAVE_HOST_CPU_pentium || HAVE_HOST_CPU_pentiummmx
    929  1.1       mrg /* Pentium bsrl takes between 10 and 72 cycles depending where the most
    930  1.1       mrg    significant 1 bit is, hence the use of the following alternatives.  bsfl
    931  1.1       mrg    is slow too, between 18 and 42 depending where the least significant 1
    932  1.1       mrg    bit is, so let the generic count_trailing_zeros below make use of the
    933  1.1       mrg    count_leading_zeros here too.  */
    934  1.1       mrg 
    935  1.1       mrg #if HAVE_HOST_CPU_pentiummmx && ! defined (LONGLONG_STANDALONE)
    936  1.1       mrg /* The following should be a fixed 14 or 15 cycles, but possibly plus an L1
    937  1.1       mrg    cache miss reading from __clz_tab.  For P55 it's favoured over the float
    938  1.1       mrg    below so as to avoid mixing MMX and x87, since the penalty for switching
    939  1.1       mrg    between the two is about 100 cycles.
    940  1.1       mrg 
    941  1.1       mrg    The asm block sets __shift to -3 if the high 24 bits are clear, -2 for
    942  1.1       mrg    16, -1 for 8, or 0 otherwise.  This could be written equivalently as
    943  1.1       mrg    follows, but as of gcc 2.95.2 it results in conditional jumps.
    944  1.1       mrg 
    945  1.1       mrg        __shift = -(__n < 0x1000000);
    946  1.1       mrg        __shift -= (__n < 0x10000);
    947  1.1       mrg        __shift -= (__n < 0x100);
    948  1.1       mrg 
    949  1.1       mrg    The middle two sbbl and cmpl's pair, and with luck something gcc
    950  1.1       mrg    generates might pair with the first cmpl and the last sbbl.  The "32+1"
    951  1.1       mrg    constant could be folded into __clz_tab[], but it doesn't seem worth
    952  1.1       mrg    making a different table just for that.  */
    953  1.1       mrg 
    954  1.1       mrg #define count_leading_zeros(c,n)					\
    955  1.1       mrg   do {									\
    956  1.1       mrg     USItype  __n = (n);							\
    957  1.1       mrg     USItype  __shift;							\
    958  1.1       mrg     __asm__ ("cmpl  $0x1000000, %1\n"					\
    959  1.1       mrg 	     "sbbl  %0, %0\n"						\
    960  1.1       mrg 	     "cmpl  $0x10000, %1\n"					\
    961  1.1       mrg 	     "sbbl  $0, %0\n"						\
    962  1.1       mrg 	     "cmpl  $0x100, %1\n"					\
    963  1.1       mrg 	     "sbbl  $0, %0\n"						\
    964  1.1       mrg 	     : "=&r" (__shift) : "r"  (__n));				\
    965  1.1       mrg     __shift = __shift*8 + 24 + 1;					\
    966  1.1       mrg     (c) = 32 + 1 - __shift - __clz_tab[__n >> __shift];			\
    967  1.1       mrg   } while (0)
    968  1.1       mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
    969  1.1       mrg #define COUNT_LEADING_ZEROS_0   31   /* n==0 indistinguishable from n==1 */
    970  1.1       mrg 
    971  1.1       mrg #else /* ! pentiummmx || LONGLONG_STANDALONE */
    972  1.1       mrg /* The following should be a fixed 14 cycles or so.  Some scheduling
    973  1.1       mrg    opportunities should be available between the float load/store too.  This
    974  1.1       mrg    sort of code is used in gcc 3 for __builtin_ffs (with "n&-n") and is
    975  1.1       mrg    apparently suggested by the Intel optimizing manual (don't know exactly
    976  1.1       mrg    where).  gcc 2.95 or up will be best for this, so the "double" is
    977  1.1       mrg    correctly aligned on the stack.  */
    978  1.1       mrg #define count_leading_zeros(c,n)					\
    979  1.1       mrg   do {									\
    980  1.1       mrg     union {								\
    981  1.1       mrg       double    d;							\
    982  1.1       mrg       unsigned  a[2];							\
    983  1.1       mrg     } __u;								\
    984  1.1       mrg     __u.d = (UWtype) (n);						\
    985  1.1       mrg     (c) = 0x3FF + 31 - (__u.a[1] >> 20);				\
    986  1.1       mrg   } while (0)
    987  1.1       mrg #define COUNT_LEADING_ZEROS_0   (0x3FF + 31)
    988  1.1       mrg #endif /* pentiummx */
    989  1.1       mrg 
    990  1.1       mrg #else /* ! pentium */
    991  1.1       mrg 
    992  1.1       mrg #if __GMP_GNUC_PREREQ (3,4)  /* using bsrl */
    993  1.1       mrg #define count_leading_zeros(count,x)  count_leading_zeros_gcc_clz(count,x)
    994  1.1       mrg #endif /* gcc clz */
    995  1.1       mrg 
    996  1.1       mrg /* On P6, gcc prior to 3.0 generates a partial register stall for
    997  1.1       mrg    __cbtmp^31, due to using "xorb $31" instead of "xorl $31", the former
    998  1.1       mrg    being 1 code byte smaller.  "31-__cbtmp" is a workaround, probably at the
    999  1.1       mrg    cost of one extra instruction.  Do this for "i386" too, since that means
   1000  1.1       mrg    generic x86.  */
   1001  1.3       mrg #if ! defined (count_leading_zeros) && __GNUC__ < 3			\
   1002  1.1       mrg   && (HAVE_HOST_CPU_i386						\
   1003  1.1       mrg       || HAVE_HOST_CPU_i686						\
   1004  1.1       mrg       || HAVE_HOST_CPU_pentiumpro					\
   1005  1.1       mrg       || HAVE_HOST_CPU_pentium2						\
   1006  1.1       mrg       || HAVE_HOST_CPU_pentium3)
   1007  1.1       mrg #define count_leading_zeros(count, x)					\
   1008  1.1       mrg   do {									\
   1009  1.1       mrg     USItype __cbtmp;							\
   1010  1.1       mrg     ASSERT ((x) != 0);							\
   1011  1.1       mrg     __asm__ ("bsrl %1,%0" : "=r" (__cbtmp) : "rm" ((USItype)(x)));	\
   1012  1.1       mrg     (count) = 31 - __cbtmp;						\
   1013  1.1       mrg   } while (0)
   1014  1.1       mrg #endif /* gcc<3 asm bsrl */
   1015  1.1       mrg 
   1016  1.1       mrg #ifndef count_leading_zeros
   1017  1.1       mrg #define count_leading_zeros(count, x)					\
   1018  1.1       mrg   do {									\
   1019  1.1       mrg     USItype __cbtmp;							\
   1020  1.1       mrg     ASSERT ((x) != 0);							\
   1021  1.1       mrg     __asm__ ("bsrl %1,%0" : "=r" (__cbtmp) : "rm" ((USItype)(x)));	\
   1022  1.1       mrg     (count) = __cbtmp ^ 31;						\
   1023  1.1       mrg   } while (0)
   1024  1.1       mrg #endif /* asm bsrl */
   1025  1.1       mrg 
   1026  1.1       mrg #if __GMP_GNUC_PREREQ (3,4)  /* using bsfl */
   1027  1.1       mrg #define count_trailing_zeros(count,x)  count_trailing_zeros_gcc_ctz(count,x)
   1028  1.1       mrg #endif /* gcc ctz */
   1029  1.1       mrg 
   1030  1.1       mrg #ifndef count_trailing_zeros
   1031  1.1       mrg #define count_trailing_zeros(count, x)					\
   1032  1.1       mrg   do {									\
   1033  1.1       mrg     ASSERT ((x) != 0);							\
   1034  1.1       mrg     __asm__ ("bsfl %1,%k0" : "=r" (count) : "rm" ((USItype)(x)));	\
   1035  1.1       mrg   } while (0)
   1036  1.1       mrg #endif /* asm bsfl */
   1037  1.1       mrg 
   1038  1.1       mrg #endif /* ! pentium */
   1039  1.1       mrg 
   1040  1.1       mrg #endif /* 80x86 */
   1041  1.1       mrg 
   1042  1.1       mrg #if defined (__amd64__) && W_TYPE_SIZE == 64
   1043  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
   1044  1.1       mrg   __asm__ ("addq %5,%q1\n\tadcq %3,%q0"					\
   1045  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
   1046  1.1       mrg 	   : "0"  ((UDItype)(ah)), "rme" ((UDItype)(bh)),		\
   1047  1.1       mrg 	     "%1" ((UDItype)(al)), "rme" ((UDItype)(bl)))
   1048  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
   1049  1.1       mrg   __asm__ ("subq %5,%q1\n\tsbbq %3,%q0"					\
   1050  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
   1051  1.1       mrg 	   : "0" ((UDItype)(ah)), "rme" ((UDItype)(bh)),		\
   1052  1.1       mrg 	     "1" ((UDItype)(al)), "rme" ((UDItype)(bl)))
   1053  1.4       mrg #if X86_ASM_MULX \
   1054  1.4       mrg    && (HAVE_HOST_CPU_haswell || HAVE_HOST_CPU_broadwell \
   1055  1.4       mrg        || HAVE_HOST_CPU_skylake || HAVE_HOST_CPU_bd4 || HAVE_HOST_CPU_zen)
   1056  1.4       mrg #define umul_ppmm(w1, w0, u, v) \
   1057  1.5       mrg   __asm__ ("mulx\t%3, %q0, %q1"						\
   1058  1.4       mrg 	   : "=r" (w0), "=r" (w1)					\
   1059  1.4       mrg 	   : "%d" ((UDItype)(u)), "rm" ((UDItype)(v)))
   1060  1.4       mrg #else
   1061  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1062  1.4       mrg   __asm__ ("mulq\t%3"							\
   1063  1.1       mrg 	   : "=a" (w0), "=d" (w1)					\
   1064  1.1       mrg 	   : "%0" ((UDItype)(u)), "rm" ((UDItype)(v)))
   1065  1.4       mrg #endif
   1066  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, dx) /* d renamed to dx avoiding "=d" */\
   1067  1.1       mrg   __asm__ ("divq %4"		     /* stringification in K&R C */	\
   1068  1.1       mrg 	   : "=a" (q), "=d" (r)						\
   1069  1.1       mrg 	   : "0" ((UDItype)(n0)), "1" ((UDItype)(n1)), "rm" ((UDItype)(dx)))
   1070  1.4       mrg 
   1071  1.4       mrg #if HAVE_HOST_CPU_haswell || HAVE_HOST_CPU_broadwell || HAVE_HOST_CPU_skylake \
   1072  1.4       mrg   || HAVE_HOST_CPU_k10 || HAVE_HOST_CPU_bd1 || HAVE_HOST_CPU_bd2	\
   1073  1.4       mrg   || HAVE_HOST_CPU_bd3 || HAVE_HOST_CPU_bd4 || HAVE_HOST_CPU_zen	\
   1074  1.4       mrg   || HAVE_HOST_CPU_bobcat || HAVE_HOST_CPU_jaguar
   1075  1.4       mrg #define count_leading_zeros(count, x)					\
   1076  1.4       mrg   do {									\
   1077  1.4       mrg     /* This is lzcnt, spelled for older assemblers.  Destination and */	\
   1078  1.4       mrg     /* source must be a 64-bit registers, hence cast and %q.         */	\
   1079  1.4       mrg     __asm__ ("rep;bsr\t%1, %q0" : "=r" (count) : "rm" ((UDItype)(x)));	\
   1080  1.4       mrg   } while (0)
   1081  1.4       mrg #define COUNT_LEADING_ZEROS_0 64
   1082  1.4       mrg #else
   1083  1.1       mrg #define count_leading_zeros(count, x)					\
   1084  1.1       mrg   do {									\
   1085  1.1       mrg     UDItype __cbtmp;							\
   1086  1.1       mrg     ASSERT ((x) != 0);							\
   1087  1.4       mrg     __asm__ ("bsr\t%1,%0" : "=r" (__cbtmp) : "rm" ((UDItype)(x)));	\
   1088  1.1       mrg     (count) = __cbtmp ^ 63;						\
   1089  1.1       mrg   } while (0)
   1090  1.4       mrg #endif
   1091  1.4       mrg 
   1092  1.4       mrg #if HAVE_HOST_CPU_bd2 || HAVE_HOST_CPU_bd3 || HAVE_HOST_CPU_bd4 \
   1093  1.4       mrg   || HAVE_HOST_CPU_zen || HAVE_HOST_CPU_jaguar
   1094  1.4       mrg #define count_trailing_zeros(count, x)					\
   1095  1.4       mrg   do {									\
   1096  1.4       mrg     /* This is tzcnt, spelled for older assemblers.  Destination and */	\
   1097  1.4       mrg     /* source must be a 64-bit registers, hence cast and %q.         */	\
   1098  1.4       mrg     __asm__ ("rep;bsf\t%1, %q0" : "=r" (count) : "rm" ((UDItype)(x)));	\
   1099  1.4       mrg   } while (0)
   1100  1.4       mrg #define COUNT_TRAILING_ZEROS_0 64
   1101  1.4       mrg #else
   1102  1.1       mrg #define count_trailing_zeros(count, x)					\
   1103  1.1       mrg   do {									\
   1104  1.1       mrg     ASSERT ((x) != 0);							\
   1105  1.4       mrg     __asm__ ("bsf\t%1, %q0" : "=r" (count) : "rm" ((UDItype)(x)));	\
   1106  1.1       mrg   } while (0)
   1107  1.4       mrg #endif
   1108  1.3       mrg #endif /* __amd64__ */
   1109  1.1       mrg 
   1110  1.1       mrg #if defined (__i860__) && W_TYPE_SIZE == 32
   1111  1.1       mrg #define rshift_rhlc(r,h,l,c) \
   1112  1.1       mrg   __asm__ ("shr %3,r0,r0\;shrd %1,%2,%0"				\
   1113  1.1       mrg 	   "=r" (r) : "r" (h), "r" (l), "rn" (c))
   1114  1.1       mrg #endif /* i860 */
   1115  1.1       mrg 
   1116  1.1       mrg #if defined (__i960__) && W_TYPE_SIZE == 32
   1117  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
   1118  1.1       mrg   __asm__ ("cmpo 1,0\;addc %5,%4,%1\;addc %3,%2,%0"			\
   1119  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
   1120  1.1       mrg 	   : "dI" (ah), "dI" (bh), "%dI" (al), "dI" (bl))
   1121  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
   1122  1.1       mrg   __asm__ ("cmpo 0,0\;subc %5,%4,%1\;subc %3,%2,%0"			\
   1123  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
   1124  1.1       mrg 	   : "dI" (ah), "dI" (bh), "dI" (al), "dI" (bl))
   1125  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1126  1.1       mrg   ({union {UDItype __ll;						\
   1127  1.1       mrg 	   struct {USItype __l, __h;} __i;				\
   1128  1.1       mrg 	  } __x;							\
   1129  1.1       mrg   __asm__ ("emul %2,%1,%0"						\
   1130  1.1       mrg 	   : "=d" (__x.__ll) : "%dI" (u), "dI" (v));			\
   1131  1.1       mrg   (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
   1132  1.1       mrg #define __umulsidi3(u, v) \
   1133  1.1       mrg   ({UDItype __w;							\
   1134  1.1       mrg     __asm__ ("emul %2,%1,%0" : "=d" (__w) : "%dI" (u), "dI" (v));	\
   1135  1.1       mrg     __w; })
   1136  1.1       mrg #define udiv_qrnnd(q, r, nh, nl, d) \
   1137  1.1       mrg   do {									\
   1138  1.1       mrg     union {UDItype __ll;						\
   1139  1.1       mrg 	   struct {USItype __l, __h;} __i;				\
   1140  1.1       mrg 	  } __nn;							\
   1141  1.1       mrg     __nn.__i.__h = (nh); __nn.__i.__l = (nl);				\
   1142  1.1       mrg     __asm__ ("ediv %d,%n,%0"						\
   1143  1.1       mrg 	   : "=d" (__rq.__ll) : "dI" (__nn.__ll), "dI" (d));		\
   1144  1.1       mrg     (r) = __rq.__i.__l; (q) = __rq.__i.__h;				\
   1145  1.1       mrg   } while (0)
   1146  1.1       mrg #define count_leading_zeros(count, x) \
   1147  1.1       mrg   do {									\
   1148  1.1       mrg     USItype __cbtmp;							\
   1149  1.1       mrg     __asm__ ("scanbit %1,%0" : "=r" (__cbtmp) : "r" (x));		\
   1150  1.1       mrg     (count) = __cbtmp ^ 31;						\
   1151  1.1       mrg   } while (0)
   1152  1.1       mrg #define COUNT_LEADING_ZEROS_0 (-32) /* sic */
   1153  1.1       mrg #if defined (__i960mx)		/* what is the proper symbol to test??? */
   1154  1.1       mrg #define rshift_rhlc(r,h,l,c) \
   1155  1.1       mrg   do {									\
   1156  1.1       mrg     union {UDItype __ll;						\
   1157  1.1       mrg 	   struct {USItype __l, __h;} __i;				\
   1158  1.1       mrg 	  } __nn;							\
   1159  1.1       mrg     __nn.__i.__h = (h); __nn.__i.__l = (l);				\
   1160  1.1       mrg     __asm__ ("shre %2,%1,%0" : "=d" (r) : "dI" (__nn.__ll), "dI" (c));	\
   1161  1.1       mrg   }
   1162  1.1       mrg #endif /* i960mx */
   1163  1.1       mrg #endif /* i960 */
   1164  1.1       mrg 
   1165  1.1       mrg #if (defined (__mc68000__) || defined (__mc68020__) || defined(mc68020) \
   1166  1.1       mrg      || defined (__m68k__) || defined (__mc5200__) || defined (__mc5206e__) \
   1167  1.1       mrg      || defined (__mc5307__)) && W_TYPE_SIZE == 32
   1168  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
   1169  1.1       mrg   __asm__ ("add%.l %5,%1\n\taddx%.l %3,%0"				\
   1170  1.1       mrg 	   : "=d" (sh), "=&d" (sl)					\
   1171  1.1       mrg 	   : "0"  ((USItype)(ah)), "d" ((USItype)(bh)),			\
   1172  1.1       mrg 	     "%1" ((USItype)(al)), "g" ((USItype)(bl)))
   1173  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
   1174  1.1       mrg   __asm__ ("sub%.l %5,%1\n\tsubx%.l %3,%0"				\
   1175  1.1       mrg 	   : "=d" (sh), "=&d" (sl)					\
   1176  1.1       mrg 	   : "0" ((USItype)(ah)), "d" ((USItype)(bh)),			\
   1177  1.1       mrg 	     "1" ((USItype)(al)), "g" ((USItype)(bl)))
   1178  1.1       mrg /* The '020, '030, '040 and CPU32 have 32x32->64 and 64/32->32q-32r.  */
   1179  1.1       mrg #if defined (__mc68020__) || defined(mc68020) \
   1180  1.1       mrg      || defined (__mc68030__) || defined (mc68030) \
   1181  1.1       mrg      || defined (__mc68040__) || defined (mc68040) \
   1182  1.1       mrg      || defined (__mcpu32__) || defined (mcpu32) \
   1183  1.1       mrg      || defined (__NeXT__)
   1184  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1185  1.1       mrg   __asm__ ("mulu%.l %3,%1:%0"						\
   1186  1.1       mrg 	   : "=d" (w0), "=d" (w1)					\
   1187  1.1       mrg 	   : "%0" ((USItype)(u)), "dmi" ((USItype)(v)))
   1188  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, d) \
   1189  1.1       mrg   __asm__ ("divu%.l %4,%1:%0"						\
   1190  1.1       mrg 	   : "=d" (q), "=d" (r)						\
   1191  1.1       mrg 	   : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "dmi" ((USItype)(d)))
   1192  1.1       mrg #define sdiv_qrnnd(q, r, n1, n0, d) \
   1193  1.1       mrg   __asm__ ("divs%.l %4,%1:%0"						\
   1194  1.1       mrg 	   : "=d" (q), "=d" (r)						\
   1195  1.1       mrg 	   : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "dmi" ((USItype)(d)))
   1196  1.1       mrg #else /* for other 68k family members use 16x16->32 multiplication */
   1197  1.1       mrg #define umul_ppmm(xh, xl, a, b) \
   1198  1.1       mrg   do { USItype __umul_tmp1, __umul_tmp2;				\
   1199  1.1       mrg 	__asm__ ("| Inlined umul_ppmm\n"				\
   1200  1.1       mrg "	move%.l	%5,%3\n"						\
   1201  1.1       mrg "	move%.l	%2,%0\n"						\
   1202  1.1       mrg "	move%.w	%3,%1\n"						\
   1203  1.1       mrg "	swap	%3\n"							\
   1204  1.1       mrg "	swap	%0\n"							\
   1205  1.1       mrg "	mulu%.w	%2,%1\n"						\
   1206  1.1       mrg "	mulu%.w	%3,%0\n"						\
   1207  1.1       mrg "	mulu%.w	%2,%3\n"						\
   1208  1.1       mrg "	swap	%2\n"							\
   1209  1.1       mrg "	mulu%.w	%5,%2\n"						\
   1210  1.1       mrg "	add%.l	%3,%2\n"						\
   1211  1.1       mrg "	jcc	1f\n"							\
   1212  1.1       mrg "	add%.l	%#0x10000,%0\n"						\
   1213  1.1       mrg "1:	move%.l	%2,%3\n"						\
   1214  1.1       mrg "	clr%.w	%2\n"							\
   1215  1.1       mrg "	swap	%2\n"							\
   1216  1.1       mrg "	swap	%3\n"							\
   1217  1.1       mrg "	clr%.w	%3\n"							\
   1218  1.1       mrg "	add%.l	%3,%1\n"						\
   1219  1.1       mrg "	addx%.l	%2,%0\n"						\
   1220  1.1       mrg "	| End inlined umul_ppmm"					\
   1221  1.1       mrg 	      : "=&d" (xh), "=&d" (xl),					\
   1222  1.1       mrg 		"=d" (__umul_tmp1), "=&d" (__umul_tmp2)			\
   1223  1.1       mrg 	      : "%2" ((USItype)(a)), "d" ((USItype)(b)));		\
   1224  1.1       mrg   } while (0)
   1225  1.1       mrg #endif /* not mc68020 */
   1226  1.1       mrg /* The '020, '030, '040 and '060 have bitfield insns.
   1227  1.1       mrg    GCC 3.4 defines __mc68020__ when in CPU32 mode, check for __mcpu32__ to
   1228  1.1       mrg    exclude bfffo on that chip (bitfield insns not available).  */
   1229  1.1       mrg #if (defined (__mc68020__) || defined (mc68020)    \
   1230  1.1       mrg      || defined (__mc68030__) || defined (mc68030) \
   1231  1.1       mrg      || defined (__mc68040__) || defined (mc68040) \
   1232  1.1       mrg      || defined (__mc68060__) || defined (mc68060) \
   1233  1.3       mrg      || defined (__NeXT__))			   \
   1234  1.1       mrg   && ! defined (__mcpu32__)
   1235  1.1       mrg #define count_leading_zeros(count, x) \
   1236  1.1       mrg   __asm__ ("bfffo %1{%b2:%b2},%0"					\
   1237  1.1       mrg 	   : "=d" (count)						\
   1238  1.1       mrg 	   : "od" ((USItype) (x)), "n" (0))
   1239  1.1       mrg #define COUNT_LEADING_ZEROS_0 32
   1240  1.1       mrg #endif
   1241  1.1       mrg #endif /* mc68000 */
   1242  1.1       mrg 
   1243  1.1       mrg #if defined (__m88000__) && W_TYPE_SIZE == 32
   1244  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
   1245  1.1       mrg   __asm__ ("addu.co %1,%r4,%r5\n\taddu.ci %0,%r2,%r3"			\
   1246  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
   1247  1.1       mrg 	   : "rJ" (ah), "rJ" (bh), "%rJ" (al), "rJ" (bl))
   1248  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
   1249  1.1       mrg   __asm__ ("subu.co %1,%r4,%r5\n\tsubu.ci %0,%r2,%r3"			\
   1250  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
   1251  1.1       mrg 	   : "rJ" (ah), "rJ" (bh), "rJ" (al), "rJ" (bl))
   1252  1.1       mrg #define count_leading_zeros(count, x) \
   1253  1.1       mrg   do {									\
   1254  1.1       mrg     USItype __cbtmp;							\
   1255  1.1       mrg     __asm__ ("ff1 %0,%1" : "=r" (__cbtmp) : "r" (x));			\
   1256  1.1       mrg     (count) = __cbtmp ^ 31;						\
   1257  1.1       mrg   } while (0)
   1258  1.1       mrg #define COUNT_LEADING_ZEROS_0 63 /* sic */
   1259  1.1       mrg #if defined (__m88110__)
   1260  1.1       mrg #define umul_ppmm(wh, wl, u, v) \
   1261  1.1       mrg   do {									\
   1262  1.1       mrg     union {UDItype __ll;						\
   1263  1.1       mrg 	   struct {USItype __h, __l;} __i;				\
   1264  1.1       mrg 	  } __x;							\
   1265  1.1       mrg     __asm__ ("mulu.d %0,%1,%2" : "=r" (__x.__ll) : "r" (u), "r" (v));	\
   1266  1.1       mrg     (wh) = __x.__i.__h;							\
   1267  1.1       mrg     (wl) = __x.__i.__l;							\
   1268  1.1       mrg   } while (0)
   1269  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, d) \
   1270  1.1       mrg   ({union {UDItype __ll;						\
   1271  1.1       mrg 	   struct {USItype __h, __l;} __i;				\
   1272  1.1       mrg 	  } __x, __q;							\
   1273  1.1       mrg   __x.__i.__h = (n1); __x.__i.__l = (n0);				\
   1274  1.1       mrg   __asm__ ("divu.d %0,%1,%2"						\
   1275  1.1       mrg 	   : "=r" (__q.__ll) : "r" (__x.__ll), "r" (d));		\
   1276  1.1       mrg   (r) = (n0) - __q.__l * (d); (q) = __q.__l; })
   1277  1.1       mrg #endif /* __m88110__ */
   1278  1.1       mrg #endif /* __m88000__ */
   1279  1.1       mrg 
   1280  1.1       mrg #if defined (__mips) && W_TYPE_SIZE == 32
   1281  1.2     joerg #if __GMP_GNUC_PREREQ (4,4) || defined(__clang__)
   1282  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1283  1.1       mrg   do {									\
   1284  1.1       mrg     UDItype __ll = (UDItype)(u) * (v);					\
   1285  1.1       mrg     w1 = __ll >> 32;							\
   1286  1.1       mrg     w0 = __ll;								\
   1287  1.1       mrg   } while (0)
   1288  1.1       mrg #endif
   1289  1.3       mrg #if !defined (umul_ppmm) && __GMP_GNUC_PREREQ (2,7) && !defined (__clang__)
   1290  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1291  1.1       mrg   __asm__ ("multu %2,%3" : "=l" (w0), "=h" (w1) : "d" (u), "d" (v))
   1292  1.1       mrg #endif
   1293  1.1       mrg #if !defined (umul_ppmm)
   1294  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1295  1.1       mrg   __asm__ ("multu %2,%3\n\tmflo %0\n\tmfhi %1"				\
   1296  1.1       mrg 	   : "=d" (w0), "=d" (w1) : "d" (u), "d" (v))
   1297  1.1       mrg #endif
   1298  1.1       mrg #endif /* __mips */
   1299  1.1       mrg 
   1300  1.1       mrg #if (defined (__mips) && __mips >= 3) && W_TYPE_SIZE == 64
   1301  1.4       mrg #if defined (_MIPS_ARCH_MIPS64R6)
   1302  1.4       mrg #define umul_ppmm(w1, w0, u, v) \
   1303  1.4       mrg   do {									\
   1304  1.4       mrg     UDItype __m0 = (u), __m1 = (v);					\
   1305  1.4       mrg     (w0) = __m0 * __m1;							\
   1306  1.4       mrg     __asm__ ("dmuhu\t%0, %1, %2" : "=d" (w1) : "d" (__m0), "d" (__m1));	\
   1307  1.4       mrg   } while (0)
   1308  1.4       mrg #endif
   1309  1.4       mrg #if !defined (umul_ppmm) && (__GMP_GNUC_PREREQ (4,4) || defined(__clang__))
   1310  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1311  1.1       mrg   do {									\
   1312  1.1       mrg     typedef unsigned int __ll_UTItype __attribute__((mode(TI)));	\
   1313  1.1       mrg     __ll_UTItype __ll = (__ll_UTItype)(u) * (v);			\
   1314  1.1       mrg     w1 = __ll >> 64;							\
   1315  1.1       mrg     w0 = __ll;								\
   1316  1.1       mrg   } while (0)
   1317  1.1       mrg #endif
   1318  1.3       mrg #if !defined (umul_ppmm) && __GMP_GNUC_PREREQ (2,7) && !defined (__clang__)
   1319  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1320  1.3       mrg   __asm__ ("dmultu %2,%3"						\
   1321  1.3       mrg 	   : "=l" (w0), "=h" (w1)					\
   1322  1.3       mrg 	   : "d" ((UDItype)(u)), "d" ((UDItype)(v)))
   1323  1.1       mrg #endif
   1324  1.1       mrg #if !defined (umul_ppmm)
   1325  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1326  1.1       mrg   __asm__ ("dmultu %2,%3\n\tmflo %0\n\tmfhi %1"				\
   1327  1.3       mrg 	   : "=d" (w0), "=d" (w1)					\
   1328  1.3       mrg 	   : "d" ((UDItype)(u)), "d" ((UDItype)(v)))
   1329  1.1       mrg #endif
   1330  1.1       mrg #endif /* __mips */
   1331  1.1       mrg 
   1332  1.1       mrg #if defined (__mmix__) && W_TYPE_SIZE == 64
   1333  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1334  1.1       mrg   __asm__ ("MULU %0,%2,%3" : "=r" (w0), "=z" (w1) : "r" (u), "r" (v))
   1335  1.1       mrg #endif
   1336  1.1       mrg 
   1337  1.1       mrg #if defined (__ns32000__) && W_TYPE_SIZE == 32
   1338  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1339  1.1       mrg   ({union {UDItype __ll;						\
   1340  1.1       mrg 	   struct {USItype __l, __h;} __i;				\
   1341  1.1       mrg 	  } __x;							\
   1342  1.1       mrg   __asm__ ("meid %2,%0"							\
   1343  1.1       mrg 	   : "=g" (__x.__ll)						\
   1344  1.1       mrg 	   : "%0" ((USItype)(u)), "g" ((USItype)(v)));			\
   1345  1.1       mrg   (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
   1346  1.1       mrg #define __umulsidi3(u, v) \
   1347  1.1       mrg   ({UDItype __w;							\
   1348  1.1       mrg     __asm__ ("meid %2,%0"						\
   1349  1.1       mrg 	     : "=g" (__w)						\
   1350  1.1       mrg 	     : "%0" ((USItype)(u)), "g" ((USItype)(v)));		\
   1351  1.1       mrg     __w; })
   1352  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, d) \
   1353  1.1       mrg   ({union {UDItype __ll;						\
   1354  1.1       mrg 	   struct {USItype __l, __h;} __i;				\
   1355  1.1       mrg 	  } __x;							\
   1356  1.1       mrg   __x.__i.__h = (n1); __x.__i.__l = (n0);				\
   1357  1.1       mrg   __asm__ ("deid %2,%0"							\
   1358  1.1       mrg 	   : "=g" (__x.__ll)						\
   1359  1.1       mrg 	   : "0" (__x.__ll), "g" ((USItype)(d)));			\
   1360  1.1       mrg   (r) = __x.__i.__l; (q) = __x.__i.__h; })
   1361  1.1       mrg #define count_trailing_zeros(count,x) \
   1362  1.1       mrg   do {									\
   1363  1.1       mrg     __asm__ ("ffsd	%2,%0"						\
   1364  1.1       mrg 	     : "=r" (count)						\
   1365  1.1       mrg 	     : "0" ((USItype) 0), "r" ((USItype) (x)));			\
   1366  1.1       mrg   } while (0)
   1367  1.1       mrg #endif /* __ns32000__ */
   1368  1.1       mrg 
   1369  1.1       mrg /* In the past we had a block of various #defines tested
   1370  1.1       mrg        _ARCH_PPC    - AIX
   1371  1.1       mrg        _ARCH_PWR    - AIX
   1372  1.1       mrg        __powerpc__  - gcc
   1373  1.1       mrg        __POWERPC__  - BEOS
   1374  1.1       mrg        __ppc__      - Darwin
   1375  1.1       mrg        PPC          - old gcc, GNU/Linux, SysV
   1376  1.1       mrg    The plain PPC test was not good for vxWorks, since PPC is defined on all
   1377  1.1       mrg    CPUs there (eg. m68k too), as a constant one is expected to compare
   1378  1.1       mrg    CPU_FAMILY against.
   1379  1.1       mrg 
   1380  1.1       mrg    At any rate, this was pretty unattractive and a bit fragile.  The use of
   1381  1.1       mrg    HAVE_HOST_CPU_FAMILY is designed to cut through it all and be sure of
   1382  1.1       mrg    getting the desired effect.
   1383  1.1       mrg 
   1384  1.1       mrg    ENHANCE-ME: We should test _IBMR2 here when we add assembly support for
   1385  1.1       mrg    the system vendor compilers.  (Is that vendor compilers with inline asm,
   1386  1.1       mrg    or what?)  */
   1387  1.1       mrg 
   1388  1.3       mrg #if (HAVE_HOST_CPU_FAMILY_power || HAVE_HOST_CPU_FAMILY_powerpc)	\
   1389  1.1       mrg   && W_TYPE_SIZE == 32
   1390  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
   1391  1.1       mrg   do {									\
   1392  1.1       mrg     if (__builtin_constant_p (bh) && (bh) == 0)				\
   1393  1.3       mrg       __asm__ ("add%I4c %1,%3,%4\n\taddze %0,%2"			\
   1394  1.4       mrg 	       : "=r" (sh), "=&r" (sl) : "r" (ah), "%r" (al), "rI" (bl)	\
   1395  1.4       mrg 		 __CLOBBER_CC);						\
   1396  1.1       mrg     else if (__builtin_constant_p (bh) && (bh) == ~(USItype) 0)		\
   1397  1.3       mrg       __asm__ ("add%I4c %1,%3,%4\n\taddme %0,%2"			\
   1398  1.4       mrg 	       : "=r" (sh), "=&r" (sl) : "r" (ah), "%r" (al), "rI" (bl)	\
   1399  1.4       mrg 		 __CLOBBER_CC);						\
   1400  1.1       mrg     else								\
   1401  1.3       mrg       __asm__ ("add%I5c %1,%4,%5\n\tadde %0,%2,%3"			\
   1402  1.4       mrg 	       : "=r" (sh), "=&r" (sl)					\
   1403  1.4       mrg 	       : "r" (ah), "r" (bh), "%r" (al), "rI" (bl)		\
   1404  1.4       mrg 		 __CLOBBER_CC);						\
   1405  1.1       mrg   } while (0)
   1406  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
   1407  1.1       mrg   do {									\
   1408  1.1       mrg     if (__builtin_constant_p (ah) && (ah) == 0)				\
   1409  1.3       mrg       __asm__ ("subf%I3c %1,%4,%3\n\tsubfze %0,%2"			\
   1410  1.4       mrg 	       : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl)	\
   1411  1.4       mrg 		 __CLOBBER_CC);						\
   1412  1.1       mrg     else if (__builtin_constant_p (ah) && (ah) == ~(USItype) 0)		\
   1413  1.3       mrg       __asm__ ("subf%I3c %1,%4,%3\n\tsubfme %0,%2"			\
   1414  1.4       mrg 	       : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl)	\
   1415  1.4       mrg 		 __CLOBBER_CC);						\
   1416  1.1       mrg     else if (__builtin_constant_p (bh) && (bh) == 0)			\
   1417  1.3       mrg       __asm__ ("subf%I3c %1,%4,%3\n\taddme %0,%2"			\
   1418  1.4       mrg 	       : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "r" (bl)	\
   1419  1.4       mrg 		 __CLOBBER_CC);						\
   1420  1.1       mrg     else if (__builtin_constant_p (bh) && (bh) == ~(USItype) 0)		\
   1421  1.3       mrg       __asm__ ("subf%I3c %1,%4,%3\n\taddze %0,%2"			\
   1422  1.4       mrg 	       : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "r" (bl)	\
   1423  1.4       mrg 		 __CLOBBER_CC);						\
   1424  1.1       mrg     else								\
   1425  1.3       mrg       __asm__ ("subf%I4c %1,%5,%4\n\tsubfe %0,%3,%2"			\
   1426  1.1       mrg 	       : "=r" (sh), "=&r" (sl)					\
   1427  1.4       mrg 	       : "r" (ah), "r" (bh), "rI" (al), "r" (bl)		\
   1428  1.4       mrg 		 __CLOBBER_CC);						\
   1429  1.1       mrg   } while (0)
   1430  1.1       mrg #define count_leading_zeros(count, x) \
   1431  1.2     joerg   __asm__ ("cntlzw %0,%1" : "=r" (count) : "r" (x))
   1432  1.1       mrg #define COUNT_LEADING_ZEROS_0 32
   1433  1.1       mrg #if HAVE_HOST_CPU_FAMILY_powerpc
   1434  1.2     joerg #if __GMP_GNUC_PREREQ (4,4) || defined(__clang__)
   1435  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1436  1.1       mrg   do {									\
   1437  1.1       mrg     UDItype __ll = (UDItype)(u) * (v);					\
   1438  1.1       mrg     w1 = __ll >> 32;							\
   1439  1.1       mrg     w0 = __ll;								\
   1440  1.1       mrg   } while (0)
   1441  1.1       mrg #endif
   1442  1.1       mrg #if !defined (umul_ppmm)
   1443  1.1       mrg #define umul_ppmm(ph, pl, m0, m1) \
   1444  1.1       mrg   do {									\
   1445  1.1       mrg     USItype __m0 = (m0), __m1 = (m1);					\
   1446  1.1       mrg     __asm__ ("mulhwu %0,%1,%2" : "=r" (ph) : "%r" (m0), "r" (m1));	\
   1447  1.1       mrg     (pl) = __m0 * __m1;							\
   1448  1.1       mrg   } while (0)
   1449  1.1       mrg #endif
   1450  1.1       mrg #define smul_ppmm(ph, pl, m0, m1) \
   1451  1.1       mrg   do {									\
   1452  1.1       mrg     SItype __m0 = (m0), __m1 = (m1);					\
   1453  1.1       mrg     __asm__ ("mulhw %0,%1,%2" : "=r" (ph) : "%r" (m0), "r" (m1));	\
   1454  1.1       mrg     (pl) = __m0 * __m1;							\
   1455  1.1       mrg   } while (0)
   1456  1.1       mrg #else
   1457  1.1       mrg #define smul_ppmm(xh, xl, m0, m1) \
   1458  1.1       mrg   __asm__ ("mul %0,%2,%3" : "=r" (xh), "=q" (xl) : "r" (m0), "r" (m1))
   1459  1.1       mrg #define sdiv_qrnnd(q, r, nh, nl, d) \
   1460  1.1       mrg   __asm__ ("div %0,%2,%4" : "=r" (q), "=q" (r) : "r" (nh), "1" (nl), "r" (d))
   1461  1.1       mrg #endif
   1462  1.1       mrg #endif /* 32-bit POWER architecture variants.  */
   1463  1.1       mrg 
   1464  1.1       mrg /* We should test _IBMR2 here when we add assembly support for the system
   1465  1.1       mrg    vendor compilers.  */
   1466  1.1       mrg #if HAVE_HOST_CPU_FAMILY_powerpc && W_TYPE_SIZE == 64
   1467  1.1       mrg #if !defined (_LONG_LONG_LIMB)
   1468  1.1       mrg /* _LONG_LONG_LIMB is ABI=mode32 where adde operates on 32-bit values.  So
   1469  1.1       mrg    use adde etc only when not _LONG_LONG_LIMB.  */
   1470  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
   1471  1.1       mrg   do {									\
   1472  1.1       mrg     if (__builtin_constant_p (bh) && (bh) == 0)				\
   1473  1.3       mrg       __asm__ ("add%I4c %1,%3,%4\n\taddze %0,%2"			\
   1474  1.3       mrg 	       : "=r" (sh), "=&r" (sl)					\
   1475  1.3       mrg 	       : "r"  ((UDItype)(ah)),					\
   1476  1.4       mrg 		 "%r" ((UDItype)(al)), "rI" ((UDItype)(bl))		\
   1477  1.4       mrg 		 __CLOBBER_CC);						\
   1478  1.1       mrg     else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0)		\
   1479  1.3       mrg       __asm__ ("add%I4c %1,%3,%4\n\taddme %0,%2"			\
   1480  1.3       mrg 	       : "=r" (sh), "=&r" (sl)					\
   1481  1.3       mrg 	       : "r"  ((UDItype)(ah)),					\
   1482  1.4       mrg 		 "%r" ((UDItype)(al)), "rI" ((UDItype)(bl))		\
   1483  1.4       mrg 		 __CLOBBER_CC);						\
   1484  1.1       mrg     else								\
   1485  1.3       mrg       __asm__ ("add%I5c %1,%4,%5\n\tadde %0,%2,%3"			\
   1486  1.3       mrg 	       : "=r" (sh), "=&r" (sl)					\
   1487  1.3       mrg 	       : "r"  ((UDItype)(ah)), "r"  ((UDItype)(bh)),		\
   1488  1.4       mrg 		 "%r" ((UDItype)(al)), "rI" ((UDItype)(bl))		\
   1489  1.4       mrg 		 __CLOBBER_CC);						\
   1490  1.1       mrg   } while (0)
   1491  1.1       mrg /* We use "*rI" for the constant operand here, since with just "I", gcc barfs.
   1492  1.1       mrg    This might seem strange, but gcc folds away the dead code late.  */
   1493  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
   1494  1.3       mrg   do {									\
   1495  1.5       mrg     if (__builtin_constant_p (bl)					\
   1496  1.5       mrg 	&& (bl) > -0x8000 && (bl) <= 0x8000 && (bl) != 0) {		\
   1497  1.3       mrg 	if (__builtin_constant_p (ah) && (ah) == 0)			\
   1498  1.3       mrg 	  __asm__ ("addic %1,%3,%4\n\tsubfze %0,%2"			\
   1499  1.3       mrg 		   : "=r" (sh), "=&r" (sl)				\
   1500  1.3       mrg 		   :                       "r" ((UDItype)(bh)),		\
   1501  1.5       mrg 		     "r" ((UDItype)(al)), "*rI" (-((UDItype)(bl)))	\
   1502  1.4       mrg 		     __CLOBBER_CC);					\
   1503  1.3       mrg 	else if (__builtin_constant_p (ah) && (ah) == ~(UDItype) 0)	\
   1504  1.3       mrg 	  __asm__ ("addic %1,%3,%4\n\tsubfme %0,%2"			\
   1505  1.3       mrg 		   : "=r" (sh), "=&r" (sl)				\
   1506  1.3       mrg 		   :                       "r" ((UDItype)(bh)),		\
   1507  1.5       mrg 		     "r" ((UDItype)(al)), "*rI" (-((UDItype)(bl)))	\
   1508  1.4       mrg 		     __CLOBBER_CC);					\
   1509  1.3       mrg 	else if (__builtin_constant_p (bh) && (bh) == 0)		\
   1510  1.3       mrg 	  __asm__ ("addic %1,%3,%4\n\taddme %0,%2"			\
   1511  1.3       mrg 		   : "=r" (sh), "=&r" (sl)				\
   1512  1.5       mrg 		   : "r" ((UDItype)(ah)),				\
   1513  1.5       mrg 		     "r" ((UDItype)(al)), "*rI" (-((UDItype)(bl)))	\
   1514  1.4       mrg 		     __CLOBBER_CC);					\
   1515  1.3       mrg 	else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0)	\
   1516  1.3       mrg 	  __asm__ ("addic %1,%3,%4\n\taddze %0,%2"			\
   1517  1.3       mrg 		   : "=r" (sh), "=&r" (sl)				\
   1518  1.5       mrg 		   : "r" ((UDItype)(ah)),				\
   1519  1.5       mrg 		     "r" ((UDItype)(al)), "*rI" (-((UDItype)(bl)))	\
   1520  1.4       mrg 		     __CLOBBER_CC);					\
   1521  1.3       mrg 	else								\
   1522  1.3       mrg 	  __asm__ ("addic %1,%4,%5\n\tsubfe %0,%3,%2"			\
   1523  1.3       mrg 		   : "=r" (sh), "=&r" (sl)				\
   1524  1.5       mrg 		   : "r" ((UDItype)(ah)), "r" ((UDItype)(bh)),		\
   1525  1.5       mrg 		     "r" ((UDItype)(al)), "*rI" (-((UDItype)(bl)))	\
   1526  1.4       mrg 		     __CLOBBER_CC);					\
   1527  1.3       mrg     } else {								\
   1528  1.3       mrg 	if (__builtin_constant_p (ah) && (ah) == 0)			\
   1529  1.3       mrg 	  __asm__ ("subf%I3c %1,%4,%3\n\tsubfze %0,%2"			\
   1530  1.3       mrg 		   : "=r" (sh), "=&r" (sl)				\
   1531  1.3       mrg 		   :                       "r" ((UDItype)(bh)),		\
   1532  1.4       mrg 		     "rI" ((UDItype)(al)), "r" ((UDItype)(bl))		\
   1533  1.4       mrg 		     __CLOBBER_CC);					\
   1534  1.3       mrg 	else if (__builtin_constant_p (ah) && (ah) == ~(UDItype) 0)	\
   1535  1.3       mrg 	  __asm__ ("subf%I3c %1,%4,%3\n\tsubfme %0,%2"			\
   1536  1.3       mrg 		   : "=r" (sh), "=&r" (sl)				\
   1537  1.3       mrg 		   :                       "r" ((UDItype)(bh)),		\
   1538  1.4       mrg 		     "rI" ((UDItype)(al)), "r" ((UDItype)(bl))		\
   1539  1.4       mrg 		     __CLOBBER_CC);					\
   1540  1.3       mrg 	else if (__builtin_constant_p (bh) && (bh) == 0)		\
   1541  1.3       mrg 	  __asm__ ("subf%I3c %1,%4,%3\n\taddme %0,%2"			\
   1542  1.3       mrg 		   : "=r" (sh), "=&r" (sl)				\
   1543  1.3       mrg 		   : "r"  ((UDItype)(ah)),				\
   1544  1.4       mrg 		     "rI" ((UDItype)(al)), "r" ((UDItype)(bl))		\
   1545  1.4       mrg 		     __CLOBBER_CC);					\
   1546  1.3       mrg 	else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0)	\
   1547  1.3       mrg 	  __asm__ ("subf%I3c %1,%4,%3\n\taddze %0,%2"			\
   1548  1.3       mrg 		   : "=r" (sh), "=&r" (sl)				\
   1549  1.3       mrg 		   : "r"  ((UDItype)(ah)),				\
   1550  1.4       mrg 		     "rI" ((UDItype)(al)), "r" ((UDItype)(bl))		\
   1551  1.4       mrg 		     __CLOBBER_CC);					\
   1552  1.3       mrg 	else								\
   1553  1.3       mrg 	  __asm__ ("subf%I4c %1,%5,%4\n\tsubfe %0,%3,%2"		\
   1554  1.3       mrg 		   : "=r" (sh), "=&r" (sl)				\
   1555  1.3       mrg 		   : "r"  ((UDItype)(ah)), "r" ((UDItype)(bh)),		\
   1556  1.4       mrg 		     "rI" ((UDItype)(al)), "r" ((UDItype)(bl))		\
   1557  1.4       mrg 		     __CLOBBER_CC);					\
   1558  1.3       mrg     }									\
   1559  1.1       mrg   } while (0)
   1560  1.1       mrg #endif /* ! _LONG_LONG_LIMB */
   1561  1.1       mrg #define count_leading_zeros(count, x) \
   1562  1.1       mrg   __asm__ ("cntlzd %0,%1" : "=r" (count) : "r" (x))
   1563  1.1       mrg #define COUNT_LEADING_ZEROS_0 64
   1564  1.4       mrg #if __GMP_GNUC_PREREQ (4,8)
   1565  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1566  1.1       mrg   do {									\
   1567  1.1       mrg     typedef unsigned int __ll_UTItype __attribute__((mode(TI)));	\
   1568  1.1       mrg     __ll_UTItype __ll = (__ll_UTItype)(u) * (v);			\
   1569  1.1       mrg     w1 = __ll >> 64;							\
   1570  1.1       mrg     w0 = __ll;								\
   1571  1.1       mrg   } while (0)
   1572  1.1       mrg #endif
   1573  1.1       mrg #if !defined (umul_ppmm)
   1574  1.1       mrg #define umul_ppmm(ph, pl, m0, m1) \
   1575  1.1       mrg   do {									\
   1576  1.1       mrg     UDItype __m0 = (m0), __m1 = (m1);					\
   1577  1.3       mrg     __asm__ ("mulhdu %0,%1,%2" : "=r" (ph) : "%r" (__m0), "r" (__m1));	\
   1578  1.1       mrg     (pl) = __m0 * __m1;							\
   1579  1.1       mrg   } while (0)
   1580  1.1       mrg #endif
   1581  1.1       mrg #define smul_ppmm(ph, pl, m0, m1) \
   1582  1.1       mrg   do {									\
   1583  1.1       mrg     DItype __m0 = (m0), __m1 = (m1);					\
   1584  1.3       mrg     __asm__ ("mulhd %0,%1,%2" : "=r" (ph) : "%r" (__m0), "r" (__m1));	\
   1585  1.1       mrg     (pl) = __m0 * __m1;							\
   1586  1.1       mrg   } while (0)
   1587  1.1       mrg #endif /* 64-bit PowerPC.  */
   1588  1.1       mrg 
   1589  1.1       mrg #if defined (__pyr__) && W_TYPE_SIZE == 32
   1590  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
   1591  1.1       mrg   __asm__ ("addw %5,%1\n\taddwc %3,%0"					\
   1592  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
   1593  1.1       mrg 	   : "0"  ((USItype)(ah)), "g" ((USItype)(bh)),			\
   1594  1.1       mrg 	     "%1" ((USItype)(al)), "g" ((USItype)(bl)))
   1595  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
   1596  1.1       mrg   __asm__ ("subw %5,%1\n\tsubwb %3,%0"					\
   1597  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
   1598  1.1       mrg 	   : "0" ((USItype)(ah)), "g" ((USItype)(bh)),			\
   1599  1.1       mrg 	     "1" ((USItype)(al)), "g" ((USItype)(bl)))
   1600  1.1       mrg /* This insn works on Pyramids with AP, XP, or MI CPUs, but not with SP.  */
   1601  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1602  1.1       mrg   ({union {UDItype __ll;						\
   1603  1.1       mrg 	   struct {USItype __h, __l;} __i;				\
   1604  1.1       mrg 	  } __x;							\
   1605  1.1       mrg   __asm__ ("movw %1,%R0\n\tuemul %2,%0"					\
   1606  1.1       mrg 	   : "=&r" (__x.__ll)						\
   1607  1.1       mrg 	   : "g" ((USItype) (u)), "g" ((USItype)(v)));			\
   1608  1.1       mrg   (w1) = __x.__i.__h; (w0) = __x.__i.__l;})
   1609  1.1       mrg #endif /* __pyr__ */
   1610  1.1       mrg 
   1611  1.1       mrg #if defined (__ibm032__) /* RT/ROMP */  && W_TYPE_SIZE == 32
   1612  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
   1613  1.1       mrg   __asm__ ("a %1,%5\n\tae %0,%3"					\
   1614  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
   1615  1.1       mrg 	   : "0"  ((USItype)(ah)), "r" ((USItype)(bh)),			\
   1616  1.1       mrg 	     "%1" ((USItype)(al)), "r" ((USItype)(bl)))
   1617  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
   1618  1.1       mrg   __asm__ ("s %1,%5\n\tse %0,%3"					\
   1619  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
   1620  1.1       mrg 	   : "0" ((USItype)(ah)), "r" ((USItype)(bh)),			\
   1621  1.1       mrg 	     "1" ((USItype)(al)), "r" ((USItype)(bl)))
   1622  1.1       mrg #define smul_ppmm(ph, pl, m0, m1) \
   1623  1.1       mrg   __asm__ (								\
   1624  1.1       mrg        "s	r2,r2\n"						\
   1625  1.1       mrg "	mts r10,%2\n"							\
   1626  1.1       mrg "	m	r2,%3\n"						\
   1627  1.1       mrg "	m	r2,%3\n"						\
   1628  1.1       mrg "	m	r2,%3\n"						\
   1629  1.1       mrg "	m	r2,%3\n"						\
   1630  1.1       mrg "	m	r2,%3\n"						\
   1631  1.1       mrg "	m	r2,%3\n"						\
   1632  1.1       mrg "	m	r2,%3\n"						\
   1633  1.1       mrg "	m	r2,%3\n"						\
   1634  1.1       mrg "	m	r2,%3\n"						\
   1635  1.1       mrg "	m	r2,%3\n"						\
   1636  1.1       mrg "	m	r2,%3\n"						\
   1637  1.1       mrg "	m	r2,%3\n"						\
   1638  1.1       mrg "	m	r2,%3\n"						\
   1639  1.1       mrg "	m	r2,%3\n"						\
   1640  1.1       mrg "	m	r2,%3\n"						\
   1641  1.1       mrg "	m	r2,%3\n"						\
   1642  1.1       mrg "	cas	%0,r2,r0\n"						\
   1643  1.1       mrg "	mfs	r10,%1"							\
   1644  1.1       mrg 	   : "=r" (ph), "=r" (pl)					\
   1645  1.1       mrg 	   : "%r" ((USItype)(m0)), "r" ((USItype)(m1))			\
   1646  1.1       mrg 	   : "r2")
   1647  1.1       mrg #define count_leading_zeros(count, x) \
   1648  1.1       mrg   do {									\
   1649  1.1       mrg     if ((x) >= 0x10000)							\
   1650  1.1       mrg       __asm__ ("clz	%0,%1"						\
   1651  1.1       mrg 	       : "=r" (count) : "r" ((USItype)(x) >> 16));		\
   1652  1.1       mrg     else								\
   1653  1.1       mrg       {									\
   1654  1.1       mrg 	__asm__ ("clz	%0,%1"						\
   1655  1.1       mrg 		 : "=r" (count) : "r" ((USItype)(x)));			\
   1656  1.1       mrg 	(count) += 16;							\
   1657  1.1       mrg       }									\
   1658  1.1       mrg   } while (0)
   1659  1.1       mrg #endif /* RT/ROMP */
   1660  1.1       mrg 
   1661  1.4       mrg #if defined (__riscv64) && W_TYPE_SIZE == 64
   1662  1.4       mrg #define umul_ppmm(ph, pl, u, v) \
   1663  1.4       mrg   do {									\
   1664  1.4       mrg     UDItype __u = (u), __v = (v);					\
   1665  1.4       mrg     (pl) = __u * __v;							\
   1666  1.4       mrg     __asm__ ("mulhu\t%2, %1, %0" : "=r" (ph) : "%r" (__u), "r" (__v));	\
   1667  1.4       mrg   } while (0)
   1668  1.4       mrg #endif
   1669  1.4       mrg 
   1670  1.2     joerg #if (defined (__SH2__) || defined (__SH3__) || defined (__SH4__)) && W_TYPE_SIZE == 32
   1671  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1672  1.1       mrg   __asm__ ("dmulu.l %2,%3\n\tsts macl,%1\n\tsts mach,%0"		\
   1673  1.1       mrg 	   : "=r" (w1), "=r" (w0) : "r" (u), "r" (v) : "macl", "mach")
   1674  1.1       mrg #endif
   1675  1.1       mrg 
   1676  1.1       mrg #if defined (__sparc__) && W_TYPE_SIZE == 32
   1677  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
   1678  1.1       mrg   __asm__ ("addcc %r4,%5,%1\n\taddx %r2,%3,%0"				\
   1679  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
   1680  1.1       mrg 	   : "rJ" (ah), "rI" (bh),"%rJ" (al), "rI" (bl)			\
   1681  1.1       mrg 	   __CLOBBER_CC)
   1682  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
   1683  1.1       mrg   __asm__ ("subcc %r4,%5,%1\n\tsubx %r2,%3,%0"				\
   1684  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
   1685  1.1       mrg 	   : "rJ" (ah), "rI" (bh), "rJ" (al), "rI" (bl)	\
   1686  1.1       mrg 	   __CLOBBER_CC)
   1687  1.1       mrg /* FIXME: When gcc -mcpu=v9 is used on solaris, gcc/config/sol2-sld-64.h
   1688  1.1       mrg    doesn't define anything to indicate that to us, it only sets __sparcv8. */
   1689  1.1       mrg #if defined (__sparc_v9__) || defined (__sparcv9)
   1690  1.1       mrg /* Perhaps we should use floating-point operations here?  */
   1691  1.1       mrg #if 0
   1692  1.1       mrg /* Triggers a bug making mpz/tests/t-gcd.c fail.
   1693  1.1       mrg    Perhaps we simply need explicitly zero-extend the inputs?  */
   1694  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1695  1.1       mrg   __asm__ ("mulx %2,%3,%%g1; srl %%g1,0,%1; srlx %%g1,32,%0" :		\
   1696  1.1       mrg 	   "=r" (w1), "=r" (w0) : "r" (u), "r" (v) : "g1")
   1697  1.1       mrg #else
   1698  1.1       mrg /* Use v8 umul until above bug is fixed.  */
   1699  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1700  1.1       mrg   __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v))
   1701  1.1       mrg #endif
   1702  1.1       mrg /* Use a plain v8 divide for v9.  */
   1703  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, d) \
   1704  1.1       mrg   do {									\
   1705  1.1       mrg     USItype __q;							\
   1706  1.1       mrg     __asm__ ("mov %1,%%y;nop;nop;nop;udiv %2,%3,%0"			\
   1707  1.1       mrg 	     : "=r" (__q) : "r" (n1), "r" (n0), "r" (d));		\
   1708  1.1       mrg     (r) = (n0) - __q * (d);						\
   1709  1.1       mrg     (q) = __q;								\
   1710  1.1       mrg   } while (0)
   1711  1.1       mrg #else
   1712  1.1       mrg #if defined (__sparc_v8__)   /* gcc normal */				\
   1713  1.1       mrg   || defined (__sparcv8)     /* gcc solaris */				\
   1714  1.1       mrg   || HAVE_HOST_CPU_supersparc
   1715  1.1       mrg /* Don't match immediate range because, 1) it is not often useful,
   1716  1.1       mrg    2) the 'I' flag thinks of the range as a 13 bit signed interval,
   1717  1.1       mrg    while we want to match a 13 bit interval, sign extended to 32 bits,
   1718  1.1       mrg    but INTERPRETED AS UNSIGNED.  */
   1719  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1720  1.1       mrg   __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v))
   1721  1.1       mrg 
   1722  1.1       mrg #if HAVE_HOST_CPU_supersparc
   1723  1.1       mrg #else
   1724  1.1       mrg /* Don't use this on SuperSPARC because its udiv only handles 53 bit
   1725  1.1       mrg    dividends and will trap to the kernel for the rest. */
   1726  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, d) \
   1727  1.1       mrg   do {									\
   1728  1.1       mrg     USItype __q;							\
   1729  1.1       mrg     __asm__ ("mov %1,%%y;nop;nop;nop;udiv %2,%3,%0"			\
   1730  1.1       mrg 	     : "=r" (__q) : "r" (n1), "r" (n0), "r" (d));		\
   1731  1.1       mrg     (r) = (n0) - __q * (d);						\
   1732  1.1       mrg     (q) = __q;								\
   1733  1.1       mrg   } while (0)
   1734  1.1       mrg #endif /* HAVE_HOST_CPU_supersparc */
   1735  1.1       mrg 
   1736  1.1       mrg #else /* ! __sparc_v8__ */
   1737  1.1       mrg #if defined (__sparclite__)
   1738  1.1       mrg /* This has hardware multiply but not divide.  It also has two additional
   1739  1.1       mrg    instructions scan (ffs from high bit) and divscc.  */
   1740  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1741  1.1       mrg   __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v))
   1742  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, d) \
   1743  1.1       mrg   __asm__ ("! Inlined udiv_qrnnd\n"					\
   1744  1.1       mrg "	wr	%%g0,%2,%%y	! Not a delayed write for sparclite\n"	\
   1745  1.1       mrg "	tst	%%g0\n"							\
   1746  1.1       mrg "	divscc	%3,%4,%%g1\n"						\
   1747  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1748  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1749  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1750  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1751  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1752  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1753  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1754  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1755  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1756  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1757  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1758  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1759  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1760  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1761  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1762  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1763  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1764  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1765  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1766  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1767  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1768  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1769  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1770  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1771  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1772  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1773  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1774  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1775  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1776  1.1       mrg "	divscc	%%g1,%4,%%g1\n"						\
   1777  1.1       mrg "	divscc	%%g1,%4,%0\n"						\
   1778  1.1       mrg "	rd	%%y,%1\n"						\
   1779  1.1       mrg "	bl,a 1f\n"							\
   1780  1.1       mrg "	add	%1,%4,%1\n"						\
   1781  1.1       mrg "1:	! End of inline udiv_qrnnd"					\
   1782  1.1       mrg 	   : "=r" (q), "=r" (r) : "r" (n1), "r" (n0), "rI" (d)		\
   1783  1.1       mrg 	   : "%g1" __AND_CLOBBER_CC)
   1784  1.1       mrg #define count_leading_zeros(count, x) \
   1785  1.1       mrg   __asm__ ("scan %1,1,%0" : "=r" (count) : "r" (x))
   1786  1.1       mrg /* Early sparclites return 63 for an argument of 0, but they warn that future
   1787  1.1       mrg    implementations might change this.  Therefore, leave COUNT_LEADING_ZEROS_0
   1788  1.1       mrg    undefined.  */
   1789  1.1       mrg #endif /* __sparclite__ */
   1790  1.1       mrg #endif /* __sparc_v8__ */
   1791  1.1       mrg #endif /* __sparc_v9__ */
   1792  1.1       mrg /* Default to sparc v7 versions of umul_ppmm and udiv_qrnnd.  */
   1793  1.1       mrg #ifndef umul_ppmm
   1794  1.1       mrg #define umul_ppmm(w1, w0, u, v) \
   1795  1.1       mrg   __asm__ ("! Inlined umul_ppmm\n"					\
   1796  1.1       mrg "	wr	%%g0,%2,%%y	! SPARC has 0-3 delay insn after a wr\n" \
   1797  1.1       mrg "	sra	%3,31,%%g2	! Don't move this insn\n"		\
   1798  1.1       mrg "	and	%2,%%g2,%%g2	! Don't move this insn\n"		\
   1799  1.1       mrg "	andcc	%%g0,0,%%g1	! Don't move this insn\n"		\
   1800  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1801  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1802  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1803  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1804  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1805  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1806  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1807  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1808  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1809  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1810  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1811  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1812  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1813  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1814  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1815  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1816  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1817  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1818  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1819  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1820  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1821  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1822  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1823  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1824  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1825  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1826  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1827  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1828  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1829  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1830  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1831  1.1       mrg "	mulscc	%%g1,%3,%%g1\n"						\
   1832  1.1       mrg "	mulscc	%%g1,0,%%g1\n"						\
   1833  1.1       mrg "	add	%%g1,%%g2,%0\n"						\
   1834  1.1       mrg "	rd	%%y,%1"							\
   1835  1.1       mrg 	   : "=r" (w1), "=r" (w0) : "%rI" (u), "r" (v)			\
   1836  1.1       mrg 	   : "%g1", "%g2" __AND_CLOBBER_CC)
   1837  1.1       mrg #endif
   1838  1.1       mrg #ifndef udiv_qrnnd
   1839  1.1       mrg #ifndef LONGLONG_STANDALONE
   1840  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, d) \
   1841  1.1       mrg   do { UWtype __r;							\
   1842  1.1       mrg     (q) = __MPN(udiv_qrnnd) (&__r, (n1), (n0), (d));			\
   1843  1.1       mrg     (r) = __r;								\
   1844  1.1       mrg   } while (0)
   1845  1.2     joerg extern UWtype __MPN(udiv_qrnnd) (UWtype *, UWtype, UWtype, UWtype);
   1846  1.1       mrg #endif /* LONGLONG_STANDALONE */
   1847  1.1       mrg #endif /* udiv_qrnnd */
   1848  1.1       mrg #endif /* __sparc__ */
   1849  1.1       mrg 
   1850  1.1       mrg #if defined (__sparc__) && W_TYPE_SIZE == 64
   1851  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
   1852  1.1       mrg   __asm__ (								\
   1853  1.1       mrg        "addcc	%r4,%5,%1\n"						\
   1854  1.1       mrg       "	addccc	%r6,%7,%%g0\n"						\
   1855  1.1       mrg       "	addc	%r2,%3,%0"						\
   1856  1.3       mrg        : "=r" (sh), "=&r" (sl)						\
   1857  1.3       mrg        : "rJ"  ((UDItype)(ah)), "rI" ((UDItype)(bh)),			\
   1858  1.3       mrg 	 "%rJ" ((UDItype)(al)), "rI" ((UDItype)(bl)),			\
   1859  1.3       mrg 	 "%rJ" ((UDItype)(al) >> 32), "rI" ((UDItype)(bl) >> 32)	\
   1860  1.1       mrg 	   __CLOBBER_CC)
   1861  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
   1862  1.1       mrg   __asm__ (								\
   1863  1.1       mrg        "subcc	%r4,%5,%1\n"						\
   1864  1.1       mrg       "	subccc	%r6,%7,%%g0\n"						\
   1865  1.1       mrg       "	subc	%r2,%3,%0"						\
   1866  1.3       mrg        : "=r" (sh), "=&r" (sl)						\
   1867  1.3       mrg        : "rJ" ((UDItype)(ah)), "rI" ((UDItype)(bh)),			\
   1868  1.3       mrg 	 "rJ" ((UDItype)(al)), "rI" ((UDItype)(bl)),			\
   1869  1.3       mrg 	 "rJ" ((UDItype)(al) >> 32), "rI" ((UDItype)(bl) >> 32)		\
   1870  1.3       mrg 	   __CLOBBER_CC)
   1871  1.3       mrg #if __VIS__ >= 0x300
   1872  1.3       mrg #undef add_ssaaaa
   1873  1.3       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
   1874  1.3       mrg   __asm__ (								\
   1875  1.3       mrg        "addcc	%r4, %5, %1\n"						\
   1876  1.3       mrg       "	addxc	%r2, %r3, %0"						\
   1877  1.1       mrg 	  : "=r" (sh), "=&r" (sl)					\
   1878  1.3       mrg        : "rJ"  ((UDItype)(ah)), "rJ" ((UDItype)(bh)),			\
   1879  1.3       mrg 	 "%rJ" ((UDItype)(al)), "rI" ((UDItype)(bl)) __CLOBBER_CC)
   1880  1.3       mrg #define umul_ppmm(ph, pl, m0, m1) \
   1881  1.3       mrg   do {									\
   1882  1.3       mrg     UDItype __m0 = (m0), __m1 = (m1);					\
   1883  1.3       mrg     (pl) = __m0 * __m1;							\
   1884  1.3       mrg     __asm__ ("umulxhi\t%2, %1, %0"					\
   1885  1.3       mrg 	     : "=r" (ph)						\
   1886  1.3       mrg 	     : "%r" (__m0), "r" (__m1));				\
   1887  1.3       mrg   } while (0)
   1888  1.3       mrg #define count_leading_zeros(count, x) \
   1889  1.3       mrg   __asm__ ("lzd\t%1,%0" : "=r" (count) : "r" (x))
   1890  1.3       mrg /* Needed by count_leading_zeros_32 in sparc64.h.  */
   1891  1.3       mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
   1892  1.3       mrg #endif
   1893  1.1       mrg #endif
   1894  1.1       mrg 
   1895  1.2     joerg #if (defined (__vax) || defined (__vax__)) && W_TYPE_SIZE == 32
   1896  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
   1897  1.1       mrg   __asm__ ("addl2 %5,%1\n\tadwc %3,%0"					\
   1898  1.1       mrg 	   : "=g" (sh), "=&g" (sl)					\
   1899  1.1       mrg 	   : "0"  ((USItype)(ah)), "g" ((USItype)(bh)),			\
   1900  1.1       mrg 	     "%1" ((USItype)(al)), "g" ((USItype)(bl)))
   1901  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
   1902  1.1       mrg   __asm__ ("subl2 %5,%1\n\tsbwc %3,%0"					\
   1903  1.1       mrg 	   : "=g" (sh), "=&g" (sl)					\
   1904  1.1       mrg 	   : "0" ((USItype)(ah)), "g" ((USItype)(bh)),			\
   1905  1.1       mrg 	     "1" ((USItype)(al)), "g" ((USItype)(bl)))
   1906  1.1       mrg #define smul_ppmm(xh, xl, m0, m1) \
   1907  1.1       mrg   do {									\
   1908  1.1       mrg     union {UDItype __ll;						\
   1909  1.1       mrg 	   struct {USItype __l, __h;} __i;				\
   1910  1.1       mrg 	  } __x;							\
   1911  1.1       mrg     USItype __m0 = (m0), __m1 = (m1);					\
   1912  1.1       mrg     __asm__ ("emul %1,%2,$0,%0"						\
   1913  1.1       mrg 	     : "=g" (__x.__ll) : "g" (__m0), "g" (__m1));		\
   1914  1.1       mrg     (xh) = __x.__i.__h; (xl) = __x.__i.__l;				\
   1915  1.1       mrg   } while (0)
   1916  1.1       mrg #define sdiv_qrnnd(q, r, n1, n0, d) \
   1917  1.1       mrg   do {									\
   1918  1.1       mrg     union {DItype __ll;							\
   1919  1.1       mrg 	   struct {SItype __l, __h;} __i;				\
   1920  1.1       mrg 	  } __x;							\
   1921  1.1       mrg     __x.__i.__h = n1; __x.__i.__l = n0;					\
   1922  1.1       mrg     __asm__ ("ediv %3,%2,%0,%1"						\
   1923  1.1       mrg 	     : "=g" (q), "=g" (r) : "g" (__x.__ll), "g" (d));		\
   1924  1.1       mrg   } while (0)
   1925  1.1       mrg #if 0
   1926  1.1       mrg /* FIXME: This instruction appears to be unimplemented on some systems (vax
   1927  1.1       mrg    8800 maybe). */
   1928  1.1       mrg #define count_trailing_zeros(count,x)					\
   1929  1.1       mrg   do {									\
   1930  1.1       mrg     __asm__ ("ffs 0, 31, %1, %0"					\
   1931  1.1       mrg 	     : "=g" (count)						\
   1932  1.1       mrg 	     : "g" ((USItype) (x)));					\
   1933  1.1       mrg   } while (0)
   1934  1.1       mrg #endif
   1935  1.2     joerg #endif /* vax */
   1936  1.1       mrg 
   1937  1.1       mrg #if defined (__z8000__) && W_TYPE_SIZE == 16
   1938  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
   1939  1.1       mrg   __asm__ ("add	%H1,%H5\n\tadc	%H0,%H3"				\
   1940  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
   1941  1.1       mrg 	   : "0"  ((unsigned int)(ah)), "r" ((unsigned int)(bh)),	\
   1942  1.1       mrg 	     "%1" ((unsigned int)(al)), "rQR" ((unsigned int)(bl)))
   1943  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
   1944  1.1       mrg   __asm__ ("sub	%H1,%H5\n\tsbc	%H0,%H3"				\
   1945  1.1       mrg 	   : "=r" (sh), "=&r" (sl)					\
   1946  1.1       mrg 	   : "0" ((unsigned int)(ah)), "r" ((unsigned int)(bh)),	\
   1947  1.1       mrg 	     "1" ((unsigned int)(al)), "rQR" ((unsigned int)(bl)))
   1948  1.1       mrg #define umul_ppmm(xh, xl, m0, m1) \
   1949  1.1       mrg   do {									\
   1950  1.1       mrg     union {long int __ll;						\
   1951  1.1       mrg 	   struct {unsigned int __h, __l;} __i;				\
   1952  1.1       mrg 	  } __x;							\
   1953  1.1       mrg     unsigned int __m0 = (m0), __m1 = (m1);				\
   1954  1.1       mrg     __asm__ ("mult	%S0,%H3"					\
   1955  1.1       mrg 	     : "=r" (__x.__i.__h), "=r" (__x.__i.__l)			\
   1956  1.1       mrg 	     : "%1" (m0), "rQR" (m1));					\
   1957  1.1       mrg     (xh) = __x.__i.__h; (xl) = __x.__i.__l;				\
   1958  1.1       mrg     (xh) += ((((signed int) __m0 >> 15) & __m1)				\
   1959  1.1       mrg 	     + (((signed int) __m1 >> 15) & __m0));			\
   1960  1.1       mrg   } while (0)
   1961  1.1       mrg #endif /* __z8000__ */
   1962  1.1       mrg 
   1963  1.1       mrg #endif /* __GNUC__ */
   1964  1.1       mrg 
   1965  1.1       mrg #endif /* NO_ASM */
   1966  1.1       mrg 
   1967  1.1       mrg 
   1968  1.2     joerg /* FIXME: "sidi" here is highly doubtful, should sometimes be "diti".  */
   1969  1.1       mrg #if !defined (umul_ppmm) && defined (__umulsidi3)
   1970  1.1       mrg #define umul_ppmm(ph, pl, m0, m1) \
   1971  1.3       mrg   do {									\
   1972  1.1       mrg     UDWtype __ll = __umulsidi3 (m0, m1);				\
   1973  1.1       mrg     ph = (UWtype) (__ll >> W_TYPE_SIZE);				\
   1974  1.1       mrg     pl = (UWtype) __ll;							\
   1975  1.3       mrg   } while (0)
   1976  1.1       mrg #endif
   1977  1.1       mrg 
   1978  1.1       mrg #if !defined (__umulsidi3)
   1979  1.1       mrg #define __umulsidi3(u, v) \
   1980  1.1       mrg   ({UWtype __hi, __lo;							\
   1981  1.1       mrg     umul_ppmm (__hi, __lo, u, v);					\
   1982  1.1       mrg     ((UDWtype) __hi << W_TYPE_SIZE) | __lo; })
   1983  1.1       mrg #endif
   1984  1.1       mrg 
   1985  1.1       mrg 
   1986  1.3       mrg #if defined (__cplusplus)
   1987  1.3       mrg #define __longlong_h_C "C"
   1988  1.3       mrg #else
   1989  1.3       mrg #define __longlong_h_C
   1990  1.3       mrg #endif
   1991  1.3       mrg 
   1992  1.1       mrg /* Use mpn_umul_ppmm or mpn_udiv_qrnnd functions, if they exist.  The "_r"
   1993  1.1       mrg    forms have "reversed" arguments, meaning the pointer is last, which
   1994  1.1       mrg    sometimes allows better parameter passing, in particular on 64-bit
   1995  1.1       mrg    hppa. */
   1996  1.1       mrg 
   1997  1.1       mrg #define mpn_umul_ppmm  __MPN(umul_ppmm)
   1998  1.3       mrg extern __longlong_h_C UWtype mpn_umul_ppmm (UWtype *, UWtype, UWtype);
   1999  1.1       mrg 
   2000  1.1       mrg #if ! defined (umul_ppmm) && HAVE_NATIVE_mpn_umul_ppmm  \
   2001  1.1       mrg   && ! defined (LONGLONG_STANDALONE)
   2002  1.3       mrg #define umul_ppmm(wh, wl, u, v)						\
   2003  1.3       mrg   do {									\
   2004  1.3       mrg     UWtype __umul_ppmm__p0;						\
   2005  1.3       mrg     (wh) = mpn_umul_ppmm (&__umul_ppmm__p0, (UWtype) (u), (UWtype) (v));\
   2006  1.3       mrg     (wl) = __umul_ppmm__p0;						\
   2007  1.1       mrg   } while (0)
   2008  1.1       mrg #endif
   2009  1.1       mrg 
   2010  1.1       mrg #define mpn_umul_ppmm_r  __MPN(umul_ppmm_r)
   2011  1.3       mrg extern __longlong_h_C UWtype mpn_umul_ppmm_r (UWtype, UWtype, UWtype *);
   2012  1.1       mrg 
   2013  1.1       mrg #if ! defined (umul_ppmm) && HAVE_NATIVE_mpn_umul_ppmm_r	\
   2014  1.1       mrg   && ! defined (LONGLONG_STANDALONE)
   2015  1.3       mrg #define umul_ppmm(wh, wl, u, v)						\
   2016  1.3       mrg   do {									\
   2017  1.3       mrg     UWtype __umul_p0;							\
   2018  1.3       mrg     (wh) = mpn_umul_ppmm_r ((UWtype) (u), (UWtype) (v), &__umul_p0);	\
   2019  1.3       mrg     (wl) = __umul_p0;							\
   2020  1.1       mrg   } while (0)
   2021  1.1       mrg #endif
   2022  1.1       mrg 
   2023  1.1       mrg #define mpn_udiv_qrnnd  __MPN(udiv_qrnnd)
   2024  1.3       mrg extern __longlong_h_C UWtype mpn_udiv_qrnnd (UWtype *, UWtype, UWtype, UWtype);
   2025  1.1       mrg 
   2026  1.1       mrg #if ! defined (udiv_qrnnd) && HAVE_NATIVE_mpn_udiv_qrnnd	\
   2027  1.1       mrg   && ! defined (LONGLONG_STANDALONE)
   2028  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, d)					\
   2029  1.1       mrg   do {									\
   2030  1.3       mrg     UWtype __udiv_qrnnd_r;						\
   2031  1.3       mrg     (q) = mpn_udiv_qrnnd (&__udiv_qrnnd_r,				\
   2032  1.1       mrg 			  (UWtype) (n1), (UWtype) (n0), (UWtype) d);	\
   2033  1.3       mrg     (r) = __udiv_qrnnd_r;						\
   2034  1.1       mrg   } while (0)
   2035  1.1       mrg #endif
   2036  1.1       mrg 
   2037  1.1       mrg #define mpn_udiv_qrnnd_r  __MPN(udiv_qrnnd_r)
   2038  1.3       mrg extern __longlong_h_C UWtype mpn_udiv_qrnnd_r (UWtype, UWtype, UWtype, UWtype *);
   2039  1.1       mrg 
   2040  1.1       mrg #if ! defined (udiv_qrnnd) && HAVE_NATIVE_mpn_udiv_qrnnd_r	\
   2041  1.1       mrg   && ! defined (LONGLONG_STANDALONE)
   2042  1.1       mrg #define udiv_qrnnd(q, r, n1, n0, d)					\
   2043  1.1       mrg   do {									\
   2044  1.3       mrg     UWtype __udiv_qrnnd_r;						\
   2045  1.1       mrg     (q) = mpn_udiv_qrnnd_r ((UWtype) (n1), (UWtype) (n0), (UWtype) d,	\
   2046  1.3       mrg 			    &__udiv_qrnnd_r);				\
   2047  1.3       mrg     (r) = __udiv_qrnnd_r;						\
   2048  1.1       mrg   } while (0)
   2049  1.1       mrg #endif
   2050  1.1       mrg 
   2051  1.1       mrg 
   2052  1.1       mrg /* If this machine has no inline assembler, use C macros.  */
   2053  1.1       mrg 
   2054  1.1       mrg #if !defined (add_ssaaaa)
   2055  1.1       mrg #define add_ssaaaa(sh, sl, ah, al, bh, bl) \
   2056  1.1       mrg   do {									\
   2057  1.1       mrg     UWtype __x;								\
   2058  1.5       mrg     UWtype __al = (al);							\
   2059  1.5       mrg     UWtype __bl = (bl);							\
   2060  1.5       mrg     __x = __al + __bl;							\
   2061  1.5       mrg     (sh) = (ah) + (bh) + (__x < __al);					\
   2062  1.1       mrg     (sl) = __x;								\
   2063  1.1       mrg   } while (0)
   2064  1.1       mrg #endif
   2065  1.1       mrg 
   2066  1.1       mrg #if !defined (sub_ddmmss)
   2067  1.1       mrg #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
   2068  1.1       mrg   do {									\
   2069  1.1       mrg     UWtype __x;								\
   2070  1.5       mrg     UWtype __al = (al);							\
   2071  1.5       mrg     UWtype __bl = (bl);							\
   2072  1.5       mrg     __x = __al - __bl;							\
   2073  1.5       mrg     (sh) = (ah) - (bh) - (__al < __bl);					\
   2074  1.1       mrg     (sl) = __x;								\
   2075  1.1       mrg   } while (0)
   2076  1.1       mrg #endif
   2077  1.1       mrg 
   2078  1.1       mrg /* If we lack umul_ppmm but have smul_ppmm, define umul_ppmm in terms of
   2079  1.1       mrg    smul_ppmm.  */
   2080  1.1       mrg #if !defined (umul_ppmm) && defined (smul_ppmm)
   2081  1.1       mrg #define umul_ppmm(w1, w0, u, v)						\
   2082  1.1       mrg   do {									\
   2083  1.1       mrg     UWtype __w1;							\
   2084  1.1       mrg     UWtype __xm0 = (u), __xm1 = (v);					\
   2085  1.1       mrg     smul_ppmm (__w1, w0, __xm0, __xm1);					\
   2086  1.1       mrg     (w1) = __w1 + (-(__xm0 >> (W_TYPE_SIZE - 1)) & __xm1)		\
   2087  1.1       mrg 		+ (-(__xm1 >> (W_TYPE_SIZE - 1)) & __xm0);		\
   2088  1.1       mrg   } while (0)
   2089  1.1       mrg #endif
   2090  1.1       mrg 
   2091  1.1       mrg /* If we still don't have umul_ppmm, define it using plain C.
   2092  1.1       mrg 
   2093  1.1       mrg    For reference, when this code is used for squaring (ie. u and v identical
   2094  1.1       mrg    expressions), gcc recognises __x1 and __x2 are the same and generates 3
   2095  1.1       mrg    multiplies, not 4.  The subsequent additions could be optimized a bit,
   2096  1.1       mrg    but the only place GMP currently uses such a square is mpn_sqr_basecase,
   2097  1.1       mrg    and chips obliged to use this generic C umul will have plenty of worse
   2098  1.1       mrg    performance problems than a couple of extra instructions on the diagonal
   2099  1.1       mrg    of sqr_basecase.  */
   2100  1.1       mrg 
   2101  1.1       mrg #if !defined (umul_ppmm)
   2102  1.1       mrg #define umul_ppmm(w1, w0, u, v)						\
   2103  1.1       mrg   do {									\
   2104  1.1       mrg     UWtype __x0, __x1, __x2, __x3;					\
   2105  1.1       mrg     UHWtype __ul, __vl, __uh, __vh;					\
   2106  1.1       mrg     UWtype __u = (u), __v = (v);					\
   2107  1.1       mrg 									\
   2108  1.1       mrg     __ul = __ll_lowpart (__u);						\
   2109  1.1       mrg     __uh = __ll_highpart (__u);						\
   2110  1.1       mrg     __vl = __ll_lowpart (__v);						\
   2111  1.1       mrg     __vh = __ll_highpart (__v);						\
   2112  1.1       mrg 									\
   2113  1.1       mrg     __x0 = (UWtype) __ul * __vl;					\
   2114  1.1       mrg     __x1 = (UWtype) __ul * __vh;					\
   2115  1.1       mrg     __x2 = (UWtype) __uh * __vl;					\
   2116  1.1       mrg     __x3 = (UWtype) __uh * __vh;					\
   2117  1.1       mrg 									\
   2118  1.1       mrg     __x1 += __ll_highpart (__x0);/* this can't give carry */		\
   2119  1.1       mrg     __x1 += __x2;		/* but this indeed can */		\
   2120  1.1       mrg     if (__x1 < __x2)		/* did we get it? */			\
   2121  1.1       mrg       __x3 += __ll_B;		/* yes, add it in the proper pos. */	\
   2122  1.1       mrg 									\
   2123  1.1       mrg     (w1) = __x3 + __ll_highpart (__x1);					\
   2124  1.1       mrg     (w0) = (__x1 << W_TYPE_SIZE/2) + __ll_lowpart (__x0);		\
   2125  1.1       mrg   } while (0)
   2126  1.1       mrg #endif
   2127  1.1       mrg 
   2128  1.1       mrg /* If we don't have smul_ppmm, define it using umul_ppmm (which surely will
   2129  1.1       mrg    exist in one form or another.  */
   2130  1.1       mrg #if !defined (smul_ppmm)
   2131  1.1       mrg #define smul_ppmm(w1, w0, u, v)						\
   2132  1.1       mrg   do {									\
   2133  1.1       mrg     UWtype __w1;							\
   2134  1.1       mrg     UWtype __xm0 = (u), __xm1 = (v);					\
   2135  1.1       mrg     umul_ppmm (__w1, w0, __xm0, __xm1);					\
   2136  1.1       mrg     (w1) = __w1 - (-(__xm0 >> (W_TYPE_SIZE - 1)) & __xm1)		\
   2137  1.1       mrg 		- (-(__xm1 >> (W_TYPE_SIZE - 1)) & __xm0);		\
   2138  1.1       mrg   } while (0)
   2139  1.1       mrg #endif
   2140  1.1       mrg 
   2141  1.1       mrg /* Define this unconditionally, so it can be used for debugging.  */
   2142  1.1       mrg #define __udiv_qrnnd_c(q, r, n1, n0, d) \
   2143  1.1       mrg   do {									\
   2144  1.1       mrg     UWtype __d1, __d0, __q1, __q0, __r1, __r0, __m;			\
   2145  1.1       mrg 									\
   2146  1.1       mrg     ASSERT ((d) != 0);							\
   2147  1.1       mrg     ASSERT ((n1) < (d));						\
   2148  1.1       mrg 									\
   2149  1.1       mrg     __d1 = __ll_highpart (d);						\
   2150  1.1       mrg     __d0 = __ll_lowpart (d);						\
   2151  1.1       mrg 									\
   2152  1.1       mrg     __q1 = (n1) / __d1;							\
   2153  1.1       mrg     __r1 = (n1) - __q1 * __d1;						\
   2154  1.1       mrg     __m = __q1 * __d0;							\
   2155  1.1       mrg     __r1 = __r1 * __ll_B | __ll_highpart (n0);				\
   2156  1.1       mrg     if (__r1 < __m)							\
   2157  1.1       mrg       {									\
   2158  1.1       mrg 	__q1--, __r1 += (d);						\
   2159  1.1       mrg 	if (__r1 >= (d)) /* i.e. we didn't get carry when adding to __r1 */\
   2160  1.1       mrg 	  if (__r1 < __m)						\
   2161  1.1       mrg 	    __q1--, __r1 += (d);					\
   2162  1.1       mrg       }									\
   2163  1.1       mrg     __r1 -= __m;							\
   2164  1.1       mrg 									\
   2165  1.1       mrg     __q0 = __r1 / __d1;							\
   2166  1.1       mrg     __r0 = __r1  - __q0 * __d1;						\
   2167  1.1       mrg     __m = __q0 * __d0;							\
   2168  1.1       mrg     __r0 = __r0 * __ll_B | __ll_lowpart (n0);				\
   2169  1.1       mrg     if (__r0 < __m)							\
   2170  1.1       mrg       {									\
   2171  1.1       mrg 	__q0--, __r0 += (d);						\
   2172  1.1       mrg 	if (__r0 >= (d))						\
   2173  1.1       mrg 	  if (__r0 < __m)						\
   2174  1.1       mrg 	    __q0--, __r0 += (d);					\
   2175  1.1       mrg       }									\
   2176  1.1       mrg     __r0 -= __m;							\
   2177  1.1       mrg 									\
   2178  1.1       mrg     (q) = __q1 * __ll_B | __q0;						\
   2179  1.1       mrg     (r) = __r0;								\
   2180  1.1       mrg   } while (0)
   2181  1.1       mrg 
   2182  1.1       mrg /* If the processor has no udiv_qrnnd but sdiv_qrnnd, go through
   2183  1.1       mrg    __udiv_w_sdiv (defined in libgcc or elsewhere).  */
   2184  1.3       mrg #if !defined (udiv_qrnnd) && defined (sdiv_qrnnd) \
   2185  1.3       mrg   && ! defined (LONGLONG_STANDALONE)
   2186  1.1       mrg #define udiv_qrnnd(q, r, nh, nl, d) \
   2187  1.1       mrg   do {									\
   2188  1.1       mrg     UWtype __r;								\
   2189  1.1       mrg     (q) = __MPN(udiv_w_sdiv) (&__r, nh, nl, d);				\
   2190  1.1       mrg     (r) = __r;								\
   2191  1.1       mrg   } while (0)
   2192  1.2     joerg __GMP_DECLSPEC UWtype __MPN(udiv_w_sdiv) (UWtype *, UWtype, UWtype, UWtype);
   2193  1.1       mrg #endif
   2194  1.1       mrg 
   2195  1.1       mrg /* If udiv_qrnnd was not defined for this processor, use __udiv_qrnnd_c.  */
   2196  1.1       mrg #if !defined (udiv_qrnnd)
   2197  1.1       mrg #define UDIV_NEEDS_NORMALIZATION 1
   2198  1.1       mrg #define udiv_qrnnd __udiv_qrnnd_c
   2199  1.1       mrg #endif
   2200  1.1       mrg 
   2201  1.1       mrg #if !defined (count_leading_zeros)
   2202  1.1       mrg #define count_leading_zeros(count, x) \
   2203  1.1       mrg   do {									\
   2204  1.1       mrg     UWtype __xr = (x);							\
   2205  1.1       mrg     UWtype __a;								\
   2206  1.1       mrg 									\
   2207  1.1       mrg     if (W_TYPE_SIZE == 32)						\
   2208  1.1       mrg       {									\
   2209  1.1       mrg 	__a = __xr < ((UWtype) 1 << 2*__BITS4)				\
   2210  1.1       mrg 	  ? (__xr < ((UWtype) 1 << __BITS4) ? 1 : __BITS4 + 1)		\
   2211  1.1       mrg 	  : (__xr < ((UWtype) 1 << 3*__BITS4) ? 2*__BITS4 + 1		\
   2212  1.1       mrg 	  : 3*__BITS4 + 1);						\
   2213  1.1       mrg       }									\
   2214  1.1       mrg     else								\
   2215  1.1       mrg       {									\
   2216  1.1       mrg 	for (__a = W_TYPE_SIZE - 8; __a > 0; __a -= 8)			\
   2217  1.1       mrg 	  if (((__xr >> __a) & 0xff) != 0)				\
   2218  1.1       mrg 	    break;							\
   2219  1.1       mrg 	++__a;								\
   2220  1.1       mrg       }									\
   2221  1.1       mrg 									\
   2222  1.1       mrg     (count) = W_TYPE_SIZE + 1 - __a - __clz_tab[__xr >> __a];		\
   2223  1.1       mrg   } while (0)
   2224  1.1       mrg /* This version gives a well-defined value for zero. */
   2225  1.1       mrg #define COUNT_LEADING_ZEROS_0 (W_TYPE_SIZE - 1)
   2226  1.1       mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
   2227  1.2     joerg #define COUNT_LEADING_ZEROS_SLOW
   2228  1.1       mrg #endif
   2229  1.1       mrg 
   2230  1.1       mrg /* clz_tab needed by mpn/x86/pentium/mod_1.asm in a fat binary */
   2231  1.1       mrg #if HAVE_HOST_CPU_FAMILY_x86 && WANT_FAT_BINARY
   2232  1.1       mrg #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB
   2233  1.1       mrg #endif
   2234  1.1       mrg 
   2235  1.1       mrg #ifdef COUNT_LEADING_ZEROS_NEED_CLZ_TAB
   2236  1.2     joerg extern const unsigned char __GMP_DECLSPEC __clz_tab[129];
   2237  1.1       mrg #endif
   2238  1.1       mrg 
   2239  1.1       mrg #if !defined (count_trailing_zeros)
   2240  1.2     joerg #if !defined (COUNT_LEADING_ZEROS_SLOW)
   2241  1.2     joerg /* Define count_trailing_zeros using an asm count_leading_zeros.  */
   2242  1.2     joerg #define count_trailing_zeros(count, x)					\
   2243  1.1       mrg   do {									\
   2244  1.1       mrg     UWtype __ctz_x = (x);						\
   2245  1.1       mrg     UWtype __ctz_c;							\
   2246  1.1       mrg     ASSERT (__ctz_x != 0);						\
   2247  1.1       mrg     count_leading_zeros (__ctz_c, __ctz_x & -__ctz_x);			\
   2248  1.1       mrg     (count) = W_TYPE_SIZE - 1 - __ctz_c;				\
   2249  1.1       mrg   } while (0)
   2250  1.2     joerg #else
   2251  1.2     joerg /* Define count_trailing_zeros in plain C, assuming small counts are common.
   2252  1.2     joerg    We use clz_tab without ado, since the C count_leading_zeros above will have
   2253  1.2     joerg    pulled it in.  */
   2254  1.2     joerg #define count_trailing_zeros(count, x)					\
   2255  1.2     joerg   do {									\
   2256  1.2     joerg     UWtype __ctz_x = (x);						\
   2257  1.2     joerg     int __ctz_c;							\
   2258  1.2     joerg 									\
   2259  1.2     joerg     if (LIKELY ((__ctz_x & 0xff) != 0))					\
   2260  1.2     joerg       (count) = __clz_tab[__ctz_x & -__ctz_x] - 2;			\
   2261  1.2     joerg     else								\
   2262  1.2     joerg       {									\
   2263  1.2     joerg 	for (__ctz_c = 8 - 2; __ctz_c < W_TYPE_SIZE - 2; __ctz_c += 8)	\
   2264  1.2     joerg 	  {								\
   2265  1.2     joerg 	    __ctz_x >>= 8;						\
   2266  1.2     joerg 	    if (LIKELY ((__ctz_x & 0xff) != 0))				\
   2267  1.2     joerg 	      break;							\
   2268  1.2     joerg 	  }								\
   2269  1.2     joerg 									\
   2270  1.2     joerg 	(count) = __ctz_c + __clz_tab[__ctz_x & -__ctz_x];		\
   2271  1.2     joerg       }									\
   2272  1.2     joerg   } while (0)
   2273  1.2     joerg #endif
   2274  1.1       mrg #endif
   2275  1.1       mrg 
   2276  1.1       mrg #ifndef UDIV_NEEDS_NORMALIZATION
   2277  1.1       mrg #define UDIV_NEEDS_NORMALIZATION 0
   2278  1.1       mrg #endif
   2279  1.1       mrg 
   2280  1.1       mrg /* Whether udiv_qrnnd is actually implemented with udiv_qrnnd_preinv, and
   2281  1.1       mrg    that hence the latter should always be used.  */
   2282  1.1       mrg #ifndef UDIV_PREINV_ALWAYS
   2283  1.1       mrg #define UDIV_PREINV_ALWAYS 0
   2284  1.1       mrg #endif
   2285