github.com/aergoio/aergo@v1.3.1/libtool/src/gmp-6.1.2/longlong.h (about) 1 /* longlong.h -- definitions for mixed size 32/64 bit arithmetic. 2 3 Copyright 1991-1994, 1996, 1997, 1999-2005, 2007-2009, 2011-2016 Free Software 4 Foundation, Inc. 5 6 This file is part of the GNU MP Library. 7 8 The GNU MP Library is free software; you can redistribute it and/or modify 9 it under the terms of either: 10 11 * the GNU Lesser General Public License as published by the Free 12 Software Foundation; either version 3 of the License, or (at your 13 option) any later version. 14 15 or 16 17 * the GNU General Public License as published by the Free Software 18 Foundation; either version 2 of the License, or (at your option) any 19 later version. 20 21 or both in parallel, as here. 22 23 The GNU MP Library is distributed in the hope that it will be useful, but 24 WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY 25 or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 26 for more details. 27 28 You should have received copies of the GNU General Public License and the 29 GNU Lesser General Public License along with the GNU MP Library. If not, 30 see https://www.gnu.org/licenses/. */ 31 32 /* You have to define the following before including this file: 33 34 UWtype -- An unsigned type, default type for operations (typically a "word") 35 UHWtype -- An unsigned type, at least half the size of UWtype 36 UDWtype -- An unsigned type, at least twice as large a UWtype 37 W_TYPE_SIZE -- size in bits of UWtype 38 39 SItype, USItype -- Signed and unsigned 32 bit types 40 DItype, UDItype -- Signed and unsigned 64 bit types 41 42 On a 32 bit machine UWtype should typically be USItype; 43 on a 64 bit machine, UWtype should typically be UDItype. 44 45 Optionally, define: 46 47 LONGLONG_STANDALONE -- Avoid code that needs machine-dependent support files 48 NO_ASM -- Disable inline asm 49 50 51 CAUTION! Using this version of longlong.h outside of GMP is not safe. You 52 need to include gmp.h and gmp-impl.h, or certain things might not work as 53 expected. 54 */ 55 56 #define __BITS4 (W_TYPE_SIZE / 4) 57 #define __ll_B ((UWtype) 1 << (W_TYPE_SIZE / 2)) 58 #define __ll_lowpart(t) ((UWtype) (t) & (__ll_B - 1)) 59 #define __ll_highpart(t) ((UWtype) (t) >> (W_TYPE_SIZE / 2)) 60 61 /* This is used to make sure no undesirable sharing between different libraries 62 that use this file takes place. */ 63 #ifndef __MPN 64 #define __MPN(x) __##x 65 #endif 66 67 /* Define auxiliary asm macros. 68 69 1) umul_ppmm(high_prod, low_prod, multiplier, multiplicand) multiplies two 70 UWtype integers MULTIPLIER and MULTIPLICAND, and generates a two UWtype 71 word product in HIGH_PROD and LOW_PROD. 72 73 2) __umulsidi3(a,b) multiplies two UWtype integers A and B, and returns a 74 UDWtype product. This is just a variant of umul_ppmm. 75 76 3) udiv_qrnnd(quotient, remainder, high_numerator, low_numerator, 77 denominator) divides a UDWtype, composed by the UWtype integers 78 HIGH_NUMERATOR and LOW_NUMERATOR, by DENOMINATOR and places the quotient 79 in QUOTIENT and the remainder in REMAINDER. HIGH_NUMERATOR must be less 80 than DENOMINATOR for correct operation. If, in addition, the most 81 significant bit of DENOMINATOR must be 1, then the pre-processor symbol 82 UDIV_NEEDS_NORMALIZATION is defined to 1. 83 84 4) sdiv_qrnnd(quotient, remainder, high_numerator, low_numerator, 85 denominator). Like udiv_qrnnd but the numbers are signed. The quotient 86 is rounded towards 0. 87 88 5) count_leading_zeros(count, x) counts the number of zero-bits from the 89 msb to the first non-zero bit in the UWtype X. This is the number of 90 steps X needs to be shifted left to set the msb. Undefined for X == 0, 91 unless the symbol COUNT_LEADING_ZEROS_0 is defined to some value. 92 93 6) count_trailing_zeros(count, x) like count_leading_zeros, but counts 94 from the least significant end. 95 96 7) add_ssaaaa(high_sum, low_sum, high_addend_1, low_addend_1, 97 high_addend_2, low_addend_2) adds two UWtype integers, composed by 98 HIGH_ADDEND_1 and LOW_ADDEND_1, and HIGH_ADDEND_2 and LOW_ADDEND_2 99 respectively. The result is placed in HIGH_SUM and LOW_SUM. Overflow 100 (i.e. carry out) is not stored anywhere, and is lost. 101 102 8) sub_ddmmss(high_difference, low_difference, high_minuend, low_minuend, 103 high_subtrahend, low_subtrahend) subtracts two two-word UWtype integers, 104 composed by HIGH_MINUEND_1 and LOW_MINUEND_1, and HIGH_SUBTRAHEND_2 and 105 LOW_SUBTRAHEND_2 respectively. The result is placed in HIGH_DIFFERENCE 106 and LOW_DIFFERENCE. Overflow (i.e. carry out) is not stored anywhere, 107 and is lost. 108 109 If any of these macros are left undefined for a particular CPU, 110 C macros are used. 111 112 113 Notes: 114 115 For add_ssaaaa the two high and two low addends can both commute, but 116 unfortunately gcc only supports one "%" commutative in each asm block. 117 This has always been so but is only documented in recent versions 118 (eg. pre-release 3.3). Having two or more "%"s can cause an internal 119 compiler error in certain rare circumstances. 120 121 Apparently it was only the last "%" that was ever actually respected, so 122 the code has been updated to leave just that. Clearly there's a free 123 choice whether high or low should get it, if there's a reason to favour 124 one over the other. Also obviously when the constraints on the two 125 operands are identical there's no benefit to the reloader in any "%" at 126 all. 127 128 */ 129 130 /* The CPUs come in alphabetical order below. 131 132 Please add support for more CPUs here, or improve the current support 133 for the CPUs below! */ 134 135 136 /* count_leading_zeros_gcc_clz is count_leading_zeros implemented with gcc 137 3.4 __builtin_clzl or __builtin_clzll, according to our limb size. 138 Similarly count_trailing_zeros_gcc_ctz using __builtin_ctzl or 139 __builtin_ctzll. 140 141 These builtins are only used when we check what code comes out, on some 142 chips they're merely libgcc calls, where we will instead want an inline 143 in that case (either asm or generic C). 144 145 These builtins are better than an asm block of the same insn, since an 146 asm block doesn't give gcc any information about scheduling or resource 147 usage. We keep an asm block for use on prior versions of gcc though. 148 149 For reference, __builtin_ffs existed in gcc prior to __builtin_clz, but 150 it's not used (for count_leading_zeros) because it generally gives extra 151 code to ensure the result is 0 when the input is 0, which we don't need 152 or want. */ 153 154 #ifdef _LONG_LONG_LIMB 155 #define count_leading_zeros_gcc_clz(count,x) \ 156 do { \ 157 ASSERT ((x) != 0); \ 158 (count) = __builtin_clzll (x); \ 159 } while (0) 160 #else 161 #define count_leading_zeros_gcc_clz(count,x) \ 162 do { \ 163 ASSERT ((x) != 0); \ 164 (count) = __builtin_clzl (x); \ 165 } while (0) 166 #endif 167 168 #ifdef _LONG_LONG_LIMB 169 #define count_trailing_zeros_gcc_ctz(count,x) \ 170 do { \ 171 ASSERT ((x) != 0); \ 172 (count) = __builtin_ctzll (x); \ 173 } while (0) 174 #else 175 #define count_trailing_zeros_gcc_ctz(count,x) \ 176 do { \ 177 ASSERT ((x) != 0); \ 178 (count) = __builtin_ctzl (x); \ 179 } while (0) 180 #endif 181 182 183 /* FIXME: The macros using external routines like __MPN(count_leading_zeros) 184 don't need to be under !NO_ASM */ 185 #if ! defined (NO_ASM) 186 187 #if defined (__alpha) && W_TYPE_SIZE == 64 188 /* Most alpha-based machines, except Cray systems. */ 189 #if defined (__GNUC__) 190 #if __GMP_GNUC_PREREQ (3,3) 191 #define umul_ppmm(ph, pl, m0, m1) \ 192 do { \ 193 UDItype __m0 = (m0), __m1 = (m1); \ 194 (ph) = __builtin_alpha_umulh (__m0, __m1); \ 195 (pl) = __m0 * __m1; \ 196 } while (0) 197 #else 198 #define umul_ppmm(ph, pl, m0, m1) \ 199 do { \ 200 UDItype __m0 = (m0), __m1 = (m1); \ 201 __asm__ ("umulh %r1,%2,%0" \ 202 : "=r" (ph) \ 203 : "%rJ" (__m0), "rI" (__m1)); \ 204 (pl) = __m0 * __m1; \ 205 } while (0) 206 #endif 207 #define UMUL_TIME 18 208 #else /* ! __GNUC__ */ 209 #include <machine/builtins.h> 210 #define umul_ppmm(ph, pl, m0, m1) \ 211 do { \ 212 UDItype __m0 = (m0), __m1 = (m1); \ 213 (ph) = __UMULH (__m0, __m1); \ 214 (pl) = __m0 * __m1; \ 215 } while (0) 216 #endif 217 #ifndef LONGLONG_STANDALONE 218 #define udiv_qrnnd(q, r, n1, n0, d) \ 219 do { UWtype __di; \ 220 __di = __MPN(invert_limb) (d); \ 221 udiv_qrnnd_preinv (q, r, n1, n0, d, __di); \ 222 } while (0) 223 #define UDIV_PREINV_ALWAYS 1 224 #define UDIV_NEEDS_NORMALIZATION 1 225 #define UDIV_TIME 220 226 #endif /* LONGLONG_STANDALONE */ 227 228 /* clz_tab is required in all configurations, since mpn/alpha/cntlz.asm 229 always goes into libgmp.so, even when not actually used. */ 230 #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB 231 232 #if defined (__GNUC__) && HAVE_HOST_CPU_alpha_CIX 233 #define count_leading_zeros(COUNT,X) \ 234 __asm__("ctlz %1,%0" : "=r"(COUNT) : "r"(X)) 235 #define count_trailing_zeros(COUNT,X) \ 236 __asm__("cttz %1,%0" : "=r"(COUNT) : "r"(X)) 237 #endif /* clz/ctz using cix */ 238 239 #if ! defined (count_leading_zeros) \ 240 && defined (__GNUC__) && ! defined (LONGLONG_STANDALONE) 241 /* ALPHA_CMPBGE_0 gives "cmpbge $31,src,dst", ie. test src bytes == 0. 242 "$31" is written explicitly in the asm, since an "r" constraint won't 243 select reg 31. There seems no need to worry about "r31" syntax for cray, 244 since gcc itself (pre-release 3.4) emits just $31 in various places. */ 245 #define ALPHA_CMPBGE_0(dst, src) \ 246 do { asm ("cmpbge $31, %1, %0" : "=r" (dst) : "r" (src)); } while (0) 247 /* Zero bytes are turned into bits with cmpbge, a __clz_tab lookup counts 248 them, locating the highest non-zero byte. A second __clz_tab lookup 249 counts the leading zero bits in that byte, giving the result. */ 250 #define count_leading_zeros(count, x) \ 251 do { \ 252 UWtype __clz__b, __clz__c, __clz__x = (x); \ 253 ALPHA_CMPBGE_0 (__clz__b, __clz__x); /* zero bytes */ \ 254 __clz__b = __clz_tab [(__clz__b >> 1) ^ 0x7F]; /* 8 to 1 byte */ \ 255 __clz__b = __clz__b * 8 - 7; /* 57 to 1 shift */ \ 256 __clz__x >>= __clz__b; \ 257 __clz__c = __clz_tab [__clz__x]; /* 8 to 1 bit */ \ 258 __clz__b = 65 - __clz__b; \ 259 (count) = __clz__b - __clz__c; \ 260 } while (0) 261 #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB 262 #endif /* clz using cmpbge */ 263 264 #if ! defined (count_leading_zeros) && ! defined (LONGLONG_STANDALONE) 265 #if HAVE_ATTRIBUTE_CONST 266 long __MPN(count_leading_zeros) (UDItype) __attribute__ ((const)); 267 #else 268 long __MPN(count_leading_zeros) (UDItype); 269 #endif 270 #define count_leading_zeros(count, x) \ 271 ((count) = __MPN(count_leading_zeros) (x)) 272 #endif /* clz using mpn */ 273 #endif /* __alpha */ 274 275 #if defined (__AVR) && W_TYPE_SIZE == 8 276 #define umul_ppmm(ph, pl, m0, m1) \ 277 do { \ 278 unsigned short __p = (unsigned short) (m0) * (m1); \ 279 (ph) = __p >> 8; \ 280 (pl) = __p; \ 281 } while (0) 282 #endif /* AVR */ 283 284 #if defined (_CRAY) && W_TYPE_SIZE == 64 285 #include <intrinsics.h> 286 #define UDIV_PREINV_ALWAYS 1 287 #define UDIV_NEEDS_NORMALIZATION 1 288 #define UDIV_TIME 220 289 long __MPN(count_leading_zeros) (UDItype); 290 #define count_leading_zeros(count, x) \ 291 ((count) = _leadz ((UWtype) (x))) 292 #if defined (_CRAYIEEE) /* I.e., Cray T90/ieee, T3D, and T3E */ 293 #define umul_ppmm(ph, pl, m0, m1) \ 294 do { \ 295 UDItype __m0 = (m0), __m1 = (m1); \ 296 (ph) = _int_mult_upper (__m0, __m1); \ 297 (pl) = __m0 * __m1; \ 298 } while (0) 299 #ifndef LONGLONG_STANDALONE 300 #define udiv_qrnnd(q, r, n1, n0, d) \ 301 do { UWtype __di; \ 302 __di = __MPN(invert_limb) (d); \ 303 udiv_qrnnd_preinv (q, r, n1, n0, d, __di); \ 304 } while (0) 305 #endif /* LONGLONG_STANDALONE */ 306 #endif /* _CRAYIEEE */ 307 #endif /* _CRAY */ 308 309 #if defined (__ia64) && W_TYPE_SIZE == 64 310 /* This form encourages gcc (pre-release 3.4 at least) to emit predicated 311 "sub r=r,r" and "sub r=r,r,1", giving a 2 cycle latency. The generic 312 code using "al<bl" arithmetically comes out making an actual 0 or 1 in a 313 register, which takes an extra cycle. */ 314 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 315 do { \ 316 UWtype __x; \ 317 __x = (al) - (bl); \ 318 if ((al) < (bl)) \ 319 (sh) = (ah) - (bh) - 1; \ 320 else \ 321 (sh) = (ah) - (bh); \ 322 (sl) = __x; \ 323 } while (0) 324 #if defined (__GNUC__) && ! defined (__INTEL_COMPILER) 325 /* Do both product parts in assembly, since that gives better code with 326 all gcc versions. Some callers will just use the upper part, and in 327 that situation we waste an instruction, but not any cycles. */ 328 #define umul_ppmm(ph, pl, m0, m1) \ 329 __asm__ ("xma.hu %0 = %2, %3, f0\n\txma.l %1 = %2, %3, f0" \ 330 : "=&f" (ph), "=f" (pl) \ 331 : "f" (m0), "f" (m1)) 332 #define UMUL_TIME 14 333 #define count_leading_zeros(count, x) \ 334 do { \ 335 UWtype _x = (x), _y, _a, _c; \ 336 __asm__ ("mux1 %0 = %1, @rev" : "=r" (_y) : "r" (_x)); \ 337 __asm__ ("czx1.l %0 = %1" : "=r" (_a) : "r" (-_y | _y)); \ 338 _c = (_a - 1) << 3; \ 339 _x >>= _c; \ 340 if (_x >= 1 << 4) \ 341 _x >>= 4, _c += 4; \ 342 if (_x >= 1 << 2) \ 343 _x >>= 2, _c += 2; \ 344 _c += _x >> 1; \ 345 (count) = W_TYPE_SIZE - 1 - _c; \ 346 } while (0) 347 /* similar to what gcc does for __builtin_ffs, but 0 based rather than 1 348 based, and we don't need a special case for x==0 here */ 349 #define count_trailing_zeros(count, x) \ 350 do { \ 351 UWtype __ctz_x = (x); \ 352 __asm__ ("popcnt %0 = %1" \ 353 : "=r" (count) \ 354 : "r" ((__ctz_x-1) & ~__ctz_x)); \ 355 } while (0) 356 #endif 357 #if defined (__INTEL_COMPILER) 358 #include <ia64intrin.h> 359 #define umul_ppmm(ph, pl, m0, m1) \ 360 do { \ 361 UWtype __m0 = (m0), __m1 = (m1); \ 362 ph = _m64_xmahu (__m0, __m1, 0); \ 363 pl = __m0 * __m1; \ 364 } while (0) 365 #endif 366 #ifndef LONGLONG_STANDALONE 367 #define udiv_qrnnd(q, r, n1, n0, d) \ 368 do { UWtype __di; \ 369 __di = __MPN(invert_limb) (d); \ 370 udiv_qrnnd_preinv (q, r, n1, n0, d, __di); \ 371 } while (0) 372 #define UDIV_PREINV_ALWAYS 1 373 #define UDIV_NEEDS_NORMALIZATION 1 374 #endif 375 #define UDIV_TIME 220 376 #endif 377 378 379 #if defined (__GNUC__) 380 381 /* We sometimes need to clobber "cc" with gcc2, but that would not be 382 understood by gcc1. Use cpp to avoid major code duplication. */ 383 #if __GNUC__ < 2 384 #define __CLOBBER_CC 385 #define __AND_CLOBBER_CC 386 #else /* __GNUC__ >= 2 */ 387 #define __CLOBBER_CC : "cc" 388 #define __AND_CLOBBER_CC , "cc" 389 #endif /* __GNUC__ < 2 */ 390 391 #if (defined (__a29k__) || defined (_AM29K)) && W_TYPE_SIZE == 32 392 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 393 __asm__ ("add %1,%4,%5\n\taddc %0,%2,%3" \ 394 : "=r" (sh), "=&r" (sl) \ 395 : "r" (ah), "rI" (bh), "%r" (al), "rI" (bl)) 396 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 397 __asm__ ("sub %1,%4,%5\n\tsubc %0,%2,%3" \ 398 : "=r" (sh), "=&r" (sl) \ 399 : "r" (ah), "rI" (bh), "r" (al), "rI" (bl)) 400 #define umul_ppmm(xh, xl, m0, m1) \ 401 do { \ 402 USItype __m0 = (m0), __m1 = (m1); \ 403 __asm__ ("multiplu %0,%1,%2" \ 404 : "=r" (xl) \ 405 : "r" (__m0), "r" (__m1)); \ 406 __asm__ ("multmu %0,%1,%2" \ 407 : "=r" (xh) \ 408 : "r" (__m0), "r" (__m1)); \ 409 } while (0) 410 #define udiv_qrnnd(q, r, n1, n0, d) \ 411 __asm__ ("dividu %0,%3,%4" \ 412 : "=r" (q), "=q" (r) \ 413 : "1" (n1), "r" (n0), "r" (d)) 414 #define count_leading_zeros(count, x) \ 415 __asm__ ("clz %0,%1" \ 416 : "=r" (count) \ 417 : "r" (x)) 418 #define COUNT_LEADING_ZEROS_0 32 419 #endif /* __a29k__ */ 420 421 #if defined (__arc__) 422 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 423 __asm__ ("add.f\t%1, %4, %5\n\tadc\t%0, %2, %3" \ 424 : "=r" (sh), \ 425 "=&r" (sl) \ 426 : "r" ((USItype) (ah)), \ 427 "rICal" ((USItype) (bh)), \ 428 "%r" ((USItype) (al)), \ 429 "rICal" ((USItype) (bl))) 430 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 431 __asm__ ("sub.f\t%1, %4, %5\n\tsbc\t%0, %2, %3" \ 432 : "=r" (sh), \ 433 "=&r" (sl) \ 434 : "r" ((USItype) (ah)), \ 435 "rICal" ((USItype) (bh)), \ 436 "r" ((USItype) (al)), \ 437 "rICal" ((USItype) (bl))) 438 #endif 439 440 #if defined (__arm__) && (defined (__thumb2__) || !defined (__thumb__)) \ 441 && W_TYPE_SIZE == 32 442 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 443 __asm__ ("adds\t%1, %4, %5\n\tadc\t%0, %2, %3" \ 444 : "=r" (sh), "=&r" (sl) \ 445 : "r" (ah), "rI" (bh), "%r" (al), "rI" (bl) __CLOBBER_CC) 446 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 447 do { \ 448 if (__builtin_constant_p (al)) \ 449 { \ 450 if (__builtin_constant_p (ah)) \ 451 __asm__ ("rsbs\t%1, %5, %4\n\trsc\t%0, %3, %2" \ 452 : "=r" (sh), "=&r" (sl) \ 453 : "rI" (ah), "r" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \ 454 else \ 455 __asm__ ("rsbs\t%1, %5, %4\n\tsbc\t%0, %2, %3" \ 456 : "=r" (sh), "=&r" (sl) \ 457 : "r" (ah), "rI" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \ 458 } \ 459 else if (__builtin_constant_p (ah)) \ 460 { \ 461 if (__builtin_constant_p (bl)) \ 462 __asm__ ("subs\t%1, %4, %5\n\trsc\t%0, %3, %2" \ 463 : "=r" (sh), "=&r" (sl) \ 464 : "rI" (ah), "r" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \ 465 else \ 466 __asm__ ("rsbs\t%1, %5, %4\n\trsc\t%0, %3, %2" \ 467 : "=r" (sh), "=&r" (sl) \ 468 : "rI" (ah), "r" (bh), "rI" (al), "r" (bl) __CLOBBER_CC); \ 469 } \ 470 else if (__builtin_constant_p (bl)) \ 471 { \ 472 if (__builtin_constant_p (bh)) \ 473 __asm__ ("subs\t%1, %4, %5\n\tsbc\t%0, %2, %3" \ 474 : "=r" (sh), "=&r" (sl) \ 475 : "r" (ah), "rI" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \ 476 else \ 477 __asm__ ("subs\t%1, %4, %5\n\trsc\t%0, %3, %2" \ 478 : "=r" (sh), "=&r" (sl) \ 479 : "rI" (ah), "r" (bh), "r" (al), "rI" (bl) __CLOBBER_CC); \ 480 } \ 481 else /* only bh might be a constant */ \ 482 __asm__ ("subs\t%1, %4, %5\n\tsbc\t%0, %2, %3" \ 483 : "=r" (sh), "=&r" (sl) \ 484 : "r" (ah), "rI" (bh), "r" (al), "rI" (bl) __CLOBBER_CC);\ 485 } while (0) 486 #if defined (__ARM_ARCH_2__) || defined (__ARM_ARCH_2A__) \ 487 || defined (__ARM_ARCH_3__) 488 #define umul_ppmm(xh, xl, a, b) \ 489 do { \ 490 register USItype __t0, __t1, __t2; \ 491 __asm__ ("%@ Inlined umul_ppmm\n" \ 492 " mov %2, %5, lsr #16\n" \ 493 " mov %0, %6, lsr #16\n" \ 494 " bic %3, %5, %2, lsl #16\n" \ 495 " bic %4, %6, %0, lsl #16\n" \ 496 " mul %1, %3, %4\n" \ 497 " mul %4, %2, %4\n" \ 498 " mul %3, %0, %3\n" \ 499 " mul %0, %2, %0\n" \ 500 " adds %3, %4, %3\n" \ 501 " addcs %0, %0, #65536\n" \ 502 " adds %1, %1, %3, lsl #16\n" \ 503 " adc %0, %0, %3, lsr #16" \ 504 : "=&r" ((USItype) (xh)), "=r" ((USItype) (xl)), \ 505 "=&r" (__t0), "=&r" (__t1), "=r" (__t2) \ 506 : "r" ((USItype) (a)), "r" ((USItype) (b)) __CLOBBER_CC); \ 507 } while (0) 508 #define UMUL_TIME 20 509 #ifndef LONGLONG_STANDALONE 510 #define udiv_qrnnd(q, r, n1, n0, d) \ 511 do { UWtype __r; \ 512 (q) = __MPN(udiv_qrnnd) (&__r, (n1), (n0), (d)); \ 513 (r) = __r; \ 514 } while (0) 515 extern UWtype __MPN(udiv_qrnnd) (UWtype *, UWtype, UWtype, UWtype); 516 #define UDIV_TIME 200 517 #endif /* LONGLONG_STANDALONE */ 518 #else /* ARMv4 or newer */ 519 #define umul_ppmm(xh, xl, a, b) \ 520 __asm__ ("umull %0,%1,%2,%3" : "=&r" (xl), "=&r" (xh) : "r" (a), "r" (b)) 521 #define UMUL_TIME 5 522 #define smul_ppmm(xh, xl, a, b) \ 523 __asm__ ("smull %0,%1,%2,%3" : "=&r" (xl), "=&r" (xh) : "r" (a), "r" (b)) 524 #ifndef LONGLONG_STANDALONE 525 #define udiv_qrnnd(q, r, n1, n0, d) \ 526 do { UWtype __di; \ 527 __di = __MPN(invert_limb) (d); \ 528 udiv_qrnnd_preinv (q, r, n1, n0, d, __di); \ 529 } while (0) 530 #define UDIV_PREINV_ALWAYS 1 531 #define UDIV_NEEDS_NORMALIZATION 1 532 #define UDIV_TIME 70 533 #endif /* LONGLONG_STANDALONE */ 534 #endif /* defined(__ARM_ARCH_2__) ... */ 535 #define count_leading_zeros(count, x) count_leading_zeros_gcc_clz(count, x) 536 #define count_trailing_zeros(count, x) count_trailing_zeros_gcc_ctz(count, x) 537 #define COUNT_LEADING_ZEROS_0 32 538 #endif /* __arm__ */ 539 540 #if defined (__aarch64__) && W_TYPE_SIZE == 64 541 /* FIXME: Extend the immediate range for the low word by using both 542 ADDS and SUBS, since they set carry in the same way. */ 543 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 544 __asm__ ("adds\t%1, %x4, %5\n\tadc\t%0, %x2, %x3" \ 545 : "=r" (sh), "=&r" (sl) \ 546 : "rZ" ((UDItype)(ah)), "rZ" ((UDItype)(bh)), \ 547 "%r" ((UDItype)(al)), "rI" ((UDItype)(bl)) __CLOBBER_CC) 548 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 549 __asm__ ("subs\t%1, %x4, %5\n\tsbc\t%0, %x2, %x3" \ 550 : "=r,r" (sh), "=&r,&r" (sl) \ 551 : "rZ,rZ" ((UDItype)(ah)), "rZ,rZ" ((UDItype)(bh)), \ 552 "r,Z" ((UDItype)(al)), "rI,r" ((UDItype)(bl)) __CLOBBER_CC) 553 #define umul_ppmm(ph, pl, m0, m1) \ 554 do { \ 555 UDItype __m0 = (m0), __m1 = (m1); \ 556 __asm__ ("umulh\t%0, %1, %2" : "=r" (ph) : "r" (__m0), "r" (__m1)); \ 557 (pl) = __m0 * __m1; \ 558 } while (0) 559 #define count_leading_zeros(count, x) count_leading_zeros_gcc_clz(count, x) 560 #define count_trailing_zeros(count, x) count_trailing_zeros_gcc_ctz(count, x) 561 #define COUNT_LEADING_ZEROS_0 64 562 #endif /* __aarch64__ */ 563 564 #if defined (__clipper__) && W_TYPE_SIZE == 32 565 #define umul_ppmm(w1, w0, u, v) \ 566 ({union {UDItype __ll; \ 567 struct {USItype __l, __h;} __i; \ 568 } __x; \ 569 __asm__ ("mulwux %2,%0" \ 570 : "=r" (__x.__ll) \ 571 : "%0" ((USItype)(u)), "r" ((USItype)(v))); \ 572 (w1) = __x.__i.__h; (w0) = __x.__i.__l;}) 573 #define smul_ppmm(w1, w0, u, v) \ 574 ({union {DItype __ll; \ 575 struct {SItype __l, __h;} __i; \ 576 } __x; \ 577 __asm__ ("mulwx %2,%0" \ 578 : "=r" (__x.__ll) \ 579 : "%0" ((SItype)(u)), "r" ((SItype)(v))); \ 580 (w1) = __x.__i.__h; (w0) = __x.__i.__l;}) 581 #define __umulsidi3(u, v) \ 582 ({UDItype __w; \ 583 __asm__ ("mulwux %2,%0" \ 584 : "=r" (__w) : "%0" ((USItype)(u)), "r" ((USItype)(v))); \ 585 __w; }) 586 #endif /* __clipper__ */ 587 588 /* Fujitsu vector computers. */ 589 #if defined (__uxp__) && W_TYPE_SIZE == 32 590 #define umul_ppmm(ph, pl, u, v) \ 591 do { \ 592 union {UDItype __ll; \ 593 struct {USItype __h, __l;} __i; \ 594 } __x; \ 595 __asm__ ("mult.lu %1,%2,%0" : "=r" (__x.__ll) : "%r" (u), "rK" (v));\ 596 (ph) = __x.__i.__h; \ 597 (pl) = __x.__i.__l; \ 598 } while (0) 599 #define smul_ppmm(ph, pl, u, v) \ 600 do { \ 601 union {UDItype __ll; \ 602 struct {USItype __h, __l;} __i; \ 603 } __x; \ 604 __asm__ ("mult.l %1,%2,%0" : "=r" (__x.__ll) : "%r" (u), "rK" (v)); \ 605 (ph) = __x.__i.__h; \ 606 (pl) = __x.__i.__l; \ 607 } while (0) 608 #endif 609 610 #if defined (__gmicro__) && W_TYPE_SIZE == 32 611 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 612 __asm__ ("add.w %5,%1\n\taddx %3,%0" \ 613 : "=g" (sh), "=&g" (sl) \ 614 : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 615 "%1" ((USItype)(al)), "g" ((USItype)(bl))) 616 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 617 __asm__ ("sub.w %5,%1\n\tsubx %3,%0" \ 618 : "=g" (sh), "=&g" (sl) \ 619 : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 620 "1" ((USItype)(al)), "g" ((USItype)(bl))) 621 #define umul_ppmm(ph, pl, m0, m1) \ 622 __asm__ ("mulx %3,%0,%1" \ 623 : "=g" (ph), "=r" (pl) \ 624 : "%0" ((USItype)(m0)), "g" ((USItype)(m1))) 625 #define udiv_qrnnd(q, r, nh, nl, d) \ 626 __asm__ ("divx %4,%0,%1" \ 627 : "=g" (q), "=r" (r) \ 628 : "1" ((USItype)(nh)), "0" ((USItype)(nl)), "g" ((USItype)(d))) 629 #define count_leading_zeros(count, x) \ 630 __asm__ ("bsch/1 %1,%0" \ 631 : "=g" (count) : "g" ((USItype)(x)), "0" ((USItype)0)) 632 #endif 633 634 #if defined (__hppa) && W_TYPE_SIZE == 32 635 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 636 __asm__ ("add%I5 %5,%r4,%1\n\taddc %r2,%r3,%0" \ 637 : "=r" (sh), "=&r" (sl) \ 638 : "rM" (ah), "rM" (bh), "%rM" (al), "rI" (bl)) 639 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 640 __asm__ ("sub%I4 %4,%r5,%1\n\tsubb %r2,%r3,%0" \ 641 : "=r" (sh), "=&r" (sl) \ 642 : "rM" (ah), "rM" (bh), "rI" (al), "rM" (bl)) 643 #if defined (_PA_RISC1_1) 644 #define umul_ppmm(wh, wl, u, v) \ 645 do { \ 646 union {UDItype __ll; \ 647 struct {USItype __h, __l;} __i; \ 648 } __x; \ 649 __asm__ ("xmpyu %1,%2,%0" : "=*f" (__x.__ll) : "*f" (u), "*f" (v)); \ 650 (wh) = __x.__i.__h; \ 651 (wl) = __x.__i.__l; \ 652 } while (0) 653 #define UMUL_TIME 8 654 #define UDIV_TIME 60 655 #else 656 #define UMUL_TIME 40 657 #define UDIV_TIME 80 658 #endif 659 #define count_leading_zeros(count, x) \ 660 do { \ 661 USItype __tmp; \ 662 __asm__ ( \ 663 "ldi 1,%0\n" \ 664 " extru,= %1,15,16,%%r0 ; Bits 31..16 zero?\n" \ 665 " extru,tr %1,15,16,%1 ; No. Shift down, skip add.\n" \ 666 " ldo 16(%0),%0 ; Yes. Perform add.\n" \ 667 " extru,= %1,23,8,%%r0 ; Bits 15..8 zero?\n" \ 668 " extru,tr %1,23,8,%1 ; No. Shift down, skip add.\n" \ 669 " ldo 8(%0),%0 ; Yes. Perform add.\n" \ 670 " extru,= %1,27,4,%%r0 ; Bits 7..4 zero?\n" \ 671 " extru,tr %1,27,4,%1 ; No. Shift down, skip add.\n" \ 672 " ldo 4(%0),%0 ; Yes. Perform add.\n" \ 673 " extru,= %1,29,2,%%r0 ; Bits 3..2 zero?\n" \ 674 " extru,tr %1,29,2,%1 ; No. Shift down, skip add.\n" \ 675 " ldo 2(%0),%0 ; Yes. Perform add.\n" \ 676 " extru %1,30,1,%1 ; Extract bit 1.\n" \ 677 " sub %0,%1,%0 ; Subtract it.\n" \ 678 : "=r" (count), "=r" (__tmp) : "1" (x)); \ 679 } while (0) 680 #endif /* hppa */ 681 682 /* These macros are for ABI=2.0w. In ABI=2.0n they can't be used, since GCC 683 (3.2) puts longlong into two adjacent 32-bit registers. Presumably this 684 is just a case of no direct support for 2.0n but treating it like 1.0. */ 685 #if defined (__hppa) && W_TYPE_SIZE == 64 && ! defined (_LONG_LONG_LIMB) 686 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 687 __asm__ ("add%I5 %5,%r4,%1\n\tadd,dc %r2,%r3,%0" \ 688 : "=r" (sh), "=&r" (sl) \ 689 : "rM" (ah), "rM" (bh), "%rM" (al), "rI" (bl)) 690 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 691 __asm__ ("sub%I4 %4,%r5,%1\n\tsub,db %r2,%r3,%0" \ 692 : "=r" (sh), "=&r" (sl) \ 693 : "rM" (ah), "rM" (bh), "rI" (al), "rM" (bl)) 694 #endif /* hppa */ 695 696 #if (defined (__i370__) || defined (__s390__) || defined (__mvs__)) && W_TYPE_SIZE == 32 697 #if defined (__zarch__) || defined (HAVE_HOST_CPU_s390_zarch) 698 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 699 do { \ 700 /* if (__builtin_constant_p (bl)) \ 701 __asm__ ("alfi\t%1,%o5\n\talcr\t%0,%3" \ 702 : "=r" (sh), "=&r" (sl) \ 703 : "0" (ah), "r" (bh), "%1" (al), "n" (bl) __CLOBBER_CC);\ 704 else \ 705 */ __asm__ ("alr\t%1,%5\n\talcr\t%0,%3" \ 706 : "=r" (sh), "=&r" (sl) \ 707 : "0" (ah), "r" (bh), "%1" (al), "r" (bl)__CLOBBER_CC); \ 708 } while (0) 709 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 710 do { \ 711 /* if (__builtin_constant_p (bl)) \ 712 __asm__ ("slfi\t%1,%o5\n\tslbr\t%0,%3" \ 713 : "=r" (sh), "=&r" (sl) \ 714 : "0" (ah), "r" (bh), "1" (al), "n" (bl) __CLOBBER_CC); \ 715 else \ 716 */ __asm__ ("slr\t%1,%5\n\tslbr\t%0,%3" \ 717 : "=r" (sh), "=&r" (sl) \ 718 : "0" (ah), "r" (bh), "1" (al), "r" (bl) __CLOBBER_CC); \ 719 } while (0) 720 #if __GMP_GNUC_PREREQ (4,5) 721 #define umul_ppmm(xh, xl, m0, m1) \ 722 do { \ 723 union {UDItype __ll; \ 724 struct {USItype __h, __l;} __i; \ 725 } __x; \ 726 __x.__ll = (UDItype) (m0) * (UDItype) (m1); \ 727 (xh) = __x.__i.__h; (xl) = __x.__i.__l; \ 728 } while (0) 729 #else 730 #if 0 731 /* FIXME: this fails if gcc knows about the 64-bit registers. Use only 732 with a new enough processor pretending we have 32-bit registers. */ 733 #define umul_ppmm(xh, xl, m0, m1) \ 734 do { \ 735 union {UDItype __ll; \ 736 struct {USItype __h, __l;} __i; \ 737 } __x; \ 738 __asm__ ("mlr\t%0,%2" \ 739 : "=r" (__x.__ll) \ 740 : "%0" (m0), "r" (m1)); \ 741 (xh) = __x.__i.__h; (xl) = __x.__i.__l; \ 742 } while (0) 743 #else 744 #define umul_ppmm(xh, xl, m0, m1) \ 745 do { \ 746 /* When we have 64-bit regs and gcc is aware of that, we cannot simply use 747 DImode for the product, since that would be allocated to a single 64-bit 748 register, whereas mlr uses the low 32-bits of an even-odd register pair. 749 */ \ 750 register USItype __r0 __asm__ ("0"); \ 751 register USItype __r1 __asm__ ("1") = (m0); \ 752 __asm__ ("mlr\t%0,%3" \ 753 : "=r" (__r0), "=r" (__r1) \ 754 : "r" (__r1), "r" (m1)); \ 755 (xh) = __r0; (xl) = __r1; \ 756 } while (0) 757 #endif /* if 0 */ 758 #endif 759 #if 0 760 /* FIXME: this fails if gcc knows about the 64-bit registers. Use only 761 with a new enough processor pretending we have 32-bit registers. */ 762 #define udiv_qrnnd(q, r, n1, n0, d) \ 763 do { \ 764 union {UDItype __ll; \ 765 struct {USItype __h, __l;} __i; \ 766 } __x; \ 767 __x.__i.__h = n1; __x.__i.__l = n0; \ 768 __asm__ ("dlr\t%0,%2" \ 769 : "=r" (__x.__ll) \ 770 : "0" (__x.__ll), "r" (d)); \ 771 (q) = __x.__i.__l; (r) = __x.__i.__h; \ 772 } while (0) 773 #else 774 #define udiv_qrnnd(q, r, n1, n0, d) \ 775 do { \ 776 register USItype __r0 __asm__ ("0") = (n1); \ 777 register USItype __r1 __asm__ ("1") = (n0); \ 778 __asm__ ("dlr\t%0,%4" \ 779 : "=r" (__r0), "=r" (__r1) \ 780 : "r" (__r0), "r" (__r1), "r" (d)); \ 781 (q) = __r1; (r) = __r0; \ 782 } while (0) 783 #endif /* if 0 */ 784 #else /* if __zarch__ */ 785 /* FIXME: this fails if gcc knows about the 64-bit registers. */ 786 #define smul_ppmm(xh, xl, m0, m1) \ 787 do { \ 788 union {DItype __ll; \ 789 struct {USItype __h, __l;} __i; \ 790 } __x; \ 791 __asm__ ("mr\t%0,%2" \ 792 : "=r" (__x.__ll) \ 793 : "%0" (m0), "r" (m1)); \ 794 (xh) = __x.__i.__h; (xl) = __x.__i.__l; \ 795 } while (0) 796 /* FIXME: this fails if gcc knows about the 64-bit registers. */ 797 #define sdiv_qrnnd(q, r, n1, n0, d) \ 798 do { \ 799 union {DItype __ll; \ 800 struct {USItype __h, __l;} __i; \ 801 } __x; \ 802 __x.__i.__h = n1; __x.__i.__l = n0; \ 803 __asm__ ("dr\t%0,%2" \ 804 : "=r" (__x.__ll) \ 805 : "0" (__x.__ll), "r" (d)); \ 806 (q) = __x.__i.__l; (r) = __x.__i.__h; \ 807 } while (0) 808 #endif /* if __zarch__ */ 809 #endif 810 811 #if defined (__s390x__) && W_TYPE_SIZE == 64 812 /* We need to cast operands with register constraints, otherwise their types 813 will be assumed to be SImode by gcc. For these machines, such operations 814 will insert a value into the low 32 bits, and leave the high 32 bits with 815 garbage. */ 816 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 817 do { \ 818 __asm__ ("algr\t%1,%5\n\talcgr\t%0,%3" \ 819 : "=r" (sh), "=&r" (sl) \ 820 : "0" ((UDItype)(ah)), "r" ((UDItype)(bh)), \ 821 "%1" ((UDItype)(al)), "r" ((UDItype)(bl)) __CLOBBER_CC); \ 822 } while (0) 823 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 824 do { \ 825 __asm__ ("slgr\t%1,%5\n\tslbgr\t%0,%3" \ 826 : "=r" (sh), "=&r" (sl) \ 827 : "0" ((UDItype)(ah)), "r" ((UDItype)(bh)), \ 828 "1" ((UDItype)(al)), "r" ((UDItype)(bl)) __CLOBBER_CC); \ 829 } while (0) 830 #define umul_ppmm(xh, xl, m0, m1) \ 831 do { \ 832 union {unsigned int __attribute__ ((mode(TI))) __ll; \ 833 struct {UDItype __h, __l;} __i; \ 834 } __x; \ 835 __asm__ ("mlgr\t%0,%2" \ 836 : "=r" (__x.__ll) \ 837 : "%0" ((UDItype)(m0)), "r" ((UDItype)(m1))); \ 838 (xh) = __x.__i.__h; (xl) = __x.__i.__l; \ 839 } while (0) 840 #define udiv_qrnnd(q, r, n1, n0, d) \ 841 do { \ 842 union {unsigned int __attribute__ ((mode(TI))) __ll; \ 843 struct {UDItype __h, __l;} __i; \ 844 } __x; \ 845 __x.__i.__h = n1; __x.__i.__l = n0; \ 846 __asm__ ("dlgr\t%0,%2" \ 847 : "=r" (__x.__ll) \ 848 : "0" (__x.__ll), "r" ((UDItype)(d))); \ 849 (q) = __x.__i.__l; (r) = __x.__i.__h; \ 850 } while (0) 851 #if 0 /* FIXME: Enable for z10 (?) */ 852 #define count_leading_zeros(cnt, x) \ 853 do { \ 854 union {unsigned int __attribute__ ((mode(TI))) __ll; \ 855 struct {UDItype __h, __l;} __i; \ 856 } __clr_cnt; \ 857 __asm__ ("flogr\t%0,%1" \ 858 : "=r" (__clr_cnt.__ll) \ 859 : "r" (x) __CLOBBER_CC); \ 860 (cnt) = __clr_cnt.__i.__h; \ 861 } while (0) 862 #endif 863 #endif 864 865 /* On x86 and x86_64, every asm implicitly clobbers "flags" and "fpsr", 866 so we don't need __CLOBBER_CC. */ 867 #if (defined (__i386__) || defined (__i486__)) && W_TYPE_SIZE == 32 868 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 869 __asm__ ("addl %5,%k1\n\tadcl %3,%k0" \ 870 : "=r" (sh), "=&r" (sl) \ 871 : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 872 "%1" ((USItype)(al)), "g" ((USItype)(bl))) 873 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 874 __asm__ ("subl %5,%k1\n\tsbbl %3,%k0" \ 875 : "=r" (sh), "=&r" (sl) \ 876 : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 877 "1" ((USItype)(al)), "g" ((USItype)(bl))) 878 #define umul_ppmm(w1, w0, u, v) \ 879 __asm__ ("mull %3" \ 880 : "=a" (w0), "=d" (w1) \ 881 : "%0" ((USItype)(u)), "rm" ((USItype)(v))) 882 #define udiv_qrnnd(q, r, n1, n0, dx) /* d renamed to dx avoiding "=d" */\ 883 __asm__ ("divl %4" /* stringification in K&R C */ \ 884 : "=a" (q), "=d" (r) \ 885 : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "rm" ((USItype)(dx))) 886 887 #if HAVE_HOST_CPU_i586 || HAVE_HOST_CPU_pentium || HAVE_HOST_CPU_pentiummmx 888 /* Pentium bsrl takes between 10 and 72 cycles depending where the most 889 significant 1 bit is, hence the use of the following alternatives. bsfl 890 is slow too, between 18 and 42 depending where the least significant 1 891 bit is, so let the generic count_trailing_zeros below make use of the 892 count_leading_zeros here too. */ 893 894 #if HAVE_HOST_CPU_pentiummmx && ! defined (LONGLONG_STANDALONE) 895 /* The following should be a fixed 14 or 15 cycles, but possibly plus an L1 896 cache miss reading from __clz_tab. For P55 it's favoured over the float 897 below so as to avoid mixing MMX and x87, since the penalty for switching 898 between the two is about 100 cycles. 899 900 The asm block sets __shift to -3 if the high 24 bits are clear, -2 for 901 16, -1 for 8, or 0 otherwise. This could be written equivalently as 902 follows, but as of gcc 2.95.2 it results in conditional jumps. 903 904 __shift = -(__n < 0x1000000); 905 __shift -= (__n < 0x10000); 906 __shift -= (__n < 0x100); 907 908 The middle two sbbl and cmpl's pair, and with luck something gcc 909 generates might pair with the first cmpl and the last sbbl. The "32+1" 910 constant could be folded into __clz_tab[], but it doesn't seem worth 911 making a different table just for that. */ 912 913 #define count_leading_zeros(c,n) \ 914 do { \ 915 USItype __n = (n); \ 916 USItype __shift; \ 917 __asm__ ("cmpl $0x1000000, %1\n" \ 918 "sbbl %0, %0\n" \ 919 "cmpl $0x10000, %1\n" \ 920 "sbbl $0, %0\n" \ 921 "cmpl $0x100, %1\n" \ 922 "sbbl $0, %0\n" \ 923 : "=&r" (__shift) : "r" (__n)); \ 924 __shift = __shift*8 + 24 + 1; \ 925 (c) = 32 + 1 - __shift - __clz_tab[__n >> __shift]; \ 926 } while (0) 927 #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB 928 #define COUNT_LEADING_ZEROS_0 31 /* n==0 indistinguishable from n==1 */ 929 930 #else /* ! pentiummmx || LONGLONG_STANDALONE */ 931 /* The following should be a fixed 14 cycles or so. Some scheduling 932 opportunities should be available between the float load/store too. This 933 sort of code is used in gcc 3 for __builtin_ffs (with "n&-n") and is 934 apparently suggested by the Intel optimizing manual (don't know exactly 935 where). gcc 2.95 or up will be best for this, so the "double" is 936 correctly aligned on the stack. */ 937 #define count_leading_zeros(c,n) \ 938 do { \ 939 union { \ 940 double d; \ 941 unsigned a[2]; \ 942 } __u; \ 943 ASSERT ((n) != 0); \ 944 __u.d = (UWtype) (n); \ 945 (c) = 0x3FF + 31 - (__u.a[1] >> 20); \ 946 } while (0) 947 #define COUNT_LEADING_ZEROS_0 (0x3FF + 31) 948 #endif /* pentiummx */ 949 950 #else /* ! pentium */ 951 952 #if __GMP_GNUC_PREREQ (3,4) /* using bsrl */ 953 #define count_leading_zeros(count,x) count_leading_zeros_gcc_clz(count,x) 954 #endif /* gcc clz */ 955 956 /* On P6, gcc prior to 3.0 generates a partial register stall for 957 __cbtmp^31, due to using "xorb $31" instead of "xorl $31", the former 958 being 1 code byte smaller. "31-__cbtmp" is a workaround, probably at the 959 cost of one extra instruction. Do this for "i386" too, since that means 960 generic x86. */ 961 #if ! defined (count_leading_zeros) && __GNUC__ < 3 \ 962 && (HAVE_HOST_CPU_i386 \ 963 || HAVE_HOST_CPU_i686 \ 964 || HAVE_HOST_CPU_pentiumpro \ 965 || HAVE_HOST_CPU_pentium2 \ 966 || HAVE_HOST_CPU_pentium3) 967 #define count_leading_zeros(count, x) \ 968 do { \ 969 USItype __cbtmp; \ 970 ASSERT ((x) != 0); \ 971 __asm__ ("bsrl %1,%0" : "=r" (__cbtmp) : "rm" ((USItype)(x))); \ 972 (count) = 31 - __cbtmp; \ 973 } while (0) 974 #endif /* gcc<3 asm bsrl */ 975 976 #ifndef count_leading_zeros 977 #define count_leading_zeros(count, x) \ 978 do { \ 979 USItype __cbtmp; \ 980 ASSERT ((x) != 0); \ 981 __asm__ ("bsrl %1,%0" : "=r" (__cbtmp) : "rm" ((USItype)(x))); \ 982 (count) = __cbtmp ^ 31; \ 983 } while (0) 984 #endif /* asm bsrl */ 985 986 #if __GMP_GNUC_PREREQ (3,4) /* using bsfl */ 987 #define count_trailing_zeros(count,x) count_trailing_zeros_gcc_ctz(count,x) 988 #endif /* gcc ctz */ 989 990 #ifndef count_trailing_zeros 991 #define count_trailing_zeros(count, x) \ 992 do { \ 993 ASSERT ((x) != 0); \ 994 __asm__ ("bsfl %1,%k0" : "=r" (count) : "rm" ((USItype)(x))); \ 995 } while (0) 996 #endif /* asm bsfl */ 997 998 #endif /* ! pentium */ 999 1000 #ifndef UMUL_TIME 1001 #define UMUL_TIME 10 1002 #endif 1003 #ifndef UDIV_TIME 1004 #define UDIV_TIME 40 1005 #endif 1006 #endif /* 80x86 */ 1007 1008 #if defined (__amd64__) && W_TYPE_SIZE == 64 1009 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1010 __asm__ ("addq %5,%q1\n\tadcq %3,%q0" \ 1011 : "=r" (sh), "=&r" (sl) \ 1012 : "0" ((UDItype)(ah)), "rme" ((UDItype)(bh)), \ 1013 "%1" ((UDItype)(al)), "rme" ((UDItype)(bl))) 1014 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1015 __asm__ ("subq %5,%q1\n\tsbbq %3,%q0" \ 1016 : "=r" (sh), "=&r" (sl) \ 1017 : "0" ((UDItype)(ah)), "rme" ((UDItype)(bh)), \ 1018 "1" ((UDItype)(al)), "rme" ((UDItype)(bl))) 1019 #define umul_ppmm(w1, w0, u, v) \ 1020 __asm__ ("mulq %3" \ 1021 : "=a" (w0), "=d" (w1) \ 1022 : "%0" ((UDItype)(u)), "rm" ((UDItype)(v))) 1023 #define udiv_qrnnd(q, r, n1, n0, dx) /* d renamed to dx avoiding "=d" */\ 1024 __asm__ ("divq %4" /* stringification in K&R C */ \ 1025 : "=a" (q), "=d" (r) \ 1026 : "0" ((UDItype)(n0)), "1" ((UDItype)(n1)), "rm" ((UDItype)(dx))) 1027 /* bsrq destination must be a 64-bit register, hence UDItype for __cbtmp. */ 1028 #define count_leading_zeros(count, x) \ 1029 do { \ 1030 UDItype __cbtmp; \ 1031 ASSERT ((x) != 0); \ 1032 __asm__ ("bsrq %1,%0" : "=r" (__cbtmp) : "rm" ((UDItype)(x))); \ 1033 (count) = __cbtmp ^ 63; \ 1034 } while (0) 1035 /* bsfq destination must be a 64-bit register, "%q0" forces this in case 1036 count is only an int. */ 1037 #define count_trailing_zeros(count, x) \ 1038 do { \ 1039 ASSERT ((x) != 0); \ 1040 __asm__ ("bsfq %1,%q0" : "=r" (count) : "rm" ((UDItype)(x))); \ 1041 } while (0) 1042 #endif /* __amd64__ */ 1043 1044 #if defined (__i860__) && W_TYPE_SIZE == 32 1045 #define rshift_rhlc(r,h,l,c) \ 1046 __asm__ ("shr %3,r0,r0\;shrd %1,%2,%0" \ 1047 "=r" (r) : "r" (h), "r" (l), "rn" (c)) 1048 #endif /* i860 */ 1049 1050 #if defined (__i960__) && W_TYPE_SIZE == 32 1051 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1052 __asm__ ("cmpo 1,0\;addc %5,%4,%1\;addc %3,%2,%0" \ 1053 : "=r" (sh), "=&r" (sl) \ 1054 : "dI" (ah), "dI" (bh), "%dI" (al), "dI" (bl)) 1055 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1056 __asm__ ("cmpo 0,0\;subc %5,%4,%1\;subc %3,%2,%0" \ 1057 : "=r" (sh), "=&r" (sl) \ 1058 : "dI" (ah), "dI" (bh), "dI" (al), "dI" (bl)) 1059 #define umul_ppmm(w1, w0, u, v) \ 1060 ({union {UDItype __ll; \ 1061 struct {USItype __l, __h;} __i; \ 1062 } __x; \ 1063 __asm__ ("emul %2,%1,%0" \ 1064 : "=d" (__x.__ll) : "%dI" (u), "dI" (v)); \ 1065 (w1) = __x.__i.__h; (w0) = __x.__i.__l;}) 1066 #define __umulsidi3(u, v) \ 1067 ({UDItype __w; \ 1068 __asm__ ("emul %2,%1,%0" : "=d" (__w) : "%dI" (u), "dI" (v)); \ 1069 __w; }) 1070 #define udiv_qrnnd(q, r, nh, nl, d) \ 1071 do { \ 1072 union {UDItype __ll; \ 1073 struct {USItype __l, __h;} __i; \ 1074 } __nn; \ 1075 __nn.__i.__h = (nh); __nn.__i.__l = (nl); \ 1076 __asm__ ("ediv %d,%n,%0" \ 1077 : "=d" (__rq.__ll) : "dI" (__nn.__ll), "dI" (d)); \ 1078 (r) = __rq.__i.__l; (q) = __rq.__i.__h; \ 1079 } while (0) 1080 #define count_leading_zeros(count, x) \ 1081 do { \ 1082 USItype __cbtmp; \ 1083 __asm__ ("scanbit %1,%0" : "=r" (__cbtmp) : "r" (x)); \ 1084 (count) = __cbtmp ^ 31; \ 1085 } while (0) 1086 #define COUNT_LEADING_ZEROS_0 (-32) /* sic */ 1087 #if defined (__i960mx) /* what is the proper symbol to test??? */ 1088 #define rshift_rhlc(r,h,l,c) \ 1089 do { \ 1090 union {UDItype __ll; \ 1091 struct {USItype __l, __h;} __i; \ 1092 } __nn; \ 1093 __nn.__i.__h = (h); __nn.__i.__l = (l); \ 1094 __asm__ ("shre %2,%1,%0" : "=d" (r) : "dI" (__nn.__ll), "dI" (c)); \ 1095 } 1096 #endif /* i960mx */ 1097 #endif /* i960 */ 1098 1099 #if (defined (__mc68000__) || defined (__mc68020__) || defined(mc68020) \ 1100 || defined (__m68k__) || defined (__mc5200__) || defined (__mc5206e__) \ 1101 || defined (__mc5307__)) && W_TYPE_SIZE == 32 1102 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1103 __asm__ ("add%.l %5,%1\n\taddx%.l %3,%0" \ 1104 : "=d" (sh), "=&d" (sl) \ 1105 : "0" ((USItype)(ah)), "d" ((USItype)(bh)), \ 1106 "%1" ((USItype)(al)), "g" ((USItype)(bl))) 1107 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1108 __asm__ ("sub%.l %5,%1\n\tsubx%.l %3,%0" \ 1109 : "=d" (sh), "=&d" (sl) \ 1110 : "0" ((USItype)(ah)), "d" ((USItype)(bh)), \ 1111 "1" ((USItype)(al)), "g" ((USItype)(bl))) 1112 /* The '020, '030, '040 and CPU32 have 32x32->64 and 64/32->32q-32r. */ 1113 #if defined (__mc68020__) || defined(mc68020) \ 1114 || defined (__mc68030__) || defined (mc68030) \ 1115 || defined (__mc68040__) || defined (mc68040) \ 1116 || defined (__mcpu32__) || defined (mcpu32) \ 1117 || defined (__NeXT__) 1118 #define umul_ppmm(w1, w0, u, v) \ 1119 __asm__ ("mulu%.l %3,%1:%0" \ 1120 : "=d" (w0), "=d" (w1) \ 1121 : "%0" ((USItype)(u)), "dmi" ((USItype)(v))) 1122 #define UMUL_TIME 45 1123 #define udiv_qrnnd(q, r, n1, n0, d) \ 1124 __asm__ ("divu%.l %4,%1:%0" \ 1125 : "=d" (q), "=d" (r) \ 1126 : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "dmi" ((USItype)(d))) 1127 #define UDIV_TIME 90 1128 #define sdiv_qrnnd(q, r, n1, n0, d) \ 1129 __asm__ ("divs%.l %4,%1:%0" \ 1130 : "=d" (q), "=d" (r) \ 1131 : "0" ((USItype)(n0)), "1" ((USItype)(n1)), "dmi" ((USItype)(d))) 1132 #else /* for other 68k family members use 16x16->32 multiplication */ 1133 #define umul_ppmm(xh, xl, a, b) \ 1134 do { USItype __umul_tmp1, __umul_tmp2; \ 1135 __asm__ ("| Inlined umul_ppmm\n" \ 1136 " move%.l %5,%3\n" \ 1137 " move%.l %2,%0\n" \ 1138 " move%.w %3,%1\n" \ 1139 " swap %3\n" \ 1140 " swap %0\n" \ 1141 " mulu%.w %2,%1\n" \ 1142 " mulu%.w %3,%0\n" \ 1143 " mulu%.w %2,%3\n" \ 1144 " swap %2\n" \ 1145 " mulu%.w %5,%2\n" \ 1146 " add%.l %3,%2\n" \ 1147 " jcc 1f\n" \ 1148 " add%.l %#0x10000,%0\n" \ 1149 "1: move%.l %2,%3\n" \ 1150 " clr%.w %2\n" \ 1151 " swap %2\n" \ 1152 " swap %3\n" \ 1153 " clr%.w %3\n" \ 1154 " add%.l %3,%1\n" \ 1155 " addx%.l %2,%0\n" \ 1156 " | End inlined umul_ppmm" \ 1157 : "=&d" (xh), "=&d" (xl), \ 1158 "=d" (__umul_tmp1), "=&d" (__umul_tmp2) \ 1159 : "%2" ((USItype)(a)), "d" ((USItype)(b))); \ 1160 } while (0) 1161 #define UMUL_TIME 100 1162 #define UDIV_TIME 400 1163 #endif /* not mc68020 */ 1164 /* The '020, '030, '040 and '060 have bitfield insns. 1165 GCC 3.4 defines __mc68020__ when in CPU32 mode, check for __mcpu32__ to 1166 exclude bfffo on that chip (bitfield insns not available). */ 1167 #if (defined (__mc68020__) || defined (mc68020) \ 1168 || defined (__mc68030__) || defined (mc68030) \ 1169 || defined (__mc68040__) || defined (mc68040) \ 1170 || defined (__mc68060__) || defined (mc68060) \ 1171 || defined (__NeXT__)) \ 1172 && ! defined (__mcpu32__) 1173 #define count_leading_zeros(count, x) \ 1174 __asm__ ("bfffo %1{%b2:%b2},%0" \ 1175 : "=d" (count) \ 1176 : "od" ((USItype) (x)), "n" (0)) 1177 #define COUNT_LEADING_ZEROS_0 32 1178 #endif 1179 #endif /* mc68000 */ 1180 1181 #if defined (__m88000__) && W_TYPE_SIZE == 32 1182 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1183 __asm__ ("addu.co %1,%r4,%r5\n\taddu.ci %0,%r2,%r3" \ 1184 : "=r" (sh), "=&r" (sl) \ 1185 : "rJ" (ah), "rJ" (bh), "%rJ" (al), "rJ" (bl)) 1186 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1187 __asm__ ("subu.co %1,%r4,%r5\n\tsubu.ci %0,%r2,%r3" \ 1188 : "=r" (sh), "=&r" (sl) \ 1189 : "rJ" (ah), "rJ" (bh), "rJ" (al), "rJ" (bl)) 1190 #define count_leading_zeros(count, x) \ 1191 do { \ 1192 USItype __cbtmp; \ 1193 __asm__ ("ff1 %0,%1" : "=r" (__cbtmp) : "r" (x)); \ 1194 (count) = __cbtmp ^ 31; \ 1195 } while (0) 1196 #define COUNT_LEADING_ZEROS_0 63 /* sic */ 1197 #if defined (__m88110__) 1198 #define umul_ppmm(wh, wl, u, v) \ 1199 do { \ 1200 union {UDItype __ll; \ 1201 struct {USItype __h, __l;} __i; \ 1202 } __x; \ 1203 __asm__ ("mulu.d %0,%1,%2" : "=r" (__x.__ll) : "r" (u), "r" (v)); \ 1204 (wh) = __x.__i.__h; \ 1205 (wl) = __x.__i.__l; \ 1206 } while (0) 1207 #define udiv_qrnnd(q, r, n1, n0, d) \ 1208 ({union {UDItype __ll; \ 1209 struct {USItype __h, __l;} __i; \ 1210 } __x, __q; \ 1211 __x.__i.__h = (n1); __x.__i.__l = (n0); \ 1212 __asm__ ("divu.d %0,%1,%2" \ 1213 : "=r" (__q.__ll) : "r" (__x.__ll), "r" (d)); \ 1214 (r) = (n0) - __q.__l * (d); (q) = __q.__l; }) 1215 #define UMUL_TIME 5 1216 #define UDIV_TIME 25 1217 #else 1218 #define UMUL_TIME 17 1219 #define UDIV_TIME 150 1220 #endif /* __m88110__ */ 1221 #endif /* __m88000__ */ 1222 1223 #if defined (__mips) && W_TYPE_SIZE == 32 1224 #if __GMP_GNUC_PREREQ (4,4) 1225 #define umul_ppmm(w1, w0, u, v) \ 1226 do { \ 1227 UDItype __ll = (UDItype)(u) * (v); \ 1228 w1 = __ll >> 32; \ 1229 w0 = __ll; \ 1230 } while (0) 1231 #endif 1232 #if !defined (umul_ppmm) && __GMP_GNUC_PREREQ (2,7) && !defined (__clang__) 1233 #define umul_ppmm(w1, w0, u, v) \ 1234 __asm__ ("multu %2,%3" : "=l" (w0), "=h" (w1) : "d" (u), "d" (v)) 1235 #endif 1236 #if !defined (umul_ppmm) 1237 #define umul_ppmm(w1, w0, u, v) \ 1238 __asm__ ("multu %2,%3\n\tmflo %0\n\tmfhi %1" \ 1239 : "=d" (w0), "=d" (w1) : "d" (u), "d" (v)) 1240 #endif 1241 #define UMUL_TIME 10 1242 #define UDIV_TIME 100 1243 #endif /* __mips */ 1244 1245 #if (defined (__mips) && __mips >= 3) && W_TYPE_SIZE == 64 1246 #if __GMP_GNUC_PREREQ (4,4) 1247 #define umul_ppmm(w1, w0, u, v) \ 1248 do { \ 1249 typedef unsigned int __ll_UTItype __attribute__((mode(TI))); \ 1250 __ll_UTItype __ll = (__ll_UTItype)(u) * (v); \ 1251 w1 = __ll >> 64; \ 1252 w0 = __ll; \ 1253 } while (0) 1254 #endif 1255 #if !defined (umul_ppmm) && __GMP_GNUC_PREREQ (2,7) && !defined (__clang__) 1256 #define umul_ppmm(w1, w0, u, v) \ 1257 __asm__ ("dmultu %2,%3" \ 1258 : "=l" (w0), "=h" (w1) \ 1259 : "d" ((UDItype)(u)), "d" ((UDItype)(v))) 1260 #endif 1261 #if !defined (umul_ppmm) 1262 #define umul_ppmm(w1, w0, u, v) \ 1263 __asm__ ("dmultu %2,%3\n\tmflo %0\n\tmfhi %1" \ 1264 : "=d" (w0), "=d" (w1) \ 1265 : "d" ((UDItype)(u)), "d" ((UDItype)(v))) 1266 #endif 1267 #define UMUL_TIME 20 1268 #define UDIV_TIME 140 1269 #endif /* __mips */ 1270 1271 #if defined (__mmix__) && W_TYPE_SIZE == 64 1272 #define umul_ppmm(w1, w0, u, v) \ 1273 __asm__ ("MULU %0,%2,%3" : "=r" (w0), "=z" (w1) : "r" (u), "r" (v)) 1274 #endif 1275 1276 #if defined (__ns32000__) && W_TYPE_SIZE == 32 1277 #define umul_ppmm(w1, w0, u, v) \ 1278 ({union {UDItype __ll; \ 1279 struct {USItype __l, __h;} __i; \ 1280 } __x; \ 1281 __asm__ ("meid %2,%0" \ 1282 : "=g" (__x.__ll) \ 1283 : "%0" ((USItype)(u)), "g" ((USItype)(v))); \ 1284 (w1) = __x.__i.__h; (w0) = __x.__i.__l;}) 1285 #define __umulsidi3(u, v) \ 1286 ({UDItype __w; \ 1287 __asm__ ("meid %2,%0" \ 1288 : "=g" (__w) \ 1289 : "%0" ((USItype)(u)), "g" ((USItype)(v))); \ 1290 __w; }) 1291 #define udiv_qrnnd(q, r, n1, n0, d) \ 1292 ({union {UDItype __ll; \ 1293 struct {USItype __l, __h;} __i; \ 1294 } __x; \ 1295 __x.__i.__h = (n1); __x.__i.__l = (n0); \ 1296 __asm__ ("deid %2,%0" \ 1297 : "=g" (__x.__ll) \ 1298 : "0" (__x.__ll), "g" ((USItype)(d))); \ 1299 (r) = __x.__i.__l; (q) = __x.__i.__h; }) 1300 #define count_trailing_zeros(count,x) \ 1301 do { \ 1302 __asm__ ("ffsd %2,%0" \ 1303 : "=r" (count) \ 1304 : "0" ((USItype) 0), "r" ((USItype) (x))); \ 1305 } while (0) 1306 #endif /* __ns32000__ */ 1307 1308 /* In the past we had a block of various #defines tested 1309 _ARCH_PPC - AIX 1310 _ARCH_PWR - AIX 1311 __powerpc__ - gcc 1312 __POWERPC__ - BEOS 1313 __ppc__ - Darwin 1314 PPC - old gcc, GNU/Linux, SysV 1315 The plain PPC test was not good for vxWorks, since PPC is defined on all 1316 CPUs there (eg. m68k too), as a constant one is expected to compare 1317 CPU_FAMILY against. 1318 1319 At any rate, this was pretty unattractive and a bit fragile. The use of 1320 HAVE_HOST_CPU_FAMILY is designed to cut through it all and be sure of 1321 getting the desired effect. 1322 1323 ENHANCE-ME: We should test _IBMR2 here when we add assembly support for 1324 the system vendor compilers. (Is that vendor compilers with inline asm, 1325 or what?) */ 1326 1327 #if (HAVE_HOST_CPU_FAMILY_power || HAVE_HOST_CPU_FAMILY_powerpc) \ 1328 && W_TYPE_SIZE == 32 1329 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1330 do { \ 1331 if (__builtin_constant_p (bh) && (bh) == 0) \ 1332 __asm__ ("add%I4c %1,%3,%4\n\taddze %0,%2" \ 1333 : "=r" (sh), "=&r" (sl) : "r" (ah), "%r" (al), "rI" (bl)); \ 1334 else if (__builtin_constant_p (bh) && (bh) == ~(USItype) 0) \ 1335 __asm__ ("add%I4c %1,%3,%4\n\taddme %0,%2" \ 1336 : "=r" (sh), "=&r" (sl) : "r" (ah), "%r" (al), "rI" (bl)); \ 1337 else \ 1338 __asm__ ("add%I5c %1,%4,%5\n\tadde %0,%2,%3" \ 1339 : "=r" (sh), "=&r" (sl) \ 1340 : "r" (ah), "r" (bh), "%r" (al), "rI" (bl)); \ 1341 } while (0) 1342 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1343 do { \ 1344 if (__builtin_constant_p (ah) && (ah) == 0) \ 1345 __asm__ ("subf%I3c %1,%4,%3\n\tsubfze %0,%2" \ 1346 : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl));\ 1347 else if (__builtin_constant_p (ah) && (ah) == ~(USItype) 0) \ 1348 __asm__ ("subf%I3c %1,%4,%3\n\tsubfme %0,%2" \ 1349 : "=r" (sh), "=&r" (sl) : "r" (bh), "rI" (al), "r" (bl));\ 1350 else if (__builtin_constant_p (bh) && (bh) == 0) \ 1351 __asm__ ("subf%I3c %1,%4,%3\n\taddme %0,%2" \ 1352 : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "r" (bl));\ 1353 else if (__builtin_constant_p (bh) && (bh) == ~(USItype) 0) \ 1354 __asm__ ("subf%I3c %1,%4,%3\n\taddze %0,%2" \ 1355 : "=r" (sh), "=&r" (sl) : "r" (ah), "rI" (al), "r" (bl));\ 1356 else \ 1357 __asm__ ("subf%I4c %1,%5,%4\n\tsubfe %0,%3,%2" \ 1358 : "=r" (sh), "=&r" (sl) \ 1359 : "r" (ah), "r" (bh), "rI" (al), "r" (bl)); \ 1360 } while (0) 1361 #define count_leading_zeros(count, x) \ 1362 __asm__ ("cntlzw %0,%1" : "=r" (count) : "r" (x)) 1363 #define COUNT_LEADING_ZEROS_0 32 1364 #if HAVE_HOST_CPU_FAMILY_powerpc 1365 #if __GMP_GNUC_PREREQ (4,4) 1366 #define umul_ppmm(w1, w0, u, v) \ 1367 do { \ 1368 UDItype __ll = (UDItype)(u) * (v); \ 1369 w1 = __ll >> 32; \ 1370 w0 = __ll; \ 1371 } while (0) 1372 #endif 1373 #if !defined (umul_ppmm) 1374 #define umul_ppmm(ph, pl, m0, m1) \ 1375 do { \ 1376 USItype __m0 = (m0), __m1 = (m1); \ 1377 __asm__ ("mulhwu %0,%1,%2" : "=r" (ph) : "%r" (m0), "r" (m1)); \ 1378 (pl) = __m0 * __m1; \ 1379 } while (0) 1380 #endif 1381 #define UMUL_TIME 15 1382 #define smul_ppmm(ph, pl, m0, m1) \ 1383 do { \ 1384 SItype __m0 = (m0), __m1 = (m1); \ 1385 __asm__ ("mulhw %0,%1,%2" : "=r" (ph) : "%r" (m0), "r" (m1)); \ 1386 (pl) = __m0 * __m1; \ 1387 } while (0) 1388 #define SMUL_TIME 14 1389 #define UDIV_TIME 120 1390 #else 1391 #define UMUL_TIME 8 1392 #define smul_ppmm(xh, xl, m0, m1) \ 1393 __asm__ ("mul %0,%2,%3" : "=r" (xh), "=q" (xl) : "r" (m0), "r" (m1)) 1394 #define SMUL_TIME 4 1395 #define sdiv_qrnnd(q, r, nh, nl, d) \ 1396 __asm__ ("div %0,%2,%4" : "=r" (q), "=q" (r) : "r" (nh), "1" (nl), "r" (d)) 1397 #define UDIV_TIME 100 1398 #endif 1399 #endif /* 32-bit POWER architecture variants. */ 1400 1401 /* We should test _IBMR2 here when we add assembly support for the system 1402 vendor compilers. */ 1403 #if HAVE_HOST_CPU_FAMILY_powerpc && W_TYPE_SIZE == 64 1404 #if !defined (_LONG_LONG_LIMB) 1405 /* _LONG_LONG_LIMB is ABI=mode32 where adde operates on 32-bit values. So 1406 use adde etc only when not _LONG_LONG_LIMB. */ 1407 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1408 do { \ 1409 if (__builtin_constant_p (bh) && (bh) == 0) \ 1410 __asm__ ("add%I4c %1,%3,%4\n\taddze %0,%2" \ 1411 : "=r" (sh), "=&r" (sl) \ 1412 : "r" ((UDItype)(ah)), \ 1413 "%r" ((UDItype)(al)), "rI" ((UDItype)(bl))); \ 1414 else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0) \ 1415 __asm__ ("add%I4c %1,%3,%4\n\taddme %0,%2" \ 1416 : "=r" (sh), "=&r" (sl) \ 1417 : "r" ((UDItype)(ah)), \ 1418 "%r" ((UDItype)(al)), "rI" ((UDItype)(bl))); \ 1419 else \ 1420 __asm__ ("add%I5c %1,%4,%5\n\tadde %0,%2,%3" \ 1421 : "=r" (sh), "=&r" (sl) \ 1422 : "r" ((UDItype)(ah)), "r" ((UDItype)(bh)), \ 1423 "%r" ((UDItype)(al)), "rI" ((UDItype)(bl))); \ 1424 } while (0) 1425 /* We use "*rI" for the constant operand here, since with just "I", gcc barfs. 1426 This might seem strange, but gcc folds away the dead code late. */ 1427 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1428 do { \ 1429 if (__builtin_constant_p (bl) && bl > -0x8000 && bl <= 0x8000) { \ 1430 if (__builtin_constant_p (ah) && (ah) == 0) \ 1431 __asm__ ("addic %1,%3,%4\n\tsubfze %0,%2" \ 1432 : "=r" (sh), "=&r" (sl) \ 1433 : "r" ((UDItype)(bh)), \ 1434 "rI" ((UDItype)(al)), "*rI" (-((UDItype)(bl)))); \ 1435 else if (__builtin_constant_p (ah) && (ah) == ~(UDItype) 0) \ 1436 __asm__ ("addic %1,%3,%4\n\tsubfme %0,%2" \ 1437 : "=r" (sh), "=&r" (sl) \ 1438 : "r" ((UDItype)(bh)), \ 1439 "rI" ((UDItype)(al)), "*rI" (-((UDItype)(bl)))); \ 1440 else if (__builtin_constant_p (bh) && (bh) == 0) \ 1441 __asm__ ("addic %1,%3,%4\n\taddme %0,%2" \ 1442 : "=r" (sh), "=&r" (sl) \ 1443 : "r" ((UDItype)(ah)), \ 1444 "rI" ((UDItype)(al)), "*rI" (-((UDItype)(bl)))); \ 1445 else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0) \ 1446 __asm__ ("addic %1,%3,%4\n\taddze %0,%2" \ 1447 : "=r" (sh), "=&r" (sl) \ 1448 : "r" ((UDItype)(ah)), \ 1449 "rI" ((UDItype)(al)), "*rI" (-((UDItype)(bl)))); \ 1450 else \ 1451 __asm__ ("addic %1,%4,%5\n\tsubfe %0,%3,%2" \ 1452 : "=r" (sh), "=&r" (sl) \ 1453 : "r" ((UDItype)(ah)), "r" ((UDItype)(bh)), \ 1454 "rI" ((UDItype)(al)), "*rI" (-((UDItype)(bl)))); \ 1455 } else { \ 1456 if (__builtin_constant_p (ah) && (ah) == 0) \ 1457 __asm__ ("subf%I3c %1,%4,%3\n\tsubfze %0,%2" \ 1458 : "=r" (sh), "=&r" (sl) \ 1459 : "r" ((UDItype)(bh)), \ 1460 "rI" ((UDItype)(al)), "r" ((UDItype)(bl))); \ 1461 else if (__builtin_constant_p (ah) && (ah) == ~(UDItype) 0) \ 1462 __asm__ ("subf%I3c %1,%4,%3\n\tsubfme %0,%2" \ 1463 : "=r" (sh), "=&r" (sl) \ 1464 : "r" ((UDItype)(bh)), \ 1465 "rI" ((UDItype)(al)), "r" ((UDItype)(bl))); \ 1466 else if (__builtin_constant_p (bh) && (bh) == 0) \ 1467 __asm__ ("subf%I3c %1,%4,%3\n\taddme %0,%2" \ 1468 : "=r" (sh), "=&r" (sl) \ 1469 : "r" ((UDItype)(ah)), \ 1470 "rI" ((UDItype)(al)), "r" ((UDItype)(bl))); \ 1471 else if (__builtin_constant_p (bh) && (bh) == ~(UDItype) 0) \ 1472 __asm__ ("subf%I3c %1,%4,%3\n\taddze %0,%2" \ 1473 : "=r" (sh), "=&r" (sl) \ 1474 : "r" ((UDItype)(ah)), \ 1475 "rI" ((UDItype)(al)), "r" ((UDItype)(bl))); \ 1476 else \ 1477 __asm__ ("subf%I4c %1,%5,%4\n\tsubfe %0,%3,%2" \ 1478 : "=r" (sh), "=&r" (sl) \ 1479 : "r" ((UDItype)(ah)), "r" ((UDItype)(bh)), \ 1480 "rI" ((UDItype)(al)), "r" ((UDItype)(bl))); \ 1481 } \ 1482 } while (0) 1483 #endif /* ! _LONG_LONG_LIMB */ 1484 #define count_leading_zeros(count, x) \ 1485 __asm__ ("cntlzd %0,%1" : "=r" (count) : "r" (x)) 1486 #define COUNT_LEADING_ZEROS_0 64 1487 #if 0 && __GMP_GNUC_PREREQ (4,4) /* Disable, this results in libcalls! */ 1488 #define umul_ppmm(w1, w0, u, v) \ 1489 do { \ 1490 typedef unsigned int __ll_UTItype __attribute__((mode(TI))); \ 1491 __ll_UTItype __ll = (__ll_UTItype)(u) * (v); \ 1492 w1 = __ll >> 64; \ 1493 w0 = __ll; \ 1494 } while (0) 1495 #endif 1496 #if !defined (umul_ppmm) 1497 #define umul_ppmm(ph, pl, m0, m1) \ 1498 do { \ 1499 UDItype __m0 = (m0), __m1 = (m1); \ 1500 __asm__ ("mulhdu %0,%1,%2" : "=r" (ph) : "%r" (__m0), "r" (__m1)); \ 1501 (pl) = __m0 * __m1; \ 1502 } while (0) 1503 #endif 1504 #define UMUL_TIME 15 1505 #define smul_ppmm(ph, pl, m0, m1) \ 1506 do { \ 1507 DItype __m0 = (m0), __m1 = (m1); \ 1508 __asm__ ("mulhd %0,%1,%2" : "=r" (ph) : "%r" (__m0), "r" (__m1)); \ 1509 (pl) = __m0 * __m1; \ 1510 } while (0) 1511 #define SMUL_TIME 14 /* ??? */ 1512 #define UDIV_TIME 120 /* ??? */ 1513 #endif /* 64-bit PowerPC. */ 1514 1515 #if defined (__pyr__) && W_TYPE_SIZE == 32 1516 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1517 __asm__ ("addw %5,%1\n\taddwc %3,%0" \ 1518 : "=r" (sh), "=&r" (sl) \ 1519 : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 1520 "%1" ((USItype)(al)), "g" ((USItype)(bl))) 1521 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1522 __asm__ ("subw %5,%1\n\tsubwb %3,%0" \ 1523 : "=r" (sh), "=&r" (sl) \ 1524 : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 1525 "1" ((USItype)(al)), "g" ((USItype)(bl))) 1526 /* This insn works on Pyramids with AP, XP, or MI CPUs, but not with SP. */ 1527 #define umul_ppmm(w1, w0, u, v) \ 1528 ({union {UDItype __ll; \ 1529 struct {USItype __h, __l;} __i; \ 1530 } __x; \ 1531 __asm__ ("movw %1,%R0\n\tuemul %2,%0" \ 1532 : "=&r" (__x.__ll) \ 1533 : "g" ((USItype) (u)), "g" ((USItype)(v))); \ 1534 (w1) = __x.__i.__h; (w0) = __x.__i.__l;}) 1535 #endif /* __pyr__ */ 1536 1537 #if defined (__ibm032__) /* RT/ROMP */ && W_TYPE_SIZE == 32 1538 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1539 __asm__ ("a %1,%5\n\tae %0,%3" \ 1540 : "=r" (sh), "=&r" (sl) \ 1541 : "0" ((USItype)(ah)), "r" ((USItype)(bh)), \ 1542 "%1" ((USItype)(al)), "r" ((USItype)(bl))) 1543 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1544 __asm__ ("s %1,%5\n\tse %0,%3" \ 1545 : "=r" (sh), "=&r" (sl) \ 1546 : "0" ((USItype)(ah)), "r" ((USItype)(bh)), \ 1547 "1" ((USItype)(al)), "r" ((USItype)(bl))) 1548 #define smul_ppmm(ph, pl, m0, m1) \ 1549 __asm__ ( \ 1550 "s r2,r2\n" \ 1551 " mts r10,%2\n" \ 1552 " m r2,%3\n" \ 1553 " m r2,%3\n" \ 1554 " m r2,%3\n" \ 1555 " m r2,%3\n" \ 1556 " m r2,%3\n" \ 1557 " m r2,%3\n" \ 1558 " m r2,%3\n" \ 1559 " m r2,%3\n" \ 1560 " m r2,%3\n" \ 1561 " m r2,%3\n" \ 1562 " m r2,%3\n" \ 1563 " m r2,%3\n" \ 1564 " m r2,%3\n" \ 1565 " m r2,%3\n" \ 1566 " m r2,%3\n" \ 1567 " m r2,%3\n" \ 1568 " cas %0,r2,r0\n" \ 1569 " mfs r10,%1" \ 1570 : "=r" (ph), "=r" (pl) \ 1571 : "%r" ((USItype)(m0)), "r" ((USItype)(m1)) \ 1572 : "r2") 1573 #define UMUL_TIME 20 1574 #define UDIV_TIME 200 1575 #define count_leading_zeros(count, x) \ 1576 do { \ 1577 if ((x) >= 0x10000) \ 1578 __asm__ ("clz %0,%1" \ 1579 : "=r" (count) : "r" ((USItype)(x) >> 16)); \ 1580 else \ 1581 { \ 1582 __asm__ ("clz %0,%1" \ 1583 : "=r" (count) : "r" ((USItype)(x))); \ 1584 (count) += 16; \ 1585 } \ 1586 } while (0) 1587 #endif /* RT/ROMP */ 1588 1589 #if (defined (__SH2__) || defined (__SH3__) || defined (__SH4__)) && W_TYPE_SIZE == 32 1590 #define umul_ppmm(w1, w0, u, v) \ 1591 __asm__ ("dmulu.l %2,%3\n\tsts macl,%1\n\tsts mach,%0" \ 1592 : "=r" (w1), "=r" (w0) : "r" (u), "r" (v) : "macl", "mach") 1593 #define UMUL_TIME 5 1594 #endif 1595 1596 #if defined (__sparc__) && W_TYPE_SIZE == 32 1597 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1598 __asm__ ("addcc %r4,%5,%1\n\taddx %r2,%3,%0" \ 1599 : "=r" (sh), "=&r" (sl) \ 1600 : "rJ" (ah), "rI" (bh),"%rJ" (al), "rI" (bl) \ 1601 __CLOBBER_CC) 1602 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1603 __asm__ ("subcc %r4,%5,%1\n\tsubx %r2,%3,%0" \ 1604 : "=r" (sh), "=&r" (sl) \ 1605 : "rJ" (ah), "rI" (bh), "rJ" (al), "rI" (bl) \ 1606 __CLOBBER_CC) 1607 /* FIXME: When gcc -mcpu=v9 is used on solaris, gcc/config/sol2-sld-64.h 1608 doesn't define anything to indicate that to us, it only sets __sparcv8. */ 1609 #if defined (__sparc_v9__) || defined (__sparcv9) 1610 /* Perhaps we should use floating-point operations here? */ 1611 #if 0 1612 /* Triggers a bug making mpz/tests/t-gcd.c fail. 1613 Perhaps we simply need explicitly zero-extend the inputs? */ 1614 #define umul_ppmm(w1, w0, u, v) \ 1615 __asm__ ("mulx %2,%3,%%g1; srl %%g1,0,%1; srlx %%g1,32,%0" : \ 1616 "=r" (w1), "=r" (w0) : "r" (u), "r" (v) : "g1") 1617 #else 1618 /* Use v8 umul until above bug is fixed. */ 1619 #define umul_ppmm(w1, w0, u, v) \ 1620 __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v)) 1621 #endif 1622 /* Use a plain v8 divide for v9. */ 1623 #define udiv_qrnnd(q, r, n1, n0, d) \ 1624 do { \ 1625 USItype __q; \ 1626 __asm__ ("mov %1,%%y;nop;nop;nop;udiv %2,%3,%0" \ 1627 : "=r" (__q) : "r" (n1), "r" (n0), "r" (d)); \ 1628 (r) = (n0) - __q * (d); \ 1629 (q) = __q; \ 1630 } while (0) 1631 #else 1632 #if defined (__sparc_v8__) /* gcc normal */ \ 1633 || defined (__sparcv8) /* gcc solaris */ \ 1634 || HAVE_HOST_CPU_supersparc 1635 /* Don't match immediate range because, 1) it is not often useful, 1636 2) the 'I' flag thinks of the range as a 13 bit signed interval, 1637 while we want to match a 13 bit interval, sign extended to 32 bits, 1638 but INTERPRETED AS UNSIGNED. */ 1639 #define umul_ppmm(w1, w0, u, v) \ 1640 __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v)) 1641 #define UMUL_TIME 5 1642 1643 #if HAVE_HOST_CPU_supersparc 1644 #define UDIV_TIME 60 /* SuperSPARC timing */ 1645 #else 1646 /* Don't use this on SuperSPARC because its udiv only handles 53 bit 1647 dividends and will trap to the kernel for the rest. */ 1648 #define udiv_qrnnd(q, r, n1, n0, d) \ 1649 do { \ 1650 USItype __q; \ 1651 __asm__ ("mov %1,%%y;nop;nop;nop;udiv %2,%3,%0" \ 1652 : "=r" (__q) : "r" (n1), "r" (n0), "r" (d)); \ 1653 (r) = (n0) - __q * (d); \ 1654 (q) = __q; \ 1655 } while (0) 1656 #define UDIV_TIME 25 1657 #endif /* HAVE_HOST_CPU_supersparc */ 1658 1659 #else /* ! __sparc_v8__ */ 1660 #if defined (__sparclite__) 1661 /* This has hardware multiply but not divide. It also has two additional 1662 instructions scan (ffs from high bit) and divscc. */ 1663 #define umul_ppmm(w1, w0, u, v) \ 1664 __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v)) 1665 #define UMUL_TIME 5 1666 #define udiv_qrnnd(q, r, n1, n0, d) \ 1667 __asm__ ("! Inlined udiv_qrnnd\n" \ 1668 " wr %%g0,%2,%%y ! Not a delayed write for sparclite\n" \ 1669 " tst %%g0\n" \ 1670 " divscc %3,%4,%%g1\n" \ 1671 " divscc %%g1,%4,%%g1\n" \ 1672 " divscc %%g1,%4,%%g1\n" \ 1673 " divscc %%g1,%4,%%g1\n" \ 1674 " divscc %%g1,%4,%%g1\n" \ 1675 " divscc %%g1,%4,%%g1\n" \ 1676 " divscc %%g1,%4,%%g1\n" \ 1677 " divscc %%g1,%4,%%g1\n" \ 1678 " divscc %%g1,%4,%%g1\n" \ 1679 " divscc %%g1,%4,%%g1\n" \ 1680 " divscc %%g1,%4,%%g1\n" \ 1681 " divscc %%g1,%4,%%g1\n" \ 1682 " divscc %%g1,%4,%%g1\n" \ 1683 " divscc %%g1,%4,%%g1\n" \ 1684 " divscc %%g1,%4,%%g1\n" \ 1685 " divscc %%g1,%4,%%g1\n" \ 1686 " divscc %%g1,%4,%%g1\n" \ 1687 " divscc %%g1,%4,%%g1\n" \ 1688 " divscc %%g1,%4,%%g1\n" \ 1689 " divscc %%g1,%4,%%g1\n" \ 1690 " divscc %%g1,%4,%%g1\n" \ 1691 " divscc %%g1,%4,%%g1\n" \ 1692 " divscc %%g1,%4,%%g1\n" \ 1693 " divscc %%g1,%4,%%g1\n" \ 1694 " divscc %%g1,%4,%%g1\n" \ 1695 " divscc %%g1,%4,%%g1\n" \ 1696 " divscc %%g1,%4,%%g1\n" \ 1697 " divscc %%g1,%4,%%g1\n" \ 1698 " divscc %%g1,%4,%%g1\n" \ 1699 " divscc %%g1,%4,%%g1\n" \ 1700 " divscc %%g1,%4,%%g1\n" \ 1701 " divscc %%g1,%4,%0\n" \ 1702 " rd %%y,%1\n" \ 1703 " bl,a 1f\n" \ 1704 " add %1,%4,%1\n" \ 1705 "1: ! End of inline udiv_qrnnd" \ 1706 : "=r" (q), "=r" (r) : "r" (n1), "r" (n0), "rI" (d) \ 1707 : "%g1" __AND_CLOBBER_CC) 1708 #define UDIV_TIME 37 1709 #define count_leading_zeros(count, x) \ 1710 __asm__ ("scan %1,1,%0" : "=r" (count) : "r" (x)) 1711 /* Early sparclites return 63 for an argument of 0, but they warn that future 1712 implementations might change this. Therefore, leave COUNT_LEADING_ZEROS_0 1713 undefined. */ 1714 #endif /* __sparclite__ */ 1715 #endif /* __sparc_v8__ */ 1716 #endif /* __sparc_v9__ */ 1717 /* Default to sparc v7 versions of umul_ppmm and udiv_qrnnd. */ 1718 #ifndef umul_ppmm 1719 #define umul_ppmm(w1, w0, u, v) \ 1720 __asm__ ("! Inlined umul_ppmm\n" \ 1721 " wr %%g0,%2,%%y ! SPARC has 0-3 delay insn after a wr\n" \ 1722 " sra %3,31,%%g2 ! Don't move this insn\n" \ 1723 " and %2,%%g2,%%g2 ! Don't move this insn\n" \ 1724 " andcc %%g0,0,%%g1 ! Don't move this insn\n" \ 1725 " mulscc %%g1,%3,%%g1\n" \ 1726 " mulscc %%g1,%3,%%g1\n" \ 1727 " mulscc %%g1,%3,%%g1\n" \ 1728 " mulscc %%g1,%3,%%g1\n" \ 1729 " mulscc %%g1,%3,%%g1\n" \ 1730 " mulscc %%g1,%3,%%g1\n" \ 1731 " mulscc %%g1,%3,%%g1\n" \ 1732 " mulscc %%g1,%3,%%g1\n" \ 1733 " mulscc %%g1,%3,%%g1\n" \ 1734 " mulscc %%g1,%3,%%g1\n" \ 1735 " mulscc %%g1,%3,%%g1\n" \ 1736 " mulscc %%g1,%3,%%g1\n" \ 1737 " mulscc %%g1,%3,%%g1\n" \ 1738 " mulscc %%g1,%3,%%g1\n" \ 1739 " mulscc %%g1,%3,%%g1\n" \ 1740 " mulscc %%g1,%3,%%g1\n" \ 1741 " mulscc %%g1,%3,%%g1\n" \ 1742 " mulscc %%g1,%3,%%g1\n" \ 1743 " mulscc %%g1,%3,%%g1\n" \ 1744 " mulscc %%g1,%3,%%g1\n" \ 1745 " mulscc %%g1,%3,%%g1\n" \ 1746 " mulscc %%g1,%3,%%g1\n" \ 1747 " mulscc %%g1,%3,%%g1\n" \ 1748 " mulscc %%g1,%3,%%g1\n" \ 1749 " mulscc %%g1,%3,%%g1\n" \ 1750 " mulscc %%g1,%3,%%g1\n" \ 1751 " mulscc %%g1,%3,%%g1\n" \ 1752 " mulscc %%g1,%3,%%g1\n" \ 1753 " mulscc %%g1,%3,%%g1\n" \ 1754 " mulscc %%g1,%3,%%g1\n" \ 1755 " mulscc %%g1,%3,%%g1\n" \ 1756 " mulscc %%g1,%3,%%g1\n" \ 1757 " mulscc %%g1,0,%%g1\n" \ 1758 " add %%g1,%%g2,%0\n" \ 1759 " rd %%y,%1" \ 1760 : "=r" (w1), "=r" (w0) : "%rI" (u), "r" (v) \ 1761 : "%g1", "%g2" __AND_CLOBBER_CC) 1762 #define UMUL_TIME 39 /* 39 instructions */ 1763 #endif 1764 #ifndef udiv_qrnnd 1765 #ifndef LONGLONG_STANDALONE 1766 #define udiv_qrnnd(q, r, n1, n0, d) \ 1767 do { UWtype __r; \ 1768 (q) = __MPN(udiv_qrnnd) (&__r, (n1), (n0), (d)); \ 1769 (r) = __r; \ 1770 } while (0) 1771 extern UWtype __MPN(udiv_qrnnd) (UWtype *, UWtype, UWtype, UWtype); 1772 #ifndef UDIV_TIME 1773 #define UDIV_TIME 140 1774 #endif 1775 #endif /* LONGLONG_STANDALONE */ 1776 #endif /* udiv_qrnnd */ 1777 #endif /* __sparc__ */ 1778 1779 #if defined (__sparc__) && W_TYPE_SIZE == 64 1780 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1781 __asm__ ( \ 1782 "addcc %r4,%5,%1\n" \ 1783 " addccc %r6,%7,%%g0\n" \ 1784 " addc %r2,%3,%0" \ 1785 : "=r" (sh), "=&r" (sl) \ 1786 : "rJ" ((UDItype)(ah)), "rI" ((UDItype)(bh)), \ 1787 "%rJ" ((UDItype)(al)), "rI" ((UDItype)(bl)), \ 1788 "%rJ" ((UDItype)(al) >> 32), "rI" ((UDItype)(bl) >> 32) \ 1789 __CLOBBER_CC) 1790 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1791 __asm__ ( \ 1792 "subcc %r4,%5,%1\n" \ 1793 " subccc %r6,%7,%%g0\n" \ 1794 " subc %r2,%3,%0" \ 1795 : "=r" (sh), "=&r" (sl) \ 1796 : "rJ" ((UDItype)(ah)), "rI" ((UDItype)(bh)), \ 1797 "rJ" ((UDItype)(al)), "rI" ((UDItype)(bl)), \ 1798 "rJ" ((UDItype)(al) >> 32), "rI" ((UDItype)(bl) >> 32) \ 1799 __CLOBBER_CC) 1800 #if __VIS__ >= 0x300 1801 #undef add_ssaaaa 1802 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1803 __asm__ ( \ 1804 "addcc %r4, %5, %1\n" \ 1805 " addxc %r2, %r3, %0" \ 1806 : "=r" (sh), "=&r" (sl) \ 1807 : "rJ" ((UDItype)(ah)), "rJ" ((UDItype)(bh)), \ 1808 "%rJ" ((UDItype)(al)), "rI" ((UDItype)(bl)) __CLOBBER_CC) 1809 #define umul_ppmm(ph, pl, m0, m1) \ 1810 do { \ 1811 UDItype __m0 = (m0), __m1 = (m1); \ 1812 (pl) = __m0 * __m1; \ 1813 __asm__ ("umulxhi\t%2, %1, %0" \ 1814 : "=r" (ph) \ 1815 : "%r" (__m0), "r" (__m1)); \ 1816 } while (0) 1817 #define count_leading_zeros(count, x) \ 1818 __asm__ ("lzd\t%1,%0" : "=r" (count) : "r" (x)) 1819 /* Needed by count_leading_zeros_32 in sparc64.h. */ 1820 #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB 1821 #endif 1822 #endif 1823 1824 #if (defined (__vax) || defined (__vax__)) && W_TYPE_SIZE == 32 1825 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1826 __asm__ ("addl2 %5,%1\n\tadwc %3,%0" \ 1827 : "=g" (sh), "=&g" (sl) \ 1828 : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 1829 "%1" ((USItype)(al)), "g" ((USItype)(bl))) 1830 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1831 __asm__ ("subl2 %5,%1\n\tsbwc %3,%0" \ 1832 : "=g" (sh), "=&g" (sl) \ 1833 : "0" ((USItype)(ah)), "g" ((USItype)(bh)), \ 1834 "1" ((USItype)(al)), "g" ((USItype)(bl))) 1835 #define smul_ppmm(xh, xl, m0, m1) \ 1836 do { \ 1837 union {UDItype __ll; \ 1838 struct {USItype __l, __h;} __i; \ 1839 } __x; \ 1840 USItype __m0 = (m0), __m1 = (m1); \ 1841 __asm__ ("emul %1,%2,$0,%0" \ 1842 : "=g" (__x.__ll) : "g" (__m0), "g" (__m1)); \ 1843 (xh) = __x.__i.__h; (xl) = __x.__i.__l; \ 1844 } while (0) 1845 #define sdiv_qrnnd(q, r, n1, n0, d) \ 1846 do { \ 1847 union {DItype __ll; \ 1848 struct {SItype __l, __h;} __i; \ 1849 } __x; \ 1850 __x.__i.__h = n1; __x.__i.__l = n0; \ 1851 __asm__ ("ediv %3,%2,%0,%1" \ 1852 : "=g" (q), "=g" (r) : "g" (__x.__ll), "g" (d)); \ 1853 } while (0) 1854 #if 0 1855 /* FIXME: This instruction appears to be unimplemented on some systems (vax 1856 8800 maybe). */ 1857 #define count_trailing_zeros(count,x) \ 1858 do { \ 1859 __asm__ ("ffs 0, 31, %1, %0" \ 1860 : "=g" (count) \ 1861 : "g" ((USItype) (x))); \ 1862 } while (0) 1863 #endif 1864 #endif /* vax */ 1865 1866 #if defined (__z8000__) && W_TYPE_SIZE == 16 1867 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1868 __asm__ ("add %H1,%H5\n\tadc %H0,%H3" \ 1869 : "=r" (sh), "=&r" (sl) \ 1870 : "0" ((unsigned int)(ah)), "r" ((unsigned int)(bh)), \ 1871 "%1" ((unsigned int)(al)), "rQR" ((unsigned int)(bl))) 1872 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1873 __asm__ ("sub %H1,%H5\n\tsbc %H0,%H3" \ 1874 : "=r" (sh), "=&r" (sl) \ 1875 : "0" ((unsigned int)(ah)), "r" ((unsigned int)(bh)), \ 1876 "1" ((unsigned int)(al)), "rQR" ((unsigned int)(bl))) 1877 #define umul_ppmm(xh, xl, m0, m1) \ 1878 do { \ 1879 union {long int __ll; \ 1880 struct {unsigned int __h, __l;} __i; \ 1881 } __x; \ 1882 unsigned int __m0 = (m0), __m1 = (m1); \ 1883 __asm__ ("mult %S0,%H3" \ 1884 : "=r" (__x.__i.__h), "=r" (__x.__i.__l) \ 1885 : "%1" (m0), "rQR" (m1)); \ 1886 (xh) = __x.__i.__h; (xl) = __x.__i.__l; \ 1887 (xh) += ((((signed int) __m0 >> 15) & __m1) \ 1888 + (((signed int) __m1 >> 15) & __m0)); \ 1889 } while (0) 1890 #endif /* __z8000__ */ 1891 1892 #endif /* __GNUC__ */ 1893 1894 #endif /* NO_ASM */ 1895 1896 1897 /* FIXME: "sidi" here is highly doubtful, should sometimes be "diti". */ 1898 #if !defined (umul_ppmm) && defined (__umulsidi3) 1899 #define umul_ppmm(ph, pl, m0, m1) \ 1900 do { \ 1901 UDWtype __ll = __umulsidi3 (m0, m1); \ 1902 ph = (UWtype) (__ll >> W_TYPE_SIZE); \ 1903 pl = (UWtype) __ll; \ 1904 } while (0) 1905 #endif 1906 1907 #if !defined (__umulsidi3) 1908 #define __umulsidi3(u, v) \ 1909 ({UWtype __hi, __lo; \ 1910 umul_ppmm (__hi, __lo, u, v); \ 1911 ((UDWtype) __hi << W_TYPE_SIZE) | __lo; }) 1912 #endif 1913 1914 1915 #if defined (__cplusplus) 1916 #define __longlong_h_C "C" 1917 #else 1918 #define __longlong_h_C 1919 #endif 1920 1921 /* Use mpn_umul_ppmm or mpn_udiv_qrnnd functions, if they exist. The "_r" 1922 forms have "reversed" arguments, meaning the pointer is last, which 1923 sometimes allows better parameter passing, in particular on 64-bit 1924 hppa. */ 1925 1926 #define mpn_umul_ppmm __MPN(umul_ppmm) 1927 extern __longlong_h_C UWtype mpn_umul_ppmm (UWtype *, UWtype, UWtype); 1928 1929 #if ! defined (umul_ppmm) && HAVE_NATIVE_mpn_umul_ppmm \ 1930 && ! defined (LONGLONG_STANDALONE) 1931 #define umul_ppmm(wh, wl, u, v) \ 1932 do { \ 1933 UWtype __umul_ppmm__p0; \ 1934 (wh) = mpn_umul_ppmm (&__umul_ppmm__p0, (UWtype) (u), (UWtype) (v));\ 1935 (wl) = __umul_ppmm__p0; \ 1936 } while (0) 1937 #endif 1938 1939 #define mpn_umul_ppmm_r __MPN(umul_ppmm_r) 1940 extern __longlong_h_C UWtype mpn_umul_ppmm_r (UWtype, UWtype, UWtype *); 1941 1942 #if ! defined (umul_ppmm) && HAVE_NATIVE_mpn_umul_ppmm_r \ 1943 && ! defined (LONGLONG_STANDALONE) 1944 #define umul_ppmm(wh, wl, u, v) \ 1945 do { \ 1946 UWtype __umul_p0; \ 1947 (wh) = mpn_umul_ppmm_r ((UWtype) (u), (UWtype) (v), &__umul_p0); \ 1948 (wl) = __umul_p0; \ 1949 } while (0) 1950 #endif 1951 1952 #define mpn_udiv_qrnnd __MPN(udiv_qrnnd) 1953 extern __longlong_h_C UWtype mpn_udiv_qrnnd (UWtype *, UWtype, UWtype, UWtype); 1954 1955 #if ! defined (udiv_qrnnd) && HAVE_NATIVE_mpn_udiv_qrnnd \ 1956 && ! defined (LONGLONG_STANDALONE) 1957 #define udiv_qrnnd(q, r, n1, n0, d) \ 1958 do { \ 1959 UWtype __udiv_qrnnd_r; \ 1960 (q) = mpn_udiv_qrnnd (&__udiv_qrnnd_r, \ 1961 (UWtype) (n1), (UWtype) (n0), (UWtype) d); \ 1962 (r) = __udiv_qrnnd_r; \ 1963 } while (0) 1964 #endif 1965 1966 #define mpn_udiv_qrnnd_r __MPN(udiv_qrnnd_r) 1967 extern __longlong_h_C UWtype mpn_udiv_qrnnd_r (UWtype, UWtype, UWtype, UWtype *); 1968 1969 #if ! defined (udiv_qrnnd) && HAVE_NATIVE_mpn_udiv_qrnnd_r \ 1970 && ! defined (LONGLONG_STANDALONE) 1971 #define udiv_qrnnd(q, r, n1, n0, d) \ 1972 do { \ 1973 UWtype __udiv_qrnnd_r; \ 1974 (q) = mpn_udiv_qrnnd_r ((UWtype) (n1), (UWtype) (n0), (UWtype) d, \ 1975 &__udiv_qrnnd_r); \ 1976 (r) = __udiv_qrnnd_r; \ 1977 } while (0) 1978 #endif 1979 1980 1981 /* If this machine has no inline assembler, use C macros. */ 1982 1983 #if !defined (add_ssaaaa) 1984 #define add_ssaaaa(sh, sl, ah, al, bh, bl) \ 1985 do { \ 1986 UWtype __x; \ 1987 __x = (al) + (bl); \ 1988 (sh) = (ah) + (bh) + (__x < (al)); \ 1989 (sl) = __x; \ 1990 } while (0) 1991 #endif 1992 1993 #if !defined (sub_ddmmss) 1994 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \ 1995 do { \ 1996 UWtype __x; \ 1997 __x = (al) - (bl); \ 1998 (sh) = (ah) - (bh) - ((al) < (bl)); \ 1999 (sl) = __x; \ 2000 } while (0) 2001 #endif 2002 2003 /* If we lack umul_ppmm but have smul_ppmm, define umul_ppmm in terms of 2004 smul_ppmm. */ 2005 #if !defined (umul_ppmm) && defined (smul_ppmm) 2006 #define umul_ppmm(w1, w0, u, v) \ 2007 do { \ 2008 UWtype __w1; \ 2009 UWtype __xm0 = (u), __xm1 = (v); \ 2010 smul_ppmm (__w1, w0, __xm0, __xm1); \ 2011 (w1) = __w1 + (-(__xm0 >> (W_TYPE_SIZE - 1)) & __xm1) \ 2012 + (-(__xm1 >> (W_TYPE_SIZE - 1)) & __xm0); \ 2013 } while (0) 2014 #endif 2015 2016 /* If we still don't have umul_ppmm, define it using plain C. 2017 2018 For reference, when this code is used for squaring (ie. u and v identical 2019 expressions), gcc recognises __x1 and __x2 are the same and generates 3 2020 multiplies, not 4. The subsequent additions could be optimized a bit, 2021 but the only place GMP currently uses such a square is mpn_sqr_basecase, 2022 and chips obliged to use this generic C umul will have plenty of worse 2023 performance problems than a couple of extra instructions on the diagonal 2024 of sqr_basecase. */ 2025 2026 #if !defined (umul_ppmm) 2027 #define umul_ppmm(w1, w0, u, v) \ 2028 do { \ 2029 UWtype __x0, __x1, __x2, __x3; \ 2030 UHWtype __ul, __vl, __uh, __vh; \ 2031 UWtype __u = (u), __v = (v); \ 2032 \ 2033 __ul = __ll_lowpart (__u); \ 2034 __uh = __ll_highpart (__u); \ 2035 __vl = __ll_lowpart (__v); \ 2036 __vh = __ll_highpart (__v); \ 2037 \ 2038 __x0 = (UWtype) __ul * __vl; \ 2039 __x1 = (UWtype) __ul * __vh; \ 2040 __x2 = (UWtype) __uh * __vl; \ 2041 __x3 = (UWtype) __uh * __vh; \ 2042 \ 2043 __x1 += __ll_highpart (__x0);/* this can't give carry */ \ 2044 __x1 += __x2; /* but this indeed can */ \ 2045 if (__x1 < __x2) /* did we get it? */ \ 2046 __x3 += __ll_B; /* yes, add it in the proper pos. */ \ 2047 \ 2048 (w1) = __x3 + __ll_highpart (__x1); \ 2049 (w0) = (__x1 << W_TYPE_SIZE/2) + __ll_lowpart (__x0); \ 2050 } while (0) 2051 #endif 2052 2053 /* If we don't have smul_ppmm, define it using umul_ppmm (which surely will 2054 exist in one form or another. */ 2055 #if !defined (smul_ppmm) 2056 #define smul_ppmm(w1, w0, u, v) \ 2057 do { \ 2058 UWtype __w1; \ 2059 UWtype __xm0 = (u), __xm1 = (v); \ 2060 umul_ppmm (__w1, w0, __xm0, __xm1); \ 2061 (w1) = __w1 - (-(__xm0 >> (W_TYPE_SIZE - 1)) & __xm1) \ 2062 - (-(__xm1 >> (W_TYPE_SIZE - 1)) & __xm0); \ 2063 } while (0) 2064 #endif 2065 2066 /* Define this unconditionally, so it can be used for debugging. */ 2067 #define __udiv_qrnnd_c(q, r, n1, n0, d) \ 2068 do { \ 2069 UWtype __d1, __d0, __q1, __q0, __r1, __r0, __m; \ 2070 \ 2071 ASSERT ((d) != 0); \ 2072 ASSERT ((n1) < (d)); \ 2073 \ 2074 __d1 = __ll_highpart (d); \ 2075 __d0 = __ll_lowpart (d); \ 2076 \ 2077 __q1 = (n1) / __d1; \ 2078 __r1 = (n1) - __q1 * __d1; \ 2079 __m = __q1 * __d0; \ 2080 __r1 = __r1 * __ll_B | __ll_highpart (n0); \ 2081 if (__r1 < __m) \ 2082 { \ 2083 __q1--, __r1 += (d); \ 2084 if (__r1 >= (d)) /* i.e. we didn't get carry when adding to __r1 */\ 2085 if (__r1 < __m) \ 2086 __q1--, __r1 += (d); \ 2087 } \ 2088 __r1 -= __m; \ 2089 \ 2090 __q0 = __r1 / __d1; \ 2091 __r0 = __r1 - __q0 * __d1; \ 2092 __m = __q0 * __d0; \ 2093 __r0 = __r0 * __ll_B | __ll_lowpart (n0); \ 2094 if (__r0 < __m) \ 2095 { \ 2096 __q0--, __r0 += (d); \ 2097 if (__r0 >= (d)) \ 2098 if (__r0 < __m) \ 2099 __q0--, __r0 += (d); \ 2100 } \ 2101 __r0 -= __m; \ 2102 \ 2103 (q) = __q1 * __ll_B | __q0; \ 2104 (r) = __r0; \ 2105 } while (0) 2106 2107 /* If the processor has no udiv_qrnnd but sdiv_qrnnd, go through 2108 __udiv_w_sdiv (defined in libgcc or elsewhere). */ 2109 #if !defined (udiv_qrnnd) && defined (sdiv_qrnnd) \ 2110 && ! defined (LONGLONG_STANDALONE) 2111 #define udiv_qrnnd(q, r, nh, nl, d) \ 2112 do { \ 2113 UWtype __r; \ 2114 (q) = __MPN(udiv_w_sdiv) (&__r, nh, nl, d); \ 2115 (r) = __r; \ 2116 } while (0) 2117 __GMP_DECLSPEC UWtype __MPN(udiv_w_sdiv) (UWtype *, UWtype, UWtype, UWtype); 2118 #endif 2119 2120 /* If udiv_qrnnd was not defined for this processor, use __udiv_qrnnd_c. */ 2121 #if !defined (udiv_qrnnd) 2122 #define UDIV_NEEDS_NORMALIZATION 1 2123 #define udiv_qrnnd __udiv_qrnnd_c 2124 #endif 2125 2126 #if !defined (count_leading_zeros) 2127 #define count_leading_zeros(count, x) \ 2128 do { \ 2129 UWtype __xr = (x); \ 2130 UWtype __a; \ 2131 \ 2132 if (W_TYPE_SIZE == 32) \ 2133 { \ 2134 __a = __xr < ((UWtype) 1 << 2*__BITS4) \ 2135 ? (__xr < ((UWtype) 1 << __BITS4) ? 1 : __BITS4 + 1) \ 2136 : (__xr < ((UWtype) 1 << 3*__BITS4) ? 2*__BITS4 + 1 \ 2137 : 3*__BITS4 + 1); \ 2138 } \ 2139 else \ 2140 { \ 2141 for (__a = W_TYPE_SIZE - 8; __a > 0; __a -= 8) \ 2142 if (((__xr >> __a) & 0xff) != 0) \ 2143 break; \ 2144 ++__a; \ 2145 } \ 2146 \ 2147 (count) = W_TYPE_SIZE + 1 - __a - __clz_tab[__xr >> __a]; \ 2148 } while (0) 2149 /* This version gives a well-defined value for zero. */ 2150 #define COUNT_LEADING_ZEROS_0 (W_TYPE_SIZE - 1) 2151 #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB 2152 #define COUNT_LEADING_ZEROS_SLOW 2153 #endif 2154 2155 /* clz_tab needed by mpn/x86/pentium/mod_1.asm in a fat binary */ 2156 #if HAVE_HOST_CPU_FAMILY_x86 && WANT_FAT_BINARY 2157 #define COUNT_LEADING_ZEROS_NEED_CLZ_TAB 2158 #endif 2159 2160 #ifdef COUNT_LEADING_ZEROS_NEED_CLZ_TAB 2161 extern const unsigned char __GMP_DECLSPEC __clz_tab[129]; 2162 #endif 2163 2164 #if !defined (count_trailing_zeros) 2165 #if !defined (COUNT_LEADING_ZEROS_SLOW) 2166 /* Define count_trailing_zeros using an asm count_leading_zeros. */ 2167 #define count_trailing_zeros(count, x) \ 2168 do { \ 2169 UWtype __ctz_x = (x); \ 2170 UWtype __ctz_c; \ 2171 ASSERT (__ctz_x != 0); \ 2172 count_leading_zeros (__ctz_c, __ctz_x & -__ctz_x); \ 2173 (count) = W_TYPE_SIZE - 1 - __ctz_c; \ 2174 } while (0) 2175 #else 2176 /* Define count_trailing_zeros in plain C, assuming small counts are common. 2177 We use clz_tab without ado, since the C count_leading_zeros above will have 2178 pulled it in. */ 2179 #define count_trailing_zeros(count, x) \ 2180 do { \ 2181 UWtype __ctz_x = (x); \ 2182 int __ctz_c; \ 2183 \ 2184 if (LIKELY ((__ctz_x & 0xff) != 0)) \ 2185 (count) = __clz_tab[__ctz_x & -__ctz_x] - 2; \ 2186 else \ 2187 { \ 2188 for (__ctz_c = 8 - 2; __ctz_c < W_TYPE_SIZE - 2; __ctz_c += 8) \ 2189 { \ 2190 __ctz_x >>= 8; \ 2191 if (LIKELY ((__ctz_x & 0xff) != 0)) \ 2192 break; \ 2193 } \ 2194 \ 2195 (count) = __ctz_c + __clz_tab[__ctz_x & -__ctz_x]; \ 2196 } \ 2197 } while (0) 2198 #endif 2199 #endif 2200 2201 #ifndef UDIV_NEEDS_NORMALIZATION 2202 #define UDIV_NEEDS_NORMALIZATION 0 2203 #endif 2204 2205 /* Whether udiv_qrnnd is actually implemented with udiv_qrnnd_preinv, and 2206 that hence the latter should always be used. */ 2207 #ifndef UDIV_PREINV_ALWAYS 2208 #define UDIV_PREINV_ALWAYS 0 2209 #endif 2210 2211 /* Give defaults for UMUL_TIME and UDIV_TIME. */ 2212 #ifndef UMUL_TIME 2213 #define UMUL_TIME 1 2214 #endif 2215 2216 #ifndef UDIV_TIME 2217 #define UDIV_TIME UMUL_TIME 2218 #endif