Home | History | Annotate | Download | only in xtensa

Lines Matching refs:a6

68 	bnall	a3, a6, .Ladd_return_nan_or_inf
83 movi a6, 0x400000 /* make it a quiet NaN */
84 or a2, a2, a6
89 slli a7, a6, 8
98 movi a6, 0x7f800000
106 ball a2, a6, .Ladd_xnan_or_inf
107 ball a3, a6, .Ladd_ynan_or_inf
118 bnone a3, a6, .Ladd_yexpzero
121 or a3, a3, a6
155 bnone a2, a6, .Ladd_bothexpzero
179 bnone a2, a6, .Ladd_xexpzero
181 or a2, a2, a6
242 ball a2, a6, .Ladd_infinity
277 bnall a3, a6, .Lsub_return_nan_or_inf
287 slli a7, a6, 8
297 slli a7, a6, 8
306 movi a6, 0x7f800000
314 ball a2, a6, .Lsub_xnan_or_inf
315 ball a3, a6, .Lsub_ynan_or_inf
325 bnone a3, a6, .Lsub_yexpzero
328 or a3, a3, a6
371 bnone a2, a6, .Lsub_yexpdiff
377 slli a7, a6, 8
384 bnone a2, a6, .Lsub_xexpzero
386 or a2, a2, a6
400 slli a11, a6, 8
428 bnone a3, a6, .Lsub_xexpdiff
445 do_nsau a6, a8, a7, a11
447 bge a6, a10, .Lsub_subnormal
448 addi a6, a6, 1
451 /* Shift the mantissa (a8/a9) left by a6. */
452 ssl a6
457 decrementing the exponent by a6. (The exponent has already
461 sub a2, a2, a6
480 movi a6, 24
481 movltz a6, a10, a11
486 Set a6 to the current exponent (which has already been
488 will be zero. Do not add 1 to a6 in this case, because: (1)
493 mov a6, a10
557 bnall a3, a6, .Lmul_returnx
603 movi a6, 0x7f800000
609 ball a2, a6, .Lmul_xnan_or_inf
610 ball a3, a6, .Lmul_ynan_or_inf
626 or a2, a2, a6
628 or a3, a3, a6
631 /* Multiply 32x32 to 64 bits. The result ends up in a2/a6. */
635 mull a6, a2, a3
734 /* Add pp1 and pp2 into a6 with carry-out in a9. */
735 do_mul(a6, a2, l, a3, h) /* pp 1 */
738 add a6, a6, a11
739 bgeu a6, a11, 1f
742 /* Shift the high half of a9/a6 into position in a9. Note that
745 src a9, a9, a6
747 /* Compute the low word into a6. */
749 sll a6, a6
750 add a6, a6, a11
751 bgeu a6, a11, 1f
774 src a2, a2, a6
775 sll a6, a6
790 bgez a6, .Lmul_rounded
792 slli a6, a6, 1
793 beqz a6, .Lmul_exactlyhalf
834 mov a9, a6
839 in a6 (combined with the shifted-out bits currently in a6) for
841 sll a6, a2
847 /* Pack any nonzero bits shifted out into a6. */
850 or a6, a6, a9
896 mul_mulsi3_body a2, a4, a3, a5, a6
979 or a2, a2, a6
1014 ball a3, a6, .Ldiv_return_nan
1036 movi a6, 0x7f800000
1042 ball a2, a6, .Ldiv_xnan_or_inf
1043 ball a3, a6, .Ldiv_ynan_or_inf
1059 or a2, a2, a6
1061 or a3, a3, a6
1152 saved in a6 for rounding the result. */
1153 sll a6, a10
1159 /* Pack any nonzero remainder (in a2) into a6. */
1162 or a6, a6, a9
1164 /* Round a10 based on the bits shifted out into a6. */
1165 1: bgez a6, .Ldiv_rounded
1167 slli a6, a6, 1
1168 bnez a6, .Ldiv_rounded
1197 movi a6, 0x7f800000
1198 ball a2, a6, 3f
1231 movi a6, 0x7f800000
1232 ball a2, a6, 2f
1233 1: bnall a3, a6, .Lle_cmp
1255 movi a6, 0x7f800000
1256 ball a2, a6, 2f
1257 1: bnall a3, a6, .Lle_cmp
1309 movi a6, 0x7f800000
1310 ball a2, a6, 2f
1311 1: bnall a3, a6, .Llt_cmp
1333 movi a6, 0x7f800000
1334 ball a2, a6, 2f
1335 1: bnall a3, a6, .Llt_cmp
1387 movi a6, 0x7f800000
1388 ball a2, a6, 3f
1389 1: ball a3, a6, 4f
1414 movi a6, 0x7f800000
1415 ball a2, a6, .Lfixsfsi_nan_or_inf
1424 or a7, a2, a6
1445 slli a4, a6, 8 /* 0x80000000 */
1466 movi a6, 0x7f800000
1467 ball a2, a6, .Lfixsfdi_nan_or_inf
1476 or a7, a2, a6
1509 slli a7, a6, 8 /* 0x80000000 */
1535 movi a6, 0x7f800000
1536 ball a2, a6, .Lfixunssfsi_nan_or_inf
1545 or a7, a2, a6
1569 slli a4, a6, 8 /* 0x80000000 */
1586 1: slli a2, a6, 8
1600 movi a6, 0x7f800000
1601 ball a2, a6, .Lfixunssfdi_nan_or_inf
1610 or a7, a2, a6
1648 2: slli xh, a6, 8 /* 0x80000000 */
1700 do_nsau a4, a2, a5, a6
1704 /* Shift the mantissa into position, with rounding bits in a6. */
1706 slli a6, a5, (32 - 8)
1719 bgez a6, .Lfloatsisf_return
1723 slli a6, a6, 1
1724 beqz a6, .Lfloatsisf_exactlyhalf
1776 do_nsau a4, xh, a5, a6
1782 /* Shift the mantissa into position, with rounding bits in a6. */
1785 src a6, xh, xl
1789 or a6, a6, a5
1802 bgez a6, 2f
1806 slli a6, a6, 1
1807 beqz a6, .Lfloatdisf_exactlyhalf
1812 do_nsau a4, xl, a5, a6