Home | History | Annotate | Download | only in xtensa

Lines Matching defs:xl

28 #define xl a3
33 #define xl a2
71 or a7, a7, xl
77 mov xl, yl
81 or a7, a7, xl
143 add xl, xl, yl
145 bgeu xl, yl, 1f
170 add xl, xl, yl
172 bgeu xl, yl, 1f
214 sll a9, xl
215 src xl, xh, xl
219 add xl, xl, yl
221 bgeu xl, yl, 1f
231 addi xl, xl, 1
232 beqz xl, .Ladd_roundcarry
244 sll a11, xl
245 src a9, xh, xl
246 srl xl, xh
254 mov xl, yl
276 /* Shift xh/xl right by one bit. Save the lsb of xl. */
277 mov a10, xl
279 src xl, xh, xl
293 addi xl, xl, 1
294 beqz xl, .Ladd_roundcarry
300 movi xl, 0
311 srli xl, xl, 1
312 slli xl, xl, 1
316 /* xl is always zero when the rounding increment overflows, so
344 mov xl, yl
348 or a7, a7, xl
406 bgeu xl, yl, 1f
408 1: sub xl, xl, yl
411 borrow from xh/xl. */
415 moveqz xh, a5, xl
416 addi xl, xl, -1
425 bltu xl, yl, .Lsub_xsmaller
426 bltu yl, xl, .Lsub_ysmaller
429 movi xl, 0
473 src a9, xl, a9
474 src xl, xh, xl
482 sub xl, yl, xl
484 bgeu yl, xl, 1f
488 borrow from xh/xl. */
492 moveqz xh, a5, xl
493 addi xl, xl, -1
502 addi xl, xl, 1
503 beqz xl, .Lsub_roundcarry
525 sll a11, xl
526 src a9, xh, xl
527 srl xl, xh
537 mov xl, yl
555 /* Shift the mantissa (a8/xl/a9) left by a6. */
557 src a8, a8, xl
558 src xl, xl, a9
573 srli xl, xl, 1
574 slli xl, xl, 1
578 /* xl is always zero when the rounding increment overflows, so
586 word are zero. Shift by "20 + (leading zero count of xl) + 1". */
587 do_nsau a6, xl, a7, a11
595 src a8, xl, a9
596 sll xl, a9
639 or a10, xh, xl
647 src xh, xh, xl
648 sll xl, xl
653 do_nsau a10, xl, a11, a12
659 sll xh, xl
660 movi xl, 0
663 srl xh, xl
664 sll xl, xl
705 movi xl, 0
722 mov xl, yl
726 or a8, a8, xl
737 bnez xl, .Lmul_returny
793 /* Multiply 64x64 to 128 bits. The result ends up in xh/xl/a6.
800 mull a6, xl, yh
806 muluh a11, xl, yl
812 mull a11, xl, yl
817 /* Compute xl with any carry-outs in a9. */
829 muluh xl, xl, yh
830 add xl, xl, a10
831 bgeu xl, a10, 1f
892 srli xlh, xl, 16
897 #define xll xl
905 extui xl, xl, 0, 16
961 do_mul(a10, xl, l, yl, h) /* pp 1 */
962 do_mul(a11, xl, h, yl, l) /* pp 2 */
974 do_mul(a11, xl, l, yl, l) /* pp 0 */
985 do_mul(a11, xl, l, yh, l) /* pp 3 */
990 do_mul(a11, xl, h, yl, h) /* pp 4 */
1006 do_mul(a7, xl, l, yh, h) /* pp 6 */
1013 do_mul(a7, xl, h, yh, l) /* pp 7 */
1031 /* Add pp10-12 into xl with carry-outs in a9. */
1033 do_mul(xl, xl, h, yh, h) /* pp 10 */
1034 add xl, xl, a10
1035 bgeu xl, a10, 1f
1039 add xl, xl, a10
1040 bgeu xl, a10, 1f
1044 add xl, xl, a10
1045 bgeu xl, a10, 1f
1060 add xl, xl, a11
1061 bgeu xl, a11, 1f
1089 src xh, xh, xl
1090 src xl, xl, a6
1107 addi xl, xl, 1
1108 beqz xl, .Lmul_roundcarry
1135 srli xl, xl, 1
1136 slli xl, xl, 1
1140 /* xl is always zero when the rounding increment overflows, so
1151 movi xl, 0
1163 /* Shift xh/xl right. Any bits that are shifted out of xl are saved
1166 sll a6, xl
1167 src xl, xh, xl
1173 sll a10, xl /* lost bits shifted out of xl */
1174 src a6, xh, xl
1175 srl xl, xh
1192 movi xl, 0
1251 wfrd f1, xh, xl
1283 rfr xl, f0
1336 or xl, xl, xh
1340 bnez xl, 1f
1343 1: movi xl, 0
1352 or a10, xh, xl
1360 src xh, xh, xl
1361 sll xl, xl
1366 do_nsau a10, xl, a11, a8
1372 sll xh, xl
1373 movi xl, 0
1376 srl xh, xl
1377 sll xl, xl
1384 movi xl, 0
1395 or a8, a8, xl
1406 mov xl, yl
1414 bltu xl, yl, 2f
1457 2: src xh, xh, xl
1458 sll xl, xl
1463 bgeu xl, yl, 1f
1465 1: sub xl, xl, yl
1466 src xh, xh, xl
1467 sll xl, xl
1490 bgeu xl, yl, 1f
1492 1: sub xl, xl, yl
1495 3: src xh, xh, xl
1496 sll xl, xl
1516 /* Round. The remainder (<< 1) is in xh/xl. */
1524 mov xl, a11
1537 bgeu xl, yl, 2b
1541 bltu xl, yl, .Ldiv_rounded
1542 bne xl, yl, .Ldiv_roundup
1556 movi xl, 0
1580 or xl, xl, a9
1585 /* Pack any nonzero remainder (in xh/xl) into a6. */
1586 or xh, xh, xl
1612 movi xl, 0
1630 bne xl, yl, 2f
1647 or a7, a7, xl
1653 or a7, a7, xl /* xl == yl here */
1683 or a7, a7, xl
1709 or a7, a7, xl
1725 bltu yl, xl, 5f
1733 bgeu xl, yl, 4b
1743 or a7, a7, xl
1771 or a7, a7, xl
1797 or a7, a7, xl
1813 bgeu xl, yl, 5f
1821 bltu yl, xl, 4b
1831 or a7, a7, xl
1853 or a7, a7, xl
1888 src a5, a7, xl
1902 or a4, a4, xl
1943 src xh, a7, xl
1944 sll xl, xl
1949 srl xl, xh
1955 neg xl, xl
1957 beqz xl, 1f
1962 src xl, xh, xl
1969 or a4, a4, xl
1979 movi xl, 0
1983 movi xl, -1
1988 movi xl, 0
2015 src a5, a7, xl
2031 or a4, a4, xl
2083 src xh, a7, xl
2084 sll xl, xl
2091 srl xl, xh
2097 neg xl, xl
2099 beqz xl, 1f
2104 src xl, xh, xl
2111 or a4, a4, xl
2116 movi xl, -1
2122 movi xl, 0
2127 movi xl, 0
2178 slli xl, a5, (32 - 11)
2206 or a4, xh, xl
2220 or a4, xh, xl
2228 neg xl, xl
2230 beqz xl, .Lfloatdidf_normalize
2238 src xh, xh, xl
2239 sll xl, xl
2244 sll a6, xl
2245 src xl, xh, xl
2260 addi xl, xl, 1
2261 beqz xl, .Lfloatdidf_roundcarry
2269 /* xh is zero. Normalize with first 1 bit of xl in the msb of xh. */
2270 do_nsau a4, xl, a5, a6
2272 sll xh, xl
2273 movi xl, 0
2279 srli xl, xl, 1
2280 slli xl, xl, 1
2284 /* xl is always zero when the rounding increment overflows, so
2314 xl << 3 into a5/a4. */
2316 src a5, a5, xl
2317 sll a4, xl
2348 or a5, a5, xl
2377 src a5, a5, xl
2378 sll a4, xl
2415 /* Shift >> 3 into a4/xl. */
2417 slli xl, a2, (32 - 3)
2439 movi xl, 0
2451 /* Shift >> 3 into a4/xl. */
2452 slli xl, a4, (32 - 3)
2478 wfrd f1, xh, xl
2515 rfr xl, f0
2533 wfrd f1, xh, xl
2547 rfr xl, f0
2565 wfrd f1, xh, xl
2585 rfr xl, f0