| /src/common/lib/libc/arch/arm/atomic/ |
| H A D | atomic_cas_64.S | 60 strexd r6, r4, r5, [ip] /* store new value */ 61 cmp r6, #0 /* succeed? */
|
| H A D | sync_bool_compare_and_swap_8.S | 52 ldrexd r6, r7, [ip] /* load current value */ 53 cmp r6, r2 /* compare to old? 1st half */
|
| H A D | sync_fetch_and_add_8.S | 47 strexd r6, r4, r5, [ip] /* try to store */ 48 cmp r6, #0 /* succeed? */
|
| H A D | sync_fetch_and_and_8.S | 47 strexd r6, r4, r5, [ip] /* try to store */ 48 cmp r6, #0 /* succeed? */
|
| H A D | sync_fetch_and_nand_8.S | 49 strexd r6, r4, r5, [ip] /* try to store */ 50 cmp r6, #0 /* succeed? */
|
| H A D | sync_fetch_and_or_8.S | 47 strexd r6, r4, r5, [ip] /* try to store */ 48 cmp r6, #0 /* succeed? */
|
| H A D | sync_fetch_and_sub_8.S | 47 strexd r6, r4, r5, [ip] /* try to store */ 48 cmp r6, #0 /* succeed? */
|
| H A D | sync_fetch_and_xor_8.S | 47 strexd r6, r4, r5, [ip] /* try to store */ 48 cmp r6, #0 /* succeed? */
|
| /src/common/lib/libc/arch/arm/quad/ |
| H A D | __aeabi_ldivmod.S | 69 push {r4-r6, lr} 139 pop {r2-r6, pc} /* grab b from stack */ 148 pop {r4-r6, pc} 200 pop {r4-r6, pc} 210 pop {r4-r6, pc}
|
| /src/common/lib/libc/arch/arm/string/ |
| H A D | memcpy_xscale.S | 82 ldr r6, [r1], #0x04 /* LD:08-0b */ 89 strd r6, r7, [r3], #0x08 /* ST:08-0f */ 90 ldr r6, [r1], #0x04 /* LD:20-23 */ 99 strd r6, r7, [r3], #0x08 /* ST:20-27 */ 100 ldr r6, [r1], #0x04 /* LD:38-3b */ 109 strd r6, r7, [r3], #0x08 /* ST:38-3f */ 110 ldr r6, [r1], #0x04 /* LD:50-53 */ 119 strd r6, r7, [r3], #0x08 /* ST:50-57 */ 120 ldr r6, [r1], #0x04 /* LD:68-6b */ 128 strd r6, r [all...] |
| H A D | strcpy_arm.S | 118 add r6, r1, #1 /* save for return (deal with NUL) */ 120 mov r6, r0 /* save for return */ 148 add r6, r6, #3 /* compensate for word post-inc */ 218 add r6, r6, #3 /* compensate for word post-inc */ 467 2: add r6, r6, #3 /* compensate for post-inc */ 499 sub r0, r1, r6 /* subtract start from finish */ 517 3: mov r0, r6 /* restor [all...] |
| H A D | strlcat_arm.S | 42 push {r4-r6, lr} 45 .save {r4-r6, lr} 55 mov r6, r2 /* save siz */ 57 mov r1, r6 /* get siz */ 60 subs r2, r6, r0 /* get remaining space in dst */ 66 pop {r4-r6, pc} /* restore registers and return */
|
| /src/common/lib/libc/arch/or1k/gen/ |
| H A D | mulsi3.S | 49 l.ori r6, r3, 0 # move r3 to r6 50 l.bf .Lloop # yes, r6 already has smaller value, 52 l.ori r6, r4, 0 # move r4 to r6 because it's smaller than r3 54 l.cmov r6, r3, r4 # choose lesser of r3 or r4 58 l.andi r5, r6, 1 # get LSB 60 l.and r8, r8, r4 # r8 = r6 & 1 ? r4 : 0 63 l.slli r6, r6, [all...] |
| /src/common/lib/libc/arch/or1k/string/ |
| H A D | memcmp.S | 74 l.xor r6, r3, r4 /* check alignment compatibility */ 75 l.andi r6, r6, 3 /* only care about the two bits */ 76 l.sfeqi r6, 0 /* same alignment? */ 139 l.sfeqi r6, 2 /* check for halfword alignment */
|
| H A D | memmove.S | 58 l.or r6, r3, r0 /* swap src/dst */ 60 l.or r4, r6, r0 89 l.sw 4(r3), r6 /* Store previous word */ 100 l.lwz r6, 4(r4) /* Load another word */ 109 l.or r7, r6, r0 /* If word count -> 0, then... */ 123 l.lbz r6, 4(r4) /* 1st byte: update addr by 4 */ 124 l.sb 4(r3), r6 /* since we pre-adjusted by 4 */ 134 l.lbz r6, 1(r4) /* But handle the rest by */ 135 l.sb 1(r3), r6 /* updating addr by 1 */ 167 l.sw -4(r3), r6 /* Stor [all...] |
| H A D | memset.S | 92 l.lwz r6, 0(r3) /* get first word */ 95 l.and r6, r6, r7 /* clear bytes to be filled */ 98 l.or r6, r6, r7 /* merge existing with new */ 100 l.sw 0(r3), r6 /* store first word */ 105 l.srli r6, r5, 2 /* clear low two bits of len */ 106 l.srli r6, r6, 2 /* ... */ 107 l.sfgeu r3, r6 /* an [all...] |
| H A D | strlen.S | 80 l.addi r6, r0, -1 /* r6 = 0xffffffff */ 82 l.srl r6, r6, r5 /* clear low (MSB) bytes */ 83 l.xori r6, r6, -1 /* complement */ 84 l.or r8, r8, r6 /* merge with loaded word */ 93 l.xori r6, r7, -1 /* t1 = ~t0 */ 102 l.or r6, r8, r13 /* t1 = x | 0x7f7f7f7f */ 104 l.or r4, r5, r6 /* t [all...] |
| /src/common/lib/libc/arch/powerpc/string/ |
| H A D | memcmp.S | 85 lwz %r6, 0(%r11) /* load 1st buffer 1 word */ 90 cmplw %r8, %r6 /* compare buffer 2 and buffer 1*/ 135 lwzu %r6, 4(%r11) /* pre-load buffer 1 word */ 141 cmplw %r8, %r6 /* Compare first buffer 2 word */ 146 lwzu %r6, 4(%r11) /* pre-load buffer 1 word */ 169 cmplw %r8, %r6 /* compare last dword */ 199 lbz %r6, 0(%r11) /* byte copy loop */ 207 cmplw %r8, %r6 237 srw %r6, %r6, [all...] |
| H A D | memcpy.S | 76 mfspr %r6, 287 /* mfpvbr %r6 PVR = 287 */ 77 srwi %r6, %r6, 0x10 /* get version field from PVR */ 78 cmpwi %r6, 0x1 /* 601 CPU = 0x0001 */ 81 or %r6, %r3, %r4 /* see if both source and dest */ 82 andi. %r6, %r6, 3 /* are 32bit aligned */ 86 li %r6, 0 88 lbzx %r7, %r4, %r6 [all...] |
| H A D | memmove.S | 60 mr %r6, %r3 /* swap src/dst */ 62 mr %r4, %r6 89 stwu %r6, 4(%r8) /* Store previous word */ 95 lwzu %r6, 4(%r4) /* Load another word */ 99 mr %r7, %r6 /* If word count -> 0, then... */ 112 lbzu %r6, 4(%r4) /* 1st byte: update addr by 4 */ 113 stbu %r6, 4(%r8) /* since we pre-adjusted by 4 */ 118 lbzu %r6, 1(%r4) /* But handle the rest by */ 119 stbu %r6, 1(%r8) /* updating addr by 1 */ 143 stwu %r6, [all...] |
| H A D | strlen.S | 97 add %r6, %r7, %r9 98 and. %r0, %r0, %r6
|
| /src/common/lib/libc/arch/sh3/string/ |
| H A D | memcpy.S | 49 #define REG_LEN r6
|
| H A D | memset.S | 46 # define REG_LEN r6
|
| /src/lib/csu/arch/earm/ |
| H A D | crt0.S | 50 movs r6, #7 52 bics r7, r7, r6
|
| /src/lib/libc/arch/hppa/gen/ |
| H A D | __setjmp14.S | 77 stwm %r6, 4(%r1)
|