1 1.1 matt /* $NetBSD: memcmp.S,v 1.1 2014/09/03 19:34:25 matt Exp $ */ 2 1.1 matt 3 1.1 matt /*- 4 1.1 matt * Copyright (c) 2014 The NetBSD Foundation, Inc. 5 1.1 matt * All rights reserved. 6 1.1 matt * 7 1.1 matt * This code is derived from software contributed to The NetBSD Foundation 8 1.1 matt * by Matt Thomas of 3am Software Foundry. 9 1.1 matt * 10 1.1 matt * Redistribution and use in source and binary forms, with or without 11 1.1 matt * modification, are permitted provided that the following conditions 12 1.1 matt * are met: 13 1.1 matt * 1. Redistributions of source code must retain the above copyright 14 1.1 matt * notice, this list of conditions and the following disclaimer. 15 1.1 matt * 2. Redistributions in binary form must reproduce the above copyright 16 1.1 matt * notice, this list of conditions and the following disclaimer in the 17 1.1 matt * documentation and/or other materials provided with the distribution. 18 1.1 matt * 19 1.1 matt * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS 20 1.1 matt * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED 21 1.1 matt * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 22 1.1 matt * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS 23 1.1 matt * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 24 1.1 matt * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 25 1.1 matt * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 26 1.1 matt * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 27 1.1 matt * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 28 1.1 matt * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 29 1.1 matt * POSSIBILITY OF SUCH DAMAGE. 30 1.1 matt */ 31 1.1 matt 32 1.1 matt /* 33 1.1 matt * 34 1.1 matt * 35 1.1 matt * int memcmp(const char *s1, const char *s2, size_t n); 36 1.1 matt * 37 1.1 matt * for (; n-- != 0; s1++, s2++) { 38 1.1 matt * if (*s1 < *s2) 39 1.1 matt * return -1; 40 1.1 matt * if (*s1 > *s2) 41 1.1 matt * return 1; 42 1.1 matt * } 43 1.1 matt * return 0; 44 1.1 matt * 45 1.1 matt * Return: ((s1 > s2) ? 1 : (s1 < s2) ? -1 : 0) 46 1.1 matt * 47 1.1 matt * ========================================================================== 48 1.1 matt */ 49 1.1 matt 50 1.1 matt #include <machine/asm.h> 51 1.1 matt 52 1.1 matt .text 53 1.1 matt .align 4 54 1.1 matt /* LINTSTUB: Func: void *memcmp(const void *, const void *, size_t) */ 55 1.1 matt ENTRY(memcmp) 56 1.1 matt 57 1.1 matt /* 58 1.1 matt * Check count passed in R5. If zero, return 0; otherwise continue. 59 1.1 matt */ 60 1.1 matt l.sfeqi r5, 0 /* nothing to compare? */ 61 1.1 matt l.bf .Lret_0 /* yes, return equality */ 62 1.1 matt l.nop 63 1.1 matt 64 1.1 matt #ifdef _KERNEL 65 1.1 matt l.sfeqi r5, 6 /* less than two words? */ 66 1.1 matt l.bnf .Lsixbyte_compare /* yes, just compare by bytes */ 67 1.1 matt l.nop 68 1.1 matt #endif 69 1.1 matt 70 1.1 matt l.sfgesi r5, 7 /* less than two words? */ 71 1.1 matt l.bnf .Lbyte_compare /* yes, just compare by bytes */ 72 1.1 matt l.nop 73 1.1 matt 74 1.1 matt l.xor r6, r3, r4 /* check alignment compatibility */ 75 1.1 matt l.andi r6, r6, 3 /* only care about the two bits */ 76 1.1 matt l.sfeqi r6, 0 /* same alignment? */ 77 1.1 matt l.bnf .Lmisaligned /* no, avoid alignment errors */ 78 1.1 matt l.nop 79 1.1 matt 80 1.1 matt /* 81 1.1 matt * At this point, we know we read the data via word accesses. 82 1.1 matt */ 83 1.1 matt 84 1.1 matt l.andi r7, r3, 3 /* check alignment */ 85 1.1 matt l.sfeqi r7, 0 /* word aligned? */ 86 1.1 matt l.bf .Lword_compare /* yes, it is. */ 87 1.1 matt 88 1.1 matt l.sub r3, r3, r7 /* align string 1 */ 89 1.1 matt l.sub r4, r4, r7 /* align string 2 */ 90 1.1 matt l.add r5, r5, r7 /* pad length */ 91 1.1 matt 92 1.1 matt l.lwz r15, 0(r3) /* load word from s1 */ 93 1.1 matt l.lwz r17, 0(r4) /* load word from s2 */ 94 1.1 matt 95 1.1 matt l.slli r7, r7, 3 /* bytes to bits */ 96 1.1 matt l.sll r15, r15, r7 /* shift away leading bytes */ 97 1.1 matt l.sll r17, r17, r7 /* shift away leading bytes */ 98 1.1 matt l.j .Lword_compare /* now we can compare them */ 99 1.1 matt l.nop 100 1.1 matt 101 1.1 matt .Lword_loop: 102 1.1 matt l.lwz r15, 0(r3) /* load s1 word */ 103 1.1 matt l.lwz r17, 0(r4) /* load s2 word */ 104 1.1 matt .Lword_compare: 105 1.1 matt l.sfeq r15, r17 /* compare s1 and s2 words */ 106 1.1 matt l.bnf .Lall_done /* different? we're done */ 107 1.1 matt 108 1.1 matt l.addi r3, r3, 4 /* advance s1 one word */ 109 1.1 matt l.addi r4, r4, 4 /* advance s2 one word */ 110 1.1 matt l.addi r5, r5, -4 /* decrement one word */ 111 1.1 matt l.sfgtsi r5, 4 /* at least more than a word? */ 112 1.1 matt l.bf .Lword_loop /* yes, loop around */ 113 1.1 matt l.nop 114 1.1 matt l.sfeqi r5, 0 /* nothing left? */ 115 1.1 matt l.bf .Lret_0 /* yes, return equality */ 116 1.1 matt l.nop 117 1.1 matt 118 1.1 matt /* 119 1.1 matt * Fall through to handle the last word 120 1.1 matt */ 121 1.1 matt 122 1.1 matt l.sub r3, r0, r5 /* If count <= 4, handle */ 123 1.1 matt l.andi r3, r3, 3 /* mask off low 2 bits */ 124 1.1 matt l.slli r3, r3, 3 /* count *= 8 */ 125 1.1 matt l.srl r15, r15, r3 /* discard extra s1 bytes */ 126 1.1 matt l.srl r17, r17, r3 /* discard extra s2 bytes */ 127 1.1 matt 128 1.1 matt l.sfeq r17, r15 /* compare result */ 129 1.1 matt l.bnf .Lall_done 130 1.1 matt .Lret_0: 131 1.1 matt l.addi r11, r0, 0 132 1.1 matt l.jr lr 133 1.1 matt l.nop 134 1.1 matt 135 1.1 matt /* 136 1.1 matt * The two string don't have the same word alignment. 137 1.1 matt */ 138 1.1 matt .Lmisaligned: 139 1.1 matt l.sfeqi r6, 2 /* check for halfword alignment */ 140 1.1 matt l.bnf .Lbyte_compare 141 1.1 matt l.nop 142 1.1 matt l.andi r7, r3, 1 143 1.1 matt l.sfeqi r7, 0 144 1.1 matt l.bf .Lhalfword_loop 145 1.1 matt l.nop 146 1.1 matt l.addi r5, r5, 1 147 1.1 matt l.addi r3, r3, -1 148 1.1 matt l.addi r4, r4, -1 149 1.1 matt l.lbz r15, 1(r3) 150 1.1 matt l.lbz r17, 1(r4) 151 1.1 matt l.j .Lhalfword_compare 152 1.1 matt l.nop 153 1.1 matt .Lhalfword_loop: 154 1.1 matt l.lhz r15, 0(r3) 155 1.1 matt l.lhz r17, 0(r4) 156 1.1 matt .Lhalfword_compare: 157 1.1 matt l.sfeq r15, r17 158 1.1 matt l.bnf .Lall_done 159 1.1 matt l.nop 160 1.1 matt l.addi r3, r3, 2 161 1.1 matt l.addi r4, r4, 2 162 1.1 matt l.addi r5, r5, -2 163 1.1 matt l.sfgesi r5, 2 164 1.1 matt l.bf .Lhalfword_loop 165 1.1 matt l.nop 166 1.1 matt 167 1.1 matt .Lbyte_compare: 168 1.1 matt l.addi r5, r5, -1 169 1.1 matt l.sfgesi r5, 0 170 1.1 matt l.bnf .Lret_0 171 1.1 matt l.nop 172 1.1 matt l.lbz r15, 0(r3) 173 1.1 matt l.lbz r17, 0(r4) 174 1.1 matt l.addi r3, r3, 1 175 1.1 matt l.addi r4, r4, 1 176 1.1 matt l.sfeq r15, r17 177 1.1 matt l.bf .Lbyte_compare 178 1.1 matt l.nop 179 1.1 matt 180 1.1 matt .Lall_done: 181 1.1 matt l.sub r11, r15, r17 /* subtract s2 from s1 */ 182 1.1 matt l.srai r11, r11, 30 /* replicate sign bit thru bit 1 */ 183 1.1 matt l.ori r11, r11, 1 /* make sure bit 0 is set */ 184 1.1 matt l.jr lr 185 1.1 matt l.nop 186 1.1 matt 187 1.1 matt #ifdef _KERNEL 188 1.1 matt .Lsixbyte_compare: 189 1.1 matt l.or r7, r3, r4 190 1.1 matt l.andi r7, r7, 1 191 1.1 matt l.sfeqi r7, 0 192 1.1 matt l.bnf .Lbyte_compare 193 1.1 matt l.nop 194 1.1 matt l.lhz r15, 0(r3) 195 1.1 matt l.lhz r17, 0(r4) 196 1.1 matt l.sfeq r15, r17 197 1.1 matt l.bnf .Lall_done 198 1.1 matt l.nop 199 1.1 matt l.lhz r15, 2(r3) 200 1.1 matt l.lhz r17, 2(r4) 201 1.1 matt l.sfeq r15, r17 202 1.1 matt l.bnf .Lall_done 203 1.1 matt l.nop 204 1.1 matt l.lhz r15, 4(r3) 205 1.1 matt l.lhz r17, 4(r4) 206 1.1 matt l.sfeq r15, r17 207 1.1 matt l.bnf .Lall_done 208 1.1 matt l.nop 209 1.1 matt l.addi r11, r0, 0 210 1.1 matt l.jr lr 211 1.1 matt l.nop 212 1.1 matt #endif 213 1.1 matt END(memcmp) 214