memcmp.S revision 1.1 1 1.1 matt /* $NetBSD: memcmp.S,v 1.1 2014/09/03 19:34:25 matt Exp $ */
2 1.1 matt
3 1.1 matt /*-
4 1.1 matt * Copyright (c) 2014 The NetBSD Foundation, Inc.
5 1.1 matt * All rights reserved.
6 1.1 matt *
7 1.1 matt * This code is derived from software contributed to The NetBSD Foundation
8 1.1 matt * by Matt Thomas of 3am Software Foundry.
9 1.1 matt *
10 1.1 matt * Redistribution and use in source and binary forms, with or without
11 1.1 matt * modification, are permitted provided that the following conditions
12 1.1 matt * are met:
13 1.1 matt * 1. Redistributions of source code must retain the above copyright
14 1.1 matt * notice, this list of conditions and the following disclaimer.
15 1.1 matt * 2. Redistributions in binary form must reproduce the above copyright
16 1.1 matt * notice, this list of conditions and the following disclaimer in the
17 1.1 matt * documentation and/or other materials provided with the distribution.
18 1.1 matt *
19 1.1 matt * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
20 1.1 matt * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
21 1.1 matt * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
22 1.1 matt * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
23 1.1 matt * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24 1.1 matt * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25 1.1 matt * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26 1.1 matt * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27 1.1 matt * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28 1.1 matt * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29 1.1 matt * POSSIBILITY OF SUCH DAMAGE.
30 1.1 matt */
31 1.1 matt
32 1.1 matt /*
33 1.1 matt *
34 1.1 matt *
35 1.1 matt * int memcmp(const char *s1, const char *s2, size_t n);
36 1.1 matt *
37 1.1 matt * for (; n-- != 0; s1++, s2++) {
38 1.1 matt * if (*s1 < *s2)
39 1.1 matt * return -1;
40 1.1 matt * if (*s1 > *s2)
41 1.1 matt * return 1;
42 1.1 matt * }
43 1.1 matt * return 0;
44 1.1 matt *
45 1.1 matt * Return: ((s1 > s2) ? 1 : (s1 < s2) ? -1 : 0)
46 1.1 matt *
47 1.1 matt * ==========================================================================
48 1.1 matt */
49 1.1 matt
50 1.1 matt #include <machine/asm.h>
51 1.1 matt
52 1.1 matt .text
53 1.1 matt .align 4
54 1.1 matt /* LINTSTUB: Func: void *memcmp(const void *, const void *, size_t) */
55 1.1 matt ENTRY(memcmp)
56 1.1 matt
57 1.1 matt /*
58 1.1 matt * Check count passed in R5. If zero, return 0; otherwise continue.
59 1.1 matt */
60 1.1 matt l.sfeqi r5, 0 /* nothing to compare? */
61 1.1 matt l.bf .Lret_0 /* yes, return equality */
62 1.1 matt l.nop
63 1.1 matt
64 1.1 matt #ifdef _KERNEL
65 1.1 matt l.sfeqi r5, 6 /* less than two words? */
66 1.1 matt l.bnf .Lsixbyte_compare /* yes, just compare by bytes */
67 1.1 matt l.nop
68 1.1 matt #endif
69 1.1 matt
70 1.1 matt l.sfgesi r5, 7 /* less than two words? */
71 1.1 matt l.bnf .Lbyte_compare /* yes, just compare by bytes */
72 1.1 matt l.nop
73 1.1 matt
74 1.1 matt l.xor r6, r3, r4 /* check alignment compatibility */
75 1.1 matt l.andi r6, r6, 3 /* only care about the two bits */
76 1.1 matt l.sfeqi r6, 0 /* same alignment? */
77 1.1 matt l.bnf .Lmisaligned /* no, avoid alignment errors */
78 1.1 matt l.nop
79 1.1 matt
80 1.1 matt /*
81 1.1 matt * At this point, we know we read the data via word accesses.
82 1.1 matt */
83 1.1 matt
84 1.1 matt l.andi r7, r3, 3 /* check alignment */
85 1.1 matt l.sfeqi r7, 0 /* word aligned? */
86 1.1 matt l.bf .Lword_compare /* yes, it is. */
87 1.1 matt
88 1.1 matt l.sub r3, r3, r7 /* align string 1 */
89 1.1 matt l.sub r4, r4, r7 /* align string 2 */
90 1.1 matt l.add r5, r5, r7 /* pad length */
91 1.1 matt
92 1.1 matt l.lwz r15, 0(r3) /* load word from s1 */
93 1.1 matt l.lwz r17, 0(r4) /* load word from s2 */
94 1.1 matt
95 1.1 matt l.slli r7, r7, 3 /* bytes to bits */
96 1.1 matt l.sll r15, r15, r7 /* shift away leading bytes */
97 1.1 matt l.sll r17, r17, r7 /* shift away leading bytes */
98 1.1 matt l.j .Lword_compare /* now we can compare them */
99 1.1 matt l.nop
100 1.1 matt
101 1.1 matt .Lword_loop:
102 1.1 matt l.lwz r15, 0(r3) /* load s1 word */
103 1.1 matt l.lwz r17, 0(r4) /* load s2 word */
104 1.1 matt .Lword_compare:
105 1.1 matt l.sfeq r15, r17 /* compare s1 and s2 words */
106 1.1 matt l.bnf .Lall_done /* different? we're done */
107 1.1 matt
108 1.1 matt l.addi r3, r3, 4 /* advance s1 one word */
109 1.1 matt l.addi r4, r4, 4 /* advance s2 one word */
110 1.1 matt l.addi r5, r5, -4 /* decrement one word */
111 1.1 matt l.sfgtsi r5, 4 /* at least more than a word? */
112 1.1 matt l.bf .Lword_loop /* yes, loop around */
113 1.1 matt l.nop
114 1.1 matt l.sfeqi r5, 0 /* nothing left? */
115 1.1 matt l.bf .Lret_0 /* yes, return equality */
116 1.1 matt l.nop
117 1.1 matt
118 1.1 matt /*
119 1.1 matt * Fall through to handle the last word
120 1.1 matt */
121 1.1 matt
122 1.1 matt l.sub r3, r0, r5 /* If count <= 4, handle */
123 1.1 matt l.andi r3, r3, 3 /* mask off low 2 bits */
124 1.1 matt l.slli r3, r3, 3 /* count *= 8 */
125 1.1 matt l.srl r15, r15, r3 /* discard extra s1 bytes */
126 1.1 matt l.srl r17, r17, r3 /* discard extra s2 bytes */
127 1.1 matt
128 1.1 matt l.sfeq r17, r15 /* compare result */
129 1.1 matt l.bnf .Lall_done
130 1.1 matt .Lret_0:
131 1.1 matt l.addi r11, r0, 0
132 1.1 matt l.jr lr
133 1.1 matt l.nop
134 1.1 matt
135 1.1 matt /*
136 1.1 matt * The two string don't have the same word alignment.
137 1.1 matt */
138 1.1 matt .Lmisaligned:
139 1.1 matt l.sfeqi r6, 2 /* check for halfword alignment */
140 1.1 matt l.bnf .Lbyte_compare
141 1.1 matt l.nop
142 1.1 matt l.andi r7, r3, 1
143 1.1 matt l.sfeqi r7, 0
144 1.1 matt l.bf .Lhalfword_loop
145 1.1 matt l.nop
146 1.1 matt l.addi r5, r5, 1
147 1.1 matt l.addi r3, r3, -1
148 1.1 matt l.addi r4, r4, -1
149 1.1 matt l.lbz r15, 1(r3)
150 1.1 matt l.lbz r17, 1(r4)
151 1.1 matt l.j .Lhalfword_compare
152 1.1 matt l.nop
153 1.1 matt .Lhalfword_loop:
154 1.1 matt l.lhz r15, 0(r3)
155 1.1 matt l.lhz r17, 0(r4)
156 1.1 matt .Lhalfword_compare:
157 1.1 matt l.sfeq r15, r17
158 1.1 matt l.bnf .Lall_done
159 1.1 matt l.nop
160 1.1 matt l.addi r3, r3, 2
161 1.1 matt l.addi r4, r4, 2
162 1.1 matt l.addi r5, r5, -2
163 1.1 matt l.sfgesi r5, 2
164 1.1 matt l.bf .Lhalfword_loop
165 1.1 matt l.nop
166 1.1 matt
167 1.1 matt .Lbyte_compare:
168 1.1 matt l.addi r5, r5, -1
169 1.1 matt l.sfgesi r5, 0
170 1.1 matt l.bnf .Lret_0
171 1.1 matt l.nop
172 1.1 matt l.lbz r15, 0(r3)
173 1.1 matt l.lbz r17, 0(r4)
174 1.1 matt l.addi r3, r3, 1
175 1.1 matt l.addi r4, r4, 1
176 1.1 matt l.sfeq r15, r17
177 1.1 matt l.bf .Lbyte_compare
178 1.1 matt l.nop
179 1.1 matt
180 1.1 matt .Lall_done:
181 1.1 matt l.sub r11, r15, r17 /* subtract s2 from s1 */
182 1.1 matt l.srai r11, r11, 30 /* replicate sign bit thru bit 1 */
183 1.1 matt l.ori r11, r11, 1 /* make sure bit 0 is set */
184 1.1 matt l.jr lr
185 1.1 matt l.nop
186 1.1 matt
187 1.1 matt #ifdef _KERNEL
188 1.1 matt .Lsixbyte_compare:
189 1.1 matt l.or r7, r3, r4
190 1.1 matt l.andi r7, r7, 1
191 1.1 matt l.sfeqi r7, 0
192 1.1 matt l.bnf .Lbyte_compare
193 1.1 matt l.nop
194 1.1 matt l.lhz r15, 0(r3)
195 1.1 matt l.lhz r17, 0(r4)
196 1.1 matt l.sfeq r15, r17
197 1.1 matt l.bnf .Lall_done
198 1.1 matt l.nop
199 1.1 matt l.lhz r15, 2(r3)
200 1.1 matt l.lhz r17, 2(r4)
201 1.1 matt l.sfeq r15, r17
202 1.1 matt l.bnf .Lall_done
203 1.1 matt l.nop
204 1.1 matt l.lhz r15, 4(r3)
205 1.1 matt l.lhz r17, 4(r4)
206 1.1 matt l.sfeq r15, r17
207 1.1 matt l.bnf .Lall_done
208 1.1 matt l.nop
209 1.1 matt l.addi r11, r0, 0
210 1.1 matt l.jr lr
211 1.1 matt l.nop
212 1.1 matt #endif
213 1.1 matt END(memcmp)
214