1 1.13 riastrad /* $NetBSD: n_support.S,v 1.13 2024/07/17 12:00:48 riastradh Exp $ */ 2 1.1 ragge /* 3 1.1 ragge * Copyright (c) 1985, 1993 4 1.1 ragge * The Regents of the University of California. All rights reserved. 5 1.1 ragge * 6 1.1 ragge * Redistribution and use in source and binary forms, with or without 7 1.1 ragge * modification, are permitted provided that the following conditions 8 1.1 ragge * are met: 9 1.1 ragge * 1. Redistributions of source code must retain the above copyright 10 1.1 ragge * notice, this list of conditions and the following disclaimer. 11 1.1 ragge * 2. Redistributions in binary form must reproduce the above copyright 12 1.1 ragge * notice, this list of conditions and the following disclaimer in the 13 1.1 ragge * documentation and/or other materials provided with the distribution. 14 1.6 agc * 3. Neither the name of the University nor the names of its contributors 15 1.1 ragge * may be used to endorse or promote products derived from this software 16 1.1 ragge * without specific prior written permission. 17 1.1 ragge * 18 1.1 ragge * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 19 1.1 ragge * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 20 1.1 ragge * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 21 1.1 ragge * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 22 1.1 ragge * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 23 1.1 ragge * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 24 1.1 ragge * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 25 1.1 ragge * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 26 1.1 ragge * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 27 1.1 ragge * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 28 1.1 ragge * SUCH DAMAGE. 29 1.1 ragge * 30 1.1 ragge * @(#)support.s 8.1 (Berkeley) 6/4/93 31 1.1 ragge */ 32 1.3 matt #include <machine/asm.h> 33 1.3 matt 34 1.7 martin WEAK_ALIAS(logbl,logb) 35 1.10 joerg WEAK_ALIAS(copysignl, _copysignl) 36 1.10 joerg WEAK_ALIAS(_copysignl, copysign) 37 1.7 martin 38 1.3 matt .text 39 1.1 ragge _sccsid: 40 1.3 matt .asciz "@(#)support.s\t1.3 (Berkeley) 8/21/85; 8.1 (ucb.elefunt) 6/4/93" 41 1.1 ragge 42 1.1 ragge /* 43 1.1 ragge * copysign(x,y), 44 1.1 ragge * logb(x), 45 1.1 ragge * scalb(x,N), 46 1.1 ragge * finite(x), 47 1.1 ragge * drem(x,y), 48 1.1 ragge * Coded in vax assembly language by K.C. Ng, 3/14/85. 49 1.1 ragge * Revised by K.C. Ng on 4/9/85. 50 1.1 ragge */ 51 1.1 ragge 52 1.1 ragge /* 53 1.3 matt * double copysign(double x,double y) 54 1.1 ragge */ 55 1.3 matt 56 1.3 matt ENTRY(copysign, 0) 57 1.4 matt movq 4(%ap),%r0 # load x into %r0 58 1.4 matt bicw3 $0x807f,%r0,%r2 # mask off the exponent of x 59 1.1 ragge beql Lz # if zero or reserved op then return x 60 1.4 matt bicw3 $0x7fff,12(%ap),%r2 # copy the sign bit of y into %r2 61 1.4 matt bicw2 $0x8000,%r0 # replace x by |x| 62 1.4 matt bisw2 %r2,%r0 # copy the sign bit of y to x 63 1.1 ragge Lz: ret 64 1.11 riastrad END(copysign) 65 1.1 ragge 66 1.8 martin ENTRY(copysignf, 0) 67 1.9 martin movl 4(%ap),%r0 # load x into %r0 68 1.8 martin bicw3 $0x807f,%r0,%r2 # mask off the exponent of x 69 1.8 martin beql 1f # if zero or reserved op then return x 70 1.8 martin bicw3 $0x7fff,8(%ap),%r2 # copy the sign bit of y into %r2 71 1.8 martin bicw2 $0x8000,%r0 # replace x by |x| 72 1.8 martin bisw2 %r2,%r0 # copy the sign bit of y to x 73 1.8 martin 1: ret 74 1.11 riastrad END(copysignf) 75 1.8 martin 76 1.1 ragge /* 77 1.7 martin * float logbf(float x); 78 1.7 martin */ 79 1.7 martin ENTRY(logbf, 0) 80 1.7 martin cvtfd 4(%ap),-(%sp) 81 1.7 martin calls $2,_C_LABEL(logb) 82 1.7 martin cvtdf %r0,%r0 83 1.7 martin ret 84 1.11 riastrad END(logbf) 85 1.7 martin 86 1.7 martin /* 87 1.3 matt * double logb(double x); 88 1.1 ragge */ 89 1.3 matt ENTRY(logb, 0) 90 1.4 matt bicl3 $0xffff807f,4(%ap),%r0 # mask off the exponent of x 91 1.1 ragge beql Ln 92 1.4 matt ashl $-7,%r0,%r0 # get the bias exponent 93 1.4 matt subl2 $129,%r0 # get the unbias exponent 94 1.4 matt cvtld %r0,%r0 # return the answer in double 95 1.1 ragge ret 96 1.4 matt Ln: movq 4(%ap),%r0 # %r0:1 = x (zero or reserved op) 97 1.1 ragge bneq 1f # simply return if reserved op 98 1.4 matt movq $0x0000fe00ffffcfff,%r0 # -2147483647.0 99 1.1 ragge 1: ret 100 1.11 riastrad END(logb) 101 1.1 ragge 102 1.1 ragge /* 103 1.3 matt * long finite(double x); 104 1.1 ragge */ 105 1.5 matt #ifndef __GFLOAT__ 106 1.12 riastrad WEAK_ALIAS(finitef, _finitef) 107 1.12 riastrad STRONG_ALIAS(_finitef, _finite) 108 1.5 matt #endif 109 1.12 riastrad WEAK_ALIAS(finite, _finite) 110 1.12 riastrad ENTRY(_finite, 0) 111 1.13 riastrad bicw3 $0x7f,4(%ap),%r0 # mask off the significand 112 1.4 matt cmpw %r0,$0x8000 # to see if x is the reserved op 113 1.1 ragge beql 1f # if so, return FALSE (0) 114 1.4 matt movl $1,%r0 # else return TRUE (1) 115 1.1 ragge ret 116 1.4 matt 1: clrl %r0 117 1.1 ragge ret 118 1.12 riastrad END(_finite) 119 1.1 ragge 120 1.3 matt /* int isnan(double x); 121 1.3 matt */ 122 1.3 matt #if 0 123 1.3 matt ENTRY(isnan, 0) 124 1.4 matt clrl %r0 125 1.3 matt ret 126 1.3 matt #endif 127 1.3 matt 128 1.3 matt /* int isnanf(float x); 129 1.3 matt */ 130 1.3 matt ENTRY(isnanf, 0) 131 1.4 matt clrl %r0 132 1.3 matt ret 133 1.11 riastrad END(isnanf) 134 1.3 matt 135 1.1 ragge /* 136 1.1 ragge * double scalb(x,N) 137 1.1 ragge * double x; double N; 138 1.1 ragge */ 139 1.1 ragge .set ERANGE,34 140 1.3 matt 141 1.3 matt ENTRY(scalb, 0) 142 1.4 matt movq 4(%ap),%r0 143 1.4 matt bicl3 $0xffff807f,%r0,%r3 144 1.1 ragge beql ret1 # 0 or reserved operand 145 1.4 matt movq 12(%ap),%r4 146 1.4 matt cvtdl %r4, %r2 147 1.4 matt cmpl %r2,$0x12c 148 1.1 ragge bgeq ovfl 149 1.4 matt cmpl %r2,$-0x12c 150 1.1 ragge bleq unfl 151 1.4 matt ashl $7,%r2,%r2 152 1.4 matt addl2 %r2,%r3 153 1.1 ragge bleq unfl 154 1.4 matt cmpl %r3,$0x8000 155 1.1 ragge bgeq ovfl 156 1.4 matt addl2 %r2,%r0 157 1.1 ragge ret 158 1.1 ragge ovfl: pushl $ERANGE 159 1.3 matt calls $1,_C_LABEL(infnan) # if it returns 160 1.4 matt bicw3 $0x7fff,4(%ap),%r2 # get the sign of input arg 161 1.4 matt bisw2 %r2,%r0 # re-attach the sign to %r0/1 162 1.1 ragge ret 163 1.4 matt unfl: movq $0,%r0 164 1.1 ragge ret1: ret 165 1.11 riastrad END(scalb) 166 1.1 ragge 167 1.1 ragge /* 168 1.1 ragge * DREM(X,Y) 169 1.1 ragge * RETURN X REM Y =X-N*Y, N=[X/Y] ROUNDED (ROUNDED TO EVEN IN THE HALF WAY CASE) 170 1.1 ragge * DOUBLE PRECISION (VAX D format 56 bits) 171 1.1 ragge * CODED IN VAX ASSEMBLY LANGUAGE BY K.C. NG, 4/8/85. 172 1.1 ragge */ 173 1.1 ragge .set EDOM,33 174 1.3 matt 175 1.3 matt ENTRY(drem, 0x0fc0) 176 1.4 matt subl2 $12,%sp 177 1.4 matt movq 4(%ap),%r0 #%r0=x 178 1.4 matt movq 12(%ap),%r2 #%r2=y 179 1.1 ragge jeql Rop #if y=0 then generate reserved op fault 180 1.4 matt bicw3 $0x007f,%r0,%r4 #check if x is Rop 181 1.4 matt cmpw %r4,$0x8000 182 1.1 ragge jeql Ret #if x is Rop then return Rop 183 1.4 matt bicl3 $0x007f,%r2,%r4 #check if y is Rop 184 1.4 matt cmpw %r4,$0x8000 185 1.1 ragge jeql Ret #if y is Rop then return Rop 186 1.4 matt bicw2 $0x8000,%r2 #y := |y| 187 1.4 matt movw $0,-4(%fp) #-4(%fp) = nx := 0 188 1.4 matt cmpw %r2,$0x1c80 #yexp ? 57 189 1.1 ragge bgtr C1 #if yexp > 57 goto C1 190 1.4 matt addw2 $0x1c80,%r2 #scale up y by 2**57 191 1.4 matt movw $0x1c80,-4(%fp) #nx := 57 (exponent field) 192 1.1 ragge C1: 193 1.4 matt movw -4(%fp),-8(%fp) #-8(%fp) = nf := nx 194 1.4 matt bicw3 $0x7fff,%r0,-12(%fp) #-12(%fp) = sign of x 195 1.4 matt bicw2 $0x8000,%r0 #x := |x| 196 1.4 matt movq %r2,%r10 #y1 := y 197 1.4 matt bicl2 $0xffff07ff,%r11 #clear the last 27 bits of y1 198 1.1 ragge loop: 199 1.4 matt cmpd %r0,%r2 #x ? y 200 1.1 ragge bleq E1 #if x <= y goto E1 201 1.1 ragge /* begin argument reduction */ 202 1.4 matt movq %r2,%r4 #t =y 203 1.4 matt movq %r10,%r6 #t1=y1 204 1.4 matt bicw3 $0x807f,%r0,%r8 #xexp= exponent of x 205 1.4 matt bicw3 $0x807f,%r2,%r9 #yexp= exponent fo y 206 1.4 matt subw2 %r9,%r8 #xexp-yexp 207 1.4 matt subw2 $0x0c80,%r8 #k=xexp-yexp-25(exponent bit field) 208 1.1 ragge blss C2 #if k<0 goto C2 209 1.4 matt addw2 %r8,%r4 #t +=k 210 1.4 matt addw2 %r8,%r6 #t1+=k, scale up t and t1 211 1.1 ragge C2: 212 1.4 matt divd3 %r4,%r0,%r8 #x/t 213 1.4 matt cvtdl %r8,%r8 #n=[x/t] truncated 214 1.4 matt cvtld %r8,%r8 #float(n) 215 1.4 matt subd2 %r6,%r4 #t:=t-t1 216 1.4 matt muld2 %r8,%r4 #n*(t-t1) 217 1.4 matt muld2 %r8,%r6 #n*t1 218 1.4 matt subd2 %r6,%r0 #x-n*t1 219 1.4 matt subd2 %r4,%r0 #(x-n*t1)-n*(t-t1) 220 1.3 matt jbr loop 221 1.1 ragge E1: 222 1.4 matt movw -4(%fp),%r6 #%r6=nx 223 1.1 ragge beql C3 #if nx=0 goto C3 224 1.4 matt addw2 %r6,%r0 #x:=x*2**57 scale up x by nx 225 1.4 matt movw $0,-4(%fp) #clear nx 226 1.3 matt jbr loop 227 1.1 ragge C3: 228 1.4 matt movq %r2,%r4 #%r4 = y 229 1.4 matt subw2 $0x80,%r4 #%r4 = y/2 230 1.4 matt cmpd %r0,%r4 #x:y/2 231 1.1 ragge blss E2 #if x < y/2 goto E2 232 1.1 ragge bgtr C4 #if x > y/2 goto C4 233 1.4 matt cvtdl %r8,%r8 #ifix(float(n)) 234 1.4 matt blbc %r8,E2 #if the last bit is zero, goto E2 235 1.1 ragge C4: 236 1.4 matt subd2 %r2,%r0 #x-y 237 1.1 ragge E2: 238 1.4 matt xorw2 -12(%fp),%r0 #x^sign (exclusive or) 239 1.4 matt movw -8(%fp),%r6 #%r6=nf 240 1.4 matt bicw3 $0x807f,%r0,%r8 #%r8=exponent of x 241 1.4 matt bicw2 $0x7f80,%r0 #clear the exponent of x 242 1.4 matt subw2 %r6,%r8 #%r8=xexp-nf 243 1.1 ragge bgtr C5 #if xexp-nf is positive goto C5 244 1.4 matt movw $0,%r8 #clear %r8 245 1.4 matt movq $0,%r0 #x underflow to zero 246 1.1 ragge C5: 247 1.4 matt bisw2 %r8,%r0 /* put %r8 into x's exponent field */ 248 1.1 ragge ret 249 1.1 ragge Rop: #Reserved operand 250 1.1 ragge pushl $EDOM 251 1.3 matt calls $1,_C_LABEL(infnan) #generate reserved op fault 252 1.1 ragge ret 253 1.1 ragge Ret: 254 1.4 matt movq $0x8000,%r0 #propagate reserved op 255 1.1 ragge ret 256 1.11 riastrad END(drem) 257