| /src/crypto/external/apache2/openssl/dist/crypto/modes/asm/ |
| ghashv8-armx.pl | 71 my ($Xl,$Xm,$Xh,$IN)=map("q$_",(0..3)); 134 vpmull.p64 $Xl,$H,$H 139 vext.8 $t1,$Xl,$Xh,#8 @ Karatsuba post-processing 140 veor $t2,$Xl,$Xh 143 vpmull.p64 $t2,$Xl,$xC2 @ 1st phase 146 vmov $Xm#hi,$Xl#lo @ Xm is rotated Xl 147 veor $Xl,$Xm,$t2 149 vext.8 $t2,$Xl,$Xl,#8 @ 2nd phas [all...] |
| ghashp8-ppc.pl | 71 my ($Xl,$Xm,$Xh,$IN)=map("v$_",(0..3)); 121 vpmsumd $Xl,$IN,$Hl # H.lo·H.lo 125 vpmsumd $t2,$Xl,$xC2 # 1st reduction phase 129 vxor $Xl,$Xl,$t0 132 vsldoi $Xl,$Xl,$Xl,8 133 vxor $Xl,$Xl,$t [all...] |
| ghash-armv4.pl | 381 my ($Xl,$Xm,$Xh,$IN)=map("q$_",(0..3)); 476 vld1.64 $Xl#hi,[$Xi]! @ load Xi 477 vld1.64 $Xl#lo,[$Xi]! 482 vrev64.8 $Xl,$Xl 493 veor $IN,$Xl @ inp^=Xi 496 &clmul64x64 ($Xl,$Hlo,"$IN#lo"); # H.lo·Xi.lo 503 veor $Xm,$Xm,$Xl @ Karatsuba post-processing 505 veor $Xl#hi,$Xl#hi,$Xm#l [all...] |
| ghash-x86_64.pl | 710 my ($Xl,$Xm,$Xh,$Hkey3,$Hkey4)=map("%xmm$_",(11..15)); 730 movdqu 0x20($inp),$Xl 732 pshufb $T3,$Xl 740 movdqa $Xl,$Xh 741 pshufd \$0b01001110,$Xl,$Xm 742 pxor $Xl,$Xm 743 pclmulqdq \$0x00,$Hkey2,$Xl 746 xorps $Xl,$Xln 751 movdqu 0x10($inp),$Xl 753 pshufb $T3,$Xl [all...] |
| /src/crypto/external/bsd/openssl/dist/crypto/modes/asm/ |
| ghashp8-ppc.pl | 71 my ($Xl,$Xm,$Xh,$IN)=map("v$_",(0..3)); 121 vpmsumd $Xl,$IN,$Hl # H.lo·H.lo 125 vpmsumd $t2,$Xl,$xC2 # 1st reduction phase 129 vxor $Xl,$Xl,$t0 132 vsldoi $Xl,$Xl,$Xl,8 133 vxor $Xl,$Xl,$t [all...] |
| ghashv8-armx.pl | 71 my ($Xl,$Xm,$Xh,$IN)=map("q$_",(0..3)); 129 vpmull.p64 $Xl,$H,$H 134 vext.8 $t1,$Xl,$Xh,#8 @ Karatsuba post-processing 135 veor $t2,$Xl,$Xh 138 vpmull.p64 $t2,$Xl,$xC2 @ 1st phase 141 vmov $Xm#hi,$Xl#lo @ Xm is rotated Xl 142 veor $Xl,$Xm,$t2 144 vext.8 $t2,$Xl,$Xl,#8 @ 2nd phas [all...] |
| ghash-armv4.pl | 381 my ($Xl,$Xm,$Xh,$IN)=map("q$_",(0..3)); 476 vld1.64 $Xl#hi,[$Xi]! @ load Xi 477 vld1.64 $Xl#lo,[$Xi]! 482 vrev64.8 $Xl,$Xl 493 veor $IN,$Xl @ inp^=Xi 496 &clmul64x64 ($Xl,$Hlo,"$IN#lo"); # H.lo·Xi.lo 503 veor $Xm,$Xm,$Xl @ Karatsuba post-processing 505 veor $Xl#hi,$Xl#hi,$Xm#l [all...] |
| ghash-x86_64.pl | 709 my ($Xl,$Xm,$Xh,$Hkey3,$Hkey4)=map("%xmm$_",(11..15)); 729 movdqu 0x20($inp),$Xl 731 pshufb $T3,$Xl 739 movdqa $Xl,$Xh 740 pshufd \$0b01001110,$Xl,$Xm 741 pxor $Xl,$Xm 742 pclmulqdq \$0x00,$Hkey2,$Xl 745 xorps $Xl,$Xln 750 movdqu 0x10($inp),$Xl 752 pshufb $T3,$Xl [all...] |
| /src/crypto/external/bsd/openssl.old/dist/crypto/modes/asm/ |
| ghashp8-ppc.pl | 68 my ($Xl,$Xm,$Xh,$IN)=map("v$_",(0..3)); 118 vpmsumd $Xl,$IN,$Hl # H.lo·H.lo 122 vpmsumd $t2,$Xl,$xC2 # 1st reduction phase 126 vxor $Xl,$Xl,$t0 129 vsldoi $Xl,$Xl,$Xl,8 130 vxor $Xl,$Xl,$t [all...] |
| ghashv8-armx.pl | 67 my ($Xl,$Xm,$Xh,$IN)=map("q$_",(0..3)); 117 vpmull.p64 $Xl,$H,$H 122 vext.8 $t1,$Xl,$Xh,#8 @ Karatsuba post-processing 123 veor $t2,$Xl,$Xh 126 vpmull.p64 $t2,$Xl,$xC2 @ 1st phase 129 vmov $Xm#hi,$Xl#lo @ Xm is rotated Xl 130 veor $Xl,$Xm,$t2 132 vext.8 $t2,$Xl,$Xl,#8 @ 2nd phas [all...] |
| ghash-armv4.pl | 378 my ($Xl,$Xm,$Xh,$IN)=map("q$_",(0..3)); 473 vld1.64 $Xl#hi,[$Xi]! @ load Xi 474 vld1.64 $Xl#lo,[$Xi]! 479 vrev64.8 $Xl,$Xl 490 veor $IN,$Xl @ inp^=Xi 493 &clmul64x64 ($Xl,$Hlo,"$IN#lo"); # H.lo·Xi.lo 500 veor $Xm,$Xm,$Xl @ Karatsuba post-processing 502 veor $Xl#hi,$Xl#hi,$Xm#l [all...] |
| ghash-x86_64.pl | 703 my ($Xl,$Xm,$Xh,$Hkey3,$Hkey4)=map("%xmm$_",(11..15)); 723 movdqu 0x20($inp),$Xl 725 pshufb $T3,$Xl 733 movdqa $Xl,$Xh 734 pshufd \$0b01001110,$Xl,$Xm 735 pxor $Xl,$Xm 736 pclmulqdq \$0x00,$Hkey2,$Xl 739 xorps $Xl,$Xln 744 movdqu 0x10($inp),$Xl 746 pshufb $T3,$Xl [all...] |
| /src/crypto/external/bsd/openssh/dist/ |
| blowfish.c | 73 Blowfish_encipher(blf_ctx *c, u_int32_t *xl, u_int32_t *xr) 75 u_int32_t Xl; 80 Xl = *xl; 83 Xl ^= p[0]; 84 BLFRND(s, p, Xr, Xl, 1); BLFRND(s, p, Xl, Xr, 2); 85 BLFRND(s, p, Xr, Xl, 3); BLFRND(s, p, Xl, Xr, 4); 86 BLFRND(s, p, Xr, Xl, 5); BLFRND(s, p, Xl, Xr, 6) [all...] |
| /src/lib/libcrypt/ |
| blowfish.c | 77 Blowfish_encipher(blf_ctx *c, u_int32_t *xl, u_int32_t *xr) 79 u_int32_t Xl; 84 Xl = *xl; 87 Xl ^= p[0]; 88 BLFRND(s, p, Xr, Xl, 1); BLFRND(s, p, Xl, Xr, 2); 89 BLFRND(s, p, Xr, Xl, 3); BLFRND(s, p, Xl, Xr, 4); 90 BLFRND(s, p, Xr, Xl, 5); BLFRND(s, p, Xl, Xr, 6) [all...] |
| /src/crypto/external/apache2/openssl/lib/libcrypto/arch/arm/ |
| ghashv8-armx.S | 51 vmov d3,d0 @ Xm is rotated Xl 90 vmov d3,d0 @ Xm is rotated Xl 186 vmov d3,d0 @ Xm is rotated Xl 223 vmov d3,d0 @ Xm is rotated Xl
|
| ghash-armv4.S | 531 veor d4,d4,d3 @ Xh|Xl - 256-bit result
|
| /src/crypto/external/bsd/openssl/lib/libcrypto/arch/arm/ |
| ghashv8-armx.S | 51 vmov d3,d0 @ Xm is rotated Xl 90 vmov d3,d0 @ Xm is rotated Xl 186 vmov d3,d0 @ Xm is rotated Xl 223 vmov d3,d0 @ Xm is rotated Xl
|
| ghash-armv4.S | 531 veor d4,d4,d3 @ Xh|Xl - 256-bit result
|
| /src/crypto/external/bsd/openssl.old/lib/libcrypto/arch/aarch64/ |
| ghashv8-armx.S | 43 vmov d3,d0 @ Xm is rotated Xl 82 vmov d3,d0 @ Xm is rotated Xl 175 vmov d3,d0 @ Xm is rotated Xl 212 vmov d3,d0 @ Xm is rotated Xl
|
| /src/crypto/external/bsd/openssl.old/lib/libcrypto/arch/arm/ |
| ghashv8-armx.S | 44 vmov d3,d0 @ Xm is rotated Xl 83 vmov d3,d0 @ Xm is rotated Xl 176 vmov d3,d0 @ Xm is rotated Xl 213 vmov d3,d0 @ Xm is rotated Xl
|
| ghash-armv4.S | 530 veor d4,d4,d3 @ Xh|Xl - 256-bit result
|
| /src/external/gpl3/gcc/dist/libgcc/config/avr/libf7/ |
| asm-defs.h | 137 X, x, XL, xl, Xl, xL, x, x, \
|
| /src/external/gpl3/gcc.old/dist/libgcc/config/avr/libf7/ |
| asm-defs.h | 137 X, x, XL, xl, Xl, xL, x, x \
|
| /src/share/mk/ |
| bsd.lib.mk | 647 # -Xl,-nostdlib is not enough because we want to tell the compiler-driver not
|