| /src/crypto/external/apache2/openssl/dist/crypto/chacha/asm/ |
| chacha-x86.pl | 826 &vmovdqa (&QWP(16*$cp-128,"ebx"),$xc_) if ($ai>0 && $ai<3); 828 &vmovdqa (&QWP(16*$bp-128,"ebx"),$xb_) if ($i!=0); 830 &vmovdqa ($xc_,&QWP(16*$cn-128,"ebx")) if ($ai>0 && $ai<3); 832 &vmovdqa ($xa_,&QWP(16*$an-128,"ebx")); 834 &vmovdqa ($xb_,&QWP(16*$bn-128,"ebx")) if ($i<7); 836 &vmovdqa ($xd_,&QWP(16*$dn-128,"ebx")) if ($di!=$dn); 840 &vmovdqa (&QWP(16*$ai-128,"ebx"),$xa); 842 &vmovdqa (&QWP(16*$di-128,"ebx"),$xd) if ($di!=$dn); 890 &vmovdqa (&QWP(16*12-128,"ebp"),"xmm0"); 891 &vmovdqa (&QWP(16*13-128,"ebp"),"xmm1") [all...] |
| chacha-x86_64.pl | 1479 vmovdqa .Lsigma(%rip),$xa3 # key[0] 1487 vmovdqa $xa0,0x40(%rsp) # ... and offload 1489 vmovdqa $xa1,0x50(%rsp) 1491 vmovdqa $xa2,0x60(%rsp) 1492 vmovdqa $xa3,0x70(%rsp) 1496 vmovdqa $xb0,0x80-0x100(%rcx) 1498 vmovdqa $xb1,0x90-0x100(%rcx) 1500 vmovdqa $xb2,0xa0-0x100(%rcx) 1501 vmovdqa $xb3,0xb0-0x100(%rcx) 1505 vmovdqa $xt0,0xc0-0x100(%rcx [all...] |
| /src/crypto/external/bsd/openssl/dist/crypto/chacha/asm/ |
| chacha-x86.pl | 826 &vmovdqa (&QWP(16*$cp-128,"ebx"),$xc_) if ($ai>0 && $ai<3); 828 &vmovdqa (&QWP(16*$bp-128,"ebx"),$xb_) if ($i!=0); 830 &vmovdqa ($xc_,&QWP(16*$cn-128,"ebx")) if ($ai>0 && $ai<3); 832 &vmovdqa ($xa_,&QWP(16*$an-128,"ebx")); 834 &vmovdqa ($xb_,&QWP(16*$bn-128,"ebx")) if ($i<7); 836 &vmovdqa ($xd_,&QWP(16*$dn-128,"ebx")) if ($di!=$dn); 840 &vmovdqa (&QWP(16*$ai-128,"ebx"),$xa); 842 &vmovdqa (&QWP(16*$di-128,"ebx"),$xd) if ($di!=$dn); 890 &vmovdqa (&QWP(16*12-128,"ebp"),"xmm0"); 891 &vmovdqa (&QWP(16*13-128,"ebp"),"xmm1") [all...] |
| chacha-x86_64.pl | 1476 vmovdqa .Lsigma(%rip),$xa3 # key[0] 1484 vmovdqa $xa0,0x40(%rsp) # ... and offload 1486 vmovdqa $xa1,0x50(%rsp) 1488 vmovdqa $xa2,0x60(%rsp) 1489 vmovdqa $xa3,0x70(%rsp) 1493 vmovdqa $xb0,0x80-0x100(%rcx) 1495 vmovdqa $xb1,0x90-0x100(%rcx) 1497 vmovdqa $xb2,0xa0-0x100(%rcx) 1498 vmovdqa $xb3,0xb0-0x100(%rcx) 1502 vmovdqa $xt0,0xc0-0x100(%rcx [all...] |
| /src/crypto/external/bsd/openssl.old/dist/crypto/chacha/asm/ |
| chacha-x86.pl | 827 &vmovdqa (&QWP(16*$cp-128,"ebx"),$xc_) if ($ai>0 && $ai<3); 829 &vmovdqa (&QWP(16*$bp-128,"ebx"),$xb_) if ($i!=0); 831 &vmovdqa ($xc_,&QWP(16*$cn-128,"ebx")) if ($ai>0 && $ai<3); 833 &vmovdqa ($xa_,&QWP(16*$an-128,"ebx")); 835 &vmovdqa ($xb_,&QWP(16*$bn-128,"ebx")) if ($i<7); 837 &vmovdqa ($xd_,&QWP(16*$dn-128,"ebx")) if ($di!=$dn); 841 &vmovdqa (&QWP(16*$ai-128,"ebx"),$xa); 843 &vmovdqa (&QWP(16*$di-128,"ebx"),$xd) if ($di!=$dn); 891 &vmovdqa (&QWP(16*12-128,"ebp"),"xmm0"); 892 &vmovdqa (&QWP(16*13-128,"ebp"),"xmm1") [all...] |
| chacha-x86_64.pl | 1474 vmovdqa .Lsigma(%rip),$xa3 # key[0] 1482 vmovdqa $xa0,0x40(%rsp) # ... and offload 1484 vmovdqa $xa1,0x50(%rsp) 1486 vmovdqa $xa2,0x60(%rsp) 1487 vmovdqa $xa3,0x70(%rsp) 1491 vmovdqa $xb0,0x80-0x100(%rcx) 1493 vmovdqa $xb1,0x90-0x100(%rcx) 1495 vmovdqa $xb2,0xa0-0x100(%rcx) 1496 vmovdqa $xb3,0xb0-0x100(%rcx) 1500 vmovdqa $xt0,0xc0-0x100(%rcx [all...] |
| /src/sys/external/isc/libsodium/dist/src/libsodium/crypto_scalarmult/curve25519/sandy2x/ |
| ladder_base.S | 27 vmovdqa v0_0(%rip),%xmm0 label 28 vmovdqa v1_0(%rip),%xmm1 label 29 vmovdqa v9_0(%rip),%xmm2 label 30 vmovdqa %xmm2,0(%rsp) label 31 vmovdqa %xmm0,16(%rsp) label 32 vmovdqa %xmm0,32(%rsp) label 33 vmovdqa %xmm0,48(%rsp) label 34 vmovdqa %xmm0,64(%rsp) label 35 vmovdqa %xmm1,80(%rsp) label 36 vmovdqa %xmm0,96(%rsp label 37 vmovdqa %xmm0,112(%rsp) label 38 vmovdqa %xmm0,128(%rsp) label 39 vmovdqa %xmm0,144(%rsp) label 40 vmovdqa %xmm1,%xmm0 label 98 vmovdqa 0(%rsp),%xmm11 label 99 vmovdqa 80(%rsp),%xmm12 label 108 vmovdqa 16(%rsp),%xmm13 label 109 vmovdqa 96(%rsp),%xmm14 label 118 vmovdqa %xmm13,0(%rsp) label 119 vmovdqa %xmm14,16(%rsp) label 120 vmovdqa 32(%rsp),%xmm13 label 121 vmovdqa 112(%rsp),%xmm14 label 130 vmovdqa %xmm13,32(%rsp) label 131 vmovdqa %xmm14,80(%rsp) label 132 vmovdqa 48(%rsp),%xmm13 label 133 vmovdqa 128(%rsp),%xmm14 label 142 vmovdqa %xmm13,48(%rsp) label 143 vmovdqa %xmm14,96(%rsp) label 144 vmovdqa 64(%rsp),%xmm13 label 145 vmovdqa 144(%rsp),%xmm14 label 154 vmovdqa %xmm13,64(%rsp) label 155 vmovdqa %xmm14,112(%rsp) label 168 vmovdqa %xmm1,128(%rsp) label 171 vmovdqa %xmm0,144(%rsp) label 174 vmovdqa %xmm1,160(%rsp) label 183 vmovdqa %xmm3,176(%rsp) label 186 vmovdqa %xmm1,192(%rsp) label 189 vmovdqa %xmm3,208(%rsp) label 198 vmovdqa %xmm5,224(%rsp) label 201 vmovdqa %xmm3,240(%rsp) label 210 vmovdqa %xmm5,256(%rsp) label 212 vmovdqa %xmm5,272(%rsp) label 215 vmovdqa %xmm7,288(%rsp) label 218 vmovdqa %xmm3,304(%rsp) label 221 vmovdqa %xmm3,320(%rsp) label 227 vmovdqa %xmm3,336(%rsp) label 229 vmovdqa %xmm7,352(%rsp) label 231 vmovdqa %xmm7,368(%rsp) label 235 vmovdqa %xmm9,384(%rsp) label 240 vmovdqa %xmm3,400(%rsp) label 243 vmovdqa %xmm9,416(%rsp) label 245 vmovdqa 0(%rsp),%xmm3 label 246 vmovdqa 16(%rsp),%xmm9 label 294 vmovdqa 32(%rsp),%xmm3 label 295 vmovdqa 80(%rsp),%xmm9 label 343 vmovdqa 48(%rsp),%xmm3 label 344 vmovdqa 96(%rsp),%xmm9 label 392 vmovdqa 64(%rsp),%xmm3 label 393 vmovdqa 112(%rsp),%xmm9 label 512 vmovdqa %xmm1,0(%rsp) label 514 vmovdqa %xmm1,16(%rsp) label 516 vmovdqa %xmm1,32(%rsp) label 530 vmovdqa %xmm6,48(%rsp) label 532 vmovdqa %xmm6,64(%rsp) label 533 vmovdqa %xmm5,80(%rsp) label 535 vmovdqa %xmm5,96(%rsp) label 554 vmovdqa %xmm8,112(%rsp) label 556 vmovdqa %xmm8,160(%rsp) label 605 vmovdqa 16(%rsp),%xmm4 label 617 vmovdqa 48(%rsp),%xmm4 label 621 vmovdqa 80(%rsp),%xmm4 label 627 vmovdqa 16(%rsp),%xmm4 label 630 vmovdqa 16(%rsp),%xmm4 label 633 vmovdqa 48(%rsp),%xmm4 label 638 vmovdqa 32(%rsp),%xmm2 label 641 vmovdqa 64(%rsp),%xmm2 label 644 vmovdqa 96(%rsp),%xmm2 label 647 vmovdqa 160(%rsp),%xmm2 label 691 vmovdqa %xmm2,0(%rsp) label 694 vmovdqa %xmm2,80(%rsp) label 696 vmovdqa %xmm2,16(%rsp) label 699 vmovdqa %xmm2,96(%rsp) label 701 vmovdqa %xmm2,32(%rsp) label 704 vmovdqa %xmm1,112(%rsp) label 706 vmovdqa %xmm1,48(%rsp) label 709 vmovdqa %xmm0,160(%rsp) label 711 vmovdqa %xmm0,64(%rsp) label 714 vmovdqa %xmm0,208(%rsp) label 715 vmovdqa 144(%rsp),%xmm0 label 718 vmovdqa 128(%rsp),%xmm2 label 720 vmovdqa 192(%rsp),%xmm4 label 722 vmovdqa 176(%rsp),%xmm6 label 724 vmovdqa 240(%rsp),%xmm8 label 730 vmovdqa 384(%rsp),%xmm14 label 795 vmovdqa 256(%rsp),%xmm4 label 806 vmovdqa 304(%rsp),%xmm4 label 812 vmovdqa 256(%rsp),%xmm4 label 815 vmovdqa 256(%rsp),%xmm4 label 818 vmovdqa 304(%rsp),%xmm4 label 821 vmovdqa 352(%rsp),%xmm4 label 826 vmovdqa 272(%rsp),%xmm2 label 829 vmovdqa 320(%rsp),%xmm2 label 832 vmovdqa 368(%rsp),%xmm2 label 835 vmovdqa 400(%rsp),%xmm2 label 880 vmovdqa %xmm1,176(%rsp) label 885 vmovdqa %xmm2,192(%rsp) label 886 vmovdqa %xmm1,224(%rsp) label 888 vmovdqa %xmm1,240(%rsp) label 890 vmovdqa 80(%rsp),%xmm1 label 895 vmovdqa %xmm4,256(%rsp) label 900 vmovdqa %xmm3,272(%rsp) label 901 vmovdqa %xmm4,288(%rsp) label 903 vmovdqa %xmm4,304(%rsp) label 905 vmovdqa 96(%rsp),%xmm3 label 910 vmovdqa %xmm6,320(%rsp) label 915 vmovdqa %xmm5,336(%rsp) label 916 vmovdqa %xmm6,352(%rsp) label 918 vmovdqa %xmm6,368(%rsp) label 920 vmovdqa 112(%rsp),%xmm5 label 925 vmovdqa %xmm8,384(%rsp) label 930 vmovdqa %xmm7,400(%rsp) label 931 vmovdqa %xmm8,416(%rsp) label 933 vmovdqa %xmm8,432(%rsp) label 935 vmovdqa 160(%rsp),%xmm7 label 940 vmovdqa %xmm0,160(%rsp) label 945 vmovdqa %xmm9,448(%rsp) label 946 vmovdqa %xmm0,464(%rsp) label 948 vmovdqa %xmm0,480(%rsp) label 950 vmovdqa 208(%rsp),%xmm0 label 994 vmovdqa %xmm10,80(%rsp) label 997 vmovdqa %xmm2,96(%rsp) label 1000 vmovdqa %xmm3,112(%rsp) label 1003 vmovdqa %xmm4,128(%rsp) label 1006 vmovdqa %xmm5,144(%rsp) label 1008 vmovdqa 176(%rsp),%xmm5 label 1043 vmovdqa 256(%rsp),%xmm1 label 1089 vmovdqa 320(%rsp),%xmm1 label 1135 vmovdqa 384(%rsp),%xmm1 label 1181 vmovdqa 160(%rsp),%xmm1 label [all...] |
| ladder.S | 28 vmovdqa v0_0(%rip),%xmm0 label 29 vmovdqa v1_0(%rip),%xmm1 label 31 vmovdqa %xmm2,0(%rsp) label 33 vmovdqa %xmm2,16(%rsp) label 35 vmovdqa %xmm2,32(%rsp) label 37 vmovdqa %xmm2,48(%rsp) label 39 vmovdqa %xmm2,64(%rsp) label 40 vmovdqa %xmm1,80(%rsp) label 41 vmovdqa %xmm0,96(%rsp) label 42 vmovdqa %xmm0,112(%rsp label 43 vmovdqa %xmm0,128(%rsp) label 44 vmovdqa %xmm0,144(%rsp) label 45 vmovdqa %xmm1,%xmm0 label 56 vmovdqa %xmm10,160(%rsp) label 58 vmovdqa %xmm10,176(%rsp) label 60 vmovdqa %xmm10,192(%rsp) label 62 vmovdqa %xmm10,208(%rsp) label 64 vmovdqa %xmm10,224(%rsp) label 66 vmovdqa %xmm10,240(%rsp) label 68 vmovdqa %xmm10,256(%rsp) label 70 vmovdqa %xmm10,272(%rsp) label 72 vmovdqa %xmm10,288(%rsp) label 75 vmovdqa %xmm10,304(%rsp) label 77 vmovdqa %xmm10,320(%rsp) label 80 vmovdqa %xmm10,336(%rsp) label 82 vmovdqa %xmm10,352(%rsp) label 85 vmovdqa %xmm10,368(%rsp) label 87 vmovdqa %xmm10,384(%rsp) label 90 vmovdqa %xmm10,400(%rsp) label 92 vmovdqa %xmm10,416(%rsp) label 98 vmovdqa %xmm10,432(%rsp) label 147 vmovdqa 0(%rsp),%xmm11 label 148 vmovdqa 80(%rsp),%xmm12 label 157 vmovdqa 16(%rsp),%xmm13 label 158 vmovdqa 96(%rsp),%xmm14 label 167 vmovdqa %xmm13,0(%rsp) label 168 vmovdqa %xmm14,16(%rsp) label 169 vmovdqa 32(%rsp),%xmm13 label 170 vmovdqa 112(%rsp),%xmm14 label 179 vmovdqa %xmm13,32(%rsp) label 180 vmovdqa %xmm14,80(%rsp) label 181 vmovdqa 48(%rsp),%xmm13 label 182 vmovdqa 128(%rsp),%xmm14 label 191 vmovdqa %xmm13,48(%rsp) label 192 vmovdqa %xmm14,96(%rsp) label 193 vmovdqa 64(%rsp),%xmm13 label 194 vmovdqa 144(%rsp),%xmm14 label 203 vmovdqa %xmm13,64(%rsp) label 204 vmovdqa %xmm14,112(%rsp) label 217 vmovdqa %xmm1,128(%rsp) label 220 vmovdqa %xmm0,144(%rsp) label 223 vmovdqa %xmm1,448(%rsp) label 232 vmovdqa %xmm3,464(%rsp) label 235 vmovdqa %xmm1,480(%rsp) label 238 vmovdqa %xmm3,496(%rsp) label 247 vmovdqa %xmm5,512(%rsp) label 250 vmovdqa %xmm3,528(%rsp) label 259 vmovdqa %xmm5,544(%rsp) label 261 vmovdqa %xmm5,560(%rsp) label 264 vmovdqa %xmm7,576(%rsp) label 267 vmovdqa %xmm3,592(%rsp) label 270 vmovdqa %xmm3,608(%rsp) label 276 vmovdqa %xmm3,624(%rsp) label 278 vmovdqa %xmm7,640(%rsp) label 280 vmovdqa %xmm7,656(%rsp) label 284 vmovdqa %xmm9,672(%rsp) label 289 vmovdqa %xmm3,688(%rsp) label 292 vmovdqa %xmm9,704(%rsp) label 294 vmovdqa 0(%rsp),%xmm3 label 295 vmovdqa 16(%rsp),%xmm9 label 343 vmovdqa 32(%rsp),%xmm3 label 344 vmovdqa 80(%rsp),%xmm9 label 392 vmovdqa 48(%rsp),%xmm3 label 393 vmovdqa 96(%rsp),%xmm9 label 441 vmovdqa 64(%rsp),%xmm3 label 442 vmovdqa 112(%rsp),%xmm9 label 561 vmovdqa %xmm1,0(%rsp) label 563 vmovdqa %xmm1,16(%rsp) label 565 vmovdqa %xmm1,32(%rsp) label 579 vmovdqa %xmm6,48(%rsp) label 581 vmovdqa %xmm6,64(%rsp) label 582 vmovdqa %xmm5,80(%rsp) label 584 vmovdqa %xmm5,96(%rsp) label 603 vmovdqa %xmm8,112(%rsp) label 605 vmovdqa %xmm8,448(%rsp) label 654 vmovdqa 16(%rsp),%xmm4 label 666 vmovdqa 48(%rsp),%xmm4 label 670 vmovdqa 80(%rsp),%xmm4 label 676 vmovdqa 16(%rsp),%xmm4 label 679 vmovdqa 16(%rsp),%xmm4 label 682 vmovdqa 48(%rsp),%xmm4 label 687 vmovdqa 32(%rsp),%xmm2 label 690 vmovdqa 64(%rsp),%xmm2 label 693 vmovdqa 96(%rsp),%xmm2 label 696 vmovdqa 448(%rsp),%xmm2 label 740 vmovdqa %xmm2,0(%rsp) label 759 vmovdqa %xmm3,16(%rsp) label 783 vmovdqa %xmm3,32(%rsp) label 807 vmovdqa %xmm1,48(%rsp) label 831 vmovdqa %xmm0,64(%rsp) label 854 vmovdqa %xmm4,80(%rsp) label 855 vmovdqa %xmm6,96(%rsp) label 856 vmovdqa %xmm7,112(%rsp) label 857 vmovdqa %xmm11,448(%rsp) label 858 vmovdqa %xmm2,496(%rsp) label 859 vmovdqa 144(%rsp),%xmm0 label 862 vmovdqa 128(%rsp),%xmm2 label 864 vmovdqa 480(%rsp),%xmm4 label 866 vmovdqa 464(%rsp),%xmm6 label 868 vmovdqa 528(%rsp),%xmm8 label 874 vmovdqa 672(%rsp),%xmm14 label 939 vmovdqa 544(%rsp),%xmm4 label 950 vmovdqa 592(%rsp),%xmm4 label 956 vmovdqa 544(%rsp),%xmm4 label 959 vmovdqa 544(%rsp),%xmm4 label 962 vmovdqa 592(%rsp),%xmm4 label 965 vmovdqa 640(%rsp),%xmm4 label 970 vmovdqa 560(%rsp),%xmm2 label 973 vmovdqa 608(%rsp),%xmm2 label 976 vmovdqa 656(%rsp),%xmm2 label 979 vmovdqa 688(%rsp),%xmm2 label 1024 vmovdqa %xmm1,464(%rsp) label 1029 vmovdqa %xmm2,480(%rsp) label 1030 vmovdqa %xmm1,512(%rsp) label 1032 vmovdqa %xmm1,528(%rsp) label 1034 vmovdqa 80(%rsp),%xmm1 label 1039 vmovdqa %xmm4,544(%rsp) label 1044 vmovdqa %xmm3,560(%rsp) label 1045 vmovdqa %xmm4,576(%rsp) label 1047 vmovdqa %xmm4,592(%rsp) label 1049 vmovdqa 96(%rsp),%xmm3 label 1054 vmovdqa %xmm6,608(%rsp) label 1059 vmovdqa %xmm5,624(%rsp) label 1060 vmovdqa %xmm6,640(%rsp) label 1062 vmovdqa %xmm6,656(%rsp) label 1064 vmovdqa 112(%rsp),%xmm5 label 1069 vmovdqa %xmm8,672(%rsp) label 1074 vmovdqa %xmm7,688(%rsp) label 1075 vmovdqa %xmm8,704(%rsp) label 1077 vmovdqa %xmm8,720(%rsp) label 1079 vmovdqa 448(%rsp),%xmm7 label 1084 vmovdqa %xmm0,448(%rsp) label 1089 vmovdqa %xmm9,736(%rsp) label 1090 vmovdqa %xmm0,752(%rsp) label 1092 vmovdqa %xmm0,768(%rsp) label 1094 vmovdqa 496(%rsp),%xmm0 label 1138 vmovdqa %xmm10,80(%rsp) label 1141 vmovdqa %xmm2,96(%rsp) label 1144 vmovdqa %xmm3,112(%rsp) label 1147 vmovdqa %xmm4,128(%rsp) label 1150 vmovdqa %xmm5,144(%rsp) label 1152 vmovdqa 464(%rsp),%xmm5 label 1187 vmovdqa 544(%rsp),%xmm1 label 1233 vmovdqa 608(%rsp),%xmm1 label 1279 vmovdqa 672(%rsp),%xmm1 label 1325 vmovdqa 448(%rsp),%xmm1 label [all...] |
| /src/crypto/external/apache2/openssl/lib/libcrypto/arch/x86_64/ |
| chacha-x86_64.S | 1190 vmovdqa .Lsigma(%rip),%xmm11 1198 vmovdqa %xmm8,64(%rsp) 1200 vmovdqa %xmm9,80(%rsp) 1202 vmovdqa %xmm10,96(%rsp) 1203 vmovdqa %xmm11,112(%rsp) 1207 vmovdqa %xmm0,128-256(%rcx) 1209 vmovdqa %xmm1,144-256(%rcx) 1211 vmovdqa %xmm2,160-256(%rcx) 1212 vmovdqa %xmm3,176-256(%rcx) 1216 vmovdqa %xmm12,192-256(%rcx [all...] |
| sha1-mb-x86_64.S | 3034 vmovdqa -32(%rbp),%xmm15 3054 vmovdqa %xmm0,0-128(%rax) 3077 vmovdqa %xmm1,16-128(%rax) 3100 vmovdqa %xmm2,32-128(%rax) 3123 vmovdqa %xmm3,48-128(%rax) 3146 vmovdqa %xmm4,64-128(%rax) 3169 vmovdqa %xmm0,80-128(%rax) 3192 vmovdqa %xmm1,96-128(%rax) 3215 vmovdqa %xmm2,112-128(%rax) 3238 vmovdqa %xmm3,128-128(%rax [all...] |
| /src/crypto/external/bsd/openssl/lib/libcrypto/arch/x86_64/ |
| chacha-x86_64.S | 1188 vmovdqa .Lsigma(%rip),%xmm11 1196 vmovdqa %xmm8,64(%rsp) 1198 vmovdqa %xmm9,80(%rsp) 1200 vmovdqa %xmm10,96(%rsp) 1201 vmovdqa %xmm11,112(%rsp) 1205 vmovdqa %xmm0,128-256(%rcx) 1207 vmovdqa %xmm1,144-256(%rcx) 1209 vmovdqa %xmm2,160-256(%rcx) 1210 vmovdqa %xmm3,176-256(%rcx) 1214 vmovdqa %xmm12,192-256(%rcx [all...] |
| sha1-mb-x86_64.S | 3034 vmovdqa -32(%rbp),%xmm15 3054 vmovdqa %xmm0,0-128(%rax) 3077 vmovdqa %xmm1,16-128(%rax) 3100 vmovdqa %xmm2,32-128(%rax) 3123 vmovdqa %xmm3,48-128(%rax) 3146 vmovdqa %xmm4,64-128(%rax) 3169 vmovdqa %xmm0,80-128(%rax) 3192 vmovdqa %xmm1,96-128(%rax) 3215 vmovdqa %xmm2,112-128(%rax) 3238 vmovdqa %xmm3,128-128(%rax [all...] |
| /src/crypto/external/bsd/openssl.old/lib/libcrypto/arch/x86_64/ |
| chacha-x86_64.S | 1188 vmovdqa .Lsigma(%rip),%xmm11 1196 vmovdqa %xmm8,64(%rsp) 1198 vmovdqa %xmm9,80(%rsp) 1200 vmovdqa %xmm10,96(%rsp) 1201 vmovdqa %xmm11,112(%rsp) 1205 vmovdqa %xmm0,128-256(%rcx) 1207 vmovdqa %xmm1,144-256(%rcx) 1209 vmovdqa %xmm2,160-256(%rcx) 1210 vmovdqa %xmm3,176-256(%rcx) 1214 vmovdqa %xmm12,192-256(%rcx [all...] |
| sha1-mb-x86_64.S | 3014 vmovdqa -32(%rbp),%xmm15 3034 vmovdqa %xmm0,0-128(%rax) 3057 vmovdqa %xmm1,16-128(%rax) 3080 vmovdqa %xmm2,32-128(%rax) 3103 vmovdqa %xmm3,48-128(%rax) 3126 vmovdqa %xmm4,64-128(%rax) 3149 vmovdqa %xmm0,80-128(%rax) 3172 vmovdqa %xmm1,96-128(%rax) 3195 vmovdqa %xmm2,112-128(%rax) 3218 vmovdqa %xmm3,128-128(%rax [all...] |
| /src/external/lgpl3/gmp/dist/mpn/x86_64/fastavx/ |
| copyd.asm | 91 vmovdqa %xmm0, 16(rp) 103 vmovdqa %ymm0, (rp) 104 vmovdqa %ymm1, -32(rp) 105 vmovdqa %ymm2, -64(rp) 106 vmovdqa %ymm3, -96(rp) 116 vmovdqa %ymm0, (rp) 117 vmovdqa %ymm1, -32(rp) 124 vmovdqa %ymm0, (rp) 131 vmovdqa %xmm0, 16(rp)
|
| copyi.asm | 88 vmovdqa %xmm0, (rp) 100 vmovdqa %ymm0, (rp) 101 vmovdqa %ymm1, 32(rp) 102 vmovdqa %ymm2, 64(rp) 103 vmovdqa %ymm3, 96(rp) 113 vmovdqa %ymm0, (rp) 114 vmovdqa %ymm1, 32(rp) 121 vmovdqa %ymm0, (rp) 128 vmovdqa %xmm0, (rp)
|
| /src/crypto/external/apache2/openssl/dist/crypto/poly1305/asm/ |
| poly1305-x86_64.pl | 791 vmovdqa %xmm6,0x50(%r11) 792 vmovdqa %xmm7,0x60(%r11) 793 vmovdqa %xmm8,0x70(%r11) 794 vmovdqa %xmm9,0x80(%r11) 795 vmovdqa %xmm10,0x90(%r11) 796 vmovdqa %xmm11,0xa0(%r11) 797 vmovdqa %xmm12,0xb0(%r11) 798 vmovdqa %xmm13,0xc0(%r11) 799 vmovdqa %xmm14,0xd0(%r11) 800 vmovdqa %xmm15,0xe0(%r11 [all...] |
| poly1305-x86.pl | 1245 &vmovdqa ($MASK,&QWP(64,"ebx")); 1260 &vmovdqa (&QWP(16*0,"esp"),$D0); 1261 &vmovdqa (&QWP(16*1,"esp"),$D1); 1262 &vmovdqa (&QWP(16*2,"esp"),$D2); 1263 &vmovdqa (&QWP(16*3,"esp"),$D3); 1264 &vmovdqa (&QWP(16*4,"esp"),$D4); 1270 &vmovdqa (&QWP(16*5,"esp"),$T1); 1271 &vmovdqa (&QWP(16*6,"esp"),$T0); 1276 &vmovdqa (&QWP(16*7,"esp"),$T1); 1277 &vmovdqa (&QWP(16*8,"esp"),$T0) [all...] |
| /src/crypto/external/bsd/openssl/dist/crypto/poly1305/asm/ |
| poly1305-x86_64.pl | 788 vmovdqa %xmm6,0x50(%r11) 789 vmovdqa %xmm7,0x60(%r11) 790 vmovdqa %xmm8,0x70(%r11) 791 vmovdqa %xmm9,0x80(%r11) 792 vmovdqa %xmm10,0x90(%r11) 793 vmovdqa %xmm11,0xa0(%r11) 794 vmovdqa %xmm12,0xb0(%r11) 795 vmovdqa %xmm13,0xc0(%r11) 796 vmovdqa %xmm14,0xd0(%r11) 797 vmovdqa %xmm15,0xe0(%r11 [all...] |
| /src/crypto/external/bsd/openssl.old/dist/crypto/poly1305/asm/ |
| poly1305-x86_64.pl | 786 vmovdqa %xmm6,0x50(%r11) 787 vmovdqa %xmm7,0x60(%r11) 788 vmovdqa %xmm8,0x70(%r11) 789 vmovdqa %xmm9,0x80(%r11) 790 vmovdqa %xmm10,0x90(%r11) 791 vmovdqa %xmm11,0xa0(%r11) 792 vmovdqa %xmm12,0xb0(%r11) 793 vmovdqa %xmm13,0xc0(%r11) 794 vmovdqa %xmm14,0xd0(%r11) 795 vmovdqa %xmm15,0xe0(%r11 [all...] |
| /src/crypto/external/apache2/openssl/dist/crypto/sha/asm/ |
| sha1-586.pl | 1126 &vmovdqa(@X[3],&QWP(0,$tmp1)); # K_00_19 1127 &vmovdqa(@X[4],&QWP(16,$tmp1)); # K_20_39 1128 &vmovdqa(@X[5],&QWP(32,$tmp1)); # K_40_59 1129 &vmovdqa(@X[6],&QWP(48,$tmp1)); # K_60_79 1130 &vmovdqa(@X[2],&QWP(64,$tmp1)); # pbswap mask 1161 &vmovdqa(&QWP(112+0,"esp"),@X[4]); # copy constants 1162 &vmovdqa(&QWP(112+16,"esp"),@X[5]); 1163 &vmovdqa(&QWP(112+32,"esp"),@X[6]); 1165 &vmovdqa(&QWP(112+48,"esp"),@X[3]); 1167 &vmovdqa(&QWP(112+64,"esp"),@X[2]) [all...] |
| /src/crypto/external/bsd/openssl/dist/crypto/sha/asm/ |
| sha1-586.pl | 1126 &vmovdqa(@X[3],&QWP(0,$tmp1)); # K_00_19 1127 &vmovdqa(@X[4],&QWP(16,$tmp1)); # K_20_39 1128 &vmovdqa(@X[5],&QWP(32,$tmp1)); # K_40_59 1129 &vmovdqa(@X[6],&QWP(48,$tmp1)); # K_60_79 1130 &vmovdqa(@X[2],&QWP(64,$tmp1)); # pbswap mask 1161 &vmovdqa(&QWP(112+0,"esp"),@X[4]); # copy constants 1162 &vmovdqa(&QWP(112+16,"esp"),@X[5]); 1163 &vmovdqa(&QWP(112+32,"esp"),@X[6]); 1165 &vmovdqa(&QWP(112+48,"esp"),@X[3]); 1167 &vmovdqa(&QWP(112+64,"esp"),@X[2]) [all...] |
| /src/crypto/external/bsd/openssl.old/dist/crypto/sha/asm/ |
| sha1-586.pl | 1127 &vmovdqa(@X[3],&QWP(0,$tmp1)); # K_00_19 1128 &vmovdqa(@X[4],&QWP(16,$tmp1)); # K_20_39 1129 &vmovdqa(@X[5],&QWP(32,$tmp1)); # K_40_59 1130 &vmovdqa(@X[6],&QWP(48,$tmp1)); # K_60_79 1131 &vmovdqa(@X[2],&QWP(64,$tmp1)); # pbswap mask 1162 &vmovdqa(&QWP(112+0,"esp"),@X[4]); # copy constants 1163 &vmovdqa(&QWP(112+16,"esp"),@X[5]); 1164 &vmovdqa(&QWP(112+32,"esp"),@X[6]); 1166 &vmovdqa(&QWP(112+48,"esp"),@X[3]); 1168 &vmovdqa(&QWP(112+64,"esp"),@X[2]) [all...] |
| /src/crypto/external/bsd/openssl.old/dist/crypto/aes/asm/ |
| vpaes-armv8.pl | 192 // vmovdqa .Lk_ipt(%rip), %xmm2 # iptlo 197 // vmovdqa .Lk_ipt+16(%rip), %xmm3 # ipthi 208 ld1 {v1.2d}, [x11], #16 // vmovdqa -0x40(%r11,%r10), %xmm1 # .Lk_mc_forward[] 214 ld1 {v4.2d}, [x10] // vmovdqa (%r11,%r10), %xmm4 # .Lk_mc_backward[] 244 // vmovdqa -0x60(%r10), %xmm4 # 3 : sbou .Lk_sbo 245 // vmovdqa -0x50(%r10), %xmm0 # 0 : sbot .Lk_sbo+16 247 ld1 {v1.2d}, [x10] // vmovdqa 0x40(%r11,%r10), %xmm1 # .Lk_sr[] 279 // vmovdqa .Lk_ipt(%rip), %xmm2 # iptlo 287 // vmovdqa .Lk_ipt+16(%rip), %xmm3 # ipthi 302 ld1 {v1.2d}, [x11], #16 // vmovdqa -0x40(%r11,%r10), %xmm1 # .Lk_mc_forward[ [all...] |
| /src/crypto/external/apache2/openssl/dist/crypto/aes/asm/ |
| aesni-xts-avx512.pl | 144 $code .= "vmovdqa $state_tweak, ($TW)\n"; 422 vmovdqa ($TW), $tw1 423 vmovdqa 0x10($TW), $tw2 424 vmovdqa 0x20($TW), $tw3 425 vmovdqa 0x30($TW), $tw4 426 vmovdqa 0x40($TW), $tw5 427 vmovdqa 0x50($TW), $tw6 428 vmovdqa 0x60($TW), $tw7 455 vmovdqa 0x0($TW), $tw[0] 471 $code .= "vmovdqa $offset($TW), $tw[$i]\n" [all...] |