| /src/crypto/external/apache2/openssl/lib/libcrypto/arch/arm/ |
| bsaes-armv7.S | 100 vldmia r6!, {q8} @ .LM0ISR 125 vmov.i8 q8,#0x55 @ compose .LBS0 131 vand q10, q10, q8 132 vand q11, q11, q8 143 vand q10, q10, q8 144 vand q11, q11, q8 151 vmov.i8 q8,#0x0f @ compose .LBS2 180 vand q10, q10, q8 181 vand q11, q11, q8 192 vand q10, q10, q8 [all...] |
| ghashv8-armx.S | 26 vext.8 q8,q10,q11,#8 @ t0=0xc2....01 29 vand q10,q10,q8 32 vand q8,q8,q9 34 veor q12,q3,q8 @ twisted H 38 vext.8 q8,q12,q12,#8 @ Karatsuba pre-processing 40 veor q8,q8,q12 42 INST(0xa0,0x2e,0xa0,0xf2) @ pmull q1,q8,q8 [all...] |
| ghash-armv4.S | 359 vmov.i8 q8,#0xe1 367 vand q8,q8,q9 369 veor q3,q3,q8 @ twisted H 417 vmull.p8 q8, d16, d6 @ F = A1*B 425 veor q8, q8, q0 @ L = E + F 441 vext.8 q8, q8, q8, #1 [all...] |
| poly1305-armv4.S | 428 vmull.u32 q8,d5,d0[1] 434 vmlal.u32 q8,d3,d1[1] 439 vmlal.u32 q8,d1,d3[1] 444 vmlal.u32 q8,d0,d5[1] 449 vmlal.u32 q8,d7,d8[1] 504 vshr.u64 q15,q8,#26 505 vmovn.i64 d16,q8 786 vmull.u32 q8,d27,d0[1] 797 vmlal.u32 q8,d25,d1[1] 803 vmlal.u32 q8,d23,d3[1 [all...] |
| armv4-gf2m.S | 186 vmull.p8 q8, d26, d16 @ G = A*B2 191 veor q2, q2, q8 @ M = G + H 198 vmull.p8 q8, d26, d16 @ K = A*B4 210 vext.8 q8, q8, q8, #12 213 veor q3, q3, q8
|
| armv4-mont.S | 230 vmull.u32 q8,d28,d1[0] 248 vmlal.u32 q8,d29,d5[0] 256 vmov q7,q8 258 vmov q8,q9 279 vmlal.u32 q8,d28,d1[0] 296 vmlal.u32 q8,d29,d5[0] 304 vmov q7,q8 306 vmov q8,q9 335 veor q8,q8,q [all...] |
| aes-gcm-armv8_64.S | 188 trn1 q8, v12.2d, v13.2d @ h2h | h1h 227 eor v16.16b, v16.16b, q8 @ h2k | h1k 376 eor q8, q8, q4 @ GHASH block 4k - mid 387 pmull v10.1q, q8, v10.1d @ GHASH block 4k - mid 417 eor q9, q9, q8 @ GHASH block 4k+2 - high 422 movi q8, #0xc2 456 pmull v31.1q, q9, q8 @ MODULO - top 64b align with mid 543 pmull v9.1q, v10.1d, q8 @ MODULO - mid 64b align with low 605 eor q8, q8, q4 @ GHASH block 4k - mi [all...] |
| /src/crypto/external/bsd/openssl/lib/libcrypto/arch/arm/ |
| bsaes-armv7.S | 100 vldmia r6!, {q8} @ .LM0ISR 125 vmov.i8 q8,#0x55 @ compose .LBS0 131 vand q10, q10, q8 132 vand q11, q11, q8 143 vand q10, q10, q8 144 vand q11, q11, q8 151 vmov.i8 q8,#0x0f @ compose .LBS2 180 vand q10, q10, q8 181 vand q11, q11, q8 192 vand q10, q10, q8 [all...] |
| ghashv8-armx.S | 26 vext.8 q8,q10,q11,#8 @ t0=0xc2....01 29 vand q10,q10,q8 32 vand q8,q8,q9 34 veor q12,q3,q8 @ twisted H 38 vext.8 q8,q12,q12,#8 @ Karatsuba pre-processing 40 veor q8,q8,q12 42 INST(0xa0,0x2e,0xa0,0xf2) @ pmull q1,q8,q8 [all...] |
| ghash-armv4.S | 359 vmov.i8 q8,#0xe1 367 vand q8,q8,q9 369 veor q3,q3,q8 @ twisted H 417 vmull.p8 q8, d16, d6 @ F = A1*B 425 veor q8, q8, q0 @ L = E + F 441 vext.8 q8, q8, q8, #1 [all...] |
| poly1305-armv4.S | 428 vmull.u32 q8,d5,d0[1] 434 vmlal.u32 q8,d3,d1[1] 439 vmlal.u32 q8,d1,d3[1] 444 vmlal.u32 q8,d0,d5[1] 449 vmlal.u32 q8,d7,d8[1] 504 vshr.u64 q15,q8,#26 505 vmovn.i64 d16,q8 786 vmull.u32 q8,d27,d0[1] 797 vmlal.u32 q8,d25,d1[1] 803 vmlal.u32 q8,d23,d3[1 [all...] |
| armv4-gf2m.S | 186 vmull.p8 q8, d26, d16 @ G = A*B2 191 veor q2, q2, q8 @ M = G + H 198 vmull.p8 q8, d26, d16 @ K = A*B4 210 vext.8 q8, q8, q8, #12 213 veor q3, q3, q8
|
| armv4-mont.S | 230 vmull.u32 q8,d28,d1[0] 248 vmlal.u32 q8,d29,d5[0] 256 vmov q7,q8 258 vmov q8,q9 279 vmlal.u32 q8,d28,d1[0] 296 vmlal.u32 q8,d29,d5[0] 304 vmov q7,q8 306 vmov q8,q9 335 veor q8,q8,q [all...] |
| /src/crypto/external/bsd/openssl.old/lib/libcrypto/arch/arm/ |
| bsaes-armv7.S | 97 vldmia r6!, {q8} @ .LM0ISR 122 vmov.i8 q8,#0x55 @ compose .LBS0 128 vand q10, q10, q8 129 vand q11, q11, q8 140 vand q10, q10, q8 141 vand q11, q11, q8 148 vmov.i8 q8,#0x0f @ compose .LBS2 177 vand q10, q10, q8 178 vand q11, q11, q8 189 vand q10, q10, q8 [all...] |
| ghashv8-armx.S | 19 vext.8 q8,q10,q11,#8 @ t0=0xc2....01 22 vand q10,q10,q8 25 vand q8,q8,q9 27 veor q12,q3,q8 @ twisted H 31 vext.8 q8,q12,q12,#8 @ Karatsuba pre-processing 33 veor q8,q8,q12 35 .byte 0xa0,0x2e,0xa0,0xf2 @ pmull q1,q8,q8 [all...] |
| ghash-armv4.S | 358 vmov.i8 q8,#0xe1 366 vand q8,q8,q9 368 veor q3,q3,q8 @ twisted H 416 vmull.p8 q8, d16, d6 @ F = A1*B 424 veor q8, q8, q0 @ L = E + F 440 vext.8 q8, q8, q8, #1 [all...] |
| poly1305-armv4.S | 434 vmull.u32 q8,d5,d0[1] 440 vmlal.u32 q8,d3,d1[1] 445 vmlal.u32 q8,d1,d3[1] 450 vmlal.u32 q8,d0,d5[1] 455 vmlal.u32 q8,d7,d8[1] 510 vshr.u64 q15,q8,#26 511 vmovn.i64 d16,q8 791 vmull.u32 q8,d27,d0[1] 802 vmlal.u32 q8,d25,d1[1] 808 vmlal.u32 q8,d23,d3[1 [all...] |
| armv4-gf2m.S | 183 vmull.p8 q8, d26, d16 @ G = A*B2 188 veor q2, q2, q8 @ M = G + H 195 vmull.p8 q8, d26, d16 @ K = A*B4 207 vext.8 q8, q8, q8, #12 210 veor q3, q3, q8
|
| armv4-mont.S | 223 vmull.u32 q8,d28,d1[0] 241 vmlal.u32 q8,d29,d5[0] 249 vmov q7,q8 251 vmov q8,q9 272 vmlal.u32 q8,d28,d1[0] 289 vmlal.u32 q8,d29,d5[0] 297 vmov q7,q8 299 vmov q8,q9 328 veor q8,q8,q [all...] |
| /src/sys/crypto/chacha/arch/arm/ |
| chacha_neon_32.S | 198 vdup.32 q8, d28[0] 210 1: ROUNDLD q0,q1,q2,q3, q5,q6,q7,q4, q10,q11,q8,q9, q15,q12,q13,q14 212 ROUND q0,q1,q2,q3, q4,q5,q6,q7, q8,q9,q10,q11, q12,q13,q14,q15, \ 214 ROUNDLD q0,q1,q2,q3, q4,q5,q6,q7, q8,q9,q10,q11, q12,q13,q14,q15 215 ROUND q0,q1,q2,q3, q5,q6,q7,q4, q10,q11,q8,q9, q15,q12,q13,q14, \ 220 * q8-q9 are free / saved on the stack. We have: 259 vdup.32 q8, r1 /* q8 := (blkno, blkno, blkno, blkno) */ 267 vadd.u32 q8, q8, q9 /* q8 := (blkno,blkno+1,blkno+2,blkno+3) * [all...] |
| /src/crypto/external/bsd/openssl.old/lib/libcrypto/arch/aarch64/ |
| ghashv8-armx.S | 18 vext.8 q8,q10,q11,#8 @ t0=0xc2....01 21 vand q10,q10,q8 24 vand q8,q8,q9 26 veor q12,q3,q8 @ twisted H 30 vext.8 q8,q12,q12,#8 @ Karatsuba pre-processing 32 veor q8,q8,q12 34 .byte 0xa0,0x2e,0xa0,0xf2 @ pmull q1,q8,q8 [all...] |
| /src/external/lgpl3/gmp/dist/mpn/arm/neon/ |
| hamdist.asm | 56 C We sum into 16 16-bit counters in q8,q9, but at the end we sum them and end 70 vmov.i64 q8, #0 C clear summation register 82 vpadal.u8 d16, d24 C d16/q8 = 0; could just splat 100 vpadal.u8 q8, q12 137 vpadal.u8 q8, q12 146 vpadal.u8 q8, q12 152 L(end): vpadal.u8 q8, q12 158 vpadal.u8 q8, q12 160 vadd.i16 q8, q8, q [all...] |
| popcount.asm | 55 C We sum into 16 16-bit counters in q8,q9, but at the end we sum them and end 69 vmov.i64 q8, #0 C clear summation register 78 vpadal.u8 d16, d24 C d16/q8 = 0; could just splat 92 vpadal.u8 q8, q12 118 vpadal.u8 q8, q12 124 vpadal.u8 q8, q12 130 L(end): vpadal.u8 q8, q12 134 vpadal.u8 q8, q12 136 vadd.i16 q8, q8, q [all...] |
| /src/crypto/external/apache2/openssl/dist/crypto/aes/asm/ |
| aesv8-armx.pl | 449 ### q8-q9 preloaded round keys except last 7 keys for big size 450 ### q5, q6, q8-q9 preloaded round keys except last 7 keys for only 16 byte 480 vld1.32 {q8-q9},[$key],#32 // load key schedule... 486 aese $dat0,q8 488 vld1.32 {q8},[$key],#16 // load key schedule... 496 aese $dat0,q8 520 vld1.32 {q8-q9},[$key],#32 // load key schedule... 526 aesd $dat0,q8 528 vld1.32 {q8},[$key],#16 // load key schedule... 536 aesd $dat0,q8 [all...] |
| /src/crypto/external/bsd/openssl/dist/crypto/aes/asm/ |
| aesv8-armx.pl | 426 ### q8-q9 preloaded round keys except last 7 keys for big size 427 ### q5, q6, q8-q9 preloaded round keys except last 7 keys for only 16 byte 456 vld1.32 {q8-q9},[$key],#32 // load key schedule... 462 aese $dat0,q8 464 vld1.32 {q8},[$key],#16 // load key schedule... 472 aese $dat0,q8 496 vld1.32 {q8-q9},[$key],#32 // load key schedule... 502 aesd $dat0,q8 504 vld1.32 {q8},[$key],#16 // load key schedule... 512 aesd $dat0,q8 [all...] |