Searched refs:inc_le128 (Results 1 – 7 of 7) sorted by relevance
/openbmc/linux/arch/arm64/crypto/ |
H A D | sm4-ce-ccm-core.S | 31 #define inc_le128(vctr) \ macro 132 inc_le128(v8) /* +0 */ 133 inc_le128(v9) /* +1 */ 134 inc_le128(v10) /* +2 */ 135 inc_le128(v11) /* +3 */ 164 inc_le128(v8) 179 inc_le128(v8) 244 inc_le128(v8) /* +0 */ 245 inc_le128(v9) /* +1 */ 246 inc_le128(v10) /* +2 */ [all …]
|
H A D | sm4-neon-core.S | 572 #define inc_le128(vctr) \ macro 580 inc_le128(v0) /* +0 */ 581 inc_le128(v1) /* +1 */ 582 inc_le128(v2) /* +2 */ 583 inc_le128(v3) /* +3 */ 584 inc_le128(v4) /* +4 */ 585 inc_le128(v5) /* +5 */ 586 inc_le128(v6) /* +6 */ 587 inc_le128(v7) /* +7 */ 619 inc_le128(v0) /* +0 */ [all …]
|
H A D | sm4-ce-core.S | 582 #define inc_le128(vctr) \ macro 590 inc_le128(v0) /* +0 */ 591 inc_le128(v1) /* +1 */ 592 inc_le128(v2) /* +2 */ 593 inc_le128(v3) /* +3 */ 594 inc_le128(v4) /* +4 */ 595 inc_le128(v5) /* +5 */ 596 inc_le128(v6) /* +6 */ 597 inc_le128(v7) /* +7 */ 627 inc_le128(v0) /* +0 */ [all …]
|
/openbmc/linux/arch/x86/crypto/ |
H A D | sm4-aesni-avx2-asm_64.S | 272 #define inc_le128(x, minus_one, tmp) \ macro 305 inc_le128(RTMP4x, RNOTx, RTMP1x); 335 inc_le128(RTMP0, RNOT, RTMP1); 336 inc_le128(RTMP0, RNOT, RTMP1); 338 inc_le128(RTMP0, RNOT, RTMP1); 339 inc_le128(RTMP0, RNOT, RTMP1); 341 inc_le128(RTMP0, RNOT, RTMP1); 342 inc_le128(RTMP0, RNOT, RTMP1); 344 inc_le128(RTMP0, RNOT, RTMP1); 345 inc_le128(RTMP0, RNOT, RTMP1); [all …]
|
H A D | aria-aesni-avx2-asm_64.S | 54 #define inc_le128(x, minus_one, tmp) \ macro 1042 inc_le128(%xmm7, %xmm0, %xmm4); 1105 inc_le128(%ymm3, %ymm0, %ymm4); 1106 inc_le128(%ymm3, %ymm0, %ymm4); 1108 inc_le128(%ymm3, %ymm0, %ymm4); 1109 inc_le128(%ymm3, %ymm0, %ymm4); 1111 inc_le128(%ymm3, %ymm0, %ymm4); 1112 inc_le128(%ymm3, %ymm0, %ymm4); 1114 inc_le128(%ymm3, %ymm0, %ymm4); 1115 inc_le128(%ymm3, %ymm0, %ymm4); [all …]
|
H A D | sm4-aesni-avx-asm_64.S | 437 #define inc_le128(x, minus_one, tmp) \ macro 444 inc_le128(RTMP0, RNOT, RTMP2); /* +1 */ 446 inc_le128(RTMP0, RNOT, RTMP2); /* +2 */ 448 inc_le128(RTMP0, RNOT, RTMP2); /* +3 */ 450 inc_le128(RTMP0, RNOT, RTMP2); /* +4 */ 452 inc_le128(RTMP0, RNOT, RTMP2); /* +5 */ 454 inc_le128(RTMP0, RNOT, RTMP2); /* +6 */ 456 inc_le128(RTMP0, RNOT, RTMP2); /* +7 */ 458 inc_le128(RTMP0, RNOT, RTMP2); /* +8 */
|
H A D | aria-aesni-avx-asm_64.S | 38 #define inc_le128(x, minus_one, tmp) \ macro 1043 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1045 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1047 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1049 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1051 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1053 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1055 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1066 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ 1068 inc_le128(%xmm3, %xmm0, %xmm5); /* +1 */ [all …]
|