Lines Matching refs:XMM8

969 … REP num_initial_blocks T1 T2 T3 T4 T5 CTR XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 T6 T_key ENC_DEC
1048 # XMM8 has the combined result here
1050 vmovdqa \XMM8, TMP1(%rsp)
1051 vmovdqa \XMM8, \T3
1087 vmovdqa \CTR, \XMM8
1088 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
1098 vpxor \T_key, \XMM8, \XMM8
1111 vaesenc \T_key, \XMM8, \XMM8
1124 vaesenclast \T_key, \XMM8, \XMM8
1176 vpxor \T1, \XMM8, \XMM8
1177 vmovdqu \XMM8, 16*7(arg3 , %r11)
1179 vmovdqa \T1, \XMM8
1192 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
1204 …_PARALLEL_AVX REP T1 T2 T3 T4 T5 T6 CTR XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 T7 loop_idx ENC_DEC
1213 vmovdqa \XMM8, TMP8(%rsp)
1223 vpaddd ONE(%rip), \XMM7, \XMM8
1224 vmovdqa \XMM8, \CTR
1233 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
1242 vpaddd ONEf(%rip), \XMM7, \XMM8
1243 vmovdqa \XMM8, \CTR
1257 vpxor \T1, \XMM8, \XMM8
1273 vaesenc \T1, \XMM8, \XMM8
1283 vaesenc \T1, \XMM8, \XMM8
1306 vaesenc \T1, \XMM8, \XMM8
1329 vaesenc \T1, \XMM8, \XMM8
1354 vaesenc \T1, \XMM8, \XMM8
1377 vaesenc \T1, \XMM8, \XMM8
1401 vaesenc \T1, \XMM8, \XMM8
1425 vaesenc \T1, \XMM8, \XMM8
1450 vaesenc \T5, \XMM8, \XMM8
1481 vaesenc \T5, \XMM8, \XMM8
1537 vmovdqu \XMM8, 16*7(arg3,%r11) # Write to the Ciphertext buffer
1560 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
1571 .macro GHASH_LAST_8_AVX T1 T2 T3 T4 T5 T6 T7 XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 argument
1677 vpshufd $0b01001110, \XMM8, \T2
1678 vpxor \XMM8, \T2, \T2
1680 vpclmulqdq $0x11, \T5, \XMM8, \T4
1683 vpclmulqdq $0x00, \T5, \XMM8, \T4
1917 … num_initial_blocks T1 T2 T3 T4 T5 CTR XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 T6 T_key ENC_DEC VER
1998 # XMM8 has the combined result here
2000 vmovdqa \XMM8, TMP1(%rsp)
2001 vmovdqa \XMM8, \T3
2037 vmovdqa \CTR, \XMM8
2038 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
2048 vpxor \T_key, \XMM8, \XMM8
2061 vaesenc \T_key, \XMM8, \XMM8
2075 vaesenclast \T_key, \XMM8, \XMM8
2127 vpxor \T1, \XMM8, \XMM8
2128 vmovdqu \XMM8, 16*7(arg3 , %r11)
2130 vmovdqa \T1, \XMM8
2144 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
2159 …PARALLEL_AVX2 REP T1 T2 T3 T4 T5 T6 CTR XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 T7 loop_idx ENC_DEC
2168 vmovdqa \XMM8, TMP8(%rsp)
2178 vpaddd ONE(%rip), \XMM7, \XMM8
2179 vmovdqa \XMM8, \CTR
2188 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
2197 vpaddd ONEf(%rip), \XMM7, \XMM8
2198 vmovdqa \XMM8, \CTR
2212 vpxor \T1, \XMM8, \XMM8
2228 vaesenc \T1, \XMM8, \XMM8
2238 vaesenc \T1, \XMM8, \XMM8
2258 vaesenc \T1, \XMM8, \XMM8
2282 vaesenc \T1, \XMM8, \XMM8
2308 vaesenc \T1, \XMM8, \XMM8
2332 vaesenc \T1, \XMM8, \XMM8
2357 vaesenc \T1, \XMM8, \XMM8
2381 vaesenc \T1, \XMM8, \XMM8
2408 vaesenc \T5, \XMM8, \XMM8
2438 vaesenc \T5, \XMM8, \XMM8
2488 vmovdqu \XMM8, 16*7(arg3,%r11) # Write to the Ciphertext buffer
2510 vpshufb SHUF_MASK(%rip), \XMM8, \XMM8 # perform a 16Byte swap
2521 .macro GHASH_LAST_8_AVX2 T1 T2 T3 T4 T5 T6 T7 XMM1 XMM2 XMM3 XMM4 XMM5 XMM6 XMM7 XMM8 argument
2648 vpshufd $0b01001110, \XMM8, \T2
2650 vpxor \XMM8, \T2, \T2
2653 vpclmulqdq $0x11, \T5, \XMM8, \T4
2656 vpclmulqdq $0x00, \T5, \XMM8, \T4