/openbmc/linux/arch/x86/crypto/ |
H A D | glue_helper-asm-avx.S | 8 #define load_8way(src, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 16 vmovdqu (7*16)(src), x7; 18 #define store_8way(dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 26 vmovdqu x7, (7*16)(dst); 28 #define store_cbc_8way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 35 vpxor (6*16)(src), x7, x7; \ 36 store_8way(dst, x0, x1, x2, x3, x4, x5, x6, x7);
|
H A D | aria-aesni-avx-asm_64.S | 200 x4, x5, x6, x7, \ 296 vpxor t2, x7, x7; 317 vgf2p8affineqb $(tf_x2_const), t4, x7, x7; \ 319 vgf2p8affineinvqb $0, t2, x7, x7 348 vpshufb t1, x7, x7; \ 362 vaesdeclast t7, x7, x7; 387 vpxor y3, x7, x7; \ 417 vpxor y3, x7, x7; 420 x4, x5, x6, x7, \ argument 475 x4, x5, x6, x7, \ argument [all …]
|
H A D | glue_helper-asm-avx2.S | 8 #define load_16way(src, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 16 vmovdqu (7*32)(src), x7; 18 #define store_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 26 vmovdqu x7, (7*32)(dst); 28 #define store_cbc_16way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7, t0) \ argument 38 vpxor (6*32+16)(src), x7, x7; \ 39 store_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7);
|
H A D | aria-aesni-avx2-asm_64.S | 216 x4, x5, x6, x7, \ 303 vpxor t0, x7, x7; 324 vgf2p8affineqb $(tf_x2_const), t4, x7, x7; \ 326 vgf2p8affineinvqb $0, t2, x7, x7 380 vpshufb t1, x7, x7; \ 401 vaesdeclast t7##_x, x7##_x, x7##_x; \ 403 vinserti128 $1, t6##_x, x7, x7; \ 428 vpxor y3, x7, x7; \ 458 vpxor y3, x7, x7; 461 x4, x5, x6, x7, \ argument [all …]
|
H A D | camellia-aesni-avx-asm_64.S | 62 vpshufb t4, x7, x7; \ 89 vaesenclast t4, x7, x7; \ 136 vpxor x7, x2, x2; \ 142 vpxor x1, x7, x7; \ 144 vpxor x7, x0, x0; \ 152 vpxor x2, x7, x7; /* note: high and low parts swapped */ \ 171 vpxor t0, x7, x7; \ 172 vpxor 3 * 16(mem_cd), x7, x7; \ 242 vmovdqu x7, 7 * 16(mem_ab); 546 vpxor x0, x7, x7; \ [all …]
|
H A D | camellia-aesni-avx2-asm_64.S | 76 vpshufb t4, x7, x7; \ 107 vaesenclast t4##_x, x7##_x, x7##_x; \ 109 vinserti128 $1, t1##_x, x7, x7; \ 162 vpxor x7, x2, x2; \ 174 vpxor x1, x7, x7; \ 180 vpxor x7, x0, x0; \ 188 vpxor x2, x7, x7; /* note: high and low parts swapped */ \ 217 vpxor t0, x7, x7; \ 218 vpxor 3 * 32(mem_cd), x7, x7; 578 vpxor x0, x7, x7; \ [all …]
|
H A D | aria-gfni-avx512-asm_64.S | 172 x4, x5, x6, x7, \ argument 200 x4, x5, x6, x7, \ 289 vpxorq t0, x7, x7; \ 325 vgf2p8affineqb $(tf_x2_const), t4, x7, x7; \ 327 vgf2p8affineinvqb $0, t2, x7, x7; 349 vgf2p8affineqb $(tf_x2_const), t4, x7, x7; \ 351 vgf2p8affineinvqb $0, t2, x7, x7; \ 389 vpxorq y3, x7, x7; \ 401 vpxorq x7, x3, x3; \ 407 vpxorq x7, y7, y7; \ [all …]
|
/openbmc/linux/arch/arm64/crypto/ |
H A D | sm4-ce-ccm-core.S | 33 mov vctr.d[0], x7; \ 36 adc x7, x7, xzr; 119 ldp x7, x8, [x3] 120 rev x7, x7 211 rev x7, x7 213 stp x7, x8, [x3] 231 ldp x7, x8, [x3] 232 rev x7, x7 323 rev x7, x7 325 stp x7, x8, [x3]
|
H A D | aes-neonbs-core.S | 135 eor \x7, \x7, \t1 236 eor \x7\().16b, \x7\().16b, v23.16b 247 tbl \x7\().16b, {\x7\().16b}, \mask\().16b 265 ext \t7\().16b, \x7\().16b, \x7\().16b, #12 268 eor \x7\().16b, \x7\().16b, \t7\().16b 284 ext \x5\().16b, \x7\().16b, \x7\().16b, #8 310 ext \t7\().16b, \x7\().16b, \x7\().16b, #8 337 eor \x7\().16b, \x7\().16b, \t5\().16b 815 adc x7, x7, xzr 827 CPU_LE( rev x7, x7 ) [all …]
|
H A D | aes-modes.S | 131 enc_prepare w8, x6, x7 384 eor x7, x7, IV_PART 389 mov v1.d[0], x7 436 2: rev x7, IV_PART 437 ins vctr.d[1], x7 441 rev x7, x7 443 mov v1.d[1], x7 555 sub x8, x7, #16 556 csel x7, x7, x8, eq 557 add IN, IN, x7 [all …]
|
H A D | sm4-ce-core.S | 333 add x7, x6, #32 335 sub x7, x7, x5 377 add x7, x6, #32 379 sub x7, x7, x5 575 rev x7, x7 587 adc x7, x7, xzr; 662 rev x7, x7 809 add x7, x6, #32 811 sub x7, x7, x5 976 add x7, x6, #32 [all …]
|
/openbmc/linux/arch/arm64/lib/ |
H A D | crc32.S | 55 and x7, x2, #0x1f 57 cbz x7, 32f // multiple of 32 bytes 59 and x8, x7, #0xf 62 add x1, x1, x7 73 tst x7, #8 77 tst x7, #4 82 tst x7, #2 87 tst x7, #1 90 tst x7, #16
|
H A D | copy_page.S | 30 ldp x6, x7, [x1, #32] 50 stnp x6, x7, [x0, #32 - 256] 51 ldp x6, x7, [x1, #32] 70 stnp x6, x7, [x0, #32 - 256]
|
H A D | mte.S | 140 multitag_transfer_size x7, x5 146 add x0, x0, x7 164 multitag_transfer_size x7, x5 169 add x0, x0, x7
|
/openbmc/u-boot/arch/arm/dts/ |
H A D | rk3288-vyasa-u-boot.dtsi | 8 0x5 0x0 0xa 0x7 0x19 0x24 0xa 0x7 10 0x1 0x7 0x7 0x4 0xc 0x43 0x100 0x0
|
H A D | rk3288-miqi.dts | 20 0x5 0x0 0xa 0x7 0x19 0x24 0xa 0x7 22 0x1 0x7 0x7 0x4 0xc 0x43 0x100 0x0
|
H A D | rk3288-popmetal.dts | 20 0x5 0x0 0xa 0x7 0x19 0x24 0xa 0x7 22 0x1 0x7 0x7 0x4 0xc 0x43 0x100 0x0
|
H A D | rk3288-firefly.dts | 25 0x5 0x0 0xa 0x7 0x19 0x24 0xa 0x7 27 0x1 0x7 0x7 0x4 0xc 0x43 0x100 0x0
|
/openbmc/linux/arch/arm/crypto/ |
H A D | aes-neonbs-core.S | 181 veor \t1, \x5, \x7 189 veor \x7, \x7, \t1 195 veor \t0, \x5, \x7 197 veor \s1, \x7, \x6 293 __tbl \x7, \x7, \mask, \t3 311 vext.8 \t7, \x7, \x7, #12 314 veor.8 \x7, \x7, \t7 330 vext.8 \x5, \x7, \x7, #8 366 veor \x7, \x7, \t7 369 vext.8 \t7, \x7, \x7, #8 [all …]
|
/openbmc/linux/Documentation/devicetree/bindings/gpio/ |
H A D | gpio-stp-xway.yaml | 50 maximum: 0x7 72 maximum: 0x7 94 lantiq,groups = <0x7>; 96 lantiq,phy1 = <0x7>; 97 lantiq,phy2 = <0x7>;
|
/openbmc/linux/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_msg_arm64.h | 62 register u64 x7 asm("x7") = ((u64)X86_IO_MAGIC << 32) | in vmw_port() 70 : "r"(x7) in vmw_port() 95 register u64 x7 asm("x7") = ((u64)X86_IO_MAGIC << 32) | in vmw_port_hb() 103 : "r"(x6), "r"(x7) in vmw_port_hb()
|
/openbmc/linux/lib/crypto/ |
H A D | curve25519-fiat32.c | 236 { const u32 x7 = in1[1]; in fe_add_impl() local 249 out[1] = (x7 + x25); in fe_add_impl() 279 { const u32 x7 = in1[1]; in fe_sub_impl() local 292 out[1] = ((0x3fffffe + x7) - x25); in fe_sub_impl() 322 { const u32 x7 = in1[1]; in fe_mul_impl() local 335 { u64 x41 = (((u64)x23 * x7) + ((u64)x25 * x5)); in fe_mul_impl() 336 { u64 x42 = ((((u64)(0x2 * x25) * x7) + ((u64)x23 * x9)) + ((u64)x27 * x5)); in fe_mul_impl() 337 { u64 x43 = (((((u64)x25 * x9) + ((u64)x27 * x7)) + ((u64)x23 * x11)) + ((u64)x29 * x5)); in fe_mul_impl() 645 { const u32 x7 = in1[1]; in fe_mul_121666_impl() local 658 { u64 x41 = (((u64)x23 * x7) + ((u64)x25 * x5)); in fe_mul_121666_impl() [all …]
|
/openbmc/linux/arch/arm64/kernel/ |
H A D | sleep.S | 79 mrs x7, mpidr_el1 88 compute_mpidr_hash x8, x3, x4, x5, x6, x7, x10 130 compute_mpidr_hash x7, x3, x4, x5, x6, x1, x2 134 ldr x0, [x0, x7, lsl #3]
|
/openbmc/linux/tools/testing/selftests/arm64/fp/ |
H A D | sve-test.S | 222 ldr x7, =scratch 224 mov x0, x7 229 mov x1, x7 233 mov x1, x7 247 ldr x7, =scratch 249 mov x0, x7 254 mov x1, x7 258 mov x1, x7
|
/openbmc/linux/arch/arm64/kvm/hyp/nvhe/ |
H A D | host.S | 29 stp x6, x7, [x0, #CPU_XREG_OFFSET(6)] 81 ldp x6, x7, [x29, #CPU_XREG_OFFSET(6)] 140 mrs x7, hpfar_el2 279 ldp x6, x7, [x18, #CPU_XREG_OFFSET(6)] 291 stp x6, x7, [x18, #CPU_XREG_OFFSET(6)]
|