/openbmc/linux/tools/perf/arch/x86/tests/ |
H A D | insn-x86-dat-src.c | 21 asm volatile("rdtsc"); /* Start here */ in main() 25 asm volatile("vcvtph2ps %xmm3,%ymm5"); in main() 31 asm volatile("cmovno %rax,%rbx"); in main() 32 asm volatile("cmovno 0x12345678(%rax),%rcx"); in main() 33 asm volatile("cmovno 0x12345678(%rax),%cx"); in main() 35 asm volatile("cmove %rax,%rbx"); in main() 36 asm volatile("cmove 0x12345678(%rax),%rcx"); in main() 37 asm volatile("cmove 0x12345678(%rax),%cx"); in main() 39 asm volatile("seto 0x12345678(%rax)"); in main() 40 asm volatile("setno 0x12345678(%rax)"); in main() [all …]
|
/openbmc/linux/lib/raid6/ |
H A D | sse2.c | 48 asm volatile("movdqa %0,%%xmm0" : : "m" (raid6_sse_constants.x1d[0])); in raid6_sse21_gen_syndrome() 49 asm volatile("pxor %xmm5,%xmm5"); /* Zero temp */ in raid6_sse21_gen_syndrome() 52 asm volatile("prefetchnta %0" : : "m" (dptr[z0][d])); in raid6_sse21_gen_syndrome() 53 asm volatile("movdqa %0,%%xmm2" : : "m" (dptr[z0][d])); /* P[0] */ in raid6_sse21_gen_syndrome() 54 asm volatile("prefetchnta %0" : : "m" (dptr[z0-1][d])); in raid6_sse21_gen_syndrome() 55 asm volatile("movdqa %xmm2,%xmm4"); /* Q[0] */ in raid6_sse21_gen_syndrome() 56 asm volatile("movdqa %0,%%xmm6" : : "m" (dptr[z0-1][d])); in raid6_sse21_gen_syndrome() 58 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_sse21_gen_syndrome() 59 asm volatile("pcmpgtb %xmm4,%xmm5"); in raid6_sse21_gen_syndrome() 60 asm volatile("paddb %xmm4,%xmm4"); in raid6_sse21_gen_syndrome() [all …]
|
H A D | avx2.c | 46 asm volatile("vmovdqa %0,%%ymm0" : : "m" (raid6_avx2_constants.x1d[0])); in raid6_avx21_gen_syndrome() 47 asm volatile("vpxor %ymm3,%ymm3,%ymm3"); /* Zero temp */ in raid6_avx21_gen_syndrome() 50 asm volatile("prefetchnta %0" : : "m" (dptr[z0][d])); in raid6_avx21_gen_syndrome() 51 asm volatile("vmovdqa %0,%%ymm2" : : "m" (dptr[z0][d]));/* P[0] */ in raid6_avx21_gen_syndrome() 52 asm volatile("prefetchnta %0" : : "m" (dptr[z0-1][d])); in raid6_avx21_gen_syndrome() 53 asm volatile("vmovdqa %ymm2,%ymm4");/* Q[0] */ in raid6_avx21_gen_syndrome() 54 asm volatile("vmovdqa %0,%%ymm6" : : "m" (dptr[z0-1][d])); in raid6_avx21_gen_syndrome() 56 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_avx21_gen_syndrome() 57 asm volatile("vpcmpgtb %ymm4,%ymm3,%ymm5"); in raid6_avx21_gen_syndrome() 58 asm volatile("vpaddb %ymm4,%ymm4,%ymm4"); in raid6_avx21_gen_syndrome() [all …]
|
H A D | recov_ssse3.c | 54 asm volatile("movdqa %0,%%xmm7" : : "m" (x0f[0])); in raid6_2data_recov_ssse3() 57 asm volatile("movdqa %0,%%xmm6" : : "m" (qmul[0])); in raid6_2data_recov_ssse3() 58 asm volatile("movdqa %0,%%xmm14" : : "m" (pbmul[0])); in raid6_2data_recov_ssse3() 59 asm volatile("movdqa %0,%%xmm15" : : "m" (pbmul[16])); in raid6_2data_recov_ssse3() 67 asm volatile("movdqa %0,%%xmm1" : : "m" (q[0])); in raid6_2data_recov_ssse3() 68 asm volatile("movdqa %0,%%xmm9" : : "m" (q[16])); in raid6_2data_recov_ssse3() 69 asm volatile("movdqa %0,%%xmm0" : : "m" (p[0])); in raid6_2data_recov_ssse3() 70 asm volatile("movdqa %0,%%xmm8" : : "m" (p[16])); in raid6_2data_recov_ssse3() 71 asm volatile("pxor %0,%%xmm1" : : "m" (dq[0])); in raid6_2data_recov_ssse3() 72 asm volatile("pxor %0,%%xmm9" : : "m" (dq[16])); in raid6_2data_recov_ssse3() [all …]
|
H A D | recov_loongarch_simd.c | 69 asm volatile("vld $vr20, %0" : : "m" (qmul[0])); in raid6_2data_recov_lsx() 70 asm volatile("vld $vr21, %0" : : "m" (qmul[16])); in raid6_2data_recov_lsx() 71 asm volatile("vld $vr22, %0" : : "m" (pbmul[0])); in raid6_2data_recov_lsx() 72 asm volatile("vld $vr23, %0" : : "m" (pbmul[16])); in raid6_2data_recov_lsx() 76 asm volatile("vld $vr4, %0" : : "m" (q[0])); in raid6_2data_recov_lsx() 77 asm volatile("vld $vr5, %0" : : "m" (q[16])); in raid6_2data_recov_lsx() 78 asm volatile("vld $vr6, %0" : : "m" (q[32])); in raid6_2data_recov_lsx() 79 asm volatile("vld $vr7, %0" : : "m" (q[48])); in raid6_2data_recov_lsx() 81 asm volatile("vld $vr8, %0" : : "m" (dq[0])); in raid6_2data_recov_lsx() 82 asm volatile("vld $vr9, %0" : : "m" (dq[16])); in raid6_2data_recov_lsx() [all …]
|
H A D | recov_avx2.c | 53 asm volatile("vpbroadcastb %0, %%ymm7" : : "m" (x0f)); in raid6_2data_recov_avx2() 57 asm volatile("vmovdqa %0, %%ymm1" : : "m" (q[0])); in raid6_2data_recov_avx2() 58 asm volatile("vmovdqa %0, %%ymm9" : : "m" (q[32])); in raid6_2data_recov_avx2() 59 asm volatile("vmovdqa %0, %%ymm0" : : "m" (p[0])); in raid6_2data_recov_avx2() 60 asm volatile("vmovdqa %0, %%ymm8" : : "m" (p[32])); in raid6_2data_recov_avx2() 61 asm volatile("vpxor %0, %%ymm1, %%ymm1" : : "m" (dq[0])); in raid6_2data_recov_avx2() 62 asm volatile("vpxor %0, %%ymm9, %%ymm9" : : "m" (dq[32])); in raid6_2data_recov_avx2() 63 asm volatile("vpxor %0, %%ymm0, %%ymm0" : : "m" (dp[0])); in raid6_2data_recov_avx2() 64 asm volatile("vpxor %0, %%ymm8, %%ymm8" : : "m" (dp[32])); in raid6_2data_recov_avx2() 73 asm volatile("vbroadcasti128 %0, %%ymm4" : : "m" (qmul[0])); in raid6_2data_recov_avx2() [all …]
|
H A D | loongarch_simd.c | 54 asm volatile("vld $vr0, %0" : : "m"(dptr[z0][d+0*NSIZE])); in raid6_lsx_gen_syndrome() 55 asm volatile("vld $vr1, %0" : : "m"(dptr[z0][d+1*NSIZE])); in raid6_lsx_gen_syndrome() 56 asm volatile("vld $vr2, %0" : : "m"(dptr[z0][d+2*NSIZE])); in raid6_lsx_gen_syndrome() 57 asm volatile("vld $vr3, %0" : : "m"(dptr[z0][d+3*NSIZE])); in raid6_lsx_gen_syndrome() 58 asm volatile("vori.b $vr4, $vr0, 0"); in raid6_lsx_gen_syndrome() 59 asm volatile("vori.b $vr5, $vr1, 0"); in raid6_lsx_gen_syndrome() 60 asm volatile("vori.b $vr6, $vr2, 0"); in raid6_lsx_gen_syndrome() 61 asm volatile("vori.b $vr7, $vr3, 0"); in raid6_lsx_gen_syndrome() 64 asm volatile("vld $vr8, %0" : : "m"(dptr[z][d+0*NSIZE])); in raid6_lsx_gen_syndrome() 65 asm volatile("vld $vr9, %0" : : "m"(dptr[z][d+1*NSIZE])); in raid6_lsx_gen_syndrome() [all …]
|
H A D | sse1.c | 52 asm volatile("movq %0,%%mm0" : : "m" (raid6_mmx_constants.x1d)); in raid6_sse11_gen_syndrome() 53 asm volatile("pxor %mm5,%mm5"); /* Zero temp */ in raid6_sse11_gen_syndrome() 56 asm volatile("prefetchnta %0" : : "m" (dptr[z0][d])); in raid6_sse11_gen_syndrome() 57 asm volatile("movq %0,%%mm2" : : "m" (dptr[z0][d])); /* P[0] */ in raid6_sse11_gen_syndrome() 58 asm volatile("prefetchnta %0" : : "m" (dptr[z0-1][d])); in raid6_sse11_gen_syndrome() 59 asm volatile("movq %mm2,%mm4"); /* Q[0] */ in raid6_sse11_gen_syndrome() 60 asm volatile("movq %0,%%mm6" : : "m" (dptr[z0-1][d])); in raid6_sse11_gen_syndrome() 62 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_sse11_gen_syndrome() 63 asm volatile("pcmpgtb %mm4,%mm5"); in raid6_sse11_gen_syndrome() 64 asm volatile("paddb %mm4,%mm4"); in raid6_sse11_gen_syndrome() [all …]
|
H A D | mmx.c | 47 asm volatile("movq %0,%%mm0" : : "m" (raid6_mmx_constants.x1d)); in raid6_mmx1_gen_syndrome() 48 asm volatile("pxor %mm5,%mm5"); /* Zero temp */ in raid6_mmx1_gen_syndrome() 51 asm volatile("movq %0,%%mm2" : : "m" (dptr[z0][d])); /* P[0] */ in raid6_mmx1_gen_syndrome() 52 asm volatile("movq %mm2,%mm4"); /* Q[0] */ in raid6_mmx1_gen_syndrome() 54 asm volatile("movq %0,%%mm6" : : "m" (dptr[z][d])); in raid6_mmx1_gen_syndrome() 55 asm volatile("pcmpgtb %mm4,%mm5"); in raid6_mmx1_gen_syndrome() 56 asm volatile("paddb %mm4,%mm4"); in raid6_mmx1_gen_syndrome() 57 asm volatile("pand %mm0,%mm5"); in raid6_mmx1_gen_syndrome() 58 asm volatile("pxor %mm5,%mm4"); in raid6_mmx1_gen_syndrome() 59 asm volatile("pxor %mm5,%mm5"); in raid6_mmx1_gen_syndrome() [all …]
|
/openbmc/linux/arch/ia64/include/uapi/asm/ |
H A D | gcc_intrin.h | 13 /* define this macro to get some asm stmts included in 'c' files */ 18 #define ia64_barrier() asm volatile ("":::"memory") 20 #define ia64_stop() asm volatile (";;"::) 22 #define ia64_invala_gr(regnum) asm volatile ("invala.e r%0" :: "i"(regnum)) 24 #define ia64_invala_fr(regnum) asm volatile ("invala.e f%0" :: "i"(regnum)) 26 #define ia64_flushrs() asm volatile ("flushrs;;":::"memory") 28 #define ia64_loadrs() asm volatile ("loadrs;;":::"memory") 38 asm volatile ("mov psr.l=%0" :: "r"(val) : "memory"); \ 41 asm volatile ("mov ar%0=%1" :: \ 46 asm volatile ("mov cr%0=%1" :: \ [all …]
|
/openbmc/qemu/scripts/ |
H A D | update-linux-headers.sh | 65 -e 'asm/setup_data.h' \ 66 -e 'asm/kvm_para.h' \ 77 # asm-* header that they care about 78 arch_cmd='/<asm\/\([^>]*\)>/d' 80 arch_cmd='s/<asm\/\([^>]*\)>/"standard-headers\/asm-'$arch'\/\1"/' 114 if ! [ -e "$linux/arch/$arch/include/asm/kvm.h" ] && 115 ! [ -e "$linux/arch/$arch/include/uapi/asm/kvm.h" ] ; then 128 rm -rf "$output/linux-headers/asm-$arch" 129 mkdir -p "$output/linux-headers/asm-$arch" 131 if test -f "$hdrdir/include/asm/$header"; then [all …]
|
/openbmc/linux/tools/perf/ |
H A D | check-headers.sh | 36 "arch/x86/include/asm/disabled-features.h" 37 "arch/x86/include/asm/required-features.h" 38 "arch/x86/include/asm/cpufeatures.h" 39 "arch/x86/include/asm/inat_types.h" 40 "arch/x86/include/asm/emulate_prefix.h" 41 "arch/x86/include/asm/irq_vectors.h" 42 "arch/x86/include/asm/msr-index.h" 43 "arch/x86/include/uapi/asm/prctl.h" 46 "arch/arm/include/uapi/asm/perf_regs.h" 47 "arch/arm64/include/uapi/asm/perf_regs.h" [all …]
|
/openbmc/linux/arch/s390/kernel/ |
H A D | fpu.c | 11 #include <asm/fpu/types.h> 12 #include <asm/fpu/api.h> 13 #include <asm/vx-insn.h> 25 asm volatile("stfpc %0" : "=Q" (state->fpc)); in __kernel_fpu_begin() 30 asm volatile("std 0,%0" : "=Q" (state->fprs[0])); in __kernel_fpu_begin() 31 asm volatile("std 1,%0" : "=Q" (state->fprs[1])); in __kernel_fpu_begin() 32 asm volatile("std 2,%0" : "=Q" (state->fprs[2])); in __kernel_fpu_begin() 33 asm volatile("std 3,%0" : "=Q" (state->fprs[3])); in __kernel_fpu_begin() 34 asm volatile("std 4,%0" : "=Q" (state->fprs[4])); in __kernel_fpu_begin() 35 asm volatile("std 5,%0" : "=Q" (state->fprs[5])); in __kernel_fpu_begin() [all …]
|
/openbmc/linux/arch/x86/kvm/ |
H A D | fpu.h | 6 #include <asm/fpu/api.h> 21 case 0: asm("movdqa %%xmm0, %0" : "=m"(*data)); break; in _kvm_read_sse_reg() 22 case 1: asm("movdqa %%xmm1, %0" : "=m"(*data)); break; in _kvm_read_sse_reg() 23 case 2: asm("movdqa %%xmm2, %0" : "=m"(*data)); break; in _kvm_read_sse_reg() 24 case 3: asm("movdqa %%xmm3, %0" : "=m"(*data)); break; in _kvm_read_sse_reg() 25 case 4: asm("movdqa %%xmm4, %0" : "=m"(*data)); break; in _kvm_read_sse_reg() 26 case 5: asm("movdqa %%xmm5, %0" : "=m"(*data)); break; in _kvm_read_sse_reg() 27 case 6: asm("movdqa %%xmm6, %0" : "=m"(*data)); break; in _kvm_read_sse_reg() 28 case 7: asm("movdqa %%xmm7, %0" : "=m"(*data)); break; in _kvm_read_sse_reg() 30 case 8: asm("movdqa %%xmm8, %0" : "=m"(*data)); break; in _kvm_read_sse_reg() [all …]
|
/openbmc/linux/arch/csky/include/asm/vdso/ |
H A D | gettimeofday.h | 8 #include <asm/barrier.h> 9 #include <asm/unistd.h> 19 register struct __kernel_old_timeval *tv asm("a0") = _tv; in gettimeofday_fallback() 20 register struct timezone *tz asm("a1") = _tz; in gettimeofday_fallback() 21 register long ret asm("a0"); in gettimeofday_fallback() 22 register long nr asm(syscallid) = __NR_gettimeofday; in gettimeofday_fallback() 24 asm volatile ("trap 0\n" in gettimeofday_fallback() 35 register clockid_t clkid asm("a0") = _clkid; in clock_gettime_fallback() 36 register struct __kernel_timespec *ts asm("a1") = _ts; in clock_gettime_fallback() 37 register long ret asm("a0"); in clock_gettime_fallback() [all …]
|
/openbmc/linux/arch/arm/include/asm/vdso/ |
H A D | gettimeofday.h | 10 #include <asm/barrier.h> 11 #include <asm/errno.h> 12 #include <asm/unistd.h> 13 #include <asm/vdso/cp15.h> 24 register struct timezone *tz asm("r1") = _tz; in gettimeofday_fallback() 25 register struct __kernel_old_timeval *tv asm("r0") = _tv; in gettimeofday_fallback() 26 register long ret asm ("r0"); in gettimeofday_fallback() 27 register long nr asm("r7") = __NR_gettimeofday; in gettimeofday_fallback() 29 asm volatile( in gettimeofday_fallback() 42 register struct __kernel_timespec *ts asm("r1") = _ts; in clock_gettime_fallback() [all …]
|
/openbmc/linux/arch/arm64/include/asm/vdso/ |
H A D | compat_gettimeofday.h | 10 #include <asm/barrier.h> 11 #include <asm/unistd.h> 12 #include <asm/errno.h> 14 #include <asm/vdso/compat_barrier.h> 24 register struct timezone *tz asm("r1") = _tz; in gettimeofday_fallback() 25 register struct __kernel_old_timeval *tv asm("r0") = _tv; in gettimeofday_fallback() 26 register long ret asm ("r0"); in gettimeofday_fallback() 27 register long nr asm("r7") = __NR_compat_gettimeofday; in gettimeofday_fallback() 29 asm volatile( in gettimeofday_fallback() 41 register struct __kernel_timespec *ts asm("r1") = _ts; in clock_gettime_fallback() [all …]
|
H A D | gettimeofday.h | 10 #include <asm/alternative.h> 11 #include <asm/barrier.h> 12 #include <asm/unistd.h> 13 #include <asm/sysreg.h> 21 register struct timezone *tz asm("x1") = _tz; in gettimeofday_fallback() 22 register struct __kernel_old_timeval *tv asm("x0") = _tv; in gettimeofday_fallback() 23 register long ret asm ("x0"); in gettimeofday_fallback() 24 register long nr asm("x8") = __NR_gettimeofday; in gettimeofday_fallback() 26 asm volatile( in gettimeofday_fallback() 38 register struct __kernel_timespec *ts asm("x1") = _ts; in clock_gettime_fallback() [all …]
|
/openbmc/linux/arch/mips/include/asm/vdso/ |
H A D | gettimeofday.h | 16 #include <asm/vdso/vdso.h> 17 #include <asm/clocksource.h> 18 #include <asm/unistd.h> 19 #include <asm/vdso.h> 33 register struct timezone *tz asm("a1") = _tz; in gettimeofday_fallback() 34 register struct __kernel_old_timeval *tv asm("a0") = _tv; in gettimeofday_fallback() 35 register long ret asm("v0"); in gettimeofday_fallback() 36 register long nr asm("v0") = __NR_gettimeofday; in gettimeofday_fallback() 37 register long error asm("a3"); in gettimeofday_fallback() 39 asm volatile( in gettimeofday_fallback() [all …]
|
/openbmc/qemu/tests/tcg/tricore/ |
H A D | Makefile.softmmu-target | 2 ASM_TESTS_PATH = $(TESTS_PATH)/asm 9 TESTS += test_abs.asm.tst 10 TESTS += test_bmerge.asm.tst 11 TESTS += test_clz.asm.tst 12 TESTS += test_crcn.asm.tst 13 TESTS += test_dextr.asm.tst 14 TESTS += test_dvstep.asm.tst 15 TESTS += test_fadd.asm.tst 16 TESTS += test_fmul.asm.tst 17 TESTS += test_ftohp.asm.tst [all …]
|
/openbmc/u-boot/arch/arm/include/asm/arch-mxs/ |
H A D | imx-regs.h | 12 #include <asm/mach-imx/regs-apbh.h> 13 #include <asm/arch/regs-base.h> 14 #include <asm/mach-imx/regs-bch.h> 15 #include <asm/arch/regs-digctl.h> 16 #include <asm/mach-imx/regs-gpmi.h> 17 #include <asm/mach-imx/regs-lcdif.h> 18 #include <asm/arch/regs-i2c.h> 19 #include <asm/arch/regs-lradc.h> 20 #include <asm/arch/regs-ocotp.h> 21 #include <asm/arch/regs-pinctrl.h> [all …]
|
/openbmc/linux/arch/x86/include/asm/ |
H A D | kexec.h | 27 #include <asm/page.h> 28 #include <asm/ptrace.h> 29 #include <asm/bootparam.h> 78 asm volatile("movl %%ebx,%0" : "=m"(newregs->bx)); in crash_setup_regs() 79 asm volatile("movl %%ecx,%0" : "=m"(newregs->cx)); in crash_setup_regs() 80 asm volatile("movl %%edx,%0" : "=m"(newregs->dx)); in crash_setup_regs() 81 asm volatile("movl %%esi,%0" : "=m"(newregs->si)); in crash_setup_regs() 82 asm volatile("movl %%edi,%0" : "=m"(newregs->di)); in crash_setup_regs() 83 asm volatile("movl %%ebp,%0" : "=m"(newregs->bp)); in crash_setup_regs() 84 asm volatile("movl %%eax,%0" : "=m"(newregs->ax)); in crash_setup_regs() [all …]
|
/openbmc/linux/tools/testing/selftests/bpf/progs/ |
H A D | verifier_sdiv.c | 15 asm volatile (" \ in sdiv32_non_zero_imm_1() 27 asm volatile (" \ in sdiv32_non_zero_imm_2() 39 asm volatile (" \ in sdiv32_non_zero_imm_3() 51 asm volatile (" \ in sdiv32_non_zero_imm_4() 63 asm volatile (" \ in sdiv32_non_zero_imm_5() 75 asm volatile (" \ in sdiv32_non_zero_imm_6() 87 asm volatile (" \ in sdiv32_non_zero_imm_7() 99 asm volatile (" \ in sdiv32_non_zero_imm_8() 111 asm volatile (" \ in sdiv32_non_zero_reg_1() 124 asm volatile (" \ in sdiv32_non_zero_reg_2() [all …]
|
/openbmc/qemu/tests/tcg/i386/ |
H A D | test-i386-ssse3.c | 20 asm volatile ("movq (%0), %%mm0" : : "r" (ehlo) : "mm0", "mm1"); in main() 21 asm volatile ("movq %0, %%mm1" : : "m" (mask)); in main() 22 asm volatile ("pshufb %mm1, %mm0"); in main() 23 asm volatile ("movq %%mm0, %0" : "=m" (hello)); in main() 27 asm volatile ("movq (%0), %%mm0" : : "r" (ehlo) : "mm0"); in main() 28 asm volatile ("pshufb %0, %%mm0" : : "m" (mask)); in main() 29 asm volatile ("movq %%mm0, %0" : "=m" (hello)); in main() 33 asm volatile ("movq %0, %%mm0" : : "r" (a) : "mm0"); in main() 34 asm volatile ("phsubsw %0, %%mm0" : : "m" (b)); in main() 35 asm volatile ("movq %%mm0, %0" : "=m" (a)); in main() [all …]
|
/openbmc/qemu/tests/tcg/m68k/ |
H A D | trap.c | 51 asm volatile(FMT2_STR("0:\tchk %0, %1") : : "d"(0), "d"(-1), FMT_INS); in main() 57 asm volatile(FMT2_STR("0:\tchk2.l %0, %1") in main() 62 asm volatile(FMT2_STR("cmp.l %0, %1\n0:\ttrapv") in main() 66 asm volatile(FMT2_STR("cmp.l %0, %0\n0:\ttrapeq") in main() 70 asm volatile(FMT2_STR("cmp.l %0, %0\n0:\ttrapeq.w #0x1234") in main() 74 asm volatile(FMT2_STR("cmp.l %0, %0\n0:\ttrapeq.l #0x12345678") in main() 78 asm volatile(FMT2_STR("fcmp.x %0, %0\n0:\tftrapeq") in main() 84 asm volatile(FMT2_STR("0:\tdivs.w %1, %0") in main() 88 asm volatile(FMT2_STR("0:\tdivsl.l %2, %1:%0") in main() 94 asm volatile(FMT0_STR("trap #1") : : FMT_INS); in main() [all …]
|