// SPDX-License-Identifier: GPL-2.0-only // Copyright (C) 2021 ARM Limited. // // Assembly portion of the syscall ABI test // // Load values from memory into registers, invoke a syscall and save the // register values back to memory for later checking. The syscall to be // invoked is configured in x8 of the input GPR data. // // x0: SVE VL, 0 for FP only // x1: SME VL // // GPRs: gpr_in, gpr_out // FPRs: fpr_in, fpr_out // Zn: z_in, z_out // Pn: p_in, p_out // FFR: ffr_in, ffr_out // ZA: za_in, za_out // SVCR: svcr_in, svcr_out #include "syscall-abi.h" .arch_extension sve /* * LDR (vector to ZA array): * LDR ZA[\nw, #\offset], [X\nxbase, #\offset, MUL VL] */ .macro _ldr_za nw, nxbase, offset=0 .inst 0xe1000000 \ | (((\nw) & 3) << 13) \ | ((\nxbase) << 5) \ | ((\offset) & 7) .endm /* * STR (vector from ZA array): * STR ZA[\nw, #\offset], [X\nxbase, #\offset, MUL VL] */ .macro _str_za nw, nxbase, offset=0 .inst 0xe1200000 \ | (((\nw) & 3) << 13) \ | ((\nxbase) << 5) \ | ((\offset) & 7) .endm .globl do_syscall do_syscall: // Store callee saved registers x19-x29 (80 bytes) plus x0 and x1 stp x29, x30, [sp, #-112]! mov x29, sp stp x0, x1, [sp, #16] stp x19, x20, [sp, #32] stp x21, x22, [sp, #48] stp x23, x24, [sp, #64] stp x25, x26, [sp, #80] stp x27, x28, [sp, #96] // Set SVCR if we're doing SME cbz x1, 1f adrp x2, svcr_in ldr x2, [x2, :lo12:svcr_in] msr S3_3_C4_C2_2, x2 1: // Load ZA if it's enabled - uses x12 as scratch due to SME LDR tbz x2, #SVCR_ZA_SHIFT, 1f mov w12, #0 ldr x2, =za_in 2: _ldr_za 12, 2 add x2, x2, x1 add x12, x12, #1 cmp x1, x12 bne 2b 1: // Load GPRs x8-x28, and save our SP/FP for later comparison ldr x2, =gpr_in add x2, x2, #64 ldp x8, x9, [x2], #16 ldp x10, x11, [x2], #16 ldp x12, x13, [x2], #16 ldp x14, x15, [x2], #16 ldp x16, x17, [x2], #16 ldp x18, x19, [x2], #16 ldp x20, x21, [x2], #16 ldp x22, x23, [x2], #16 ldp x24, x25, [x2], #16 ldp x26, x27, [x2], #16 ldr x28, [x2], #8 str x29, [x2], #8 // FP str x30, [x2], #8 // LR // Load FPRs if we're not doing SVE cbnz x0, 1f ldr x2, =fpr_in ldp q0, q1, [x2] ldp q2, q3, [x2, #16 * 2] ldp q4, q5, [x2, #16 * 4] ldp q6, q7, [x2, #16 * 6] ldp q8, q9, [x2, #16 * 8] ldp q10, q11, [x2, #16 * 10] ldp q12, q13, [x2, #16 * 12] ldp q14, q15, [x2, #16 * 14] ldp q16, q17, [x2, #16 * 16] ldp q18, q19, [x2, #16 * 18] ldp q20, q21, [x2, #16 * 20] ldp q22, q23, [x2, #16 * 22] ldp q24, q25, [x2, #16 * 24] ldp q26, q27, [x2, #16 * 26] ldp q28, q29, [x2, #16 * 28] ldp q30, q31, [x2, #16 * 30] 1: // Load the SVE registers if we're doing SVE/SME cbz x0, 1f ldr x2, =z_in ldr z0, [x2, #0, MUL VL] ldr z1, [x2, #1, MUL VL] ldr z2, [x2, #2, MUL VL] ldr z3, [x2, #3, MUL VL] ldr z4, [x2, #4, MUL VL] ldr z5, [x2, #5, MUL VL] ldr z6, [x2, #6, MUL VL] ldr z7, [x2, #7, MUL VL] ldr z8, [x2, #8, MUL VL] ldr z9, [x2, #9, MUL VL] ldr z10, [x2, #10, MUL VL] ldr z11, [x2, #11, MUL VL] ldr z12, [x2, #12, MUL VL] ldr z13, [x2, #13, MUL VL] ldr z14, [x2, #14, MUL VL] ldr z15, [x2, #15, MUL VL] ldr z16, [x2, #16, MUL VL] ldr z17, [x2, #17, MUL VL] ldr z18, [x2, #18, MUL VL] ldr z19, [x2, #19, MUL VL] ldr z20, [x2, #20, MUL VL] ldr z21, [x2, #21, MUL VL] ldr z22, [x2, #22, MUL VL] ldr z23, [x2, #23, MUL VL] ldr z24, [x2, #24, MUL VL] ldr z25, [x2, #25, MUL VL] ldr z26, [x2, #26, MUL VL] ldr z27, [x2, #27, MUL VL] ldr z28, [x2, #28, MUL VL] ldr z29, [x2, #29, MUL VL] ldr z30, [x2, #30, MUL VL] ldr z31, [x2, #31, MUL VL] // Only set a non-zero FFR, test patterns must be zero since the // syscall should clear it - this lets us handle FA64. ldr x2, =ffr_in ldr p0, [x2] ldr x2, [x2, #0] cbz x2, 2f wrffr p0.b 2: ldr x2, =p_in ldr p0, [x2, #0, MUL VL] ldr p1, [x2, #1, MUL VL] ldr p2, [x2, #2, MUL VL] ldr p3, [x2, #3, MUL VL] ldr p4, [x2, #4, MUL VL] ldr p5, [x2, #5, MUL VL] ldr p6, [x2, #6, MUL VL] ldr p7, [x2, #7, MUL VL] ldr p8, [x2, #8, MUL VL] ldr p9, [x2, #9, MUL VL] ldr p10, [x2, #10, MUL VL] ldr p11, [x2, #11, MUL VL] ldr p12, [x2, #12, MUL VL] ldr p13, [x2, #13, MUL VL] ldr p14, [x2, #14, MUL VL] ldr p15, [x2, #15, MUL VL] 1: // Do the syscall svc #0 // Save GPRs x8-x30 ldr x2, =gpr_out add x2, x2, #64 stp x8, x9, [x2], #16 stp x10, x11, [x2], #16 stp x12, x13, [x2], #16 stp x14, x15, [x2], #16 stp x16, x17, [x2], #16 stp x18, x19, [x2], #16 stp x20, x21, [x2], #16 stp x22, x23, [x2], #16 stp x24, x25, [x2], #16 stp x26, x27, [x2], #16 stp x28, x29, [x2], #16 str x30, [x2] // Restore x0 and x1 for feature checks ldp x0, x1, [sp, #16] // Save FPSIMD state ldr x2, =fpr_out stp q0, q1, [x2] stp q2, q3, [x2, #16 * 2] stp q4, q5, [x2, #16 * 4] stp q6, q7, [x2, #16 * 6] stp q8, q9, [x2, #16 * 8] stp q10, q11, [x2, #16 * 10] stp q12, q13, [x2, #16 * 12] stp q14, q15, [x2, #16 * 14] stp q16, q17, [x2, #16 * 16] stp q18, q19, [x2, #16 * 18] stp q20, q21, [x2, #16 * 20] stp q22, q23, [x2, #16 * 22] stp q24, q25, [x2, #16 * 24] stp q26, q27, [x2, #16 * 26] stp q28, q29, [x2, #16 * 28] stp q30, q31, [x2, #16 * 30] // Save SVCR if we're doing SME cbz x1, 1f mrs x2, S3_3_C4_C2_2 adrp x3, svcr_out str x2, [x3, :lo12:svcr_out] 1: // Save ZA if it's enabled - uses x12 as scratch due to SME STR tbz x2, #SVCR_ZA_SHIFT, 1f mov w12, #0 ldr x2, =za_out 2: _str_za 12, 2 add x2, x2, x1 add x12, x12, #1 cmp x1, x12 bne 2b 1: // Save the SVE state if we have some cbz x0, 1f ldr x2, =z_out str z0, [x2, #0, MUL VL] str z1, [x2, #1, MUL VL] str z2, [x2, #2, MUL VL] str z3, [x2, #3, MUL VL] str z4, [x2, #4, MUL VL] str z5, [x2, #5, MUL VL] str z6, [x2, #6, MUL VL] str z7, [x2, #7, MUL VL] str z8, [x2, #8, MUL VL] str z9, [x2, #9, MUL VL] str z10, [x2, #10, MUL VL] str z11, [x2, #11, MUL VL] str z12, [x2, #12, MUL VL] str z13, [x2, #13, MUL VL] str z14, [x2, #14, MUL VL] str z15, [x2, #15, MUL VL] str z16, [x2, #16, MUL VL] str z17, [x2, #17, MUL VL] str z18, [x2, #18, MUL VL] str z19, [x2, #19, MUL VL] str z20, [x2, #20, MUL VL] str z21, [x2, #21, MUL VL] str z22, [x2, #22, MUL VL] str z23, [x2, #23, MUL VL] str z24, [x2, #24, MUL VL] str z25, [x2, #25, MUL VL] str z26, [x2, #26, MUL VL] str z27, [x2, #27, MUL VL] str z28, [x2, #28, MUL VL] str z29, [x2, #29, MUL VL] str z30, [x2, #30, MUL VL] str z31, [x2, #31, MUL VL] ldr x2, =p_out str p0, [x2, #0, MUL VL] str p1, [x2, #1, MUL VL] str p2, [x2, #2, MUL VL] str p3, [x2, #3, MUL VL] str p4, [x2, #4, MUL VL] str p5, [x2, #5, MUL VL] str p6, [x2, #6, MUL VL] str p7, [x2, #7, MUL VL] str p8, [x2, #8, MUL VL] str p9, [x2, #9, MUL VL] str p10, [x2, #10, MUL VL] str p11, [x2, #11, MUL VL] str p12, [x2, #12, MUL VL] str p13, [x2, #13, MUL VL] str p14, [x2, #14, MUL VL] str p15, [x2, #15, MUL VL] // Only save FFR if we wrote a value for SME ldr x2, =ffr_in ldr x2, [x2, #0] cbz x2, 1f ldr x2, =ffr_out rdffr p0.b str p0, [x2] 1: // Restore callee saved registers x19-x30 ldp x19, x20, [sp, #32] ldp x21, x22, [sp, #48] ldp x23, x24, [sp, #64] ldp x25, x26, [sp, #80] ldp x27, x28, [sp, #96] ldp x29, x30, [sp], #112 // Clear SVCR if we were doing SME so future tests don't have ZA cbz x1, 1f msr S3_3_C4_C2_2, xzr 1: ret