/*
* Copyright(c) 2019-2021 Qualcomm Innovation Center, Inc. All Rights Reserved.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, see .
*/
#include
#include
typedef unsigned char uint8_t;
typedef unsigned short uint16_t;
typedef unsigned int uint32_t;
static inline void S4_storerhnew_rr(void *p, int index, uint16_t v)
{
asm volatile("{\n\t"
" r0 = %0\n\n"
" memh(%1+%2<<#2) = r0.new\n\t"
"}\n"
:: "r"(v), "r"(p), "r"(index)
: "r0", "memory");
}
static uint32_t data;
static inline void *S4_storerbnew_ap(uint8_t v)
{
void *ret;
asm volatile("{\n\t"
" r0 = %1\n\n"
" memb(%0 = ##data) = r0.new\n\t"
"}\n"
: "=r"(ret)
: "r"(v)
: "r0", "memory");
return ret;
}
static inline void *S4_storerhnew_ap(uint16_t v)
{
void *ret;
asm volatile("{\n\t"
" r0 = %1\n\n"
" memh(%0 = ##data) = r0.new\n\t"
"}\n"
: "=r"(ret)
: "r"(v)
: "r0", "memory");
return ret;
}
static inline void *S4_storerinew_ap(uint32_t v)
{
void *ret;
asm volatile("{\n\t"
" r0 = %1\n\n"
" memw(%0 = ##data) = r0.new\n\t"
"}\n"
: "=r"(ret)
: "r"(v)
: "r0", "memory");
return ret;
}
static inline void S4_storeirbt_io(void *p, int pred)
{
asm volatile("p0 = cmp.eq(%0, #1)\n\t"
"if (p0) memb(%1+#4)=#27\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirbf_io(void *p, int pred)
{
asm volatile("p0 = cmp.eq(%0, #1)\n\t"
"if (!p0) memb(%1+#4)=#27\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirbtnew_io(void *p, int pred)
{
asm volatile("{\n\t"
" p0 = cmp.eq(%0, #1)\n\t"
" if (p0.new) memb(%1+#4)=#27\n\t"
"}\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirbfnew_io(void *p, int pred)
{
asm volatile("{\n\t"
" p0 = cmp.eq(%0, #1)\n\t"
" if (!p0.new) memb(%1+#4)=#27\n\t"
"}\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirht_io(void *p, int pred)
{
asm volatile("p0 = cmp.eq(%0, #1)\n\t"
"if (p0) memh(%1+#4)=#27\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirhf_io(void *p, int pred)
{
asm volatile("p0 = cmp.eq(%0, #1)\n\t"
"if (!p0) memh(%1+#4)=#27\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirhtnew_io(void *p, int pred)
{
asm volatile("{\n\t"
" p0 = cmp.eq(%0, #1)\n\t"
" if (p0.new) memh(%1+#4)=#27\n\t"
"}\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirhfnew_io(void *p, int pred)
{
asm volatile("{\n\t"
" p0 = cmp.eq(%0, #1)\n\t"
" if (!p0.new) memh(%1+#4)=#27\n\t"
"}\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirit_io(void *p, int pred)
{
asm volatile("p0 = cmp.eq(%0, #1)\n\t"
"if (p0) memw(%1+#4)=#27\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirif_io(void *p, int pred)
{
asm volatile("p0 = cmp.eq(%0, #1)\n\t"
"if (!p0) memw(%1+#4)=#27\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeiritnew_io(void *p, int pred)
{
asm volatile("{\n\t"
" p0 = cmp.eq(%0, #1)\n\t"
" if (p0.new) memw(%1+#4)=#27\n\t"
"}\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static inline void S4_storeirifnew_io(void *p, int pred)
{
asm volatile("{\n\t"
" p0 = cmp.eq(%0, #1)\n\t"
" if (!p0.new) memw(%1+#4)=#27\n\t"
"}\n\t"
:: "r"(pred), "r"(p)
: "p0", "memory");
}
static int L2_ploadrifnew_pi(void *p, int pred)
{
int result;
asm volatile("%0 = #31\n\t"
"{\n\t"
" p0 = cmp.eq(%2, #1)\n\t"
" if (!p0.new) %0 = memw(%1++#4)\n\t"
"}\n\t"
: "=&r"(result), "+r"(p) : "r"(pred)
: "p0");
return result;
}
/*
* Test that compound-compare-jump is executed in 2 parts
* First we have to do all the compares in the packet and
* account for auto-anding. Then, we can do the predicated
* jump.
*/
static inline int cmpnd_cmp_jump(void)
{
int retval;
asm ("r5 = #7\n\t"
"r6 = #9\n\t"
"{\n\t"
" p0 = cmp.eq(r5, #7)\n\t"
" if (p0.new) jump:nt 1f\n\t"
" p0 = cmp.eq(r6, #7)\n\t"
"}\n\t"
"%0 = #12\n\t"
"jump 2f\n\t"
"1:\n\t"
"%0 = #13\n\t"
"2:\n\t"
: "=r"(retval) :: "r5", "r6", "p0");
return retval;
}
static inline int test_clrtnew(int arg1, int old_val)
{
int ret;
asm volatile("r5 = %2\n\t"
"{\n\t"
"p0 = cmp.eq(%1, #1)\n\t"
"if (p0.new) r5=#0\n\t"
"}\n\t"
"%0 = r5\n\t"
: "=r"(ret)
: "r"(arg1), "r"(old_val)
: "p0", "r5");
return ret;
}
int err;
static void check(int val, int expect)
{
if (val != expect) {
printf("ERROR: 0x%04x != 0x%04x\n", val, expect);
err++;
}
}
static void check64(long long val, long long expect)
{
if (val != expect) {
printf("ERROR: 0x%016llx != 0x%016llx\n", val, expect);
err++;
}
}
uint32_t init[10] = { 0, 1, 2, 3, 4, 5, 6, 7, 8, 9 };
uint32_t array[10];
uint32_t early_exit;
/*
* Write this as a function because we can't guarantee the compiler will
* allocate a frame with just the SL2_return_tnew packet.
*/
static void SL2_return_tnew(int x);
asm ("SL2_return_tnew:\n\t"
" allocframe(#0)\n\t"
" r1 = #1\n\t"
" memw(##early_exit) = r1\n\t"
" {\n\t"
" p0 = cmp.eq(r0, #1)\n\t"
" if (p0.new) dealloc_return:nt\n\t" /* SL2_return_tnew */
" }\n\t"
" r1 = #0\n\t"
" memw(##early_exit) = r1\n\t"
" dealloc_return\n\t"
);
static long long creg_pair(int x, int y)
{
long long retval;
asm ("m0 = %1\n\t"
"m1 = %2\n\t"
"%0 = c7:6\n\t"
: "=r"(retval) : "r"(x), "r"(y) : "m0", "m1");
return retval;
}
static long long decbin(long long x, long long y, int *pred)
{
long long retval;
asm ("%0 = decbin(%2, %3)\n\t"
"%1 = p0\n\t"
: "=r"(retval), "=r"(*pred)
: "r"(x), "r"(y));
return retval;
}
/* Check that predicates are auto-and'ed in a packet */
static int auto_and(void)
{
int retval;
asm ("r5 = #1\n\t"
"{\n\t"
" p0 = cmp.eq(r1, #1)\n\t"
" p0 = cmp.eq(r1, #2)\n\t"
"}\n\t"
"%0 = p0\n\t"
: "=r"(retval)
:
: "r5", "p0");
return retval;
}
void test_lsbnew(void)
{
int result;
asm("r0 = #2\n\t"
"r1 = #5\n\t"
"{\n\t"
" p0 = r0\n\t"
" if (p0.new) r1 = #3\n\t"
"}\n\t"
"%0 = r1\n\t"
: "=r"(result) :: "r0", "r1", "p0");
check(result, 5);
}
void test_l2fetch(void)
{
/* These don't do anything in qemu, just make sure they don't assert */
asm volatile ("l2fetch(r0, r1)\n\t"
"l2fetch(r0, r3:2)\n\t");
}
int main()
{
int res;
long long res64;
int pred;
memcpy(array, init, sizeof(array));
S4_storerhnew_rr(array, 4, 0xffff);
check(array[4], 0xffff);
data = ~0;
check((uint32_t)S4_storerbnew_ap(0x12), (uint32_t)&data);
check(data, 0xffffff12);
data = ~0;
check((uint32_t)S4_storerhnew_ap(0x1234), (uint32_t)&data);
check(data, 0xffff1234);
data = ~0;
check((uint32_t)S4_storerinew_ap(0x12345678), (uint32_t)&data);
check(data, 0x12345678);
/* Byte */
memcpy(array, init, sizeof(array));
S4_storeirbt_io(&array[1], 1);
check(array[2], 27);
S4_storeirbt_io(&array[2], 0);
check(array[3], 3);
memcpy(array, init, sizeof(array));
S4_storeirbf_io(&array[3], 0);
check(array[4], 27);
S4_storeirbf_io(&array[4], 1);
check(array[5], 5);
memcpy(array, init, sizeof(array));
S4_storeirbtnew_io(&array[5], 1);
check(array[6], 27);
S4_storeirbtnew_io(&array[6], 0);
check(array[7], 7);
memcpy(array, init, sizeof(array));
S4_storeirbfnew_io(&array[7], 0);
check(array[8], 27);
S4_storeirbfnew_io(&array[8], 1);
check(array[9], 9);
/* Half word */
memcpy(array, init, sizeof(array));
S4_storeirht_io(&array[1], 1);
check(array[2], 27);
S4_storeirht_io(&array[2], 0);
check(array[3], 3);
memcpy(array, init, sizeof(array));
S4_storeirhf_io(&array[3], 0);
check(array[4], 27);
S4_storeirhf_io(&array[4], 1);
check(array[5], 5);
memcpy(array, init, sizeof(array));
S4_storeirhtnew_io(&array[5], 1);
check(array[6], 27);
S4_storeirhtnew_io(&array[6], 0);
check(array[7], 7);
memcpy(array, init, sizeof(array));
S4_storeirhfnew_io(&array[7], 0);
check(array[8], 27);
S4_storeirhfnew_io(&array[8], 1);
check(array[9], 9);
/* Word */
memcpy(array, init, sizeof(array));
S4_storeirit_io(&array[1], 1);
check(array[2], 27);
S4_storeirit_io(&array[2], 0);
check(array[3], 3);
memcpy(array, init, sizeof(array));
S4_storeirif_io(&array[3], 0);
check(array[4], 27);
S4_storeirif_io(&array[4], 1);
check(array[5], 5);
memcpy(array, init, sizeof(array));
S4_storeiritnew_io(&array[5], 1);
check(array[6], 27);
S4_storeiritnew_io(&array[6], 0);
check(array[7], 7);
memcpy(array, init, sizeof(array));
S4_storeirifnew_io(&array[7], 0);
check(array[8], 27);
S4_storeirifnew_io(&array[8], 1);
check(array[9], 9);
memcpy(array, init, sizeof(array));
res = L2_ploadrifnew_pi(&array[6], 0);
check(res, 6);
res = L2_ploadrifnew_pi(&array[7], 1);
check(res, 31);
int x = cmpnd_cmp_jump();
check(x, 12);
SL2_return_tnew(0);
check(early_exit, 0);
SL2_return_tnew(1);
check(early_exit, 1);
long long pair = creg_pair(5, 7);
check((int)pair, 5);
check((int)(pair >> 32), 7);
res = test_clrtnew(1, 7);
check(res, 0);
res = test_clrtnew(2, 7);
check(res, 7);
res64 = decbin(0xf0f1f2f3f4f5f6f7LL, 0x7f6f5f4f3f2f1f0fLL, &pred);
check64(res64, 0x357980003700010cLL);
check(pred, 0);
res64 = decbin(0xfLL, 0x1bLL, &pred);
check64(res64, 0x78000100LL);
check(pred, 1);
res = auto_and();
check(res, 0);
test_lsbnew();
test_l2fetch();
puts(err ? "FAIL" : "PASS");
return err;
}