Lines Matching +full:5 +full:vs2

505         uint32_t idx, void *vs2);
509 uint32_t idx, void *vs2) \
511 return (base + *((ETYPE *)vs2 + H(idx))); \
521 void *vs2, CPURISCVState *env, uint32_t desc,
546 abi_ptr addr = get_index_addr(base, i, vs2) + (k << log2_esz);
558 void *vs2, CPURISCVState *env, uint32_t desc) \
560 vext_ldst_index(vd, v0, base, vs2, env, desc, INDEX_FN, \
583 void *vs2, CPURISCVState *env, uint32_t desc) \
585 vext_ldst_index(vd, v0, base, vs2, env, desc, INDEX_FN, \
1082 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
1096 ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
1117 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
1129 ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
1154 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
1167 ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
1195 void *vs2, CPURISCVState *env, uint32_t desc) \
1206 ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
1291 void *vs2, CPURISCVState *env, uint32_t desc) \
1310 TS2 s2 = *((TS2 *)vs2 + HS2(i)); \
1338 void *vs2, CPURISCVState *env, \
1359 TS2 s2 = *((TS2 *)vs2 + HS2(i)); \
1404 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
1418 ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
1471 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
1484 ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
1888 static void do_##NAME(void *vd, void *vs1, void *vs2, int i) \
1891 TX2 s2 = *((T2 *)vs2 + HS2(i)); \
1934 static void do_##NAME(void *vd, target_long s1, void *vs2, int i) \
1936 TX2 s2 = *((T2 *)vs2 + HS2(i)); \
2072 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
2084 ETYPE *vt = (!vext_elem_mask(v0, i) ? vs2 : vs1); \
2099 void *vs2, CPURISCVState *env, uint32_t desc) \
2110 ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
2135 typedef void opivv2_rm_fn(void *vd, void *vs1, void *vs2, int i,
2140 do_##NAME(void *vd, void *vs1, void *vs2, int i, \
2144 TX2 s2 = *((T2 *)vs2 + HS2(i)); \
2149 vext_vv_rm_1(void *vd, void *v0, void *vs1, void *vs2,
2160 fn(vd, vs1, vs2, i, env, vxrm);
2166 vext_vv_rm_2(void *vd, void *v0, void *vs1, void *vs2,
2181 vext_vv_rm_1(vd, v0, vs1, vs2,
2185 vext_vv_rm_1(vd, v0, vs1, vs2,
2189 vext_vv_rm_1(vd, v0, vs1, vs2,
2193 vext_vv_rm_1(vd, v0, vs1, vs2,
2203 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
2206 vext_vv_rm_2(vd, v0, vs1, vs2, env, desc, \
2263 typedef void opivx2_rm_fn(void *vd, target_long s1, void *vs2, int i,
2268 do_##NAME(void *vd, target_long s1, void *vs2, int i, \
2271 TX2 s2 = *((T2 *)vs2 + HS2(i)); \
2276 vext_vx_rm_1(void *vd, void *v0, target_long s1, void *vs2,
2287 fn(vd, s1, vs2, i, env, vxrm);
2293 vext_vx_rm_2(void *vd, void *v0, target_long s1, void *vs2,
2308 vext_vx_rm_1(vd, v0, s1, vs2,
2312 vext_vx_rm_1(vd, v0, s1, vs2,
2316 vext_vx_rm_1(vd, v0, s1, vs2,
2320 vext_vx_rm_1(vd, v0, s1, vs2,
2331 void *vs2, CPURISCVState *env, \
2334 vext_vx_rm_2(vd, v0, s1, vs2, env, desc, \
3059 static void do_##NAME(void *vd, void *vs1, void *vs2, int i, \
3063 TX2 s2 = *((T2 *)vs2 + HS2(i)); \
3069 void *vs2, CPURISCVState *env, \
3089 do_##NAME(vd, vs1, vs2, i, env); \
3105 static void do_##NAME(void *vd, uint64_t s1, void *vs2, int i, \
3108 TX2 s2 = *((T2 *)vs2 + HS2(i)); \
3114 void *vs2, CPURISCVState *env, \
3134 do_##NAME(vd, s1, vs2, i, env); \
3340 static void do_##NAME(void *vd, void *vs1, void *vs2, int i, \
3344 TX2 s2 = *((T2 *)vs2 + HS2(i)); \
3372 static void do_##NAME(void *vd, uint64_t s1, void *vs2, int i, \
3375 TX2 s2 = *((T2 *)vs2 + HS2(i)); \
3694 static void do_##NAME(void *vd, void *vs2, int i, \
3697 TX2 s2 = *((T2 *)vs2 + HS2(i)); \
3702 void HELPER(NAME)(void *vd, void *v0, void *vs2, \
3725 do_##NAME(vd, vs2, i, env); \
3758 9, 8, 7, 7, 6, 6, 5, 4,
3796 int exp_size = 5, frac_size = 10;
3949 8, 8, 7, 7, 6, 5, 5, 4,
4011 int exp_size = 5, frac_size = 10;
4226 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
4240 ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
4268 void HELPER(NAME)(void *vd, void *v0, uint64_t s1, void *vs2, \
4281 ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
4405 return sign ? 1 << 2 : 1 << 5;
4424 return sign ? 1 << 2 : 1 << 5;
4443 return sign ? 1 << 2 : 1 << 5;
4462 void HELPER(NAME)(void *vd, void *v0, uint64_t s1, void *vs2, \
4476 ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
4490 /* vfcvt.xu.f.v vd, vs2, vm # Convert float to unsigned integer. */
4498 /* vfcvt.x.f.v vd, vs2, vm # Convert float to signed integer. */
4506 /* vfcvt.f.xu.v vd, vs2, vm # Convert unsigned integer to float. */
4514 /* vfcvt.f.x.v vd, vs2, vm # Convert integer to float. */
4528 * vfwcvt.xu.f.v vd, vs2, vm # Convert float to double-width unsigned integer.
4535 /* vfwcvt.x.f.v vd, vs2, vm # Convert float to double-width signed integer. */
4542 * vfwcvt.f.xu.v vd, vs2, vm # Convert unsigned integer to double-width float.
4551 /* vfwcvt.f.x.v vd, vs2, vm # Convert integer to double-width float. */
4560 * vfwcvt.f.f.v vd, vs2, vm # Convert single-width float to double-width float.
4580 /* vfncvt.xu.f.v vd, vs2, vm # Convert float to unsigned integer. */
4588 /* vfncvt.x.f.v vd, vs2, vm # Convert double-width float to signed integer. */
4597 * vfncvt.f.xu.v vd, vs2, vm # Convert double-width unsigned integer to float.
4604 /* vfncvt.f.x.v vd, vs2, vm # Convert double-width integer to float. */
4610 /* vfncvt.f.f.v vd, vs2, vm # Convert double float to single-width float. */
4630 void *vs2, CPURISCVState *env, \
4644 TS2 s2 = *((TS2 *)vs2 + HS2(i)); \
4658 /* vd[0] = sum(vs1[0], vs2[*]) */
4664 /* vd[0] = maxu(vs1[0], vs2[*]) */
4670 /* vd[0] = max(vs1[0], vs2[*]) */
4676 /* vd[0] = minu(vs1[0], vs2[*]) */
4682 /* vd[0] = min(vs1[0], vs2[*]) */
4688 /* vd[0] = and(vs1[0], vs2[*]) */
4694 /* vd[0] = or(vs1[0], vs2[*]) */
4700 /* vd[0] = xor(vs1[0], vs2[*]) */
4720 void *vs2, CPURISCVState *env, \
4734 TS2 s2 = *((TS2 *)vs2 + HS2(i)); \
4798 void *vs2, CPURISCVState *env, \
4811 b = vext_elem_mask(vs2, i); \
4842 target_ulong HELPER(vcpop_m)(void *v0, void *vs2, CPURISCVState *env,
4852 if (vext_elem_mask(vs2, i)) {
4862 target_ulong HELPER(vfirst_m)(void *v0, void *vs2, CPURISCVState *env,
4871 if (vext_elem_mask(vs2, i)) {
4886 static void vmsetm(void *vd, void *v0, void *vs2, CPURISCVState *env,
4912 if (vext_elem_mask(vs2, i)) {
4939 void HELPER(vmsbf_m)(void *vd, void *v0, void *vs2, CPURISCVState *env,
4942 vmsetm(vd, v0, vs2, env, desc, BEFORE_FIRST);
4945 void HELPER(vmsif_m)(void *vd, void *v0, void *vs2, CPURISCVState *env,
4948 vmsetm(vd, v0, vs2, env, desc, INCLUDE_FIRST);
4951 void HELPER(vmsof_m)(void *vd, void *v0, void *vs2, CPURISCVState *env,
4954 vmsetm(vd, v0, vs2, env, desc, ONLY_FIRST);
4959 void HELPER(NAME)(void *vd, void *v0, void *vs2, CPURISCVState *env, \
4980 if (vext_elem_mask(vs2, i)) { \
5032 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
5052 *((ETYPE *)vd + H(i)) = *((ETYPE *)vs2 + H(i - offset)); \
5059 /* vslideup.vx vd, vs2, rs1, vm # vd[i+rs1] = vs2[i] */
5066 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
5088 *((ETYPE *)vd + H(i)) = *((ETYPE *)vs2 + H(i + s1)); \
5102 /* vslidedown.vx vd, vs2, rs1, vm # vd[i] = vs2[i+rs1] */
5110 void *vs2, CPURISCVState *env, \
5133 *((ETYPE *)vd + H(i)) = *((ETYPE *)vs2 + H(i - 1)); \
5147 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
5150 vslide1up_##BITWIDTH(vd, v0, s1, vs2, env, desc); \
5153 /* vslide1up.vx vd, vs2, rs1, vm # vd[0]=x[rs1], vd[i+1] = vs2[i] */
5161 void *vs2, CPURISCVState *env, \
5184 *((ETYPE *)vd + H(i)) = *((ETYPE *)vs2 + H(i + 1)); \
5198 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
5201 vslide1down_##BITWIDTH(vd, v0, s1, vs2, env, desc); \
5204 /* vslide1down.vx vd, vs2, rs1, vm # vd[i] = vs2[i+1], vd[vl-1]=x[rs1] */
5212 void HELPER(NAME)(void *vd, void *v0, uint64_t s1, void *vs2, \
5215 vslide1up_##BITWIDTH(vd, v0, s1, vs2, env, desc); \
5218 /* vfslide1up.vf vd, vs2, rs1, vm # vd[0]=f[rs1], vd[i+1] = vs2[i] */
5224 void HELPER(NAME)(void *vd, void *v0, uint64_t s1, void *vs2, \
5227 vslide1down_##BITWIDTH(vd, v0, s1, vs2, env, desc); \
5230 /* vfslide1down.vf vd, vs2, rs1, vm # vd[i] = vs2[i+1], vd[vl-1]=f[rs1] */
5237 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
5262 *((TS2 *)vd + HS2(i)) = *((TS2 *)vs2 + HS2(index)); \
5270 /* vd[i] = (vs1[i] >= VLMAX) ? 0 : vs2[vs1[i]]; */
5282 void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
5306 *((ETYPE *)vd + H(i)) = *((ETYPE *)vs2 + H(index)); \
5314 /* vd[i] = (x[rs1] >= VLMAX) ? 0 : vs2[rs1] */
5322 void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
5337 *((ETYPE *)vd + H(num)) = *((ETYPE *)vs2 + H(i)); \
5345 /* Compress into vd elements of vs2 where vs1 is enabled */
5352 void HELPER(vmvr_v)(void *vd, void *vs2, CPURISCVState *env, uint32_t desc)
5368 (uint8_t *)vs2 + H1(j - 1),
5374 (uint8_t *)vs2 + H1(i),
5382 void HELPER(NAME)(void *vd, void *v0, void *vs2, \
5401 *((ETYPE *)vd + HD(i)) = *((DTYPE *)vs2 + HS1(i)); \