Lines Matching refs:d

38 	int d, z, z0;  in raid6_avx21_gen_syndrome()  local
49 for (d = 0; d < bytes; d += 32) { in raid6_avx21_gen_syndrome()
50 asm volatile("prefetchnta %0" : : "m" (dptr[z0][d])); in raid6_avx21_gen_syndrome()
51 asm volatile("vmovdqa %0,%%ymm2" : : "m" (dptr[z0][d]));/* P[0] */ in raid6_avx21_gen_syndrome()
52 asm volatile("prefetchnta %0" : : "m" (dptr[z0-1][d])); in raid6_avx21_gen_syndrome()
54 asm volatile("vmovdqa %0,%%ymm6" : : "m" (dptr[z0-1][d])); in raid6_avx21_gen_syndrome()
56 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_avx21_gen_syndrome()
63 asm volatile("vmovdqa %0,%%ymm6" : : "m" (dptr[z][d])); in raid6_avx21_gen_syndrome()
72 asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d])); in raid6_avx21_gen_syndrome()
74 asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d])); in raid6_avx21_gen_syndrome()
87 int d, z, z0; in raid6_avx21_xor_syndrome() local
97 for (d = 0 ; d < bytes ; d += 32) { in raid6_avx21_xor_syndrome()
98 asm volatile("vmovdqa %0,%%ymm4" :: "m" (dptr[z0][d])); in raid6_avx21_xor_syndrome()
99 asm volatile("vmovdqa %0,%%ymm2" : : "m" (p[d])); in raid6_avx21_xor_syndrome()
108 asm volatile("vmovdqa %0,%%ymm5" :: "m" (dptr[z][d])); in raid6_avx21_xor_syndrome()
120 asm volatile("vpxor %0,%%ymm4,%%ymm4" : : "m" (q[d])); in raid6_avx21_xor_syndrome()
122 asm volatile("vmovdqa %%ymm4,%0" : "=m" (q[d])); in raid6_avx21_xor_syndrome()
123 asm volatile("vmovdqa %%ymm2,%0" : "=m" (p[d])); in raid6_avx21_xor_syndrome()
145 int d, z, z0; in raid6_avx22_gen_syndrome() local
157 for (d = 0; d < bytes; d += 64) { in raid6_avx22_gen_syndrome()
158 asm volatile("prefetchnta %0" : : "m" (dptr[z0][d])); in raid6_avx22_gen_syndrome()
159 asm volatile("prefetchnta %0" : : "m" (dptr[z0][d+32])); in raid6_avx22_gen_syndrome()
160 asm volatile("vmovdqa %0,%%ymm2" : : "m" (dptr[z0][d]));/* P[0] */ in raid6_avx22_gen_syndrome()
161 asm volatile("vmovdqa %0,%%ymm3" : : "m" (dptr[z0][d+32]));/* P[1] */ in raid6_avx22_gen_syndrome()
165 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_avx22_gen_syndrome()
166 asm volatile("prefetchnta %0" : : "m" (dptr[z][d+32])); in raid6_avx22_gen_syndrome()
175 asm volatile("vmovdqa %0,%%ymm5" : : "m" (dptr[z][d])); in raid6_avx22_gen_syndrome()
176 asm volatile("vmovdqa %0,%%ymm7" : : "m" (dptr[z][d+32])); in raid6_avx22_gen_syndrome()
182 asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d])); in raid6_avx22_gen_syndrome()
183 asm volatile("vmovntdq %%ymm3,%0" : "=m" (p[d+32])); in raid6_avx22_gen_syndrome()
184 asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d])); in raid6_avx22_gen_syndrome()
185 asm volatile("vmovntdq %%ymm6,%0" : "=m" (q[d+32])); in raid6_avx22_gen_syndrome()
197 int d, z, z0; in raid6_avx22_xor_syndrome() local
207 for (d = 0 ; d < bytes ; d += 64) { in raid6_avx22_xor_syndrome()
208 asm volatile("vmovdqa %0,%%ymm4" :: "m" (dptr[z0][d])); in raid6_avx22_xor_syndrome()
209 asm volatile("vmovdqa %0,%%ymm6" :: "m" (dptr[z0][d+32])); in raid6_avx22_xor_syndrome()
210 asm volatile("vmovdqa %0,%%ymm2" : : "m" (p[d])); in raid6_avx22_xor_syndrome()
211 asm volatile("vmovdqa %0,%%ymm3" : : "m" (p[d+32])); in raid6_avx22_xor_syndrome()
226 asm volatile("vmovdqa %0,%%ymm5" :: "m" (dptr[z][d])); in raid6_avx22_xor_syndrome()
228 :: "m" (dptr[z][d+32])); in raid6_avx22_xor_syndrome()
247 asm volatile("vpxor %0,%%ymm4,%%ymm4" : : "m" (q[d])); in raid6_avx22_xor_syndrome()
248 asm volatile("vpxor %0,%%ymm6,%%ymm6" : : "m" (q[d+32])); in raid6_avx22_xor_syndrome()
250 asm volatile("vmovdqa %%ymm4,%0" : "=m" (q[d])); in raid6_avx22_xor_syndrome()
251 asm volatile("vmovdqa %%ymm6,%0" : "=m" (q[d+32])); in raid6_avx22_xor_syndrome()
252 asm volatile("vmovdqa %%ymm2,%0" : "=m" (p[d])); in raid6_avx22_xor_syndrome()
253 asm volatile("vmovdqa %%ymm3,%0" : "=m" (p[d+32])); in raid6_avx22_xor_syndrome()
277 int d, z, z0; in raid6_avx24_gen_syndrome() local
296 for (d = 0; d < bytes; d += 128) { in raid6_avx24_gen_syndrome()
298 asm volatile("prefetchnta %0" : : "m" (dptr[z][d])); in raid6_avx24_gen_syndrome()
299 asm volatile("prefetchnta %0" : : "m" (dptr[z][d+32])); in raid6_avx24_gen_syndrome()
300 asm volatile("prefetchnta %0" : : "m" (dptr[z][d+64])); in raid6_avx24_gen_syndrome()
301 asm volatile("prefetchnta %0" : : "m" (dptr[z][d+96])); in raid6_avx24_gen_syndrome()
318 asm volatile("vmovdqa %0,%%ymm5" : : "m" (dptr[z][d])); in raid6_avx24_gen_syndrome()
319 asm volatile("vmovdqa %0,%%ymm7" : : "m" (dptr[z][d+32])); in raid6_avx24_gen_syndrome()
320 asm volatile("vmovdqa %0,%%ymm13" : : "m" (dptr[z][d+64])); in raid6_avx24_gen_syndrome()
321 asm volatile("vmovdqa %0,%%ymm15" : : "m" (dptr[z][d+96])); in raid6_avx24_gen_syndrome()
331 asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d])); in raid6_avx24_gen_syndrome()
333 asm volatile("vmovntdq %%ymm3,%0" : "=m" (p[d+32])); in raid6_avx24_gen_syndrome()
335 asm volatile("vmovntdq %%ymm10,%0" : "=m" (p[d+64])); in raid6_avx24_gen_syndrome()
337 asm volatile("vmovntdq %%ymm11,%0" : "=m" (p[d+96])); in raid6_avx24_gen_syndrome()
339 asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d])); in raid6_avx24_gen_syndrome()
341 asm volatile("vmovntdq %%ymm6,%0" : "=m" (q[d+32])); in raid6_avx24_gen_syndrome()
343 asm volatile("vmovntdq %%ymm12,%0" : "=m" (q[d+64])); in raid6_avx24_gen_syndrome()
345 asm volatile("vmovntdq %%ymm14,%0" : "=m" (q[d+96])); in raid6_avx24_gen_syndrome()
358 int d, z, z0; in raid6_avx24_xor_syndrome() local
368 for (d = 0 ; d < bytes ; d += 128) { in raid6_avx24_xor_syndrome()
369 asm volatile("vmovdqa %0,%%ymm4" :: "m" (dptr[z0][d])); in raid6_avx24_xor_syndrome()
370 asm volatile("vmovdqa %0,%%ymm6" :: "m" (dptr[z0][d+32])); in raid6_avx24_xor_syndrome()
371 asm volatile("vmovdqa %0,%%ymm12" :: "m" (dptr[z0][d+64])); in raid6_avx24_xor_syndrome()
372 asm volatile("vmovdqa %0,%%ymm14" :: "m" (dptr[z0][d+96])); in raid6_avx24_xor_syndrome()
373 asm volatile("vmovdqa %0,%%ymm2" : : "m" (p[d])); in raid6_avx24_xor_syndrome()
374 asm volatile("vmovdqa %0,%%ymm3" : : "m" (p[d+32])); in raid6_avx24_xor_syndrome()
375 asm volatile("vmovdqa %0,%%ymm10" : : "m" (p[d+64])); in raid6_avx24_xor_syndrome()
376 asm volatile("vmovdqa %0,%%ymm11" : : "m" (p[d+96])); in raid6_avx24_xor_syndrome()
383 asm volatile("prefetchnta %0" :: "m" (dptr[z][d])); in raid6_avx24_xor_syndrome()
384 asm volatile("prefetchnta %0" :: "m" (dptr[z][d+64])); in raid6_avx24_xor_syndrome()
405 asm volatile("vmovdqa %0,%%ymm5" :: "m" (dptr[z][d])); in raid6_avx24_xor_syndrome()
407 :: "m" (dptr[z][d+32])); in raid6_avx24_xor_syndrome()
409 :: "m" (dptr[z][d+64])); in raid6_avx24_xor_syndrome()
411 :: "m" (dptr[z][d+96])); in raid6_avx24_xor_syndrome()
421 asm volatile("prefetchnta %0" :: "m" (q[d])); in raid6_avx24_xor_syndrome()
422 asm volatile("prefetchnta %0" :: "m" (q[d+64])); in raid6_avx24_xor_syndrome()
446 asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d])); in raid6_avx24_xor_syndrome()
447 asm volatile("vmovntdq %%ymm3,%0" : "=m" (p[d+32])); in raid6_avx24_xor_syndrome()
448 asm volatile("vmovntdq %%ymm10,%0" : "=m" (p[d+64])); in raid6_avx24_xor_syndrome()
449 asm volatile("vmovntdq %%ymm11,%0" : "=m" (p[d+96])); in raid6_avx24_xor_syndrome()
450 asm volatile("vpxor %0,%%ymm4,%%ymm4" : : "m" (q[d])); in raid6_avx24_xor_syndrome()
451 asm volatile("vpxor %0,%%ymm6,%%ymm6" : : "m" (q[d+32])); in raid6_avx24_xor_syndrome()
452 asm volatile("vpxor %0,%%ymm12,%%ymm12" : : "m" (q[d+64])); in raid6_avx24_xor_syndrome()
453 asm volatile("vpxor %0,%%ymm14,%%ymm14" : : "m" (q[d+96])); in raid6_avx24_xor_syndrome()
454 asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d])); in raid6_avx24_xor_syndrome()
455 asm volatile("vmovntdq %%ymm6,%0" : "=m" (q[d+32])); in raid6_avx24_xor_syndrome()
456 asm volatile("vmovntdq %%ymm12,%0" : "=m" (q[d+64])); in raid6_avx24_xor_syndrome()
457 asm volatile("vmovntdq %%ymm14,%0" : "=m" (q[d+96])); in raid6_avx24_xor_syndrome()