Lines Matching +full:0 +full:x0f
22 static const u8 __aligned(16) x0f[16] = { in raid6_2data_recov_ssse3()
23 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, in raid6_2data_recov_ssse3()
24 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f}; in raid6_2data_recov_ssse3()
54 asm volatile("movdqa %0,%%xmm7" : : "m" (x0f[0])); in raid6_2data_recov_ssse3()
57 asm volatile("movdqa %0,%%xmm6" : : "m" (qmul[0])); in raid6_2data_recov_ssse3()
58 asm volatile("movdqa %0,%%xmm14" : : "m" (pbmul[0])); in raid6_2data_recov_ssse3()
59 asm volatile("movdqa %0,%%xmm15" : : "m" (pbmul[16])); in raid6_2data_recov_ssse3()
67 asm volatile("movdqa %0,%%xmm1" : : "m" (q[0])); in raid6_2data_recov_ssse3()
68 asm volatile("movdqa %0,%%xmm9" : : "m" (q[16])); in raid6_2data_recov_ssse3()
69 asm volatile("movdqa %0,%%xmm0" : : "m" (p[0])); in raid6_2data_recov_ssse3()
70 asm volatile("movdqa %0,%%xmm8" : : "m" (p[16])); in raid6_2data_recov_ssse3()
71 asm volatile("pxor %0,%%xmm1" : : "m" (dq[0])); in raid6_2data_recov_ssse3()
72 asm volatile("pxor %0,%%xmm9" : : "m" (dq[16])); in raid6_2data_recov_ssse3()
73 asm volatile("pxor %0,%%xmm0" : : "m" (dp[0])); in raid6_2data_recov_ssse3()
74 asm volatile("pxor %0,%%xmm8" : : "m" (dp[16])); in raid6_2data_recov_ssse3()
79 asm volatile("movdqa %0,%%xmm5" : : "m" (qmul[16])); in raid6_2data_recov_ssse3()
124 asm volatile("movdqa %%xmm1,%0" : "=m" (dq[0])); in raid6_2data_recov_ssse3()
125 asm volatile("movdqa %%xmm9,%0" : "=m" (dq[16])); in raid6_2data_recov_ssse3()
129 asm volatile("movdqa %%xmm0,%0" : "=m" (dp[0])); in raid6_2data_recov_ssse3()
130 asm volatile("movdqa %%xmm8,%0" : "=m" (dp[16])); in raid6_2data_recov_ssse3()
138 asm volatile("movdqa %0,%%xmm1" : : "m" (*q)); in raid6_2data_recov_ssse3()
139 asm volatile("movdqa %0,%%xmm0" : : "m" (*p)); in raid6_2data_recov_ssse3()
140 asm volatile("pxor %0,%%xmm1" : : "m" (*dq)); in raid6_2data_recov_ssse3()
141 asm volatile("pxor %0,%%xmm0" : : "m" (*dp)); in raid6_2data_recov_ssse3()
144 * 0 = dp ^ p in raid6_2data_recov_ssse3()
146 asm volatile("movdqa %0,%%xmm4" : : "m" (qmul[0])); in raid6_2data_recov_ssse3()
147 asm volatile("movdqa %0,%%xmm5" : : "m" (qmul[16])); in raid6_2data_recov_ssse3()
161 asm volatile("movdqa %0,%%xmm4" : : "m" (pbmul[0])); in raid6_2data_recov_ssse3()
162 asm volatile("movdqa %0,%%xmm1" : : "m" (pbmul[16])); in raid6_2data_recov_ssse3()
174 asm volatile("movdqa %%xmm1,%0" : "=m" (*dq)); in raid6_2data_recov_ssse3()
177 asm volatile("movdqa %%xmm0,%0" : "=m" (*dp)); in raid6_2data_recov_ssse3()
196 static const u8 __aligned(16) x0f[16] = { in raid6_datap_recov_ssse3()
197 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, in raid6_datap_recov_ssse3()
198 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f, 0x0f}; in raid6_datap_recov_ssse3()
220 asm volatile("movdqa %0, %%xmm7" : : "m" (x0f[0])); in raid6_datap_recov_ssse3()
224 asm volatile("movdqa %0, %%xmm3" : : "m" (dq[0])); in raid6_datap_recov_ssse3()
225 asm volatile("movdqa %0, %%xmm4" : : "m" (dq[16])); in raid6_datap_recov_ssse3()
226 asm volatile("pxor %0, %%xmm3" : : "m" (q[0])); in raid6_datap_recov_ssse3()
227 asm volatile("movdqa %0, %%xmm0" : : "m" (qmul[0])); in raid6_datap_recov_ssse3()
229 /* xmm3 = q[0] ^ dq[0] */ in raid6_datap_recov_ssse3()
231 asm volatile("pxor %0, %%xmm4" : : "m" (q[16])); in raid6_datap_recov_ssse3()
232 asm volatile("movdqa %0, %%xmm1" : : "m" (qmul[16])); in raid6_datap_recov_ssse3()
246 asm volatile("movdqa %0, %%xmm10" : : "m" (qmul[0])); in raid6_datap_recov_ssse3()
248 asm volatile("movdqa %0, %%xmm11" : : "m" (qmul[16])); in raid6_datap_recov_ssse3()
250 /* xmm1 = qmul[q[0] ^ dq[0]] */ in raid6_datap_recov_ssse3()
257 asm volatile("movdqa %0, %%xmm2" : : "m" (p[0])); in raid6_datap_recov_ssse3()
259 asm volatile("movdqa %0, %%xmm12" : : "m" (p[16])); in raid6_datap_recov_ssse3()
265 /* xmm2 = p[0] ^ qmul[q[0] ^ dq[0]] */ in raid6_datap_recov_ssse3()
271 asm volatile("movdqa %%xmm1, %0" : "=m" (dq[0])); in raid6_datap_recov_ssse3()
272 asm volatile("movdqa %%xmm11, %0" : "=m" (dq[16])); in raid6_datap_recov_ssse3()
274 asm volatile("movdqa %%xmm2, %0" : "=m" (p[0])); in raid6_datap_recov_ssse3()
275 asm volatile("movdqa %%xmm12, %0" : "=m" (p[16])); in raid6_datap_recov_ssse3()
283 asm volatile("movdqa %0, %%xmm3" : : "m" (dq[0])); in raid6_datap_recov_ssse3()
284 asm volatile("movdqa %0, %%xmm0" : : "m" (qmul[0])); in raid6_datap_recov_ssse3()
285 asm volatile("pxor %0, %%xmm3" : : "m" (q[0])); in raid6_datap_recov_ssse3()
286 asm volatile("movdqa %0, %%xmm1" : : "m" (qmul[16])); in raid6_datap_recov_ssse3()
291 asm volatile("movdqa %0, %%xmm2" : : "m" (p[0])); in raid6_datap_recov_ssse3()
305 asm volatile("movdqa %%xmm1, %0" : "=m" (dq[0])); in raid6_datap_recov_ssse3()
306 asm volatile("movdqa %%xmm2, %0" : "=m" (p[0])); in raid6_datap_recov_ssse3()