Lines Matching full:volatile

54 	asm volatile("movdqa %0,%%xmm7" : : "m" (x0f[0]));  in raid6_2data_recov_ssse3()
57 asm volatile("movdqa %0,%%xmm6" : : "m" (qmul[0])); in raid6_2data_recov_ssse3()
58 asm volatile("movdqa %0,%%xmm14" : : "m" (pbmul[0])); in raid6_2data_recov_ssse3()
59 asm volatile("movdqa %0,%%xmm15" : : "m" (pbmul[16])); in raid6_2data_recov_ssse3()
67 asm volatile("movdqa %0,%%xmm1" : : "m" (q[0])); in raid6_2data_recov_ssse3()
68 asm volatile("movdqa %0,%%xmm9" : : "m" (q[16])); in raid6_2data_recov_ssse3()
69 asm volatile("movdqa %0,%%xmm0" : : "m" (p[0])); in raid6_2data_recov_ssse3()
70 asm volatile("movdqa %0,%%xmm8" : : "m" (p[16])); in raid6_2data_recov_ssse3()
71 asm volatile("pxor %0,%%xmm1" : : "m" (dq[0])); in raid6_2data_recov_ssse3()
72 asm volatile("pxor %0,%%xmm9" : : "m" (dq[16])); in raid6_2data_recov_ssse3()
73 asm volatile("pxor %0,%%xmm0" : : "m" (dp[0])); in raid6_2data_recov_ssse3()
74 asm volatile("pxor %0,%%xmm8" : : "m" (dp[16])); in raid6_2data_recov_ssse3()
78 asm volatile("movdqa %xmm6,%xmm4"); in raid6_2data_recov_ssse3()
79 asm volatile("movdqa %0,%%xmm5" : : "m" (qmul[16])); in raid6_2data_recov_ssse3()
80 asm volatile("movdqa %xmm6,%xmm12"); in raid6_2data_recov_ssse3()
81 asm volatile("movdqa %xmm5,%xmm13"); in raid6_2data_recov_ssse3()
82 asm volatile("movdqa %xmm1,%xmm3"); in raid6_2data_recov_ssse3()
83 asm volatile("movdqa %xmm9,%xmm11"); in raid6_2data_recov_ssse3()
84 asm volatile("movdqa %xmm0,%xmm2"); /* xmm2/10 = px */ in raid6_2data_recov_ssse3()
85 asm volatile("movdqa %xmm8,%xmm10"); in raid6_2data_recov_ssse3()
86 asm volatile("psraw $4,%xmm1"); in raid6_2data_recov_ssse3()
87 asm volatile("psraw $4,%xmm9"); in raid6_2data_recov_ssse3()
88 asm volatile("pand %xmm7,%xmm3"); in raid6_2data_recov_ssse3()
89 asm volatile("pand %xmm7,%xmm11"); in raid6_2data_recov_ssse3()
90 asm volatile("pand %xmm7,%xmm1"); in raid6_2data_recov_ssse3()
91 asm volatile("pand %xmm7,%xmm9"); in raid6_2data_recov_ssse3()
92 asm volatile("pshufb %xmm3,%xmm4"); in raid6_2data_recov_ssse3()
93 asm volatile("pshufb %xmm11,%xmm12"); in raid6_2data_recov_ssse3()
94 asm volatile("pshufb %xmm1,%xmm5"); in raid6_2data_recov_ssse3()
95 asm volatile("pshufb %xmm9,%xmm13"); in raid6_2data_recov_ssse3()
96 asm volatile("pxor %xmm4,%xmm5"); in raid6_2data_recov_ssse3()
97 asm volatile("pxor %xmm12,%xmm13"); in raid6_2data_recov_ssse3()
101 asm volatile("movdqa %xmm14,%xmm4"); in raid6_2data_recov_ssse3()
102 asm volatile("movdqa %xmm15,%xmm1"); in raid6_2data_recov_ssse3()
103 asm volatile("movdqa %xmm14,%xmm12"); in raid6_2data_recov_ssse3()
104 asm volatile("movdqa %xmm15,%xmm9"); in raid6_2data_recov_ssse3()
105 asm volatile("movdqa %xmm2,%xmm3"); in raid6_2data_recov_ssse3()
106 asm volatile("movdqa %xmm10,%xmm11"); in raid6_2data_recov_ssse3()
107 asm volatile("psraw $4,%xmm2"); in raid6_2data_recov_ssse3()
108 asm volatile("psraw $4,%xmm10"); in raid6_2data_recov_ssse3()
109 asm volatile("pand %xmm7,%xmm3"); in raid6_2data_recov_ssse3()
110 asm volatile("pand %xmm7,%xmm11"); in raid6_2data_recov_ssse3()
111 asm volatile("pand %xmm7,%xmm2"); in raid6_2data_recov_ssse3()
112 asm volatile("pand %xmm7,%xmm10"); in raid6_2data_recov_ssse3()
113 asm volatile("pshufb %xmm3,%xmm4"); in raid6_2data_recov_ssse3()
114 asm volatile("pshufb %xmm11,%xmm12"); in raid6_2data_recov_ssse3()
115 asm volatile("pshufb %xmm2,%xmm1"); in raid6_2data_recov_ssse3()
116 asm volatile("pshufb %xmm10,%xmm9"); in raid6_2data_recov_ssse3()
117 asm volatile("pxor %xmm4,%xmm1"); in raid6_2data_recov_ssse3()
118 asm volatile("pxor %xmm12,%xmm9"); in raid6_2data_recov_ssse3()
121 asm volatile("pxor %xmm5,%xmm1"); in raid6_2data_recov_ssse3()
122 asm volatile("pxor %xmm13,%xmm9"); in raid6_2data_recov_ssse3()
124 asm volatile("movdqa %%xmm1,%0" : "=m" (dq[0])); in raid6_2data_recov_ssse3()
125 asm volatile("movdqa %%xmm9,%0" : "=m" (dq[16])); in raid6_2data_recov_ssse3()
127 asm volatile("pxor %xmm1,%xmm0"); in raid6_2data_recov_ssse3()
128 asm volatile("pxor %xmm9,%xmm8"); in raid6_2data_recov_ssse3()
129 asm volatile("movdqa %%xmm0,%0" : "=m" (dp[0])); in raid6_2data_recov_ssse3()
130 asm volatile("movdqa %%xmm8,%0" : "=m" (dp[16])); in raid6_2data_recov_ssse3()
138 asm volatile("movdqa %0,%%xmm1" : : "m" (*q)); in raid6_2data_recov_ssse3()
139 asm volatile("movdqa %0,%%xmm0" : : "m" (*p)); in raid6_2data_recov_ssse3()
140 asm volatile("pxor %0,%%xmm1" : : "m" (*dq)); in raid6_2data_recov_ssse3()
141 asm volatile("pxor %0,%%xmm0" : : "m" (*dp)); in raid6_2data_recov_ssse3()
146 asm volatile("movdqa %0,%%xmm4" : : "m" (qmul[0])); in raid6_2data_recov_ssse3()
147 asm volatile("movdqa %0,%%xmm5" : : "m" (qmul[16])); in raid6_2data_recov_ssse3()
149 asm volatile("movdqa %xmm1,%xmm3"); in raid6_2data_recov_ssse3()
150 asm volatile("psraw $4,%xmm1"); in raid6_2data_recov_ssse3()
151 asm volatile("pand %xmm7,%xmm3"); in raid6_2data_recov_ssse3()
152 asm volatile("pand %xmm7,%xmm1"); in raid6_2data_recov_ssse3()
153 asm volatile("pshufb %xmm3,%xmm4"); in raid6_2data_recov_ssse3()
154 asm volatile("pshufb %xmm1,%xmm5"); in raid6_2data_recov_ssse3()
155 asm volatile("pxor %xmm4,%xmm5"); in raid6_2data_recov_ssse3()
157 asm volatile("movdqa %xmm0,%xmm2"); /* xmm2 = px */ in raid6_2data_recov_ssse3()
161 asm volatile("movdqa %0,%%xmm4" : : "m" (pbmul[0])); in raid6_2data_recov_ssse3()
162 asm volatile("movdqa %0,%%xmm1" : : "m" (pbmul[16])); in raid6_2data_recov_ssse3()
163 asm volatile("movdqa %xmm2,%xmm3"); in raid6_2data_recov_ssse3()
164 asm volatile("psraw $4,%xmm2"); in raid6_2data_recov_ssse3()
165 asm volatile("pand %xmm7,%xmm3"); in raid6_2data_recov_ssse3()
166 asm volatile("pand %xmm7,%xmm2"); in raid6_2data_recov_ssse3()
167 asm volatile("pshufb %xmm3,%xmm4"); in raid6_2data_recov_ssse3()
168 asm volatile("pshufb %xmm2,%xmm1"); in raid6_2data_recov_ssse3()
169 asm volatile("pxor %xmm4,%xmm1"); in raid6_2data_recov_ssse3()
172 asm volatile("pxor %xmm5,%xmm1"); in raid6_2data_recov_ssse3()
174 asm volatile("movdqa %%xmm1,%0" : "=m" (*dq)); in raid6_2data_recov_ssse3()
176 asm volatile("pxor %xmm1,%xmm0"); in raid6_2data_recov_ssse3()
177 asm volatile("movdqa %%xmm0,%0" : "=m" (*dp)); in raid6_2data_recov_ssse3()
220 asm volatile("movdqa %0, %%xmm7" : : "m" (x0f[0])); in raid6_datap_recov_ssse3()
224 asm volatile("movdqa %0, %%xmm3" : : "m" (dq[0])); in raid6_datap_recov_ssse3()
225 asm volatile("movdqa %0, %%xmm4" : : "m" (dq[16])); in raid6_datap_recov_ssse3()
226 asm volatile("pxor %0, %%xmm3" : : "m" (q[0])); in raid6_datap_recov_ssse3()
227 asm volatile("movdqa %0, %%xmm0" : : "m" (qmul[0])); in raid6_datap_recov_ssse3()
231 asm volatile("pxor %0, %%xmm4" : : "m" (q[16])); in raid6_datap_recov_ssse3()
232 asm volatile("movdqa %0, %%xmm1" : : "m" (qmul[16])); in raid6_datap_recov_ssse3()
236 asm volatile("movdqa %xmm3, %xmm6"); in raid6_datap_recov_ssse3()
237 asm volatile("movdqa %xmm4, %xmm8"); in raid6_datap_recov_ssse3()
241 asm volatile("psraw $4, %xmm3"); in raid6_datap_recov_ssse3()
242 asm volatile("pand %xmm7, %xmm6"); in raid6_datap_recov_ssse3()
243 asm volatile("pand %xmm7, %xmm3"); in raid6_datap_recov_ssse3()
244 asm volatile("pshufb %xmm6, %xmm0"); in raid6_datap_recov_ssse3()
245 asm volatile("pshufb %xmm3, %xmm1"); in raid6_datap_recov_ssse3()
246 asm volatile("movdqa %0, %%xmm10" : : "m" (qmul[0])); in raid6_datap_recov_ssse3()
247 asm volatile("pxor %xmm0, %xmm1"); in raid6_datap_recov_ssse3()
248 asm volatile("movdqa %0, %%xmm11" : : "m" (qmul[16])); in raid6_datap_recov_ssse3()
252 asm volatile("psraw $4, %xmm4"); in raid6_datap_recov_ssse3()
253 asm volatile("pand %xmm7, %xmm8"); in raid6_datap_recov_ssse3()
254 asm volatile("pand %xmm7, %xmm4"); in raid6_datap_recov_ssse3()
255 asm volatile("pshufb %xmm8, %xmm10"); in raid6_datap_recov_ssse3()
256 asm volatile("pshufb %xmm4, %xmm11"); in raid6_datap_recov_ssse3()
257 asm volatile("movdqa %0, %%xmm2" : : "m" (p[0])); in raid6_datap_recov_ssse3()
258 asm volatile("pxor %xmm10, %xmm11"); in raid6_datap_recov_ssse3()
259 asm volatile("movdqa %0, %%xmm12" : : "m" (p[16])); in raid6_datap_recov_ssse3()
263 asm volatile("pxor %xmm1, %xmm2"); in raid6_datap_recov_ssse3()
267 asm volatile("pxor %xmm11, %xmm12"); in raid6_datap_recov_ssse3()
271 asm volatile("movdqa %%xmm1, %0" : "=m" (dq[0])); in raid6_datap_recov_ssse3()
272 asm volatile("movdqa %%xmm11, %0" : "=m" (dq[16])); in raid6_datap_recov_ssse3()
274 asm volatile("movdqa %%xmm2, %0" : "=m" (p[0])); in raid6_datap_recov_ssse3()
275 asm volatile("movdqa %%xmm12, %0" : "=m" (p[16])); in raid6_datap_recov_ssse3()
283 asm volatile("movdqa %0, %%xmm3" : : "m" (dq[0])); in raid6_datap_recov_ssse3()
284 asm volatile("movdqa %0, %%xmm0" : : "m" (qmul[0])); in raid6_datap_recov_ssse3()
285 asm volatile("pxor %0, %%xmm3" : : "m" (q[0])); in raid6_datap_recov_ssse3()
286 asm volatile("movdqa %0, %%xmm1" : : "m" (qmul[16])); in raid6_datap_recov_ssse3()
290 asm volatile("movdqa %xmm3, %xmm6"); in raid6_datap_recov_ssse3()
291 asm volatile("movdqa %0, %%xmm2" : : "m" (p[0])); in raid6_datap_recov_ssse3()
292 asm volatile("psraw $4, %xmm3"); in raid6_datap_recov_ssse3()
293 asm volatile("pand %xmm7, %xmm6"); in raid6_datap_recov_ssse3()
294 asm volatile("pand %xmm7, %xmm3"); in raid6_datap_recov_ssse3()
295 asm volatile("pshufb %xmm6, %xmm0"); in raid6_datap_recov_ssse3()
296 asm volatile("pshufb %xmm3, %xmm1"); in raid6_datap_recov_ssse3()
297 asm volatile("pxor %xmm0, %xmm1"); in raid6_datap_recov_ssse3()
301 asm volatile("pxor %xmm1, %xmm2"); in raid6_datap_recov_ssse3()
305 asm volatile("movdqa %%xmm1, %0" : "=m" (dq[0])); in raid6_datap_recov_ssse3()
306 asm volatile("movdqa %%xmm2, %0" : "=m" (p[0])); in raid6_datap_recov_ssse3()