Lines Matching +full:1 +full:- +full:9

1 /* SPDX-License-Identifier: GPL-2.0-or-later */
5 # Copyright 2023- IBM Corp. All rights reserved
14 # 1. a += b; d ^= a; d <<<= 16;
43 #include <asm/asm-offsets.h>
44 #include <asm/asm-compat.h>
80 std 0, 16(1)
81 stdu 1,-752(1)
83 SAVE_GPR 14, 112, 1
84 SAVE_GPR 15, 120, 1
85 SAVE_GPR 16, 128, 1
86 SAVE_GPR 17, 136, 1
87 SAVE_GPR 18, 144, 1
88 SAVE_GPR 19, 152, 1
89 SAVE_GPR 20, 160, 1
90 SAVE_GPR 21, 168, 1
91 SAVE_GPR 22, 176, 1
92 SAVE_GPR 23, 184, 1
93 SAVE_GPR 24, 192, 1
94 SAVE_GPR 25, 200, 1
95 SAVE_GPR 26, 208, 1
96 SAVE_GPR 27, 216, 1
97 SAVE_GPR 28, 224, 1
98 SAVE_GPR 29, 232, 1
99 SAVE_GPR 30, 240, 1
100 SAVE_GPR 31, 248, 1
102 addi 9, 1, 256
103 SAVE_VRS 20, 0, 9
104 SAVE_VRS 21, 16, 9
105 SAVE_VRS 22, 32, 9
106 SAVE_VRS 23, 48, 9
107 SAVE_VRS 24, 64, 9
108 SAVE_VRS 25, 80, 9
109 SAVE_VRS 26, 96, 9
110 SAVE_VRS 27, 112, 9
111 SAVE_VRS 28, 128, 9
112 SAVE_VRS 29, 144, 9
113 SAVE_VRS 30, 160, 9
114 SAVE_VRS 31, 176, 9
116 SAVE_VSX 14, 192, 9
117 SAVE_VSX 15, 208, 9
118 SAVE_VSX 16, 224, 9
119 SAVE_VSX 17, 240, 9
120 SAVE_VSX 18, 256, 9
121 SAVE_VSX 19, 272, 9
122 SAVE_VSX 20, 288, 9
123 SAVE_VSX 21, 304, 9
124 SAVE_VSX 22, 320, 9
125 SAVE_VSX 23, 336, 9
126 SAVE_VSX 24, 352, 9
127 SAVE_VSX 25, 368, 9
128 SAVE_VSX 26, 384, 9
129 SAVE_VSX 27, 400, 9
130 SAVE_VSX 28, 416, 9
131 SAVE_VSX 29, 432, 9
132 SAVE_VSX 30, 448, 9
133 SAVE_VSX 31, 464, 9
137 addi 9, 1, 256
138 RESTORE_VRS 20, 0, 9
139 RESTORE_VRS 21, 16, 9
140 RESTORE_VRS 22, 32, 9
141 RESTORE_VRS 23, 48, 9
142 RESTORE_VRS 24, 64, 9
143 RESTORE_VRS 25, 80, 9
144 RESTORE_VRS 26, 96, 9
145 RESTORE_VRS 27, 112, 9
146 RESTORE_VRS 28, 128, 9
147 RESTORE_VRS 29, 144, 9
148 RESTORE_VRS 30, 160, 9
149 RESTORE_VRS 31, 176, 9
151 RESTORE_VSX 14, 192, 9
152 RESTORE_VSX 15, 208, 9
153 RESTORE_VSX 16, 224, 9
154 RESTORE_VSX 17, 240, 9
155 RESTORE_VSX 18, 256, 9
156 RESTORE_VSX 19, 272, 9
157 RESTORE_VSX 20, 288, 9
158 RESTORE_VSX 21, 304, 9
159 RESTORE_VSX 22, 320, 9
160 RESTORE_VSX 23, 336, 9
161 RESTORE_VSX 24, 352, 9
162 RESTORE_VSX 25, 368, 9
163 RESTORE_VSX 26, 384, 9
164 RESTORE_VSX 27, 400, 9
165 RESTORE_VSX 28, 416, 9
166 RESTORE_VSX 29, 432, 9
167 RESTORE_VSX 30, 448, 9
168 RESTORE_VSX 31, 464, 9
170 RESTORE_GPR 14, 112, 1
171 RESTORE_GPR 15, 120, 1
172 RESTORE_GPR 16, 128, 1
173 RESTORE_GPR 17, 136, 1
174 RESTORE_GPR 18, 144, 1
175 RESTORE_GPR 19, 152, 1
176 RESTORE_GPR 20, 160, 1
177 RESTORE_GPR 21, 168, 1
178 RESTORE_GPR 22, 176, 1
179 RESTORE_GPR 23, 184, 1
180 RESTORE_GPR 24, 192, 1
181 RESTORE_GPR 25, 200, 1
182 RESTORE_GPR 26, 208, 1
183 RESTORE_GPR 27, 216, 1
184 RESTORE_GPR 28, 224, 1
185 RESTORE_GPR 29, 232, 1
186 RESTORE_GPR 30, 240, 1
187 RESTORE_GPR 31, 248, 1
189 addi 1, 1, 752
190 ld 0, 16(1)
199 vadduwm 1, 1, 5
208 vpermxor 13, 13, 1, 25
217 vadduwm 9, 9, 13
225 vxor 5, 5, 9
245 vadduwm 1, 1, 5
256 vpermxor 13, 13, 1, 25
265 vadduwm 9, 9, 13
275 vxor 5, 5, 9
296 vadduwm 1, 1, 6
305 vpermxor 12, 12, 1, 25
317 vadduwm 9, 9, 14
325 vxor 4, 4, 9
344 vadduwm 1, 1, 6
355 vpermxor 12, 12, 1, 25
367 vadduwm 9, 9, 14
378 vxor 4, 4, 9
397 vadduwm 1, 1, 5
401 vpermxor 13, 13, 1, 20
405 vadduwm 9, 9, 13
409 vxor 5, 5, 9
417 vadduwm 1, 1, 5
421 vpermxor 13, 13, 1, 22
425 vadduwm 9, 9, 13
429 vxor 5, 5, 9
439 vadduwm 1, 1, 6
443 vpermxor 12, 12, 1, 20
449 vadduwm 9, 9, 14
453 vxor 4, 4, 9
459 vadduwm 1, 1, 6
463 vpermxor 12, 12, 1, 22
469 vadduwm 9, 9, 14
473 vxor 4, 4, 9
494 vadduwm \S+0, \S+0, 16-\S
495 vadduwm \S+4, \S+4, 17-\S
496 vadduwm \S+8, \S+8, 18-\S
497 vadduwm \S+12, \S+12, 19-\S
499 vadduwm \S+1, \S+1, 16-\S
500 vadduwm \S+5, \S+5, 17-\S
501 vadduwm \S+9, \S+9, 18-\S
502 vadduwm \S+13, \S+13, 19-\S
504 vadduwm \S+2, \S+2, 16-\S
505 vadduwm \S+6, \S+6, 17-\S
506 vadduwm \S+10, \S+10, 18-\S
507 vadduwm \S+14, \S+14, 19-\S
509 vadduwm \S+3, \S+3, 16-\S
510 vadduwm \S+7, \S+7, 17-\S
511 vadduwm \S+11, \S+11, 18-\S
512 vadduwm \S+15, \S+15, 19-\S
519 add 9, 14, 5
521 lxvw4x 0, 0, 9
522 lxvw4x 1, 17, 9
523 lxvw4x 2, 18, 9
524 lxvw4x 3, 19, 9
525 lxvw4x 4, 20, 9
526 lxvw4x 5, 21, 9
527 lxvw4x 6, 22, 9
528 lxvw4x 7, 23, 9
529 lxvw4x 8, 24, 9
530 lxvw4x 9, 25, 9
531 lxvw4x 10, 26, 9
532 lxvw4x 11, 27, 9
533 lxvw4x 12, 28, 9
534 lxvw4x 13, 29, 9
535 lxvw4x 14, 30, 9
536 lxvw4x 15, 31, 9
539 xxlxor \S+36, \S+36, 1
547 xxlxor \S+38, \S+38, 9
587 # r17 - r31 mainly for Write_256 macro.
608 lxvw4x 49, 17, 3 # vr17, key 1
612 # create (0, 1, 2, 3) counters
614 vspltisw 1, 1
617 vmrghw 4, 0, 1
619 vsldoi 30, 4, 5, 8 # vr30 counter, 4 (0, 1, 2, 3)
629 sradi 8, 7, 1
645 vadduwm 31, 30, 25 # counter = (0, 1, 2, 3) + (4, 4, 4, 4)
659 xxspltw 32+1, 16, 1
664 xxspltw 32+5, 17, 1
668 xxspltw 32+9, 18, 1
672 xxspltw 32+13, 19, 1
678 xxspltw 32+17, 16, 1
683 xxspltw 32+21, 17, 1
687 xxspltw 32+25, 18, 1
691 xxspltw 32+29, 19, 1
706 TP_4x 0, 1, 2, 3
708 TP_4x 8, 9, 10, 11
712 xxlor 1, 49, 49
721 xxlor 49, 1, 1
726 addi 15, 15, -256 # len -=256
732 TP_4x 16+0, 16+1, 16+2, 16+3
734 TP_4x 16+8, 16+9, 16+10, 16+11
744 addi 15, 15, -256 # len +=256
765 lxvw4x 49, 17, 3 # vr17, key 1
776 sradi 8, 7, 1
781 vspltw 1, 16, 1
786 vspltw 5, 17, 1
790 vspltw 9, 18, 1
795 vspltw 13, 19, 1
806 TP_4x 0, 1, 2, 3
808 TP_4x 8, 9, 10, 11
814 addi 15, 15, -256 # len += 256