/linux-6.12.1/arch/x86/crypto/ |
D | nh-avx2-x86_64.S | 25 #define T0 %ymm8 macro 43 vpaddd \k0, T3, T0 49 vpshufd $0x10, T0, T4 50 vpshufd $0x32, T0, T0 57 vpmuludq T4, T0, T0 61 vpaddq T0, PASS0_SUMS, PASS0_SUMS 143 vpunpcklqdq PASS1_SUMS, PASS0_SUMS, T0 // T0 = (0A 1A 0C 1C) 148 vinserti128 $0x1, T2_XMM, T0, T4 // T4 = (0A 1A 2A 3A) 150 vperm2i128 $0x31, T2, T0, T0 // T0 = (0C 1C 2C 3C) 154 vpaddq T1, T0, T0 [all …]
|
D | poly1305-x86_64-cryptogams.pl | 419 my ($H0,$H1,$H2,$H3,$H4, $T0,$T1,$T2,$T3,$T4, $D0,$D1,$D2,$D3,$D4, $MASK) = 886 vmovdqu 16*2($inp),$T0 890 vpsrldq \$6,$T0,$T2 # splat input 892 vpunpckhqdq $T1,$T0,$T4 # 4 893 vpunpcklqdq $T1,$T0,$T0 # 0:1 897 vpsrlq \$26,$T0,$T1 898 vpand $MASK,$T0,$T0 # 0 979 vpmuludq $T0,$D4,$D0 # d0 = h0*r0 1000 vpmuludq $T0,$H2,$H2 # h0*r1 1007 vpmuludq $T0,$H3,$H3 # h0*r2 [all …]
|
D | aegis128-aesni-asm.S | 20 #define T0 %xmm6 macro 55 movdqa STATE4, T0 60 aesenc T0, STATE3 122 movq (%r8), T0 123 pxor T0, MSG 145 movq T0, %r10 151 psrldq $8, T0 152 movq T0, %r10 386 movdqa MSG, T0 387 pxor \s1, T0 [all …]
|
D | nh-sse2-x86_64.S | 21 #define T0 %xmm8 macro 113 movdqa PASS0_SUMS, T0 115 punpcklqdq PASS1_SUMS, T0 // => (PASS0_SUM_A PASS1_SUM_A) 119 paddq PASS0_SUMS, T0 121 movdqu T0, 0x00(HASH)
|
D | twofish-x86_64-asm_64-3way.S | 77 #define do16bit_ror(rot, op1, op2, T0, T1, tmp1, tmp2, ab, dst) \ argument 81 op1##l T0(CTX, tmp2, 4), dst ## d; \
|
D | camellia-x86_64-asm_64.S | 79 #define xor2ror16(T0, T1, tmp1, tmp2, ab, dst) \ argument 80 leaq T0(%rip), tmp1; \
|
/linux-6.12.1/arch/sparc/crypto/ |
D | aes_asm.S | 7 #define ENCRYPT_TWO_ROUNDS(KEY_BASE, I0, I1, T0, T1) \ argument 8 AES_EROUND01(KEY_BASE + 0, I0, I1, T0) \ 10 AES_EROUND01(KEY_BASE + 4, T0, T1, I0) \ 11 AES_EROUND23(KEY_BASE + 6, T0, T1, I1) 13 #define ENCRYPT_TWO_ROUNDS_2(KEY_BASE, I0, I1, I2, I3, T0, T1, T2, T3) \ argument 14 AES_EROUND01(KEY_BASE + 0, I0, I1, T0) \ 18 AES_EROUND01(KEY_BASE + 4, T0, T1, I0) \ 19 AES_EROUND23(KEY_BASE + 6, T0, T1, I1) \ 23 #define ENCRYPT_TWO_ROUNDS_LAST(KEY_BASE, I0, I1, T0, T1) \ argument 24 AES_EROUND01(KEY_BASE + 0, I0, I1, T0) \ [all …]
|
/linux-6.12.1/arch/arm64/crypto/ |
D | sha512-armv8.pl | 109 my ($T0,$T1,$T2)=(@X[($i-8)&15],@X[($i-9)&15],@X[($i-10)&15]); 110 $T0=@X[$i+3] if ($i<11); 142 eor $T0,$e,$e,ror#`$Sigma1[2]-$Sigma1[1]` 148 eor $t0,$t0,$T0,ror#$Sigma1[1] // Sigma1(e) 149 ror $T0,$a,#$Sigma0[0] 156 eor $t1,$T0,$t1,ror#$Sigma0[1] // Sigma0(a) 168 ror $T0,$a,#$Sigma0[0] 175 eor $T0,$T0,$a,ror#$Sigma0[1] 182 eor $t1,$T0,$a,ror#$Sigma0[2] // Sigma0(a) 463 my ($T0,$T1,$T2,$T3,$T4,$T5,$T6,$T7) = map("q$_",(4..7,16..19)); [all …]
|
D | sm4-ce-gcm-core.S | 39 #define PMUL_128x128(r0, r1, m0, m1, T0, T1) \ argument 40 ext T0.16b, m1.16b, m1.16b, #8; \ 42 pmull T1.1q, m0.1d, T0.1d; \ 43 pmull2 T0.1q, m0.2d, T0.2d; \ 45 eor T0.16b, T0.16b, T1.16b; \ 46 ext T1.16b, RZERO.16b, T0.16b, #8; \ 47 ext T0.16b, T0.16b, RZERO.16b, #8; \ 49 eor r1.16b, r1.16b, T0.16b; 51 #define PMUL_128x128_4x(r0, r1, m0, m1, T0, T1, \ argument 55 ext T0.16b, m1.16b, m1.16b, #8; \ [all …]
|
D | poly1305-armv8.pl | 267 my ($T0,$T1,$MASK) = map("v$_",(29..31)); 706 ushr $T0.2d,$ACC3,#26 710 add $ACC4,$ACC4,$T0.2d // h3 -> h4 714 ushr $T0.2d,$ACC4,#26 721 add $ACC0,$ACC0,$T0.2d 722 shl $T0.2d,$T0.2d,#2 725 add $ACC0,$ACC0,$T0.2d // h4 -> h0 730 shrn $T0.2s,$ACC0,#26 735 add $H1,$H1,$T0.2s // h0 -> h1 848 ushr $T0.2d,$ACC3,#26 [all …]
|
D | nh-neon-core.S | 26 T0 .req v8 44 add T0.4s, T3.4s, \k0\().4s 50 mov T4.d[0], T0.d[1] 54 umlal PASS0_SUMS.2d, T0.2s, T4.2s 100 addp T0.2d, PASS0_SUMS.2d, PASS1_SUMS.2d 102 st1 {T0.16b,T1.16b}, [HASH]
|
/linux-6.12.1/arch/arm/crypto/ |
D | sha256-armv4.pl | 292 my ($T0,$T1,$T2,$T3,$T4,$T5)=("q8","q9","q10","q11","d24","d25"); 312 &vext_8 ($T0,@X[0],@X[1],4); # X[1..4] 320 &vshr_u32 ($T2,$T0,$sigma0[0]); 326 &vshr_u32 ($T1,$T0,$sigma0[2]); 329 &vsli_32 ($T2,$T0,32-$sigma0[0]); 332 &vshr_u32 ($T3,$T0,$sigma0[1]); 338 &vsli_32 ($T3,$T0,32-$sigma0[1]); 386 &vld1_32 ("{$T0}","[$Ktbl,:128]!"); 398 &vadd_i32 ($T0,$T0,@X[0]); 400 &vst1_32 ("{$T0}","[$Xfer,:128]!"); [all …]
|
D | poly1305-armv4.pl | 497 my ($T0,$T1,$MASK) = map("q$_",(15,4,0)); 633 vshr.u64 $T0,$D3,#26 637 vadd.i64 $D4,$D4,$T0 @ h3 -> h4 642 vshrn.u64 $T0#lo,$D4,#26 650 vadd.i32 $D0#lo,$D0#lo,$T0#lo 651 vshl.u32 $T0#lo,$T0#lo,#2 654 vadd.i32 $D0#lo,$D0#lo,$T0#lo @ h4 -> h0 658 vshr.u32 $T0#lo,$D0#lo,#26 662 vadd.i32 $D1#lo,$D1#lo,$T0#lo @ h0 -> h1 996 vshr.u64 $T0,$D3,#26 [all …]
|
D | nh-neon-core.S | 36 T0 .req q8 58 vadd.u32 T0, T3, \k0 114 vst1.8 {T0-T1}, [HASH]
|
/linux-6.12.1/crypto/ |
D | anubis.c | 52 static const u32 T0[256] = { variable 534 inter[i] = T0[(kappa[j--] >> 24) ]; in anubis_setkey() 562 T0[T4[(v >> 24) ] & 0xff] ^ in anubis_setkey() 594 T0[(state[0] >> 24) ] ^ in anubis_crypt() 600 T0[(state[0] >> 16) & 0xff] ^ in anubis_crypt() 606 T0[(state[0] >> 8) & 0xff] ^ in anubis_crypt() 612 T0[(state[0] ) & 0xff] ^ in anubis_crypt() 628 (T0[(state[0] >> 24) ] & 0xff000000U) ^ in anubis_crypt() 634 (T0[(state[0] >> 16) & 0xff] & 0xff000000U) ^ in anubis_crypt() 640 (T0[(state[0] >> 8) & 0xff] & 0xff000000U) ^ in anubis_crypt() [all …]
|
D | khazad.c | 38 static const u64 T0[256] = { variable 771 ctx->E[r] = T0[(int)(K1 >> 56) ] ^ in khazad_setkey() 787 ctx->D[r] = T0[(int)S[(int)(K1 >> 56) ] & 0xff] ^ in khazad_setkey() 813 state = T0[(int)(state >> 56) ] ^ in khazad_crypt() 824 state = (T0[(int)(state >> 56) ] & 0xff00000000000000ULL) ^ in khazad_crypt()
|
/linux-6.12.1/arch/mips/crypto/ |
D | chacha-core.S | 28 #define T0 $s1 macro 130 lw T0, (x*4)(STATE); \ 137 addu X ## x, T0; \ 147 lw T0, (x*4)(STATE); \ 153 addu X ## x, T0; \ 322 lui T0, %hi(.Lchacha_mips_jmptbl_aligned_0) 325 ins T0, $at, 1, 6 331 addiu T0, %lo(.Lchacha_mips_jmptbl_aligned_0) 339 jr T0 402 lui T0, %hi(.Lchacha_mips_jmptbl_unaligned_0) [all …]
|
/linux-6.12.1/tools/perf/arch/riscv/util/ |
D | unwind-libdw.c | 27 dwarf_regs[5] = REG(T0); in libdw__arch_set_initial_registers()
|
/linux-6.12.1/kernel/trace/ |
D | trace_irqsoff.c | 329 u64 T0, T1, delta; in check_critical_timing() local 333 T0 = data->preempt_timestamp; in check_critical_timing() 335 delta = T1-T0; in check_critical_timing()
|
D | trace_sched_wakeup.c | 437 u64 T0, T1, delta; in probe_wakeup_sched_switch() local 482 T0 = data->preempt_timestamp; in probe_wakeup_sched_switch() 484 delta = T1-T0; in probe_wakeup_sched_switch()
|
/linux-6.12.1/tools/perf/ |
D | builtin-sched.c | 300 u64 T0 = get_nsecs(), T1; in burn_nsecs() local 304 } while (T1 + sched->run_measurement_overhead < T0 + nsecs); in burn_nsecs() 319 u64 T0, T1, delta, min_delta = NSEC_PER_SEC; in calibrate_run_measurement_overhead() local 323 T0 = get_nsecs(); in calibrate_run_measurement_overhead() 326 delta = T1-T0; in calibrate_run_measurement_overhead() 336 u64 T0, T1, delta, min_delta = NSEC_PER_SEC; in calibrate_sleep_measurement_overhead() local 340 T0 = get_nsecs(); in calibrate_sleep_measurement_overhead() 343 delta = T1-T0; in calibrate_sleep_measurement_overhead() 763 u64 T0, T1, delta, avg_delta, fluct; in run_one_test() local 765 T0 = get_nsecs(); in run_one_test() [all …]
|
/linux-6.12.1/arch/arm/boot/dts/aspeed/ |
D | aspeed-bmc-qcom-dc-scm-v1.dts | 114 /*T0-T7*/ "","","","","","","","",
|
D | aspeed-bmc-vegman-n110.dts | 34 /*T0-T7*/ "","","","","","","","",
|
D | aspeed-bmc-vegman-sx20.dts | 34 /*T0-T7*/ "","","","","","","","",
|
/linux-6.12.1/drivers/ata/ |
D | pata_octeon_cf.c | 221 unsigned int T0, Tkr, Td; in octeon_cf_set_dmamode() local 228 T0 = timing->cycle; in octeon_cf_set_dmamode() 244 oe_n = max(T0 - oe_a, Tkr); in octeon_cf_set_dmamode()
|