/linux-master/arch/riscv/lib/ |
H A D | memcpy.S | 11 move t6, a0 /* Preserve return value */ 17 andi a3, t6, SZREG-1 33 sb a5, 0(t6) 34 addi t6, t6, 1 53 REG_S a4, 0(t6) 54 REG_S a5, SZREG(t6) 55 REG_S a6, 2*SZREG(t6) 56 REG_S a7, 3*SZREG(t6) 57 REG_S t0, 4*SZREG(t6) [all...] |
H A D | uaccess_vector.S | 25 li t6, SR_SUM 26 csrs CSR_STATUS, t6 40 csrc CSR_STATUS, t6
|
H A D | uaccess.S | 25 li t6, SR_SUM 26 csrs CSR_STATUS, t6 182 csrc CSR_STATUS, t6 189 csrc CSR_STATUS, t6 202 li t6, SR_SUM 203 csrs CSR_STATUS, t6 225 csrc CSR_STATUS, t6 242 csrc CSR_STATUS, t6
|
H A D | memmove.S | 38 * Both Copy Modes: t6 - Non-Inclusive last multibyte/aligned of dest 50 * Byte copy does not need t5 or t6. 65 * Now solve for t5 and t6. 68 andi t6, t4, -SZREG 102 addi a2, t6, SZREG /* The other breakpoint for the unrolled loop*/ 150 bne t3, t6, 1b 152 mv t3, t6 /* Fix the dest pointer in case the loop was broken */ 235 bne t3, t6, 1b 272 beq t4, t6, 2f 278 bne t4, t6, [all...] |
H A D | strncmp.S | 65 * t0, t1, t2, t3, t4, t5, t6 75 andi t6, t4, -SZREG 80 bge a0, t6, 3f
|
/linux-master/scripts/ |
H A D | makelst | 29 t6=`printf "%lu" $((0x$t4 - 0x$t5))` 32 $3 -r --source --adjust-vma=${t6:-0} $1
|
/linux-master/arch/alpha/lib/ |
H A D | stxcpy.S | 73 negq t8, t6 # e0 : find low bit set 74 and t8, t6, t12 # e1 (stall) 78 and t12, 0x80, t6 # e0 : 79 bne t6, 1f # .. e1 (zdb) 84 subq t12, 1, t6 # .. e1 : 85 zapnot t1, t6, t1 # e0 : clear src bytes >= null 86 or t12, t6, t8 # .. e1 : 130 t6 == bytemask that is -1 in dest word bytes */ 142 or t1, t6, t6 # e [all...] |
H A D | ev6-stxcpy.S | 89 negq t8, t6 # E : find low bit set 90 and t8, t6, t12 # E : (stall) 93 and t12, 0x80, t6 # E : (stall) 94 bne t6, 1f # U : (stall) 99 subq t12, 1, t6 # E : 100 zapnot t1, t6, t1 # U : clear src bytes >= null (stall) 101 or t12, t6, t8 # E : (stall) 153 t6 == bytemask that is -1 in dest word bytes */ 165 or t1, t6, t6 # [all...] |
H A D | stxncpy.S | 91 and t12, 0x80, t6 # e0 : 92 bne t6, 1f # .. e1 (zdb) 97 subq t12, 1, t6 # .. e1 : 98 or t12, t6, t8 # e0 : 156 t6 == bytemask that is -1 in dest word bytes */ 165 or t0, t6, t6 # e1 : mask original data for zero test 166 cmpbge zero, t6, t8 # e0 : 168 lda t6, -1 # e0 : 171 mskql t6, a [all...] |
H A D | ev6-stxncpy.S | 117 and t12, 0x80, t6 # E : (stall) 118 bne t6, 1f # U : (stall) 123 subq t12, 1, t6 # E : 124 or t12, t6, t8 # E : (stall) 195 t6 == bytemask that is -1 in dest word bytes */ 205 or t0, t6, t6 # E : mask original data for zero test (stall) 207 cmpbge zero, t6, t8 # E : 209 lda t6, -1 # E : 213 mskql t6, a [all...] |
H A D | strrchr.S | 23 mov zero, t6 # .. e1 : t6 is last match aligned addr 46 cmovne t3, v0, t6 # .. e1 : save previous comparisons match 63 cmovne t3, v0, t6 # e0 : 80 addq t6, t0, v0 # .. e1 : add our aligned base ptr to the mix
|
H A D | ev67-strrchr.S | 40 mov zero, t6 # E : t6 is last match aligned addr 68 cmovne t3, v0, t6 # E : save previous comparisons match 94 cmovne t3, v0, t6 # E : 105 addq t6, t5, v0 # E : and add to quadword address
|
/linux-master/arch/alpha/include/uapi/asm/ |
H A D | regdef.h | 13 #define t6 $7 macro
|
/linux-master/arch/arm64/crypto/ |
H A D | crct10dif-ce-core.S | 86 t6 .req v20 138 ext t6.8b, ad.8b, ad.8b, #3 // A3 144 pmull t6.8h, t6.8b, fold_consts.8b // J = A3*B 152 tbl t6.16b, {ad.16b}, perm3.16b // A3 158 pmull2 t6.8h, t6.16b, fold_consts.16b // J = A3*B 164 eor t6.16b, t6.16b, t9.16b // N = I + J 168 uzp1 t7.2d, t6 [all...] |
/linux-master/arch/loongarch/include/asm/ |
H A D | regdef.h | 26 #define t6 $r18 macro
|
/linux-master/arch/loongarch/mm/ |
H A D | page.S | 50 ld.d t6, a1, 48 65 st.d t6, a0, 48 67 ld.d t6, a1, 112 78 st.d t6, a0, -16
|
/linux-master/arch/riscv/include/asm/ |
H A D | compat.h | 71 compat_ulong_t t6; member in struct:compat_user_regs_struct 108 cregs->t6 = (compat_ulong_t) regs->t6; 145 regs->t6 = (unsigned long) cregs->t6;
|
/linux-master/arch/mips/kernel/ |
H A D | scall32-o32.S | 62 load_a5: user_lw(t6, 20(t0)) # argument #6 from usp 68 sw t6, 20(sp) # argument #6 to ksp 161 li t6, 0 197 lw t6, 28(sp) 200 sw t6, 24(sp)
|
/linux-master/arch/sparc/lib/ |
H A D | blockops.S | 28 #define MIRROR_BLOCK(dst, src, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ 32 ldd [src + offset + 0x00], t6; \ 36 std t6, [dst + offset + 0x00];
|
/linux-master/arch/x86/crypto/ |
H A D | aria-aesni-avx2-asm_64.S | 309 t4, t5, t6, t7) \ 332 t4, t5, t6, t7) \ 334 vpxor t6, t6, t6; \ 342 vextracti128 $1, x0, t6##_x; \ 344 vaesenclast t7##_x, t6##_x, t6##_x; \ 345 vinserti128 $1, t6##_x, x0, x0; \ 347 vextracti128 $1, x4, t6##_ [all...] |
H A D | camellia-aesni-avx-asm_64.S | 50 #define roundsm16(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \ 73 filter_8bit(x0, t0, t1, t7, t6); \ 74 filter_8bit(x7, t0, t1, t7, t6); \ 75 filter_8bit(x1, t0, t1, t7, t6); \ 76 filter_8bit(x4, t0, t1, t7, t6); \ 77 filter_8bit(x2, t0, t1, t7, t6); \ 78 filter_8bit(x5, t0, t1, t7, t6); \ 82 filter_8bit(x3, t2, t3, t7, t6); \ 83 filter_8bit(x6, t2, t3, t7, t6); \ 100 filter_8bit(x0, t0, t1, t7, t6); \ [all...] |
H A D | camellia-aesni-avx2-asm_64.S | 62 #define roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \ 70 vbroadcasti128 .Lpre_tf_hi_s1(%rip), t6; \ 86 filter_8bit(x0, t5, t6, t7, t4); \ 87 filter_8bit(x7, t5, t6, t7, t4); \ 94 filter_8bit(x2, t5, t6, t7, t4); \ 95 filter_8bit(x5, t5, t6, t7, t4); \ 96 filter_8bit(x1, t5, t6, t7, t4); \ 97 filter_8bit(x4, t5, t6, t7, t4); \ 102 vextracti128 $1, x2, t6##_x; \ 121 vaesenclast t4##_x, t6##_ [all...] |
/linux-master/arch/riscv/include/uapi/asm/ |
H A D | ptrace.h | 56 unsigned long t6; member in struct:user_regs_struct
|
/linux-master/arch/riscv/kernel/ |
H A D | mcount.S | 92 la t6, ftrace_graph_entry_stub 93 bne t2, t6, .Ldo_ftrace_graph_caller
|
/linux-master/arch/loongarch/kernel/ |
H A D | rethook_trampoline.S | 24 cfi_st t6, PT_R18 59 cfi_ld t6, PT_R18
|