/linux-6.12.1/arch/x86/tools/ |
D | insn_sanity.c | 82 unsigned char *insn_buff, struct insn *insn) in dump_stream() argument 95 fprintf(fp, " %02x", insn_buff[i]); in dump_stream() 123 static int read_next_insn(unsigned char *insn_buff) in read_next_insn() argument 133 insn_buff[i] = (unsigned char)strtoul(tmp, &tmp, 16); in read_next_insn() 141 static int generate_insn(unsigned char *insn_buff) in generate_insn() argument 146 return read_next_insn(insn_buff); in generate_insn() 150 *(unsigned short *)(&insn_buff[i]) = random() & 0xffff; in generate_insn() 153 insn_buff[i++] = random() & 0xff; in generate_insn() 225 unsigned char insn_buff[MAX_INSN_SIZE * 2]; in main() local 230 memset(insn_buff + MAX_INSN_SIZE, INSN_NOP, MAX_INSN_SIZE); in main() [all …]
|
D | insn_decoder_test.c | 114 unsigned char insn_buff[16]; in main() local 133 memset(insn_buff, 0, 16); in main() 146 insn_buff[nb++] = (unsigned char) b; in main() 153 ret = insn_decode(&insn, insn_buff, sizeof(insn_buff), in main()
|
/linux-6.12.1/arch/x86/kernel/ |
D | callthunks.c | 184 u8 insn_buff[MAX_PATCH_LEN]; in patch_dest() local 187 memcpy(insn_buff, skl_call_thunk_template, tsize); in patch_dest() 188 apply_relocation(insn_buff, pad, tsize, skl_call_thunk_template, tsize); in patch_dest() 191 if (!bcmp(pad, insn_buff, tsize)) in patch_dest() 201 memcpy(pad, insn_buff, tsize); in patch_dest() 203 text_poke_copy_locked(pad, insn_buff, tsize, true); in patch_dest() 299 u8 insn_buff[MAX_PATCH_LEN]; in is_callthunk() local 309 memcpy(insn_buff, skl_call_thunk_template, tmpl_size); in is_callthunk() 310 apply_relocation(insn_buff, pad, tmpl_size, skl_call_thunk_template, tmpl_size); in is_callthunk() 312 return !bcmp(pad, insn_buff, tmpl_size); in is_callthunk() [all …]
|
D | alternative.c | 395 static int alt_replace_call(u8 *instr, u8 *insn_buff, struct alt_instr *a) in alt_replace_call() argument 400 if (a->replacementlen != 5 || insn_buff[0] != CALL_INSN_OPCODE) { in alt_replace_call() 427 *(s32 *)(insn_buff + 1) += target - bug; in alt_replace_call() 453 u8 insn_buff[MAX_PATCH_LEN]; in apply_alternatives() local 494 BUG_ON(a->instrlen > sizeof(insn_buff)); in apply_alternatives() 504 memcpy(insn_buff, instr, a->instrlen); in apply_alternatives() 505 optimize_nops(instr, insn_buff, a->instrlen); in apply_alternatives() 506 text_poke_early(instr, insn_buff, a->instrlen); in apply_alternatives() 516 memcpy(insn_buff, replacement, a->replacementlen); in apply_alternatives() 520 insn_buff_sz = alt_replace_call(instr, insn_buff, a); in apply_alternatives() [all …]
|
/linux-6.12.1/arch/x86/kernel/kprobes/ |
D | opt.c | 476 u8 insn_buff[JMP32_INSN_SIZE]; in arch_optimize_kprobes() local 488 insn_buff[0] = JMP32_INSN_OPCODE; in arch_optimize_kprobes() 489 *(s32 *)(&insn_buff[1]) = rel; in arch_optimize_kprobes() 491 text_poke_bp(op->kp.addr, insn_buff, JMP32_INSN_SIZE, NULL); in arch_optimize_kprobes()
|
/linux-6.12.1/arch/x86/events/intel/ |
D | ds.c | 604 void *buffer, *insn_buff, *cea; in alloc_pebs_buffer() local 618 insn_buff = kzalloc_node(PEBS_FIXUP_SIZE, GFP_KERNEL, node); in alloc_pebs_buffer() 619 if (!insn_buff) { in alloc_pebs_buffer() 623 per_cpu(insn_buffer, cpu) = insn_buff; in alloc_pebs_buffer()
|
/linux-6.12.1/kernel/bpf/ |
D | core.c | 1468 struct bpf_insn insn_buff[16], aux[2]; in bpf_jit_blind_constants() local 1503 rewritten = bpf_jit_blind_insn(insn, aux, insn_buff, in bpf_jit_blind_constants() 1508 tmp = bpf_patch_insn_single(clone, i, insn_buff, rewritten); in bpf_jit_blind_constants()
|