/linux-6.12.1/drivers/mtd/ |
D | mtdoops.c | 68 static void mark_page_used(struct mtdoops_context *cxt, int page) in mark_page_used() argument 70 set_bit(page, cxt->oops_page_used); in mark_page_used() 73 static void mark_page_unused(struct mtdoops_context *cxt, int page) in mark_page_unused() argument 75 clear_bit(page, cxt->oops_page_used); in mark_page_unused() 78 static int page_is_used(struct mtdoops_context *cxt, int page) in page_is_used() argument 80 return test_bit(page, cxt->oops_page_used); in page_is_used() 83 static int mtdoops_erase_block(struct mtdoops_context *cxt, int offset) in mtdoops_erase_block() argument 85 struct mtd_info *mtd = cxt->mtd; in mtdoops_erase_block() 106 mark_page_unused(cxt, page); in mtdoops_erase_block() 111 static void mtdoops_erase(struct mtdoops_context *cxt) in mtdoops_erase() argument [all …]
|
D | mtdpstore.c | 27 static int mtdpstore_block_isbad(struct mtdpstore_context *cxt, loff_t off) in mtdpstore_block_isbad() argument 30 struct mtd_info *mtd = cxt->mtd; in mtdpstore_block_isbad() 36 if (test_bit(blknum, cxt->badmap)) in mtdpstore_block_isbad() 43 set_bit(blknum, cxt->badmap); in mtdpstore_block_isbad() 49 static inline int mtdpstore_panic_block_isbad(struct mtdpstore_context *cxt, in mtdpstore_panic_block_isbad() argument 52 struct mtd_info *mtd = cxt->mtd; in mtdpstore_panic_block_isbad() 57 return test_bit(blknum, cxt->badmap); in mtdpstore_panic_block_isbad() 60 static inline void mtdpstore_mark_used(struct mtdpstore_context *cxt, in mtdpstore_mark_used() argument 63 struct mtd_info *mtd = cxt->mtd; in mtdpstore_mark_used() 64 u64 zonenum = div_u64(off, cxt->info.kmsg_size); in mtdpstore_mark_used() [all …]
|
/linux-6.12.1/fs/pstore/ |
D | ram.c | 113 struct ramoops_context *cxt = psi->data; in ramoops_pstore_open() local 115 cxt->dump_read_cnt = 0; in ramoops_pstore_open() 116 cxt->console_read_cnt = 0; in ramoops_pstore_open() 117 cxt->ftrace_read_cnt = 0; in ramoops_pstore_open() 118 cxt->pmsg_read_cnt = 0; in ramoops_pstore_open() 185 struct ramoops_context *cxt = record->psi->data; in ramoops_pstore_read() local 200 while (cxt->dump_read_cnt < cxt->max_dump_cnt && !prz) { in ramoops_pstore_read() 201 prz = ramoops_get_next_prz(cxt->dprzs, cxt->dump_read_cnt++, in ramoops_pstore_read() 216 if (!prz_ok(prz) && !cxt->console_read_cnt++) in ramoops_pstore_read() 217 prz = ramoops_get_next_prz(&cxt->cprz, 0 /* single */, record); in ramoops_pstore_read() [all …]
|
D | zone.c | 317 struct psz_context *cxt = &pstore_zone_cxt; in psz_flush_all_dirty_zones() local 320 if (cxt->ppsz) in psz_flush_all_dirty_zones() 321 ret |= psz_flush_dirty_zone(cxt->ppsz); in psz_flush_all_dirty_zones() 322 if (cxt->cpsz) in psz_flush_all_dirty_zones() 323 ret |= psz_flush_dirty_zone(cxt->cpsz); in psz_flush_all_dirty_zones() 324 if (cxt->kpszs) in psz_flush_all_dirty_zones() 325 ret |= psz_flush_dirty_zones(cxt->kpszs, cxt->kmsg_max_cnt); in psz_flush_all_dirty_zones() 326 if (cxt->fpszs) in psz_flush_all_dirty_zones() 327 ret |= psz_flush_dirty_zones(cxt->fpszs, cxt->ftrace_max_cnt); in psz_flush_all_dirty_zones() 328 if (ret && cxt->pstore_zone_info) in psz_flush_all_dirty_zones() [all …]
|
/linux-6.12.1/arch/arm64/kvm/hyp/nvhe/ |
D | tlb.c | 20 struct tlb_inv_context *cxt, in enter_vmid_context() argument 29 cxt->mmu = NULL; in enter_vmid_context() 65 cxt->mmu = vcpu->arch.hw_mmu; in enter_vmid_context() 71 cxt->mmu = host_s2_mmu; in enter_vmid_context() 86 val = cxt->tcr = read_sysreg_el1(SYS_TCR); in enter_vmid_context() 92 val = cxt->sctlr = read_sysreg_el1(SYS_SCTLR); in enter_vmid_context() 100 cxt->sctlr = SCTLR_ELx_M; in enter_vmid_context() 118 static void exit_vmid_context(struct tlb_inv_context *cxt) in exit_vmid_context() argument 120 struct kvm_s2_mmu *mmu = cxt->mmu; in exit_vmid_context() 139 if (!(cxt->sctlr & SCTLR_ELx_M)) { in exit_vmid_context() [all …]
|
/linux-6.12.1/kernel/locking/ |
D | locktorture.c | 163 static struct lock_torture_cxt cxt = { 0, 0, false, false, variable 178 if (long_hold && !(torture_random(trsp) % (cxt.nrealwriters_stress * 2000 * long_hold))) in torture_lock_busted_write_delay() 180 if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 20000))) in torture_lock_busted_write_delay() 200 (cxt.nrealwriters_stress * factor))) { in __torture_rt_boost() 214 (cxt.nrealwriters_stress * factor * 2))) { in __torture_rt_boost() 257 if (long_hold && !(torture_random(trsp) % (cxt.nrealwriters_stress * 2000 * long_hold))) { in torture_spin_lock_write_delay() 262 if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 200 * shortdelay_us))) in torture_spin_lock_write_delay() 264 if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 20000))) in torture_spin_lock_write_delay() 291 cxt.cur_ops->flags = flags; in torture_spin_lock_write_lock_irq() 298 spin_unlock_irqrestore(&torture_spinlock, cxt.cur_ops->flags); in torture_lock_spin_write_unlock_irq() [all …]
|
/linux-6.12.1/arch/arm64/kvm/hyp/vhe/ |
D | tlb.c | 21 struct tlb_inv_context *cxt) in enter_vmid_context() argument 26 local_irq_save(cxt->flags); in enter_vmid_context() 29 cxt->mmu = vcpu->arch.hw_mmu; in enter_vmid_context() 31 cxt->mmu = NULL; in enter_vmid_context() 43 val = cxt->tcr = read_sysreg_el1(SYS_TCR); in enter_vmid_context() 46 val = cxt->sctlr = read_sysreg_el1(SYS_SCTLR); in enter_vmid_context() 70 static void exit_vmid_context(struct tlb_inv_context *cxt) in exit_vmid_context() argument 80 if (cxt->mmu) in exit_vmid_context() 81 __load_stage2(cxt->mmu, cxt->mmu->arch); in exit_vmid_context() 85 write_sysreg_el1(cxt->tcr, SYS_TCR); in exit_vmid_context() [all …]
|
/linux-6.12.1/drivers/scsi/qedi/ |
D | qedi_fw_api.c | 490 struct iscsi_task_context *cxt; in init_rw_iscsi_task() local 501 cxt = task_params->context; in init_rw_iscsi_task() 505 set_local_completion_context(cxt); in init_rw_iscsi_task() 509 cxt->ystorm_st_context.pdu_hdr.r2t.desired_data_trns_len = val; in init_rw_iscsi_task() 510 cxt->mstorm_st_context.expected_itt = in init_rw_iscsi_task() 514 cxt->ystorm_st_context.pdu_hdr.cmd.expected_transfer_length = in init_rw_iscsi_task() 516 init_initiator_rw_cdb_ystorm_context(&cxt->ystorm_st_context, in init_rw_iscsi_task() 519 cxt->mstorm_st_context.sense_db.lo = val; in init_rw_iscsi_task() 522 cxt->mstorm_st_context.sense_db.hi = val; in init_rw_iscsi_task() 526 init_dif_context_flags(&cxt->ystorm_st_context.state.dif_flags, in init_rw_iscsi_task() [all …]
|
/linux-6.12.1/drivers/net/ethernet/qlogic/qed/ |
D | qed_ll2.h | 131 int qed_ll2_acquire_connection(void *cxt, struct qed_ll2_acquire_data *data); 142 int qed_ll2_establish_connection(void *cxt, u8 connection_handle); 157 int qed_ll2_post_rx_buffer(void *cxt, 173 int qed_ll2_prepare_tx_packet(void *cxt, 188 void qed_ll2_release_connection(void *cxt, u8 connection_handle); 203 int qed_ll2_set_fragment_of_tx_packet(void *cxt, 216 int qed_ll2_terminate_connection(void *cxt, u8 connection_handle); 228 int qed_ll2_get_stats(void *cxt,
|
D | qed_ll2.c | 91 static void qed_ll2b_complete_tx_packet(void *cxt, in qed_ll2b_complete_tx_packet() argument 98 struct qed_hwfn *p_hwfn = cxt; in qed_ll2b_complete_tx_packet() 164 static void qed_ll2b_complete_rx_packet(void *cxt, in qed_ll2b_complete_rx_packet() argument 167 struct qed_hwfn *p_hwfn = cxt; in qed_ll2b_complete_rx_packet() 1378 int qed_ll2_acquire_connection(void *cxt, struct qed_ll2_acquire_data *data) in qed_ll2_acquire_connection() argument 1380 struct qed_hwfn *p_hwfn = cxt; in qed_ll2_acquire_connection() 1561 int qed_ll2_establish_connection(void *cxt, u8 connection_handle) in qed_ll2_establish_connection() argument 1566 struct qed_hwfn *p_hwfn = cxt; in qed_ll2_establish_connection() 1774 int qed_ll2_post_rx_buffer(void *cxt, in qed_ll2_post_rx_buffer() argument 1779 struct qed_hwfn *p_hwfn = cxt; in qed_ll2_post_rx_buffer() [all …]
|
D | qed_iwarp.c | 2309 qed_iwarp_ll2_comp_mpa_pkt(void *cxt, struct qed_ll2_comp_rx_data *data) in qed_iwarp_ll2_comp_mpa_pkt() argument 2313 struct qed_hwfn *p_hwfn = cxt; in qed_iwarp_ll2_comp_mpa_pkt() 2355 qed_iwarp_ll2_comp_syn_pkt(void *cxt, struct qed_ll2_comp_rx_data *data) in qed_iwarp_ll2_comp_syn_pkt() argument 2361 struct qed_hwfn *p_hwfn = cxt; in qed_iwarp_ll2_comp_syn_pkt() 2467 static void qed_iwarp_ll2_rel_rx_pkt(void *cxt, u8 connection_handle, in qed_iwarp_ll2_rel_rx_pkt() argument 2472 struct qed_hwfn *p_hwfn = cxt; in qed_iwarp_ll2_rel_rx_pkt() 2479 static void qed_iwarp_ll2_comp_tx_pkt(void *cxt, u8 connection_handle, in qed_iwarp_ll2_comp_tx_pkt() argument 2485 struct qed_hwfn *p_hwfn = cxt; in qed_iwarp_ll2_comp_tx_pkt() 2505 static void qed_iwarp_ll2_rel_tx_pkt(void *cxt, u8 connection_handle, in qed_iwarp_ll2_rel_tx_pkt() argument 2510 struct qed_hwfn *p_hwfn = cxt; in qed_iwarp_ll2_rel_tx_pkt() [all …]
|
/linux-6.12.1/include/linux/qed/ |
D | qed_ll2_if.h | 108 void (*qed_ll2_complete_rx_packet_cb)(void *cxt, 112 void (*qed_ll2_release_rx_packet_cb)(void *cxt, 119 void (*qed_ll2_complete_tx_packet_cb)(void *cxt, 127 void (*qed_ll2_release_tx_packet_cb)(void *cxt, 134 void (*qed_ll2_slowpath_cb)(void *cxt, u8 connection_handle,
|
/linux-6.12.1/tools/perf/util/bpf_skel/ |
D | kwork_top.bpf.c | 195 int on_irq_handler_entry(u64 *cxt) in on_irq_handler_entry() argument 229 int on_irq_handler_exit(u64 *cxt) in on_irq_handler_exit() argument 267 int on_softirq_entry(u64 *cxt) in on_softirq_entry() argument 301 int on_softirq_exit(u64 *cxt) in on_softirq_exit() argument
|
/linux-6.12.1/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_sriov.c | 189 init_p->cxts[0] = q->cxt; in bnx2x_vfop_qctor_prep() 307 if (bnx2x_vfq(vf, qid, cxt)) { in bnx2x_vf_queue_destroy() 308 bnx2x_vfq(vf, qid, cxt)->ustorm_ag_context.cdu_usage = 0; in bnx2x_vf_queue_destroy() 309 bnx2x_vfq(vf, qid, cxt)->xstorm_ag_context.cdu_reserved = 0; in bnx2x_vf_queue_destroy() 1338 struct hw_dma *cxt = &bp->vfdb->context[i]; in bnx2x_iov_free_mem() local 1339 BNX2X_PCI_FREE(cxt->addr, cxt->mapping, cxt->size); in bnx2x_iov_free_mem() 1368 struct hw_dma *cxt = BP_VF_CXT_PAGE(bp, i); in bnx2x_iov_alloc_mem() local 1369 cxt->size = min_t(size_t, tot_size, CDU_ILT_PAGE_SZ); in bnx2x_iov_alloc_mem() 1371 if (cxt->size) { in bnx2x_iov_alloc_mem() 1372 cxt->addr = BNX2X_PCI_ALLOC(&cxt->mapping, cxt->size); in bnx2x_iov_alloc_mem() [all …]
|
D | bnx2x_sriov.h | 73 struct eth_context *cxt; member
|
D | bnx2x_cmn.c | 5072 void bnx2x_set_ctx_validation(struct bnx2x *bp, struct eth_context *cxt, in bnx2x_set_ctx_validation() argument 5075 if (!cxt) { in bnx2x_set_ctx_validation() 5076 BNX2X_ERR("bad context pointer %p\n", cxt); in bnx2x_set_ctx_validation() 5081 cxt->ustorm_ag_context.cdu_usage = in bnx2x_set_ctx_validation() 5085 cxt->xstorm_ag_context.cdu_reserved = in bnx2x_set_ctx_validation()
|
D | bnx2x_cmn.h | 1270 void bnx2x_set_ctx_validation(struct bnx2x *bp, struct eth_context *cxt,
|
/linux-6.12.1/drivers/infiniband/hw/qedr/ |
D | qedr_roce_cm.c | 67 static void qedr_ll2_complete_tx_packet(void *cxt, u8 connection_handle, in qedr_ll2_complete_tx_packet() argument 73 struct qedr_dev *dev = (struct qedr_dev *)cxt; in qedr_ll2_complete_tx_packet() 96 static void qedr_ll2_complete_rx_packet(void *cxt, in qedr_ll2_complete_rx_packet() argument 99 struct qedr_dev *dev = (struct qedr_dev *)cxt; in qedr_ll2_complete_rx_packet() 125 static void qedr_ll2_release_rx_packet(void *cxt, u8 connection_handle, in qedr_ll2_release_rx_packet() argument
|
/linux-6.12.1/drivers/staging/media/meson/vdec/ |
D | codec_vp9.c | 1492 int cxt; in adapt_coef_probs_cxt() local 1494 for (cxt = 0; cxt < cxt_num; cxt++) { in adapt_coef_probs_cxt()
|