Home
last modified time | relevance | path

Searched refs:cxt (Results 1 – 19 of 19) sorted by relevance

/linux-6.12.1/drivers/mtd/
Dmtdoops.c68 static void mark_page_used(struct mtdoops_context *cxt, int page) in mark_page_used() argument
70 set_bit(page, cxt->oops_page_used); in mark_page_used()
73 static void mark_page_unused(struct mtdoops_context *cxt, int page) in mark_page_unused() argument
75 clear_bit(page, cxt->oops_page_used); in mark_page_unused()
78 static int page_is_used(struct mtdoops_context *cxt, int page) in page_is_used() argument
80 return test_bit(page, cxt->oops_page_used); in page_is_used()
83 static int mtdoops_erase_block(struct mtdoops_context *cxt, int offset) in mtdoops_erase_block() argument
85 struct mtd_info *mtd = cxt->mtd; in mtdoops_erase_block()
106 mark_page_unused(cxt, page); in mtdoops_erase_block()
111 static void mtdoops_erase(struct mtdoops_context *cxt) in mtdoops_erase() argument
[all …]
Dmtdpstore.c27 static int mtdpstore_block_isbad(struct mtdpstore_context *cxt, loff_t off) in mtdpstore_block_isbad() argument
30 struct mtd_info *mtd = cxt->mtd; in mtdpstore_block_isbad()
36 if (test_bit(blknum, cxt->badmap)) in mtdpstore_block_isbad()
43 set_bit(blknum, cxt->badmap); in mtdpstore_block_isbad()
49 static inline int mtdpstore_panic_block_isbad(struct mtdpstore_context *cxt, in mtdpstore_panic_block_isbad() argument
52 struct mtd_info *mtd = cxt->mtd; in mtdpstore_panic_block_isbad()
57 return test_bit(blknum, cxt->badmap); in mtdpstore_panic_block_isbad()
60 static inline void mtdpstore_mark_used(struct mtdpstore_context *cxt, in mtdpstore_mark_used() argument
63 struct mtd_info *mtd = cxt->mtd; in mtdpstore_mark_used()
64 u64 zonenum = div_u64(off, cxt->info.kmsg_size); in mtdpstore_mark_used()
[all …]
/linux-6.12.1/fs/pstore/
Dram.c113 struct ramoops_context *cxt = psi->data; in ramoops_pstore_open() local
115 cxt->dump_read_cnt = 0; in ramoops_pstore_open()
116 cxt->console_read_cnt = 0; in ramoops_pstore_open()
117 cxt->ftrace_read_cnt = 0; in ramoops_pstore_open()
118 cxt->pmsg_read_cnt = 0; in ramoops_pstore_open()
185 struct ramoops_context *cxt = record->psi->data; in ramoops_pstore_read() local
200 while (cxt->dump_read_cnt < cxt->max_dump_cnt && !prz) { in ramoops_pstore_read()
201 prz = ramoops_get_next_prz(cxt->dprzs, cxt->dump_read_cnt++, in ramoops_pstore_read()
216 if (!prz_ok(prz) && !cxt->console_read_cnt++) in ramoops_pstore_read()
217 prz = ramoops_get_next_prz(&cxt->cprz, 0 /* single */, record); in ramoops_pstore_read()
[all …]
Dzone.c317 struct psz_context *cxt = &pstore_zone_cxt; in psz_flush_all_dirty_zones() local
320 if (cxt->ppsz) in psz_flush_all_dirty_zones()
321 ret |= psz_flush_dirty_zone(cxt->ppsz); in psz_flush_all_dirty_zones()
322 if (cxt->cpsz) in psz_flush_all_dirty_zones()
323 ret |= psz_flush_dirty_zone(cxt->cpsz); in psz_flush_all_dirty_zones()
324 if (cxt->kpszs) in psz_flush_all_dirty_zones()
325 ret |= psz_flush_dirty_zones(cxt->kpszs, cxt->kmsg_max_cnt); in psz_flush_all_dirty_zones()
326 if (cxt->fpszs) in psz_flush_all_dirty_zones()
327 ret |= psz_flush_dirty_zones(cxt->fpszs, cxt->ftrace_max_cnt); in psz_flush_all_dirty_zones()
328 if (ret && cxt->pstore_zone_info) in psz_flush_all_dirty_zones()
[all …]
/linux-6.12.1/arch/arm64/kvm/hyp/nvhe/
Dtlb.c20 struct tlb_inv_context *cxt, in enter_vmid_context() argument
29 cxt->mmu = NULL; in enter_vmid_context()
65 cxt->mmu = vcpu->arch.hw_mmu; in enter_vmid_context()
71 cxt->mmu = host_s2_mmu; in enter_vmid_context()
86 val = cxt->tcr = read_sysreg_el1(SYS_TCR); in enter_vmid_context()
92 val = cxt->sctlr = read_sysreg_el1(SYS_SCTLR); in enter_vmid_context()
100 cxt->sctlr = SCTLR_ELx_M; in enter_vmid_context()
118 static void exit_vmid_context(struct tlb_inv_context *cxt) in exit_vmid_context() argument
120 struct kvm_s2_mmu *mmu = cxt->mmu; in exit_vmid_context()
139 if (!(cxt->sctlr & SCTLR_ELx_M)) { in exit_vmid_context()
[all …]
/linux-6.12.1/kernel/locking/
Dlocktorture.c163 static struct lock_torture_cxt cxt = { 0, 0, false, false, variable
178 if (long_hold && !(torture_random(trsp) % (cxt.nrealwriters_stress * 2000 * long_hold))) in torture_lock_busted_write_delay()
180 if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 20000))) in torture_lock_busted_write_delay()
200 (cxt.nrealwriters_stress * factor))) { in __torture_rt_boost()
214 (cxt.nrealwriters_stress * factor * 2))) { in __torture_rt_boost()
257 if (long_hold && !(torture_random(trsp) % (cxt.nrealwriters_stress * 2000 * long_hold))) { in torture_spin_lock_write_delay()
262 if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 200 * shortdelay_us))) in torture_spin_lock_write_delay()
264 if (!(torture_random(trsp) % (cxt.nrealwriters_stress * 20000))) in torture_spin_lock_write_delay()
291 cxt.cur_ops->flags = flags; in torture_spin_lock_write_lock_irq()
298 spin_unlock_irqrestore(&torture_spinlock, cxt.cur_ops->flags); in torture_lock_spin_write_unlock_irq()
[all …]
/linux-6.12.1/arch/arm64/kvm/hyp/vhe/
Dtlb.c21 struct tlb_inv_context *cxt) in enter_vmid_context() argument
26 local_irq_save(cxt->flags); in enter_vmid_context()
29 cxt->mmu = vcpu->arch.hw_mmu; in enter_vmid_context()
31 cxt->mmu = NULL; in enter_vmid_context()
43 val = cxt->tcr = read_sysreg_el1(SYS_TCR); in enter_vmid_context()
46 val = cxt->sctlr = read_sysreg_el1(SYS_SCTLR); in enter_vmid_context()
70 static void exit_vmid_context(struct tlb_inv_context *cxt) in exit_vmid_context() argument
80 if (cxt->mmu) in exit_vmid_context()
81 __load_stage2(cxt->mmu, cxt->mmu->arch); in exit_vmid_context()
85 write_sysreg_el1(cxt->tcr, SYS_TCR); in exit_vmid_context()
[all …]
/linux-6.12.1/drivers/scsi/qedi/
Dqedi_fw_api.c490 struct iscsi_task_context *cxt; in init_rw_iscsi_task() local
501 cxt = task_params->context; in init_rw_iscsi_task()
505 set_local_completion_context(cxt); in init_rw_iscsi_task()
509 cxt->ystorm_st_context.pdu_hdr.r2t.desired_data_trns_len = val; in init_rw_iscsi_task()
510 cxt->mstorm_st_context.expected_itt = in init_rw_iscsi_task()
514 cxt->ystorm_st_context.pdu_hdr.cmd.expected_transfer_length = in init_rw_iscsi_task()
516 init_initiator_rw_cdb_ystorm_context(&cxt->ystorm_st_context, in init_rw_iscsi_task()
519 cxt->mstorm_st_context.sense_db.lo = val; in init_rw_iscsi_task()
522 cxt->mstorm_st_context.sense_db.hi = val; in init_rw_iscsi_task()
526 init_dif_context_flags(&cxt->ystorm_st_context.state.dif_flags, in init_rw_iscsi_task()
[all …]
/linux-6.12.1/drivers/net/ethernet/qlogic/qed/
Dqed_ll2.h131 int qed_ll2_acquire_connection(void *cxt, struct qed_ll2_acquire_data *data);
142 int qed_ll2_establish_connection(void *cxt, u8 connection_handle);
157 int qed_ll2_post_rx_buffer(void *cxt,
173 int qed_ll2_prepare_tx_packet(void *cxt,
188 void qed_ll2_release_connection(void *cxt, u8 connection_handle);
203 int qed_ll2_set_fragment_of_tx_packet(void *cxt,
216 int qed_ll2_terminate_connection(void *cxt, u8 connection_handle);
228 int qed_ll2_get_stats(void *cxt,
Dqed_ll2.c91 static void qed_ll2b_complete_tx_packet(void *cxt, in qed_ll2b_complete_tx_packet() argument
98 struct qed_hwfn *p_hwfn = cxt; in qed_ll2b_complete_tx_packet()
164 static void qed_ll2b_complete_rx_packet(void *cxt, in qed_ll2b_complete_rx_packet() argument
167 struct qed_hwfn *p_hwfn = cxt; in qed_ll2b_complete_rx_packet()
1378 int qed_ll2_acquire_connection(void *cxt, struct qed_ll2_acquire_data *data) in qed_ll2_acquire_connection() argument
1380 struct qed_hwfn *p_hwfn = cxt; in qed_ll2_acquire_connection()
1561 int qed_ll2_establish_connection(void *cxt, u8 connection_handle) in qed_ll2_establish_connection() argument
1566 struct qed_hwfn *p_hwfn = cxt; in qed_ll2_establish_connection()
1774 int qed_ll2_post_rx_buffer(void *cxt, in qed_ll2_post_rx_buffer() argument
1779 struct qed_hwfn *p_hwfn = cxt; in qed_ll2_post_rx_buffer()
[all …]
Dqed_iwarp.c2309 qed_iwarp_ll2_comp_mpa_pkt(void *cxt, struct qed_ll2_comp_rx_data *data) in qed_iwarp_ll2_comp_mpa_pkt() argument
2313 struct qed_hwfn *p_hwfn = cxt; in qed_iwarp_ll2_comp_mpa_pkt()
2355 qed_iwarp_ll2_comp_syn_pkt(void *cxt, struct qed_ll2_comp_rx_data *data) in qed_iwarp_ll2_comp_syn_pkt() argument
2361 struct qed_hwfn *p_hwfn = cxt; in qed_iwarp_ll2_comp_syn_pkt()
2467 static void qed_iwarp_ll2_rel_rx_pkt(void *cxt, u8 connection_handle, in qed_iwarp_ll2_rel_rx_pkt() argument
2472 struct qed_hwfn *p_hwfn = cxt; in qed_iwarp_ll2_rel_rx_pkt()
2479 static void qed_iwarp_ll2_comp_tx_pkt(void *cxt, u8 connection_handle, in qed_iwarp_ll2_comp_tx_pkt() argument
2485 struct qed_hwfn *p_hwfn = cxt; in qed_iwarp_ll2_comp_tx_pkt()
2505 static void qed_iwarp_ll2_rel_tx_pkt(void *cxt, u8 connection_handle, in qed_iwarp_ll2_rel_tx_pkt() argument
2510 struct qed_hwfn *p_hwfn = cxt; in qed_iwarp_ll2_rel_tx_pkt()
[all …]
/linux-6.12.1/include/linux/qed/
Dqed_ll2_if.h108 void (*qed_ll2_complete_rx_packet_cb)(void *cxt,
112 void (*qed_ll2_release_rx_packet_cb)(void *cxt,
119 void (*qed_ll2_complete_tx_packet_cb)(void *cxt,
127 void (*qed_ll2_release_tx_packet_cb)(void *cxt,
134 void (*qed_ll2_slowpath_cb)(void *cxt, u8 connection_handle,
/linux-6.12.1/tools/perf/util/bpf_skel/
Dkwork_top.bpf.c195 int on_irq_handler_entry(u64 *cxt) in on_irq_handler_entry() argument
229 int on_irq_handler_exit(u64 *cxt) in on_irq_handler_exit() argument
267 int on_softirq_entry(u64 *cxt) in on_softirq_entry() argument
301 int on_softirq_exit(u64 *cxt) in on_softirq_exit() argument
/linux-6.12.1/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_sriov.c189 init_p->cxts[0] = q->cxt; in bnx2x_vfop_qctor_prep()
307 if (bnx2x_vfq(vf, qid, cxt)) { in bnx2x_vf_queue_destroy()
308 bnx2x_vfq(vf, qid, cxt)->ustorm_ag_context.cdu_usage = 0; in bnx2x_vf_queue_destroy()
309 bnx2x_vfq(vf, qid, cxt)->xstorm_ag_context.cdu_reserved = 0; in bnx2x_vf_queue_destroy()
1338 struct hw_dma *cxt = &bp->vfdb->context[i]; in bnx2x_iov_free_mem() local
1339 BNX2X_PCI_FREE(cxt->addr, cxt->mapping, cxt->size); in bnx2x_iov_free_mem()
1368 struct hw_dma *cxt = BP_VF_CXT_PAGE(bp, i); in bnx2x_iov_alloc_mem() local
1369 cxt->size = min_t(size_t, tot_size, CDU_ILT_PAGE_SZ); in bnx2x_iov_alloc_mem()
1371 if (cxt->size) { in bnx2x_iov_alloc_mem()
1372 cxt->addr = BNX2X_PCI_ALLOC(&cxt->mapping, cxt->size); in bnx2x_iov_alloc_mem()
[all …]
Dbnx2x_sriov.h73 struct eth_context *cxt; member
Dbnx2x_cmn.c5072 void bnx2x_set_ctx_validation(struct bnx2x *bp, struct eth_context *cxt, in bnx2x_set_ctx_validation() argument
5075 if (!cxt) { in bnx2x_set_ctx_validation()
5076 BNX2X_ERR("bad context pointer %p\n", cxt); in bnx2x_set_ctx_validation()
5081 cxt->ustorm_ag_context.cdu_usage = in bnx2x_set_ctx_validation()
5085 cxt->xstorm_ag_context.cdu_reserved = in bnx2x_set_ctx_validation()
Dbnx2x_cmn.h1270 void bnx2x_set_ctx_validation(struct bnx2x *bp, struct eth_context *cxt,
/linux-6.12.1/drivers/infiniband/hw/qedr/
Dqedr_roce_cm.c67 static void qedr_ll2_complete_tx_packet(void *cxt, u8 connection_handle, in qedr_ll2_complete_tx_packet() argument
73 struct qedr_dev *dev = (struct qedr_dev *)cxt; in qedr_ll2_complete_tx_packet()
96 static void qedr_ll2_complete_rx_packet(void *cxt, in qedr_ll2_complete_rx_packet() argument
99 struct qedr_dev *dev = (struct qedr_dev *)cxt; in qedr_ll2_complete_rx_packet()
125 static void qedr_ll2_release_rx_packet(void *cxt, u8 connection_handle, in qedr_ll2_release_rx_packet() argument
/linux-6.12.1/drivers/staging/media/meson/vdec/
Dcodec_vp9.c1492 int cxt; in adapt_coef_probs_cxt() local
1494 for (cxt = 0; cxt < cxt_num; cxt++) { in adapt_coef_probs_cxt()