/linux-6.12.1/block/ |
D | blk-mq-sched.c | 22 void blk_mq_sched_mark_restart_hctx(struct blk_mq_hw_ctx *hctx) in blk_mq_sched_mark_restart_hctx() argument 24 if (test_bit(BLK_MQ_S_SCHED_RESTART, &hctx->state)) in blk_mq_sched_mark_restart_hctx() 27 set_bit(BLK_MQ_S_SCHED_RESTART, &hctx->state); in blk_mq_sched_mark_restart_hctx() 31 void __blk_mq_sched_restart(struct blk_mq_hw_ctx *hctx) in __blk_mq_sched_restart() argument 33 clear_bit(BLK_MQ_S_SCHED_RESTART, &hctx->state); in __blk_mq_sched_restart() 44 blk_mq_run_hw_queue(hctx, true); in __blk_mq_sched_restart() 58 struct blk_mq_hw_ctx *hctx = in blk_mq_dispatch_hctx_list() local 65 if (rq->mq_hctx != hctx) { in blk_mq_dispatch_hctx_list() 74 return blk_mq_dispatch_rq_list(hctx, &hctx_list, count); in blk_mq_dispatch_hctx_list() 87 static int __blk_mq_do_dispatch_sched(struct blk_mq_hw_ctx *hctx) in __blk_mq_do_dispatch_sched() argument [all …]
|
D | blk-mq.h | 50 bool blk_mq_dispatch_rq_list(struct blk_mq_hw_ctx *hctx, struct list_head *, 52 void blk_mq_flush_busy_ctxs(struct blk_mq_hw_ctx *hctx, struct list_head *list); 53 struct request *blk_mq_dequeue_from_ctx(struct blk_mq_hw_ctx *hctx, 123 extern void blk_mq_hctx_kobj_init(struct blk_mq_hw_ctx *hctx); 162 struct blk_mq_hw_ctx *hctx; member 178 int blk_mq_tag_update_depth(struct blk_mq_hw_ctx *hctx, 191 struct blk_mq_hw_ctx *hctx) in bt_wait_ptr() argument 193 if (!hctx) in bt_wait_ptr() 195 return sbq_wait_ptr(bt, &hctx->wait_index); in bt_wait_ptr() 201 static inline void blk_mq_tag_busy(struct blk_mq_hw_ctx *hctx) in blk_mq_tag_busy() argument [all …]
|
D | blk-mq-sysfs.c | 34 struct blk_mq_hw_ctx *hctx = container_of(kobj, struct blk_mq_hw_ctx, in blk_mq_hw_sysfs_release() local 37 blk_free_flush_queue(hctx->fq); in blk_mq_hw_sysfs_release() 38 sbitmap_free(&hctx->ctx_map); in blk_mq_hw_sysfs_release() 39 free_cpumask_var(hctx->cpumask); in blk_mq_hw_sysfs_release() 40 kfree(hctx->ctxs); in blk_mq_hw_sysfs_release() 41 kfree(hctx); in blk_mq_hw_sysfs_release() 53 struct blk_mq_hw_ctx *hctx; in blk_mq_hw_sysfs_show() local 58 hctx = container_of(kobj, struct blk_mq_hw_ctx, kobj); in blk_mq_hw_sysfs_show() 59 q = hctx->queue; in blk_mq_hw_sysfs_show() 65 res = entry->show(hctx, page); in blk_mq_hw_sysfs_show() [all …]
|
D | blk-mq.c | 50 static void blk_mq_try_issue_list_directly(struct blk_mq_hw_ctx *hctx, 52 static int blk_hctx_poll(struct request_queue *q, struct blk_mq_hw_ctx *hctx, 59 static bool blk_mq_hctx_has_pending(struct blk_mq_hw_ctx *hctx) in blk_mq_hctx_has_pending() argument 61 return !list_empty_careful(&hctx->dispatch) || in blk_mq_hctx_has_pending() 62 sbitmap_any_bit_set(&hctx->ctx_map) || in blk_mq_hctx_has_pending() 63 blk_mq_sched_has_work(hctx); in blk_mq_hctx_has_pending() 69 static void blk_mq_hctx_mark_pending(struct blk_mq_hw_ctx *hctx, in blk_mq_hctx_mark_pending() argument 72 const int bit = ctx->index_hw[hctx->type]; in blk_mq_hctx_mark_pending() 74 if (!sbitmap_test_bit(&hctx->ctx_map, bit)) in blk_mq_hctx_mark_pending() 75 sbitmap_set_bit(&hctx->ctx_map, bit); in blk_mq_hctx_mark_pending() [all …]
|
D | blk-mq-debugfs.c | 166 struct blk_mq_hw_ctx *hctx = data; in hctx_state_show() local 169 blk_flags_show(m, hctx->state, hctx_state_name, in hctx_state_show() 196 struct blk_mq_hw_ctx *hctx = data; in hctx_flags_show() local 197 const int alloc_policy = BLK_MQ_FLAG_TO_ALLOC_POLICY(hctx->flags); in hctx_flags_show() 211 hctx->flags ^ BLK_ALLOC_POLICY_TO_MQ_FLAG(alloc_policy), in hctx_flags_show() 315 __acquires(&hctx->lock) in hctx_dispatch_start() 317 struct blk_mq_hw_ctx *hctx = m->private; in hctx_dispatch_start() local 319 spin_lock(&hctx->lock); in hctx_dispatch_start() 320 return seq_list_start(&hctx->dispatch, *pos); in hctx_dispatch_start() 325 struct blk_mq_hw_ctx *hctx = m->private; in hctx_dispatch_next() local [all …]
|
D | blk-mq-tag.c | 38 void __blk_mq_tag_busy(struct blk_mq_hw_ctx *hctx) in __blk_mq_tag_busy() argument 42 struct blk_mq_tags *tags = hctx->tags; in __blk_mq_tag_busy() 48 if (blk_mq_is_shared_tags(hctx->flags)) { in __blk_mq_tag_busy() 49 struct request_queue *q = hctx->queue; in __blk_mq_tag_busy() 55 if (test_bit(BLK_MQ_S_TAG_ACTIVE, &hctx->state) || in __blk_mq_tag_busy() 56 test_and_set_bit(BLK_MQ_S_TAG_ACTIVE, &hctx->state)) in __blk_mq_tag_busy() 81 void __blk_mq_tag_idle(struct blk_mq_hw_ctx *hctx) in __blk_mq_tag_idle() argument 83 struct blk_mq_tags *tags = hctx->tags; in __blk_mq_tag_idle() 86 if (blk_mq_is_shared_tags(hctx->flags)) { in __blk_mq_tag_idle() 87 struct request_queue *q = hctx->queue; in __blk_mq_tag_idle() [all …]
|
D | blk-mq-sched.h | 16 void blk_mq_sched_mark_restart_hctx(struct blk_mq_hw_ctx *hctx); 17 void __blk_mq_sched_restart(struct blk_mq_hw_ctx *hctx); 19 void blk_mq_sched_dispatch_requests(struct blk_mq_hw_ctx *hctx); 25 static inline void blk_mq_sched_restart(struct blk_mq_hw_ctx *hctx) in blk_mq_sched_restart() argument 27 if (test_bit(BLK_MQ_S_SCHED_RESTART, &hctx->state)) in blk_mq_sched_restart() 28 __blk_mq_sched_restart(hctx); in blk_mq_sched_restart() 70 static inline bool blk_mq_sched_has_work(struct blk_mq_hw_ctx *hctx) in blk_mq_sched_has_work() argument 72 struct elevator_queue *e = hctx->queue->elevator; in blk_mq_sched_has_work() 75 return e->type->ops.has_work(hctx); in blk_mq_sched_has_work() 80 static inline bool blk_mq_sched_needs_restart(struct blk_mq_hw_ctx *hctx) in blk_mq_sched_needs_restart() argument [all …]
|
D | kyber-iosched.c | 453 static void kyber_depth_updated(struct blk_mq_hw_ctx *hctx) in kyber_depth_updated() argument 455 struct kyber_queue_data *kqd = hctx->queue->elevator->elevator_data; in kyber_depth_updated() 456 struct blk_mq_tags *tags = hctx->sched_tags; in kyber_depth_updated() 464 static int kyber_init_hctx(struct blk_mq_hw_ctx *hctx, unsigned int hctx_idx) in kyber_init_hctx() argument 469 khd = kmalloc_node(sizeof(*khd), GFP_KERNEL, hctx->numa_node); in kyber_init_hctx() 473 khd->kcqs = kmalloc_array_node(hctx->nr_ctx, in kyber_init_hctx() 475 GFP_KERNEL, hctx->numa_node); in kyber_init_hctx() 479 for (i = 0; i < hctx->nr_ctx; i++) in kyber_init_hctx() 483 if (sbitmap_init_node(&khd->kcq_map[i], hctx->nr_ctx, in kyber_init_hctx() 484 ilog2(8), GFP_KERNEL, hctx->numa_node, in kyber_init_hctx() [all …]
|
D | blk-mq-debugfs.h | 25 struct blk_mq_hw_ctx *hctx); 26 void blk_mq_debugfs_unregister_hctx(struct blk_mq_hw_ctx *hctx); 33 struct blk_mq_hw_ctx *hctx); 34 void blk_mq_debugfs_unregister_sched_hctx(struct blk_mq_hw_ctx *hctx); 44 struct blk_mq_hw_ctx *hctx) in blk_mq_debugfs_register_hctx() argument 48 static inline void blk_mq_debugfs_unregister_hctx(struct blk_mq_hw_ctx *hctx) in blk_mq_debugfs_unregister_hctx() argument 69 struct blk_mq_hw_ctx *hctx) in blk_mq_debugfs_register_sched_hctx() argument 73 static inline void blk_mq_debugfs_unregister_sched_hctx(struct blk_mq_hw_ctx *hctx) in blk_mq_debugfs_unregister_sched_hctx() argument
|
D | mq-deadline.c | 462 static struct request *dd_dispatch_request(struct blk_mq_hw_ctx *hctx) in dd_dispatch_request() argument 464 struct deadline_data *dd = hctx->queue->elevator->elevator_data; in dd_dispatch_request() 496 static int dd_to_word_depth(struct blk_mq_hw_ctx *hctx, unsigned int qdepth) in dd_to_word_depth() argument 498 struct sbitmap_queue *bt = &hctx->sched_tags->bitmap_tags; in dd_to_word_depth() 499 const unsigned int nrr = hctx->queue->nr_requests; in dd_to_word_depth() 520 data->shallow_depth = dd_to_word_depth(data->hctx, dd->async_depth); in dd_limit_depth() 524 static void dd_depth_updated(struct blk_mq_hw_ctx *hctx) in dd_depth_updated() argument 526 struct request_queue *q = hctx->queue; in dd_depth_updated() 528 struct blk_mq_tags *tags = hctx->sched_tags; in dd_depth_updated() 536 static int dd_init_hctx(struct blk_mq_hw_ctx *hctx, unsigned int hctx_idx) in dd_init_hctx() argument [all …]
|
/linux-6.12.1/samples/hid/ |
D | hid_mouse.bpf.c | 8 static int hid_y_event(struct hid_bpf_ctx *hctx) in hid_y_event() argument 11 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 9 /* size */); in hid_y_event() 16 bpf_printk("event: size: %d", hctx->size); in hid_y_event() 53 static int hid_x_event(struct hid_bpf_ctx *hctx) in hid_x_event() argument 56 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 9 /* size */); in hid_x_event() 71 int BPF_PROG(hid_event, struct hid_bpf_ctx *hctx, enum hid_report_type type) in BPF_PROG() argument 73 int ret = hid_y_event(hctx); in BPF_PROG() 78 return hid_x_event(hctx); in BPF_PROG() 83 int BPF_PROG(hid_rdesc_fixup, struct hid_bpf_ctx *hctx) in BPF_PROG() argument 85 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 4096 /* size */); in BPF_PROG()
|
D | hid_surface_dial.bpf.c | 14 int BPF_PROG(hid_event, struct hid_bpf_ctx *hctx) in BPF_PROG() argument 16 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 9 /* size */); in BPF_PROG() 105 int BPF_PROG(hid_rdesc_fixup, struct hid_bpf_ctx *hctx) in BPF_PROG() argument 107 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 4096 /* size */); in BPF_PROG()
|
/linux-6.12.1/drivers/hid/bpf/progs/ |
D | XPPen__ArtistPro16Gen2.bpf.c | 86 int BPF_PROG(hid_fix_rdesc_xppen_artistpro16gen2, struct hid_bpf_ctx *hctx) in SEC() 88 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 4096 /* size */); in SEC() 98 if (hctx->hid->product == PID_ARTIST_PRO14_GEN2) { in SEC() 108 static int xppen_16_fix_eraser(struct hid_bpf_ctx *hctx) in xppen_16_fix_eraser() argument 110 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 10 /* size */); in xppen_16_fix_eraser() 209 static int xppen_16_fix_angle_offset(struct hid_bpf_ctx *hctx) in xppen_16_fix_angle_offset() argument 211 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 10 /* size */); in xppen_16_fix_angle_offset() 244 if (hctx->hid->product == PID_ARTIST_PRO14_GEN2) { in xppen_16_fix_angle_offset() 247 } else if (hctx->hid->product == PID_ARTIST_PRO16_GEN2) { in xppen_16_fix_angle_offset() 256 int BPF_PROG(xppen_artist_pro_16_device_event, struct hid_bpf_ctx *hctx) in SEC() [all …]
|
D | Huion__Kamvas-Pro-19.bpf.c | 195 int BPF_PROG(hid_fix_rdesc_huion_kamvas_pro_19, struct hid_bpf_ctx *hctx) in SEC() 197 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, HID_MAX_DESCRIPTOR_SIZE /* size */); in SEC() 219 int BPF_PROG(kamvas_pro_19_fix_3rd_button, struct hid_bpf_ctx *hctx) in SEC() 221 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 10 /* size */); in SEC() 274 struct hid_bpf_ctx *hctx = hid_bpf_allocate_context(ctx->hid); in probe() local 276 if (!hctx) { in probe() 281 const char *name = hctx->hid->name; in probe() 290 hid_bpf_release_context(hctx); in probe()
|
D | XPPen__DecoMini4.bpf.c | 153 int BPF_PROG(hid_rdesc_fixup_xppen_deco_mini_4, struct hid_bpf_ctx *hctx) in SEC() 155 __u8 *data = hid_bpf_get_data(hctx, 0, HID_MAX_DESCRIPTOR_SIZE); in SEC() 160 if (hctx->size == RDESC_SIZE_PAD) { in SEC() 163 } else if (hctx->size == RDESC_SIZE_PEN) { in SEC() 172 int BPF_PROG(hid_device_event_xppen_deco_mini_4, struct hid_bpf_ctx *hctx) in SEC() 174 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 8 /* size */); in SEC()
|
D | FR-TEC__Raptor-Mach-2.bpf.c | 137 int BPF_PROG(hid_fix_rdesc_raptor_mach_2, struct hid_bpf_ctx *hctx) in SEC() 139 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, HID_MAX_DESCRIPTOR_SIZE /* size */); in SEC() 156 int BPF_PROG(raptor_mach_2_fix_hat_switch, struct hid_bpf_ctx *hctx) in SEC() 158 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 64 /* size */); in SEC()
|
D | XPPen__Artist24.bpf.c | 93 int BPF_PROG(hid_fix_rdesc_xppen_artist24, struct hid_bpf_ctx *hctx) in SEC() 95 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 4096 /* size */); in SEC() 154 int BPF_PROG(xppen_24_fix_eraser, struct hid_bpf_ctx *hctx) in SEC() 156 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 10 /* size */); in SEC()
|
D | Thrustmaster__TCA-Yoke-Boeing.bpf.c | 100 int BPF_PROG(hid_fix_rdesc_tca_yoke, struct hid_bpf_ctx *hctx) in SEC() 104 if (hctx->size != expected_length) in SEC() 107 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, HID_MAX_DESCRIPTOR_SIZE /* size */); in SEC()
|
D | Huion__Inspiroy-2-S.bpf.c | 359 int BPF_PROG(hid_fix_rdesc, struct hid_bpf_ctx *hctx) in SEC() 361 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, HID_MAX_DESCRIPTOR_SIZE /* size */); in SEC() 362 __s32 rdesc_size = hctx->size; in SEC() 405 int BPF_PROG(inspiroy_2_fix_events, struct hid_bpf_ctx *hctx) in SEC() 407 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 10 /* size */); in SEC()
|
D | Huion__Dial-2.bpf.c | 420 int BPF_PROG(dial_2_fix_rdesc, struct hid_bpf_ctx *hctx) in SEC() 422 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, HID_MAX_DESCRIPTOR_SIZE /* size */); in SEC() 423 __s32 rdesc_size = hctx->size; in SEC() 466 int BPF_PROG(dial_2_fix_events, struct hid_bpf_ctx *hctx) in SEC() 468 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 16 /* size */); in SEC()
|
/linux-6.12.1/drivers/crypto/marvell/octeontx2/ |
D | cn10k_cpt.c | 115 void cn10k_cpt_hw_ctx_set(union cn10k_cpt_hw_ctx *hctx, u16 ctx_sz) in cn10k_cpt_hw_ctx_set() argument 117 hctx->w0.aop_valid = 1; in cn10k_cpt_hw_ctx_set() 118 hctx->w0.ctx_hdr_sz = 0; in cn10k_cpt_hw_ctx_set() 119 hctx->w0.ctx_sz = ctx_sz; in cn10k_cpt_hw_ctx_set() 120 hctx->w0.ctx_push_sz = 1; in cn10k_cpt_hw_ctx_set() 127 union cn10k_cpt_hw_ctx *hctx; in cn10k_cpt_hw_ctx_init() local 136 hctx = kmalloc(CN10K_CPT_HW_CTX_SIZE, GFP_KERNEL); in cn10k_cpt_hw_ctx_init() 137 if (unlikely(!hctx)) in cn10k_cpt_hw_ctx_init() 139 cptr_dma = dma_map_single(&pdev->dev, hctx, CN10K_CPT_HW_CTX_SIZE, in cn10k_cpt_hw_ctx_init() 142 kfree(hctx); in cn10k_cpt_hw_ctx_init() [all …]
|
/linux-6.12.1/tools/testing/selftests/hid/progs/ |
D | hid.c | 311 int BPF_PROG(hid_test_filter_raw_request, struct hid_bpf_ctx *hctx, unsigned char reportnum, in BPF_PROG() argument 332 int BPF_PROG(hid_test_hidraw_raw_request, struct hid_bpf_ctx *hctx, unsigned char reportnum, in BPF_PROG() argument 335 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 3 /* size */); in BPF_PROG() 345 ret = hid_bpf_hw_request(hctx, data, 2, rtype, reqtype); in BPF_PROG() 363 int BPF_PROG(hid_test_infinite_loop_raw_request, struct hid_bpf_ctx *hctx, unsigned char reportnum, in BPF_PROG() argument 366 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 3 /* size */); in BPF_PROG() 377 ret = hid_bpf_hw_request(hctx, data, 2, rtype, reqtype); in BPF_PROG() 390 int BPF_PROG(hid_test_filter_output_report, struct hid_bpf_ctx *hctx, unsigned char reportnum, in BPF_PROG() argument 402 int BPF_PROG(hid_test_hidraw_output_report, struct hid_bpf_ctx *hctx, __u64 source) in BPF_PROG() argument 404 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 3 /* size */); in BPF_PROG() [all …]
|
/linux-6.12.1/net/dccp/ccids/ |
D | ccid3.h | 104 struct ccid3_hc_tx_sock *hctx = ccid_priv(dccp_sk(sk)->dccps_hc_tx_ccid); in ccid3_hc_tx_sk() local 105 BUG_ON(hctx == NULL); in ccid3_hc_tx_sk() 106 return hctx; in ccid3_hc_tx_sk()
|
/linux-6.12.1/include/linux/ |
D | blk-mq.h | 904 void blk_mq_stop_hw_queue(struct blk_mq_hw_ctx *hctx); 905 void blk_mq_start_hw_queue(struct blk_mq_hw_ctx *hctx); 908 void blk_mq_start_stopped_hw_queue(struct blk_mq_hw_ctx *hctx, bool async); 915 void blk_mq_delay_run_hw_queue(struct blk_mq_hw_ctx *hctx, unsigned long msecs); 916 void blk_mq_run_hw_queue(struct blk_mq_hw_ctx *hctx, bool async); 973 #define queue_for_each_hw_ctx(q, hctx, i) \ argument 974 xa_for_each(&(q)->hctx_table, (i), (hctx)) 976 #define hctx_for_each_ctx(hctx, ctx, i) \ argument 977 for ((i) = 0; (i) < (hctx)->nr_ctx && \ 978 ({ ctx = (hctx)->ctxs[(i)]; 1; }); (i)++) [all …]
|
/linux-6.12.1/drivers/s390/block/ |
D | scm_blk.c | 283 static blk_status_t scm_blk_request(struct blk_mq_hw_ctx *hctx, in scm_blk_request() argument 286 struct scm_device *scmdev = hctx->queue->queuedata; in scm_blk_request() 288 struct scm_queue *sq = hctx->driver_data; in scm_blk_request() 332 static int scm_blk_init_hctx(struct blk_mq_hw_ctx *hctx, void *data, in scm_blk_init_hctx() argument 341 hctx->driver_data = qd; in scm_blk_init_hctx() 346 static void scm_blk_exit_hctx(struct blk_mq_hw_ctx *hctx, unsigned int idx) in scm_blk_exit_hctx() argument 348 struct scm_queue *qd = hctx->driver_data; in scm_blk_exit_hctx() 351 kfree(hctx->driver_data); in scm_blk_exit_hctx() 352 hctx->driver_data = NULL; in scm_blk_exit_hctx()
|