/linux-6.12.1/fs/xfs/scrub/ |
D | quotacheck.c | 43 xfs_dqtype_t q_type; member 187 xfs_dqtype_t q_type, in xqcheck_get_dqtrx() argument 193 if (dqa->dqtrx[i].q_type == 0 || in xqcheck_get_dqtrx() 194 (dqa->dqtrx[i].q_type == q_type && in xqcheck_get_dqtrx() 233 switch (p->q_type) { in xqcheck_mod_live_ino_dqtrx() 271 dqtrx = xqcheck_get_dqtrx(dqa, p->q_type, p->q_id); in xqcheck_mod_live_ino_dqtrx() 274 if (dqtrx->q_type == 0) { in xqcheck_mod_live_ino_dqtrx() 275 dqtrx->q_type = p->q_type; in xqcheck_mod_live_ino_dqtrx() 328 switch (p->q_type) { in xqcheck_apply_live_dqtrx() 354 dqtrx = xqcheck_get_dqtrx(dqa, p->q_type, p->q_id); in xqcheck_apply_live_dqtrx() [all …]
|
D | quota_repair.c | 82 trace_xrep_dquot_item_fill_bmap_hole(sc->mp, dq->q_type, dq->q_id); in xrep_quota_item_fill_bmap_hole() 91 xfs_qm_init_dquot_blk(sc->tp, dq->q_id, dq->q_type, bp); in xrep_quota_item_fill_bmap_hole() 251 trace_xrep_dquot_item(sc->mp, dq->q_type, dq->q_id); in xrep_quota_item()
|
D | quotacheck_repair.c | 111 trace_xrep_quotacheck_dquot(xqc->sc->mp, dq->q_type, dq->q_id); in xqcheck_commit_dquot()
|
/linux-6.12.1/net/core/ |
D | netdev-genl.c | 296 u32 q_idx, u32 q_type, const struct genl_info *info) in netdev_nl_queue_fill_one() argument 308 nla_put_u32(rsp, NETDEV_A_QUEUE_TYPE, q_type) || in netdev_nl_queue_fill_one() 312 switch (q_type) { in netdev_nl_queue_fill_one() 342 u32 q_type) in netdev_nl_queue_validate() argument 344 switch (q_type) { in netdev_nl_queue_validate() 358 u32 q_type, const struct genl_info *info) in netdev_nl_queue_fill() argument 365 err = netdev_nl_queue_validate(netdev, q_idx, q_type); in netdev_nl_queue_fill() 369 return netdev_nl_queue_fill_one(rsp, netdev, q_idx, q_type, info); in netdev_nl_queue_fill() 374 u32 q_id, q_type, ifindex; in netdev_nl_queue_get_doit() local 385 q_type = nla_get_u32(info->attrs[NETDEV_A_QUEUE_TYPE]); in netdev_nl_queue_get_doit() [all …]
|
/linux-6.12.1/drivers/crypto/ |
D | n2_core.c | 58 u8 q_type; member 82 if (q->q_type == HV_NCS_QTYPE_MAU) { in spu_next_offset() 1536 static void *new_queue(unsigned long q_type) in new_queue() argument 1538 return kmem_cache_zalloc(queue_cache[q_type - 1], GFP_KERNEL); in new_queue() 1541 static void free_queue(void *p, unsigned long q_type) in free_queue() argument 1543 kmem_cache_free(queue_cache[q_type - 1], p); in free_queue() 1583 unsigned long q_type = qr->type; in spu_queue_register_workfn() local 1586 hv_ret = sun4v_ncs_qconf(q_type, __pa(p->q), in spu_queue_register_workfn() 1594 static int spu_queue_register(struct spu_queue *p, unsigned long q_type) in spu_queue_register() argument 1597 struct spu_qreg qr = { .queue = p, .type = q_type }; in spu_queue_register() [all …]
|
/linux-6.12.1/fs/xfs/ |
D | xfs_dquot.h | 65 xfs_dqtype_t q_type; member 136 return dqp->q_type & XFS_DQTYPE_REC_MASK; in xfs_dquot_type()
|
D | xfs_trans_dquot.c | 63 !(dqp->q_type & XFS_DQTYPE_BIGTIME)) in xfs_trans_log_dquot() 64 dqp->q_type |= XFS_DQTYPE_BIGTIME; in xfs_trans_log_dquot() 165 .q_type = xfs_dquot_type(dqp), in xfs_trans_mod_ino_dquot() 450 .q_type = xfs_dquot_type(dqp), in xfs_trans_apply_dquot_deltas_hook() 619 .q_type = xfs_dquot_type(dqp), in xfs_trans_unreserve_and_mod_dquots_hook()
|
D | xfs_dquot.c | 55 switch (dqp->q_type) { in xfs_dquot_mark_sick() 499 dqp->q_type = type; in xfs_dquot_alloc() 604 dqp->q_type = ddqp->d_type; in xfs_dquot_from_disk() 641 ddqp->d_type = dqp->q_type; in xfs_dquot_to_disk() 1225 if (dqp->q_type & XFS_DQTYPE_BIGTIME) { in xfs_qm_dqflush_check()
|
D | xfs_qm.h | 118 xfs_dqtype_t q_type; member
|
D | xfs_quota.h | 89 xfs_dqtype_t q_type; member
|
/linux-6.12.1/drivers/net/ethernet/huawei/hinic/ |
D | hinic_hw_cmdq.c | 508 enum hinic_set_arm_qtype q_type, u32 q_id) in hinic_set_arm_bit() argument 516 arm_bit.q_type = q_type; in hinic_set_arm_bit() 739 enum hinic_cmdq_type q_type, void __iomem *db_area) in init_cmdq() argument 744 cmdq->cmdq_type = q_type; in init_cmdq()
|
D | hinic_hw_cmdq.h | 117 u32 q_type; member
|
/linux-6.12.1/drivers/net/ethernet/microsoft/mana/ |
D | hw_channel.c | 490 enum gdma_queue_type q_type, u16 q_depth, in mana_hwc_create_wq() argument 499 WARN_ON(q_type != GDMA_SQ && q_type != GDMA_RQ); in mana_hwc_create_wq() 501 if (q_type == GDMA_RQ) in mana_hwc_create_wq() 513 err = mana_hwc_create_gdma_wq(hwc, q_type, queue_size, &queue); in mana_hwc_create_wq()
|
D | gdma_main.c | 274 enum gdma_queue_type q_type, u32 qid, in mana_gd_ring_doorbell() argument 280 switch (q_type) { in mana_gd_ring_doorbell() 1027 enum gdma_queue_type q_type, in mana_gd_write_client_oob() argument 1049 if (q_type == GDMA_SQ) in mana_gd_write_client_oob()
|
/linux-6.12.1/drivers/net/ethernet/intel/idpf/ |
D | idpf_virtchnl.c | 1058 static int idpf_vport_get_q_reg(u32 *reg_vals, int num_regs, u32 q_type, in idpf_vport_get_q_reg() argument 1070 if (le32_to_cpu(chunk->type) != q_type) in idpf_vport_get_q_reg() 1094 int num_regs, u32 q_type) in __idpf_queue_reg_init() argument 1099 switch (q_type) { in __idpf_queue_reg_init() 3239 static int idpf_vport_get_queue_ids(u32 *qids, int num_qids, u16 q_type, in idpf_vport_get_queue_ids() argument 3250 if (le32_to_cpu(chunk->type) != q_type) in idpf_vport_get_queue_ids() 3283 u32 q_type) in __idpf_vport_queue_ids_init() argument 3287 switch (q_type) { in __idpf_vport_queue_ids_init() 3358 u16 q_type; in idpf_vport_queue_ids_init() local 3406 q_type = VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION; in idpf_vport_queue_ids_init() [all …]
|
/linux-6.12.1/drivers/net/ethernet/intel/ice/ |
D | ice_controlq.c | 606 static int ice_init_ctrlq(struct ice_hw *hw, enum ice_ctl_q q_type) in ice_init_ctrlq() argument 611 switch (q_type) { in ice_init_ctrlq() 627 cq->qtype = q_type; in ice_init_ctrlq() 687 static void ice_shutdown_ctrlq(struct ice_hw *hw, enum ice_ctl_q q_type, in ice_shutdown_ctrlq() argument 692 switch (q_type) { in ice_shutdown_ctrlq()
|
/linux-6.12.1/drivers/net/ethernet/freescale/dpaa2/ |
D | dpni.h | 952 enum dpni_queue_type q_type, 961 enum dpni_queue_type q_type,
|
/linux-6.12.1/drivers/infiniband/hw/irdma/ |
D | puda.h | 31 u8 q_type; member
|
D | user.h | 248 u8 q_type; member
|
D | puda.c | 277 info->q_type = (u8)FIELD_GET(IRDMA_CQ_SQ, qword3); in irdma_puda_poll_info() 286 if (info->q_type == IRDMA_CQE_QTYPE_RQ) { in irdma_puda_poll_info() 361 if (info.q_type == IRDMA_CQE_QTYPE_RQ) { in irdma_puda_poll_cmpl()
|
/linux-6.12.1/fs/xfs/libxfs/ |
D | xfs_dquot_buf.c | 321 if (timer != 0 && (dqp->q_type & XFS_DQTYPE_BIGTIME)) in xfs_dquot_to_disk_ts()
|
/linux-6.12.1/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_sriov.c | 170 unsigned long q_type) in bnx2x_vfop_qctor_prep() argument 216 if (test_bit(BNX2X_Q_TYPE_HAS_RX, &q_type)) { in bnx2x_vfop_qctor_prep() 228 if (test_bit(BNX2X_Q_TYPE_HAS_TX, &q_type)) { in bnx2x_vfop_qctor_prep() 1419 unsigned long q_type = 0; in bnx2x_vfq_init() local 1421 set_bit(BNX2X_Q_TYPE_HAS_TX, &q_type); in bnx2x_vfq_init() 1422 set_bit(BNX2X_Q_TYPE_HAS_RX, &q_type); in bnx2x_vfq_init() 1429 q_type); in bnx2x_vfq_init()
|
D | bnx2x_vfpf.c | 1540 unsigned long q_type = 0; in bnx2x_vf_mbx_setup_q() local 1561 __set_bit(BNX2X_Q_TYPE_HAS_TX, &q_type); in bnx2x_vf_mbx_setup_q() 1592 __set_bit(BNX2X_Q_TYPE_HAS_RX, &q_type); in bnx2x_vf_mbx_setup_q() 1638 bnx2x_vfop_qctor_prep(bp, vf, q, &qctor, q_type); in bnx2x_vf_mbx_setup_q()
|
/linux-6.12.1/drivers/accel/habanalabs/common/ |
D | command_submission.c | 1940 enum hl_queue_type q_type, u32 q_idx, u32 encaps_signal_offset) in cs_ioctl_signal_wait_create_jobs() argument 1949 job = hl_cs_allocate_job(hdev, q_type, true); in cs_ioctl_signal_wait_create_jobs() 1962 cb = hl_cb_kernel_create(hdev, cb_size, q_type == QUEUE_TYPE_HW); in cs_ioctl_signal_wait_create_jobs() 2216 enum hl_queue_type q_type; in cs_ioctl_signal_wait() local 2243 q_type = hw_queue_prop->type; in cs_ioctl_signal_wait() 2406 rc = cs_ioctl_signal_wait_create_jobs(hdev, ctx, cs, q_type, in cs_ioctl_signal_wait() 2421 if (q_type == QUEUE_TYPE_HW) in cs_ioctl_signal_wait()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlxsw/ |
D | pci.c | 215 enum mlxsw_pci_queue_type q_type) in mlxsw_pci_queue_type_group_get() argument 217 return &mlxsw_pci->queues[q_type]; in mlxsw_pci_queue_type_group_get() 222 enum mlxsw_pci_queue_type q_type, u8 q_num) in __mlxsw_pci_queue_get() argument 224 return &mlxsw_pci->queues[q_type].q[q_num]; in __mlxsw_pci_queue_get()
|