/linux-6.12.1/drivers/gpu/drm/xe/ |
D | xe_gt_sriov_pf_control.c | 39 static int guc_action_vf_control_cmd(struct xe_guc *guc, u32 vfid, u32 cmd) in guc_action_vf_control_cmd() argument 45 FIELD_PREP(PF2GUC_VF_CONTROL_REQUEST_MSG_1_VFID, vfid), in guc_action_vf_control_cmd() 54 static int pf_send_vf_control_cmd(struct xe_gt *gt, unsigned int vfid, u32 cmd) in pf_send_vf_control_cmd() argument 58 xe_gt_assert(gt, vfid != PFID); in pf_send_vf_control_cmd() 60 vfid, control_cmd_to_string(cmd)); in pf_send_vf_control_cmd() 62 err = guc_action_vf_control_cmd(>->uc.guc, vfid, cmd); in pf_send_vf_control_cmd() 65 vfid, control_cmd_to_string(cmd), ERR_PTR(err)); in pf_send_vf_control_cmd() 69 static int pf_send_vf_pause(struct xe_gt *gt, unsigned int vfid) in pf_send_vf_pause() argument 71 return pf_send_vf_control_cmd(gt, vfid, GUC_PF_TRIGGER_VF_PAUSE); in pf_send_vf_pause() 74 static int pf_send_vf_resume(struct xe_gt *gt, unsigned int vfid) in pf_send_vf_resume() argument [all …]
|
D | xe_gt_sriov_pf_config.h | 15 u64 xe_gt_sriov_pf_config_get_ggtt(struct xe_gt *gt, unsigned int vfid); 16 int xe_gt_sriov_pf_config_set_ggtt(struct xe_gt *gt, unsigned int vfid, u64 size); 18 unsigned int vfid, unsigned int num_vfs); 20 unsigned int vfid, unsigned int num_vfs, u64 size); 22 u32 xe_gt_sriov_pf_config_get_ctxs(struct xe_gt *gt, unsigned int vfid); 23 int xe_gt_sriov_pf_config_set_ctxs(struct xe_gt *gt, unsigned int vfid, u32 num_ctxs); 24 int xe_gt_sriov_pf_config_set_fair_ctxs(struct xe_gt *gt, unsigned int vfid, unsigned int num_vfs); 25 int xe_gt_sriov_pf_config_bulk_set_ctxs(struct xe_gt *gt, unsigned int vfid, unsigned int num_vfs, 28 u32 xe_gt_sriov_pf_config_get_dbs(struct xe_gt *gt, unsigned int vfid); 29 int xe_gt_sriov_pf_config_set_dbs(struct xe_gt *gt, unsigned int vfid, u32 num_dbs); [all …]
|
D | xe_gt_sriov_pf_config.c | 41 static int guc_action_update_vf_cfg(struct xe_guc *guc, u32 vfid, in guc_action_update_vf_cfg() argument 46 vfid, in guc_action_update_vf_cfg() 58 static int pf_send_vf_cfg_reset(struct xe_gt *gt, u32 vfid) in pf_send_vf_cfg_reset() argument 63 ret = guc_action_update_vf_cfg(guc, vfid, 0, 0); in pf_send_vf_cfg_reset() 72 static int pf_send_vf_cfg_klvs(struct xe_gt *gt, u32 vfid, const u32 *klvs, u32 num_dwords) in pf_send_vf_cfg_klvs() argument 92 ret = guc_action_update_vf_cfg(guc, vfid, xe_bo_ggtt_addr(bo), num_dwords); in pf_send_vf_cfg_klvs() 103 static int pf_push_vf_cfg_klvs(struct xe_gt *gt, unsigned int vfid, u32 num_klvs, in pf_push_vf_cfg_klvs() argument 110 ret = pf_send_vf_cfg_klvs(gt, vfid, klvs, num_dwords); in pf_push_vf_cfg_klvs() 118 xe_sriov_function_name(vfid, name, sizeof(name)), in pf_push_vf_cfg_klvs() 133 static int pf_push_vf_cfg_u32(struct xe_gt *gt, unsigned int vfid, u16 key, u32 value) in pf_push_vf_cfg_u32() argument [all …]
|
D | xe_gt_sriov_pf_monitor.c | 24 void xe_gt_sriov_pf_monitor_flr(struct xe_gt *gt, u32 vfid) in xe_gt_sriov_pf_monitor_flr() argument 29 xe_gt_sriov_pf_assert_vfid(gt, vfid); in xe_gt_sriov_pf_monitor_flr() 32 gt->sriov.pf.vfs[vfid].monitor.guc.events[e] = 0; in xe_gt_sriov_pf_monitor_flr() 35 static void pf_update_event_counter(struct xe_gt *gt, u32 vfid, in pf_update_event_counter() argument 41 gt->sriov.pf.vfs[vfid].monitor.guc.events[e]++; in pf_update_event_counter() 44 static int pf_handle_vf_threshold_event(struct xe_gt *gt, u32 vfid, u32 threshold) in pf_handle_vf_threshold_event() argument 50 xe_sriov_function_name(vfid, origin, sizeof(origin)); in pf_handle_vf_threshold_event() 60 origin, xe_gt_sriov_pf_config_get_threshold(gt, vfid, e), in pf_handle_vf_threshold_event() 63 pf_update_event_counter(gt, vfid, e); in pf_handle_vf_threshold_event() 81 u32 vfid; in xe_gt_sriov_pf_monitor_process_guc2pf() local [all …]
|
D | xe_gt_sriov_pf_debugfs.c | 175 unsigned int vfid = extract_vfid(data); \ 183 err = xe_gt_sriov_pf_config_set_##CONFIG(gt, vfid, val); \ 192 unsigned int vfid = extract_vfid(data); \ 194 *val = xe_gt_sriov_pf_config_get_##CONFIG(gt, vfid); \ 229 unsigned int vfid = extract_vfid(data); in set_threshold() local 237 err = xe_gt_sriov_pf_config_set_threshold(gt, vfid, index, val); in set_threshold() 246 unsigned int vfid = extract_vfid(data); in get_threshold() local 248 *val = xe_gt_sriov_pf_config_get_threshold(gt, vfid, index); in get_threshold() 272 static void pf_add_config_attrs(struct xe_gt *gt, struct dentry *parent, unsigned int vfid) in MAKE_XE_GUC_KLV_THRESHOLDS_SET() 275 xe_gt_assert(gt, vfid == extract_vfid(parent)); in MAKE_XE_GUC_KLV_THRESHOLDS_SET() [all …]
|
D | xe_lmtt.c | 257 static void lmtt_drop_pages(struct xe_lmtt *lmtt, unsigned int vfid) in lmtt_drop_pages() argument 262 pt = pd->entries[vfid]; in lmtt_drop_pages() 263 pd->entries[vfid] = NULL; in lmtt_drop_pages() 267 lmtt_write_pte(lmtt, pd, LMTT_PTE_INVALID, vfid); in lmtt_drop_pages() 316 static int lmtt_alloc_range(struct xe_lmtt *lmtt, unsigned int vfid, u64 start, u64 end) in lmtt_alloc_range() argument 325 lmtt_assert(lmtt, vfid <= lmtt->ops->lmtt_pte_num(pd->level)); in lmtt_alloc_range() 329 if (pd->entries[vfid]) in lmtt_alloc_range() 340 lmtt_write_pte(lmtt, pd, pde, vfid); in lmtt_alloc_range() 342 pd->entries[vfid] = pt; in lmtt_alloc_range() 357 static struct xe_lmtt_pt *lmtt_leaf_pt(struct xe_lmtt *lmtt, unsigned int vfid, u64 addr) in lmtt_leaf_pt() argument [all …]
|
D | xe_gt_sriov_pf_service.c | 83 static void pf_connect(struct xe_gt *gt, u32 vfid, u32 major, u32 minor) in pf_connect() argument 85 xe_gt_sriov_pf_assert_vfid(gt, vfid); in pf_connect() 88 gt->sriov.pf.vfs[vfid].version.major = major; in pf_connect() 89 gt->sriov.pf.vfs[vfid].version.minor = minor; in pf_connect() 92 static void pf_disconnect(struct xe_gt *gt, u32 vfid) in pf_disconnect() argument 94 xe_gt_sriov_pf_assert_vfid(gt, vfid); in pf_disconnect() 96 gt->sriov.pf.vfs[vfid].version.major = 0; in pf_disconnect() 97 gt->sriov.pf.vfs[vfid].version.minor = 0; in pf_disconnect() 100 static bool pf_is_negotiated(struct xe_gt *gt, u32 vfid, u32 major, u32 minor) in pf_is_negotiated() argument 102 xe_gt_sriov_pf_assert_vfid(gt, vfid); in pf_is_negotiated() [all …]
|
D | xe_gt_sriov_pf_control.h | 17 int xe_gt_sriov_pf_control_pause_vf(struct xe_gt *gt, unsigned int vfid); 18 int xe_gt_sriov_pf_control_resume_vf(struct xe_gt *gt, unsigned int vfid); 19 int xe_gt_sriov_pf_control_stop_vf(struct xe_gt *gt, unsigned int vfid); 20 int xe_gt_sriov_pf_control_trigger_flr(struct xe_gt *gt, unsigned int vfid);
|
D | xe_lmtt.h | 18 int xe_lmtt_prepare_pages(struct xe_lmtt *lmtt, unsigned int vfid, u64 range); 19 int xe_lmtt_populate_pages(struct xe_lmtt *lmtt, unsigned int vfid, struct xe_bo *bo, u64 offset); 20 void xe_lmtt_drop_pages(struct xe_lmtt *lmtt, unsigned int vfid);
|
D | xe_gt_sriov_pf_helpers.h | 23 #define xe_gt_sriov_pf_assert_vfid(gt, vfid) xe_sriov_pf_assert_vfid(gt_to_xe(gt), (vfid)) argument
|
D | xe_sriov_pf_helpers.h | 25 #define xe_sriov_pf_assert_vfid(xe, vfid) \ argument 26 xe_assert((xe), (vfid) <= xe_sriov_pf_get_totalvfs(xe))
|
D | xe_gt_pagefault.c | 29 u8 vfid; member 57 u8 vfid; member 269 pf->asid, pf->vfid, pf->pdata, upper_32_bits(pf->page_addr), in print_pagefault() 295 pf->vfid = FIELD_GET(PFD_VFID, desc->dw2); in get_pagefault() 384 reply.dw1 = FIELD_PREP(PFR_VFID, pf.vfid) | in pf_queue_work_func() 532 acc->asid, acc->vfid, acc->engine_class, acc->engine_instance, in print_acc() 615 acc->vfid = FIELD_GET(ACC_VFID, desc->dw2); in get_acc()
|
D | xe_ggtt.c | 736 static u64 xe_encode_vfid_pte(u16 vfid) in xe_encode_vfid_pte() argument 738 return FIELD_PREP(GGTT_PTE_VFID, vfid) | XE_PAGE_PRESENT; in xe_encode_vfid_pte() 741 static void xe_ggtt_assign_locked(struct xe_ggtt *ggtt, const struct drm_mm_node *node, u16 vfid) in xe_ggtt_assign_locked() argument 746 u64 pte = xe_encode_vfid_pte(vfid); in xe_ggtt_assign_locked() 770 void xe_ggtt_assign(const struct xe_ggtt_node *node, u16 vfid) in xe_ggtt_assign() argument 773 xe_ggtt_assign_locked(node->ggtt, &node->base, vfid); in xe_ggtt_assign()
|
/linux-6.12.1/drivers/net/ethernet/marvell/octeon_ep/ |
D | octep_ctrl_net.h | 262 int octep_ctrl_net_get_link_status(struct octep_device *oct, int vfid); 274 int octep_ctrl_net_set_link_status(struct octep_device *oct, int vfid, bool up, 287 int octep_ctrl_net_set_rx_state(struct octep_device *oct, int vfid, bool up, 299 int octep_ctrl_net_get_mac_addr(struct octep_device *oct, int vfid, u8 *addr); 311 int octep_ctrl_net_set_mac_addr(struct octep_device *oct, int vfid, u8 *addr, 322 int octep_ctrl_net_get_mtu(struct octep_device *oct, int vfid); 334 int octep_ctrl_net_set_mtu(struct octep_device *oct, int vfid, int mtu, 347 int octep_ctrl_net_get_if_stats(struct octep_device *oct, int vfid, 360 int octep_ctrl_net_get_link_info(struct octep_device *oct, int vfid, 374 int vfid, [all …]
|
D | octep_ctrl_net.c | 44 u16 sz, int vfid) in init_send_req() argument 53 if (vfid != OCTEP_CTRL_NET_INVALID_VFID) { in init_send_req() 55 msg->hdr.s.vf_idx = vfid; in init_send_req() 127 int octep_ctrl_net_get_link_status(struct octep_device *oct, int vfid) in octep_ctrl_net_get_link_status() argument 133 init_send_req(&d.msg, (void *)req, state_sz, vfid); in octep_ctrl_net_get_link_status() 143 int octep_ctrl_net_set_link_status(struct octep_device *oct, int vfid, bool up, in octep_ctrl_net_set_link_status() argument 149 init_send_req(&d.msg, req, state_sz, vfid); in octep_ctrl_net_set_link_status() 158 int octep_ctrl_net_set_rx_state(struct octep_device *oct, int vfid, bool up, in octep_ctrl_net_set_rx_state() argument 164 init_send_req(&d.msg, req, state_sz, vfid); in octep_ctrl_net_set_rx_state() 173 int octep_ctrl_net_get_mac_addr(struct octep_device *oct, int vfid, u8 *addr) in octep_ctrl_net_get_mac_addr() argument [all …]
|
/linux-6.12.1/include/linux/qed/ |
D | qed_iov_if.h | 16 int (*set_mac) (struct qed_dev *cdev, u8 *mac, int vfid); 18 int (*set_vlan) (struct qed_dev *cdev, u16 vid, int vfid); 26 int (*set_spoof) (struct qed_dev *cdev, int vfid, bool val); 28 int (*set_rate) (struct qed_dev *cdev, int vfid, 31 int (*set_trust) (struct qed_dev *cdev, int vfid, bool trust);
|
/linux-6.12.1/arch/s390/pci/ |
D | pci_iov.c | 48 static int zpci_iov_link_virtfn(struct pci_dev *pdev, struct pci_dev *virtfn, int vfid) in zpci_iov_link_virtfn() argument 52 rc = pci_iov_sysfs_link(pdev, virtfn, vfid); in zpci_iov_link_virtfn() 68 int vfid = vfn - 1; /* Linux' vfid's start at 0 vfn at 1*/ in zpci_iov_setup_virtfn() local 87 cand_devfn = pci_iov_virtfn_devfn(pdev, vfid); in zpci_iov_setup_virtfn() 89 rc = zpci_iov_link_virtfn(pdev, virtfn, vfid); in zpci_iov_setup_virtfn()
|
/linux-6.12.1/drivers/net/ethernet/hisilicon/hns3/hns3vf/ |
D | hclgevf_trace.h | 23 __field(u8, vfid) 31 __entry->vfid = req->dest_vfid; 41 __get_str(pciname), __get_str(devname), __entry->vfid, 54 __field(u8, vfid) 63 __entry->vfid = req->mbx_src_vfid; 74 __get_str(pciname), __get_str(devname), __entry->vfid,
|
/linux-6.12.1/drivers/net/ethernet/qlogic/qed/ |
D | qed_sriov.c | 23 static int qed_iov_bulletin_set_mac(struct qed_hwfn *p_hwfn, u8 *mac, int vfid); 306 int vfid, struct qed_ptt *p_ptt) in qed_iov_post_vf_bulletin() argument 313 p_vf = qed_iov_get_vf_info(p_hwfn, (u16)vfid, true); in qed_iov_post_vf_bulletin() 658 int vfid, bool b_fail_malicious) in _qed_iov_pf_sanity_check() argument 666 if (!qed_iov_is_valid_vfid(p_hwfn, vfid, true, b_fail_malicious)) in _qed_iov_pf_sanity_check() 672 static bool qed_iov_pf_sanity_check(struct qed_hwfn *p_hwfn, int vfid) in qed_iov_pf_sanity_check() argument 674 return _qed_iov_pf_sanity_check(p_hwfn, vfid, true); in qed_iov_pf_sanity_check() 939 u16 vfid, in qed_iov_set_link() argument 945 vfid, in qed_iov_set_link() 1332 static void qed_iov_clean_vf(struct qed_hwfn *p_hwfn, u8 vfid) in qed_iov_clean_vf() argument [all …]
|
/linux-6.12.1/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
D | hclge_trace.h | 24 __field(u8, vfid) 33 __entry->vfid = req->mbx_src_vfid; 44 __get_str(pciname), __get_str(devname), __entry->vfid, 57 __field(u8, vfid) 65 __entry->vfid = req->dest_vfid; 75 __get_str(pciname), __get_str(devname), __entry->vfid,
|
/linux-6.12.1/drivers/crypto/cavium/cpt/ |
D | cptvf_mbox.c | 34 cptvf->vfid = mbx.data; in cptvf_handle_mbox_intr() 35 dev_dbg(&cptvf->pdev->dev, "Received VFID %d\n", cptvf->vfid); in cptvf_handle_mbox_intr() 42 cptvf->vfid, ((mbx.data == SE_TYPES) ? "SE" : "AE"), in cptvf_handle_mbox_intr() 77 (mbx->msg & 0xFF), cptvf->vfid); in cptvf_send_msg_to_pf_timeout()
|
D | cptvf_main.c | 524 intr, cptvf->vfid); in cptvf_misc_intr_handler() 532 intr, cptvf->vfid); in cptvf_misc_intr_handler() 536 intr, cptvf->vfid); in cptvf_misc_intr_handler() 540 intr, cptvf->vfid); in cptvf_misc_intr_handler() 544 intr, cptvf->vfid); in cptvf_misc_intr_handler() 547 cptvf->vfid); in cptvf_misc_intr_handler() 602 cptvf->vfid); in cptvf_done_intr_handler() 619 cptvf->vfid); in cptvf_set_irq_affinity() 623 cpu = cptvf->vfid % num_online_cpus(); in cptvf_set_irq_affinity()
|
/linux-6.12.1/drivers/crypto/marvell/octeontx/ |
D | otx_cptvf_mbox.c | 104 cptvf->vfid = mbx.data; in otx_cptvf_handle_mbox_intr() 105 dev_dbg(&cptvf->pdev->dev, "Received VFID %d\n", cptvf->vfid); in otx_cptvf_handle_mbox_intr() 111 cptvf->vfid, in otx_cptvf_handle_mbox_intr() 148 mbx->msg, cptvf->vfid); in cptvf_send_msg_to_pf_timeout()
|
D | otx_cptvf_main.c | 503 intr, cptvf->vfid); in cptvf_misc_intr_handler() 512 intr, cptvf->vfid); in cptvf_misc_intr_handler() 517 intr, cptvf->vfid); in cptvf_misc_intr_handler() 522 intr, cptvf->vfid); in cptvf_misc_intr_handler() 527 intr, cptvf->vfid); in cptvf_misc_intr_handler() 530 cptvf->vfid); in cptvf_misc_intr_handler() 585 cptvf->vfid); in cptvf_done_intr_handler() 603 cptvf->vfid); in cptvf_set_irq_affinity() 607 cpu = cptvf->vfid % num_online_cpus(); in cptvf_set_irq_affinity()
|
/linux-6.12.1/drivers/net/ethernet/intel/ice/ |
D | ice_vf_mbx.h | 18 ice_aq_send_msg_to_vf(struct ice_hw *hw, u16 vfid, u32 v_opcode, u32 v_retval, 31 u16 __always_unused vfid, u32 __always_unused v_opcode, in ice_aq_send_msg_to_vf() argument
|