/linux-6.12.1/drivers/infiniband/hw/mlx5/ |
D | umr.h | 33 if (MLX5_CAP_GEN(dev->mdev, umr_modify_entity_size_disabled)) in mlx5r_umr_can_load_pas() 40 if (!MLX5_CAP_GEN(dev->mdev, umr_extended_translation_offset) && in mlx5r_umr_can_load_pas() 58 MLX5_CAP_GEN(dev->mdev, atomic) && in mlx5r_umr_can_reconfig() 59 MLX5_CAP_GEN(dev->mdev, umr_modify_atomic_disabled)) in mlx5r_umr_can_reconfig() 63 MLX5_CAP_GEN(dev->mdev, relaxed_ordering_write) && in mlx5r_umr_can_reconfig() 64 !MLX5_CAP_GEN(dev->mdev, relaxed_ordering_write_umr)) in mlx5r_umr_can_reconfig() 68 (MLX5_CAP_GEN(dev->mdev, relaxed_ordering_read) || in mlx5r_umr_can_reconfig() 69 MLX5_CAP_GEN(dev->mdev, relaxed_ordering_read_pci_enabled)) && in mlx5r_umr_can_reconfig() 70 !MLX5_CAP_GEN(dev->mdev, relaxed_ordering_read_umr)) in mlx5r_umr_can_reconfig()
|
D | main.c | 110 int port_type_cap = MLX5_CAP_GEN(dev->mdev, port_type); in mlx5_ib_port_link_layer() 569 props->max_msg_sz = 1 << MLX5_CAP_GEN(dev->mdev, log_max_msg); in mlx5_query_port_roce() 687 if (MLX5_CAP_GEN(dev->mdev, port_type) == MLX5_CAP_PORT_TYPE_IB) in mlx5_use_mad_ifc() 688 return !MLX5_CAP_GEN(dev->mdev, ib_virt); in mlx5_use_mad_ifc() 784 *max_pkeys = mlx5_to_sw_pkey_sz(MLX5_CAP_GEN(mdev, in mlx5_query_max_pkeys() 879 u64 min_page_size = 1ull << MLX5_CAP_GEN(mdev, log_pg_sz); in mlx5_ib_query_device() 914 if (MLX5_CAP_GEN(mdev, pkv)) in mlx5_ib_query_device() 916 if (MLX5_CAP_GEN(mdev, qkv)) in mlx5_ib_query_device() 918 if (MLX5_CAP_GEN(mdev, apm)) in mlx5_ib_query_device() 920 if (MLX5_CAP_GEN(mdev, xrc)) in mlx5_ib_query_device() [all …]
|
D | counters.c | 207 return MLX5_CAP_GEN(dev->mdev, q_counter_other_vport) && in vport_qcounters_supported() 208 MLX5_CAP_GEN(dev->mdev, q_counter_aggregation); in vport_qcounters_supported() 403 if (MLX5_CAP_GEN(dev->mdev, cc_query_allowed)) { in do_get_hw_stats() 603 if (MLX5_CAP_GEN(dev->mdev, out_of_seq_cnt)) { in mlx5_ib_fill_counters() 613 if (MLX5_CAP_GEN(dev->mdev, retransmission_q_counters)) { in mlx5_ib_fill_counters() 623 if (MLX5_CAP_GEN(dev->mdev, enhanced_error_q_counters)) { in mlx5_ib_fill_counters() 633 if (MLX5_CAP_GEN(dev->mdev, roce_accl)) { in mlx5_ib_fill_counters() 643 if (MLX5_CAP_GEN(dev->mdev, cc_query_allowed)) { in mlx5_ib_fill_counters() 696 if (MLX5_CAP_GEN(dev->mdev, out_of_seq_cnt)) in __mlx5_ib_alloc_counters() 701 if (MLX5_CAP_GEN(dev->mdev, retransmission_q_counters)) in __mlx5_ib_alloc_counters() [all …]
|
D | qp.c | 326 if (!MLX5_CAP_GEN(dev->mdev, qpc_extension) || in mlx5_ib_qp_err_syndrome() 327 !MLX5_CAP_GEN(dev->mdev, qp_error_syndrome)) in mlx5_ib_qp_err_syndrome() 440 if (cap->max_recv_wr > (1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz))) in set_rq_size() 475 if (wqe_size > MLX5_CAP_GEN(dev->mdev, max_wqe_sz_rq)) { in set_rq_size() 478 MLX5_CAP_GEN(dev->mdev, in set_rq_size() 605 if (wqe_size > MLX5_CAP_GEN(dev->mdev, max_wqe_sz_sq)) { in calc_sq_size() 607 wqe_size, MLX5_CAP_GEN(dev->mdev, max_wqe_sz_sq)); in calc_sq_size() 617 if (qp->sq.wqe_cnt > (1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz))) { in calc_sq_size() 621 1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz)); in calc_sq_size() 644 if (desc_sz > MLX5_CAP_GEN(dev->mdev, max_wqe_sz_sq)) { in set_user_buf_size() [all …]
|
D | umr.c | 55 if (MLX5_CAP_GEN(dev->mdev, atomic)) in get_umr_update_access_mask() 58 if (MLX5_CAP_GEN(dev->mdev, relaxed_ordering_write_umr)) in get_umr_update_access_mask() 61 if (MLX5_CAP_GEN(dev->mdev, relaxed_ordering_read_umr)) in get_umr_update_access_mask() 79 MLX5_CAP_GEN(dev->mdev, umr_modify_entity_size_disabled)) in umr_check_mkey_mask() 83 MLX5_CAP_GEN(dev->mdev, umr_modify_atomic_disabled)) in umr_check_mkey_mask() 87 !MLX5_CAP_GEN(dev->mdev, relaxed_ordering_write_umr)) in umr_check_mkey_mask() 91 !MLX5_CAP_GEN(dev->mdev, relaxed_ordering_read_umr)) in umr_check_mkey_mask() 416 (MLX5_CAP_GEN(dev->mdev, relaxed_ordering_read) || in mlx5r_umr_set_access_flags() 750 return !MLX5_CAP_GEN(dev->mdev, umr_indirect_mkey_disabled); in umr_can_use_indirect_mkey()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
D | sf.h | 11 return MLX5_CAP_GEN(dev, sf_base_id); in mlx5_sf_start_function_id() 18 return MLX5_CAP_GEN(dev, sf); in mlx5_sf_supported() 25 if (MLX5_CAP_GEN(dev, max_num_sf)) in mlx5_sf_max_functions() 26 return MLX5_CAP_GEN(dev, max_num_sf); in mlx5_sf_max_functions() 28 return 1 << MLX5_CAP_GEN(dev, log_max_sf); in mlx5_sf_max_functions()
|
D | sd.c | 108 if (!MLX5_CAP_GEN(dev, eswitch_manager)) in mlx5_sd_is_supported() 110 if (!MLX5_CAP_GEN(dev, silent_mode)) in mlx5_sd_is_supported() 114 if (!MLX5_CAP_GEN(dev, cross_vhca_rqt)) in mlx5_sd_is_supported() 146 return (u32)((MLX5_CAP_GEN(dev, native_port_num) << 8) | sd_group); in mlx5_sd_group_id() 322 u16 vhca_id_to_be_accessed = MLX5_CAP_GEN(primary, vhca_id); in sd_secondary_create_alias_ft() 393 MLX5_CAP_GEN(primary, vhca_id)); in sd_print_group() 397 MLX5_CAP_GEN(pos, vhca_id)); in sd_print_group() 409 MLX5_CAP_GEN(dev, vhca_id)); in dev_read()
|
D | clock.h | 38 u8 rq_ts_format_cap = MLX5_CAP_GEN(mdev, rq_ts_format); in mlx5_is_real_time_rq() 47 u8 sq_ts_format_cap = MLX5_CAP_GEN(mdev, sq_ts_format); in mlx5_is_real_time_sq()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/diag/ |
D | reporter_vnic.c | 32 if (MLX5_CAP_GEN(dev, vnic_env_queue_counters)) { in mlx5_reporter_vnic_diagnose_counters() 38 if (MLX5_CAP_GEN(dev, eq_overrun_count)) { in mlx5_reporter_vnic_diagnose_counters() 44 if (MLX5_CAP_GEN(dev, vnic_env_cq_overrun)) in mlx5_reporter_vnic_diagnose_counters() 47 if (MLX5_CAP_GEN(dev, invalid_command_count)) in mlx5_reporter_vnic_diagnose_counters() 50 if (MLX5_CAP_GEN(dev, quota_exceeded_count)) in mlx5_reporter_vnic_diagnose_counters() 53 if (MLX5_CAP_GEN(dev, nic_receive_steering_discard)) in mlx5_reporter_vnic_diagnose_counters() 56 if (MLX5_CAP_GEN(dev, vnic_env_cnt_steering_fail)) { in mlx5_reporter_vnic_diagnose_counters()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/ |
D | fw.c | 150 if (MLX5_CAP_GEN(dev, port_selection_cap)) { in mlx5_query_hca_caps() 156 if (MLX5_CAP_GEN(dev, hca_cap_2)) { in mlx5_query_hca_caps() 162 if (MLX5_CAP_GEN(dev, eth_net_offloads)) { in mlx5_query_hca_caps() 169 if (MLX5_CAP_GEN(dev, ipoib_enhanced_offloads)) { in mlx5_query_hca_caps() 176 if (MLX5_CAP_GEN(dev, pg)) { in mlx5_query_hca_caps() 182 if (MLX5_CAP_GEN(dev, atomic)) { in mlx5_query_hca_caps() 188 if (MLX5_CAP_GEN(dev, roce)) { in mlx5_query_hca_caps() 194 if (MLX5_CAP_GEN(dev, nic_flow_table) || in mlx5_query_hca_caps() 195 MLX5_CAP_GEN(dev, ipoib_enhanced_offloads)) { in mlx5_query_hca_caps() 212 if (MLX5_CAP_GEN(dev, qos)) { in mlx5_query_hca_caps() [all …]
|
D | vport.c | 271 1 << MLX5_CAP_GEN(dev, log_max_current_uc_list) : in mlx5_query_nic_vport_mac_list() 272 1 << MLX5_CAP_GEN(dev, log_max_current_mc_list); in mlx5_query_nic_vport_mac_list() 330 1 << MLX5_CAP_GEN(dev, log_max_current_uc_list) : in mlx5_modify_nic_vport_mac_list() 331 1 << MLX5_CAP_GEN(dev, log_max_current_mc_list); in mlx5_modify_nic_vport_mac_list() 381 max_list_size = 1 << MLX5_CAP_GEN(dev, log_max_vlan_list); in mlx5_modify_nic_vport_vlans() 492 if (!MLX5_CAP_GEN(mdev, vport_group_manager)) in mlx5_modify_nic_vport_node_guid() 552 is_group_manager = MLX5_CAP_GEN(dev, vport_group_manager); in mlx5_query_hca_vport_gid() 553 tbsz = mlx5_get_gid_table_len(MLX5_CAP_GEN(dev, gid_table_size)); in mlx5_query_hca_vport_gid() 586 if (MLX5_CAP_GEN(dev, num_ports) == 2) in mlx5_query_hca_vport_gid() 619 is_group_manager = MLX5_CAP_GEN(dev, vport_group_manager); in mlx5_query_hca_vport_pkey() [all …]
|
D | en_dcbnl.c | 60 #define MLX5_DSCP_SUPPORTED(mdev) (MLX5_CAP_GEN(mdev, qcam_reg) && \ 92 if (!MLX5_CAP_GEN(priv->mdev, dcbx)) in mlx5e_dcbnl_switch_to_host_mode() 117 if (!MLX5_CAP_GEN(priv->mdev, ets)) in mlx5e_dcbnl_ieee_getets() 328 if (!MLX5_CAP_GEN(priv->mdev, ets)) in mlx5e_dcbnl_ieee_setets() 424 if ((!mode) && MLX5_CAP_GEN(priv->mdev, dcbx)) { in mlx5e_dcbnl_setdcbx() 456 if (!MLX5_CAP_GEN(priv->mdev, vport_group_manager) || in mlx5e_dcbnl_ieee_setapp() 509 if (!MLX5_CAP_GEN(priv->mdev, vport_group_manager) || in mlx5e_dcbnl_ieee_delapp() 631 if (!MLX5_CAP_GEN(mdev, ets)) in mlx5e_dcbnl_setall() 740 if (!MLX5_CAP_GEN(priv->mdev, ets)) { in mlx5e_dcbnl_getpgtccfgtx() 1028 if (MLX5_CAP_GEN(mdev, vport_group_manager) && MLX5_CAP_GEN(mdev, qos)) in mlx5e_dcbnl_build_netdev() [all …]
|
D | uar.c | 63 if (MLX5_CAP_GEN(mdev, uar_4k)) in uars_per_sys_page() 64 return MLX5_CAP_GEN(mdev, num_of_uars_per_page); in uars_per_sys_page() 73 if (MLX5_CAP_GEN(mdev, uar_4k)) in uar2pfn() 201 (1 << MLX5_CAP_GEN(mdev, log_bf_reg_size)) + MLX5_BF_OFFSET; in map_offset() 281 bf_reg_size = 1 << MLX5_CAP_GEN(dev, log_bf_reg_size); in addr_to_dbi_in_syspage()
|
D | en_common.c | 42 bool ro_write = MLX5_CAP_GEN(mdev, relaxed_ordering_write); in mlx5e_mkey_set_relaxed_ordering() 43 bool ro_read = MLX5_CAP_GEN(mdev, relaxed_ordering_read) || in mlx5e_mkey_set_relaxed_ordering() 45 MLX5_CAP_GEN(mdev, relaxed_ordering_read_pci_enabled)); in mlx5e_mkey_set_relaxed_ordering() 105 return MLX5_CAP_GEN(mdev, lag_tx_port_affinity) && mlx5e_get_num_lag_ports(mdev) > 1; in mlx5_lag_should_assign_affinity()
|
D | dev.c | 62 if (MLX5_CAP_GEN(dev, port_type) != MLX5_CAP_PORT_TYPE_ETH) in mlx5_eth_supported() 65 if (!MLX5_CAP_GEN(dev, eth_net_offloads)) { in mlx5_eth_supported() 70 if (!MLX5_CAP_GEN(dev, nic_flow_table)) { in mlx5_eth_supported() 103 if (!MLX5_CAP_GEN(dev, cq_moderation)) in mlx5_eth_supported() 170 if (MLX5_CAP_GEN(dev, port_type) != MLX5_CAP_PORT_TYPE_ETH) in is_mp_supported()
|
D | devlink.c | 410 if (new_state && !MLX5_CAP_GEN(dev, roce) && in mlx5_devlink_enable_roce_validate() 411 !(MLX5_CAP_GEN(dev, roce_rw_supported) && MLX5_CAP_GEN_MAX(dev, roce))) { in mlx5_devlink_enable_roce_validate() 468 if (val32 > BIT(MLX5_CAP_GEN(dev, log_max_hairpin_num_packets))) { in mlx5_devlink_hairpin_queue_size_validate() 471 BIT(MLX5_CAP_GEN(dev, log_max_hairpin_num_packets))); in mlx5_devlink_hairpin_queue_size_validate() 497 MLX5_CAP_GEN(dev, log_max_hairpin_num_packets))); in mlx5_devlink_hairpin_params_init_values() 523 value.vbool = MLX5_CAP_GEN(dev, roce) && !mlx5_dev_is_lightweight(dev); in mlx5_devlink_set_params_init_values() 753 value.vu32 = 1 << MLX5_CAP_GEN(dev, log_max_current_uc_list); in mlx5_devlink_max_uc_list_params_register()
|
D | mlx5_core.h | 281 #define MLX5_PPS_CAP(mdev) (MLX5_CAP_GEN((mdev), pps) && \ 282 MLX5_CAP_GEN((mdev), pps_modify) && \ 397 if (MLX5_CAP_GEN(dev, max_num_eqs)) in mlx5_max_eq_cap_get() 398 return MLX5_CAP_GEN(dev, max_num_eqs); in mlx5_max_eq_cap_get() 400 return 1 << MLX5_CAP_GEN(dev, log_max_eq); in mlx5_max_eq_cap_get()
|
D | main.c | 229 if (!MLX5_CAP_GEN(dev, driver_version)) in mlx5_set_driver_version() 432 if (!MLX5_CAP_GEN(dev, atomic)) in handle_hca_cap_atomic() 462 !MLX5_CAP_GEN(dev, pg)) in handle_hca_cap_odp() 546 return MLX5_CAP_GEN(dev, roce); in mlx5_is_roce_on() 592 mlx5_to_sw_pkey_sz(MLX5_CAP_GEN(dev, pkey_table_size)), in handle_hca_cap() 658 if (MLX5_CAP_GEN(dev, roce_rw_supported) && MLX5_CAP_GEN_MAX(dev, roce)) in handle_hca_cap() 686 return (MLX5_CAP_GEN(dev, roce_rw_supported) && !mlx5_is_roce_on(dev)) || in is_roce_fw_disabled() 687 (!MLX5_CAP_GEN(dev, roce_rw_supported) && !MLX5_CAP_GEN(dev, roce)); in is_roce_fw_disabled() 724 if (!MLX5_CAP_GEN(dev, port_selection_cap)) in handle_hca_cap_port_selection() 823 if (MLX5_CAP_GEN(dev, port_type) == MLX5_CAP_PORT_TYPE_ETH) in mlx5_core_set_hca_defaults() [all …]
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | monitor_stats.c | 25 if (!MLX5_CAP_GEN(mdev, max_num_of_monitor_counters)) in mlx5e_monitor_counter_cap() 28 MLX5_CAP_GEN(mdev, num_ppcnt_monitor_counters) < in mlx5e_monitor_counter_cap() 31 if (MLX5_CAP_GEN(mdev, num_q_monitor_counters) < in mlx5e_monitor_counter_cap() 114 int max_num_of_counters = MLX5_CAP_GEN(mdev, max_num_of_monitor_counters); in mlx5e_set_monitor_counter() 115 int num_q_counters = MLX5_CAP_GEN(mdev, num_q_monitor_counters); in mlx5e_set_monitor_counter() 117 MLX5_CAP_GEN(mdev, num_ppcnt_monitor_counters); in mlx5e_set_monitor_counter()
|
D | channels.c | 34 *vhca_id = MLX5_CAP_GEN(c->mdev, vhca_id); in mlx5e_channels_get_regular_rqn() 46 *vhca_id = MLX5_CAP_GEN(c->mdev, vhca_id); in mlx5e_channels_get_xsk_rqn()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/lag/ |
D | lag.h | 119 if (!MLX5_CAP_GEN(dev, vport_group_manager) || in mlx5_lag_is_supported() 120 !MLX5_CAP_GEN(dev, lag_master) || in mlx5_lag_is_supported() 121 MLX5_CAP_GEN(dev, num_lag_ports) < 2 || in mlx5_lag_is_supported() 122 MLX5_CAP_GEN(dev, num_lag_ports) > MLX5_MAX_PORTS) in mlx5_lag_is_supported()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
D | ktls.h | 26 if (!MLX5_CAP_GEN(mdev, tls_tx) && !MLX5_CAP_GEN(mdev, tls_rx)) in mlx5e_is_ktls_device() 29 if (!MLX5_CAP_GEN(mdev, log_max_dek)) in mlx5e_is_ktls_device() 65 return !is_kdump_kernel() && MLX5_CAP_GEN(mdev, tls_tx) && in mlx5e_is_ktls_tx()
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
D | dr_cmd.c | 117 caps->prio_tag_required = MLX5_CAP_GEN(mdev, prio_tag_required); in mlx5dr_cmd_query_device() 118 caps->eswitch_manager = MLX5_CAP_GEN(mdev, eswitch_manager); in mlx5dr_cmd_query_device() 119 caps->gvmi = MLX5_CAP_GEN(mdev, vhca_id); in mlx5dr_cmd_query_device() 120 caps->flex_protocols = MLX5_CAP_GEN(mdev, flex_parser_protocols); in mlx5dr_cmd_query_device() 121 caps->sw_format_ver = MLX5_CAP_GEN(mdev, steering_format_version); in mlx5dr_cmd_query_device() 123 MLX5_CAP_GEN(mdev, fl_rc_qp_when_roce_disabled); in mlx5dr_cmd_query_device() 125 if (MLX5_CAP_GEN(mdev, roce)) { in mlx5dr_cmd_query_device() 137 caps->isolate_vl_tc = MLX5_CAP_GEN(mdev, isolate_vl_tc_new); in mlx5dr_cmd_query_device() 145 MLX5_CAP_GEN(mdev, log_header_modify_argument_granularity); in mlx5dr_cmd_query_device() 147 MLX5_CAP_GEN(mdev, log_header_modify_argument_max_alloc); in mlx5dr_cmd_query_device() [all …]
|
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/esw/acl/ |
D | egress_ofld.c | 77 if (MLX5_CAP_GEN(esw->dev, prio_tag_required)) { in esw_acl_egress_ofld_rules_create() 125 if (MLX5_CAP_GEN(esw->dev, prio_tag_required)) { in esw_acl_egress_ofld_groups_create() 192 !MLX5_CAP_GEN(esw->dev, prio_tag_required)) in esw_acl_egress_ofld_setup() 202 if (MLX5_CAP_GEN(esw->dev, prio_tag_required)) in esw_acl_egress_ofld_setup() 259 fwd_dest.vport.vhca_id = MLX5_CAP_GEN(esw->dev, vhca_id); in mlx5_esw_acl_egress_vport_bond()
|
/linux-6.12.1/include/linux/mlx5/ |
D | driver.h | 1244 return dev->caps.embedded_cpu && MLX5_CAP_GEN(dev, eswitch_manager); in mlx5_core_is_ecpf_esw_manager() 1264 return MLX5_CAP_GEN(dev, vport_group_manager) && in mlx5_lag_is_lacp_owner() 1265 (MLX5_CAP_GEN(dev, num_lag_ports) > 1) && in mlx5_lag_is_lacp_owner() 1266 MLX5_CAP_GEN(dev, lag_master); in mlx5_lag_is_lacp_owner() 1291 return MLX5_CAP_GEN(dev, affiliate_nic_vport_criteria) && in mlx5_core_is_mp_slave() 1292 MLX5_CAP_GEN(dev, num_vhca_ports) <= 1; in mlx5_core_is_mp_slave() 1297 return MLX5_CAP_GEN(dev, num_vhca_ports) > 1; in mlx5_core_is_mp_master() 1311 return MLX5_CAP_GEN(dev, native_port_num); in mlx5_core_native_port_num() 1316 int idx = MLX5_CAP_GEN(dev, native_port_num); in mlx5_get_dev_index() 1332 if (MLX5_CAP_GEN(dev, roce_rw_supported)) in mlx5_get_roce_state() [all …]
|