/linux-6.12.1/drivers/thunderbolt/ |
D | path.c | 163 path->hops = kcalloc(num_hops, sizeof(*path->hops), GFP_KERNEL); in tb_path_discover() 164 if (!path->hops) { in tb_path_discover() 198 path->hops[i].in_port = p; in tb_path_discover() 199 path->hops[i].in_hop_index = h; in tb_path_discover() 200 path->hops[i].in_counter_index = -1; in tb_path_discover() 201 path->hops[i].out_port = out_port; in tb_path_discover() 202 path->hops[i].next_hop_index = next_hop; in tb_path_discover() 204 tb_dump_hop(&path->hops[i], &hop); in tb_path_discover() 270 path->hops = kcalloc(num_hops, sizeof(*path->hops), GFP_KERNEL); in tb_path_alloc() 271 if (!path->hops) { in tb_path_alloc() [all …]
|
D | test.c | 913 in_port = path->hops[i].in_port; in tb_test_path_not_bonded_lane0() 914 out_port = path->hops[i].out_port; in tb_test_path_not_bonded_lane0() 975 in_port = path->hops[i].in_port; in tb_test_path_not_bonded_lane1() 976 out_port = path->hops[i].out_port; in tb_test_path_not_bonded_lane1() 1055 in_port = path->hops[i].in_port; in tb_test_path_not_bonded_lane1_chain() 1056 out_port = path->hops[i].out_port; in tb_test_path_not_bonded_lane1_chain() 1135 in_port = path->hops[i].in_port; in tb_test_path_not_bonded_lane1_chain_reverse() 1136 out_port = path->hops[i].out_port; in tb_test_path_not_bonded_lane1_chain_reverse() 1227 in_port = path->hops[i].in_port; in tb_test_path_mixed_chain() 1228 out_port = path->hops[i].out_port; in tb_test_path_mixed_chain() [all …]
|
D | tunnel.c | 904 paths[TB_DP_VIDEO_PATH_OUT]->hops[0].in_hop_index, in tb_dp_activate() 905 paths[TB_DP_AUX_PATH_OUT]->hops[0].in_hop_index, in tb_dp_activate() 906 paths[TB_DP_AUX_PATH_IN]->hops[last].next_hop_index); in tb_dp_activate() 909 paths[TB_DP_VIDEO_PATH_OUT]->hops[last].next_hop_index, in tb_dp_activate() 910 paths[TB_DP_AUX_PATH_IN]->hops[0].in_hop_index, in tb_dp_activate() 911 paths[TB_DP_AUX_PATH_OUT]->hops[last].next_hop_index); in tb_dp_activate() 1571 hop = &path->hops[0]; in tb_dma_init_rx_path() 1579 ret = tb_dma_reserve_credits(&path->hops[i], credits); in tb_dma_init_rx_path() 1748 if (tb_port_is_nhi(path->hops[0].in_port)) in tb_tunnel_match_dma() 1750 else if (tb_port_is_nhi(path->hops[path->path_length - 1].out_port)) in tb_tunnel_match_dma() [all …]
|
/linux-6.12.1/drivers/accel/habanalabs/common/mmu/ |
D | mmu_v2.c | 245 static int hl_mmu_v2_get_tlb_info(struct hl_ctx *ctx, u64 virt_addr, struct hl_mmu_hop_info *hops) in hl_mmu_v2_get_tlb_info() argument 262 hops->range_type = HL_VA_RANGE_TYPE_DRAM; in hl_mmu_v2_get_tlb_info() 264 hops->scrambled_vaddr = hdev->asic_funcs->scramble_addr(hdev, virt_addr); in hl_mmu_v2_get_tlb_info() 266 hops->hop_info[0].hop_addr = hl_mmu_dr_get_phys_hop0_addr(ctx); in hl_mmu_v2_get_tlb_info() 267 hops->hop_info[0].hop_pte_addr = hl_mmu_get_hop_pte_phys_addr(ctx, mmu_prop, 0, in hl_mmu_v2_get_tlb_info() 268 hops->hop_info[0].hop_addr, in hl_mmu_v2_get_tlb_info() 269 hops->scrambled_vaddr); in hl_mmu_v2_get_tlb_info() 270 if (hops->hop_info[0].hop_pte_addr == U64_MAX) in hl_mmu_v2_get_tlb_info() 273 hops->hop_info[0].hop_pte_val = hdev->asic_funcs->read_pte(hdev, in hl_mmu_v2_get_tlb_info() 274 hops->hop_info[0].hop_pte_addr); in hl_mmu_v2_get_tlb_info() [all …]
|
D | mmu.c | 477 struct hl_mmu_hop_info *hops, in hl_mmu_pa_page_with_offset() argument 485 if (hops->unscrambled_paddr) in hl_mmu_pa_page_with_offset() 486 tmp_phys_addr = hops->unscrambled_paddr; in hl_mmu_pa_page_with_offset() 488 tmp_phys_addr = hops->hop_info[hops->used_hops - 1].hop_pte_val; in hl_mmu_pa_page_with_offset() 490 if (hops->range_type == HL_VA_RANGE_TYPE_HOST_HUGE) in hl_mmu_pa_page_with_offset() 492 else if (hops->range_type == HL_VA_RANGE_TYPE_HOST) in hl_mmu_pa_page_with_offset() 497 if ((hops->range_type == HL_VA_RANGE_TYPE_DRAM) && in hl_mmu_pa_page_with_offset() 524 hop_shift = mmu_prop->hop_shifts[hops->used_hops - 1]; in hl_mmu_pa_page_with_offset() 534 struct hl_mmu_hop_info hops; in hl_mmu_va_to_pa() local 537 memset(&hops, 0, sizeof(hops)); in hl_mmu_va_to_pa() [all …]
|
D | mmu_v1.c | 454 struct hl_mmu_hop_info *hops) in hl_mmu_v1_get_tlb_info() argument 491 hops->hop_info[0].hop_addr = hl_mmu_dr_get_phys_hop0_addr(ctx); in hl_mmu_v1_get_tlb_info() 492 hops->hop_info[0].hop_pte_addr = in hl_mmu_v1_get_tlb_info() 494 hops->hop_info[0].hop_addr, virt_addr); in hl_mmu_v1_get_tlb_info() 495 hops->hop_info[0].hop_pte_val = in hl_mmu_v1_get_tlb_info() 497 hops->hop_info[0].hop_pte_addr); in hl_mmu_v1_get_tlb_info() 500 hops->hop_info[i].hop_addr = in hl_mmu_v1_get_tlb_info() 502 hops->hop_info[i - 1].hop_pte_val); in hl_mmu_v1_get_tlb_info() 503 if (hops->hop_info[i].hop_addr == ULLONG_MAX) in hl_mmu_v1_get_tlb_info() 506 hops->hop_info[i].hop_pte_addr = in hl_mmu_v1_get_tlb_info() [all …]
|
D | mmu_v2_hr.c | 335 struct hl_mmu_hop_info *hops, in hl_mmu_v2_hr_get_tlb_mapping_params() argument 354 hops->range_type = HL_VA_RANGE_TYPE_DRAM; in hl_mmu_v2_hr_get_tlb_mapping_params() 358 hops->range_type = HL_VA_RANGE_TYPE_HOST; in hl_mmu_v2_hr_get_tlb_mapping_params() 362 hops->range_type = HL_VA_RANGE_TYPE_HOST_HUGE; in hl_mmu_v2_hr_get_tlb_mapping_params() 371 struct hl_mmu_hop_info *hops) in hl_mmu_v2_hr_get_tlb_info() argument 373 return hl_mmu_hr_get_tlb_info(ctx, virt_addr, hops, in hl_mmu_v2_hr_get_tlb_info()
|
/linux-6.12.1/tools/testing/selftests/bpf/prog_tests/ |
D | cls_redirect.c | 144 enum hops { enum 169 enum hops hops; member 176 const char *family_str, *type, *conn, *hops, *result, *flags; in test_str() local 190 hops = "no hops"; in test_str() 191 if (test->hops == ONE_HOP) in test_str() 192 hops = "one hop"; in test_str() 205 type, result, conn, hops, flags); in test_str() 261 encap_init(&encap, test->hops == ONE_HOP ? 1 : 0, proto); in build_input() 264 if (test->hops == ONE_HOP) { in build_input()
|
/linux-6.12.1/Documentation/networking/ |
D | nexthop-group-resilient.rst | 9 weights of constituent next hops. 48 constituent next hops: a hash table. The selection algorithm uses SKB hash 55 the individual next hops is arbitrary. Therefore when a next hop is deleted 56 the buckets that held it are simply reassigned to other next hops:: 69 When weights of next hops in a group are altered, it may be possible to 83 cause bucket allocation change, the wants counts for individual next hops 86 Next hops that have fewer buckets than their wants count, are called 88 overweight (and therefore no underweight) next hops in the group, it is 97 After assigning wants counts to next hops, an "upkeep" algorithm runs. For 105 underweight next hops. If, after considering all buckets in this manner, [all …]
|
/linux-6.12.1/block/ |
D | bdev.c | 500 const struct blk_holder_ops *hops) in bd_may_claim() argument 511 if (WARN_ON_ONCE(bdev->bd_holder_ops != hops)) in bd_may_claim() 542 const struct blk_holder_ops *hops) in bd_prepare_to_claim() argument 551 if (!bd_may_claim(bdev, holder, hops)) { in bd_prepare_to_claim() 594 const struct blk_holder_ops *hops) in bd_finish_claiming() argument 599 BUG_ON(!bd_may_claim(bdev, holder, hops)); in bd_finish_claiming() 609 bdev->bd_holder_ops = hops; in bd_finish_claiming() 886 const struct blk_holder_ops *hops, struct file *bdev_file) in bdev_open() argument 894 ret = bd_prepare_to_claim(bdev, holder, hops); in bdev_open() 921 bd_finish_claiming(bdev, holder, hops); in bdev_open() [all …]
|
/linux-6.12.1/drivers/firmware/arm_scmi/ |
D | sensors.c | 253 if (!ph->hops->protocol_msg_check(ph, in scmi_sensor_attributes_get() 257 if (!ph->hops->protocol_msg_check(ph, in scmi_sensor_attributes_get() 368 iter = ph->hops->iter_response_init(ph, &ops, s->intervals.count, in scmi_sensor_update_intervals() 375 return ph->hops->iter_response_run(iter); in scmi_sensor_update_intervals() 506 iter = ph->hops->iter_response_init(ph, &ops, s->num_axis, in scmi_sensor_axis_extended_names_get() 517 ret = ph->hops->iter_response_run(iter); in scmi_sensor_axis_extended_names_get() 547 iter = ph->hops->iter_response_init(ph, &ops, s->num_axis, in scmi_sensor_axis_description() 554 ret = ph->hops->iter_response_run(iter); in scmi_sensor_axis_description() 664 ph->hops->extended_name_get(ph, SENSOR_NAME_GET, s->id, in iter_sens_descr_process_response() 703 iter = ph->hops->iter_response_init(ph, &ops, si->num_sensors, in scmi_sensor_description_get() [all …]
|
D | perf.c | 230 if (!ph->hops->protocol_msg_check(ph, PERF_NOTIFY_LEVEL, NULL)) in scmi_perf_attributes_get() 233 if (!ph->hops->protocol_msg_check(ph, PERF_NOTIFY_LIMITS, NULL)) in scmi_perf_attributes_get() 328 ph->hops->extended_name_get(ph, PERF_DOMAIN_NAME_GET, in scmi_perf_domain_attributes_get() 487 iter = ph->hops->iter_response_init(ph, &ops, MAX_OPPS, in scmi_perf_describe_levels_get() 494 ret = ph->hops->iter_response_run(iter); in scmi_perf_describe_levels_get() 569 ph->hops->fastchannel_db_ring(fci->set_db); in __scmi_perf_limits_set() 723 ph->hops->fastchannel_db_ring(fci->set_db); in __scmi_perf_level_set() 849 ph->hops->fastchannel_init(ph, PERF_DESCRIBE_FASTCHANNEL, in scmi_perf_domain_init_fc() 854 ph->hops->fastchannel_init(ph, PERF_DESCRIBE_FASTCHANNEL, in scmi_perf_domain_init_fc() 860 ph->hops->fastchannel_init(ph, PERF_DESCRIBE_FASTCHANNEL, in scmi_perf_domain_init_fc() [all …]
|
D | powercap.c | 164 if (!ph->hops->protocol_msg_check(ph, in scmi_powercap_attributes_get() 168 if (!ph->hops->protocol_msg_check(ph, in scmi_powercap_attributes_get() 291 ph->hops->extended_name_get(ph, POWERCAP_DOMAIN_NAME_GET, in scmi_powercap_domain_attributes_get() 427 ph->hops->fastchannel_db_ring(fci->set_db); in __scmi_powercap_cap_set() 547 ph->hops->fastchannel_db_ring(fci->set_db); in scmi_powercap_pai_set() 719 ph->hops->fastchannel_init(ph, POWERCAP_DESCRIBE_FASTCHANNEL, in scmi_powercap_domain_init_fc() 725 ph->hops->fastchannel_init(ph, POWERCAP_DESCRIBE_FASTCHANNEL, in scmi_powercap_domain_init_fc() 730 ph->hops->fastchannel_init(ph, POWERCAP_DESCRIBE_FASTCHANNEL, in scmi_powercap_domain_init_fc() 736 ph->hops->fastchannel_init(ph, POWERCAP_DESCRIBE_FASTCHANNEL, in scmi_powercap_domain_init_fc()
|
D | pinctrl.c | 237 ret = ph->hops->extended_name_get(ph, PINCTRL_NAME_GET, in scmi_pinctrl_attributes() 311 iter = ph->hops->iter_response_init(ph, &ops, size, in scmi_pinctrl_list_associations() 318 return ph->hops->iter_response_run(iter); in scmi_pinctrl_list_associations() 426 iter = ph->hops->iter_response_init(ph, &ops, max_configs, in scmi_pinctrl_settings_get() 433 return ph->hops->iter_response_run(iter); in scmi_pinctrl_settings_get() 476 int max_msg_size = ph->hops->get_max_msg_size(ph); in scmi_pinctrl_settings_conf()
|
D | clock.c | 214 if (!ph->hops->protocol_msg_check(ph, CLOCK_RATE_NOTIFY, NULL)) in scmi_clock_protocol_attributes_get() 217 if (!ph->hops->protocol_msg_check(ph, in scmi_clock_protocol_attributes_get() 306 iter = ph->hops->iter_response_init(ph, &ops, 0, in scmi_clock_possible_parents() 313 ret = ph->hops->iter_response_run(iter); in scmi_clock_possible_parents() 385 ph->hops->extended_name_get(ph, CLOCK_NAME_GET, clk_id, in scmi_clock_attributes_get() 522 iter = ph->hops->iter_response_init(ph, &ops, SCMI_MAX_NUM_RATES, in scmi_clock_describe_rates_get() 529 ret = ph->hops->iter_response_run(iter); in scmi_clock_describe_rates_get()
|
D | voltage.c | 196 iter = ph->hops->iter_response_init(ph, &ops, v->num_levels, in scmi_voltage_levels_get() 203 ret = ph->hops->iter_response_run(iter); in scmi_voltage_levels_get() 248 ph->hops->extended_name_get(ph, in scmi_voltage_descriptors_get()
|
D | power.c | 103 if (!ph->hops->protocol_msg_check(ph, POWER_STATE_NOTIFY, NULL)) in scmi_power_attributes_get() 146 ph->hops->extended_name_get(ph, POWER_DOMAIN_NAME_GET, in scmi_power_domain_attributes_get()
|
D | reset.c | 95 if (!ph->hops->protocol_msg_check(ph, RESET_NOTIFY, NULL)) in scmi_reset_attributes_get() 142 ph->hops->extended_name_get(ph, RESET_DOMAIN_NAME_GET, domain, in scmi_reset_domain_attributes_get()
|
D | system.c | 163 if (!ph->hops->protocol_msg_check(ph, SYSTEM_POWER_STATE_NOTIFY, NULL)) in scmi_system_protocol_init()
|
/linux-6.12.1/net/ipv6/ |
D | exthdrs.c | 1082 int hops; in ipv6_push_rthdr0() local 1089 hops = ihdr->rt_hdr.hdrlen >> 1; in ipv6_push_rthdr0() 1091 if (hops > 1) in ipv6_push_rthdr0() 1093 (hops - 1) * sizeof(struct in6_addr)); in ipv6_push_rthdr0() 1095 phdr->addr[hops - 1] = **addr_p; in ipv6_push_rthdr0() 1107 int plen, hops; in ipv6_push_rthdr4() local 1115 hops = sr_ihdr->first_segment + 1; in ipv6_push_rthdr4() 1117 (hops - 1) * sizeof(struct in6_addr)); in ipv6_push_rthdr4() 1122 if (sr_ihdr->hdrlen > hops * 2) { in ipv6_push_rthdr4() 1125 tlvs_offset = (1 + hops * 2) << 3; in ipv6_push_rthdr4() [all …]
|
/linux-6.12.1/arch/s390/kernel/ |
D | dis.c | 510 int start, end, opsize, hops, i; in show_code() local 530 for (i = 0, hops = 0; start + i < 32 && hops < 3; hops++) { in show_code() 543 hops = 0; in show_code() 544 while (start < end && hops < 8) { in show_code() 566 hops++; in show_code()
|
/linux-6.12.1/include/linux/ |
D | topology.h | 250 extern const struct cpumask *sched_numa_hop_mask(unsigned int node, unsigned int hops); 258 sched_numa_hop_mask(unsigned int node, unsigned int hops) in sched_numa_hop_mask() argument
|
/linux-6.12.1/include/dt-bindings/usb/ |
D | pd.h | 404 #define VDO_ACABLE2(mtemp, stemp, u3p, trans, phy, ele, u4, hops, u2, u32, lane, iso, gen) \ argument 407 | ((hops) & 0x3) << 6 | (u2) << 5 | (u32) << 4 | (lane) << 3 \
|
/linux-6.12.1/include/linux/usb/ |
D | pd_vdo.h | 437 #define VDO_ACABLE2(mtemp, stemp, u3p, trans, phy, ele, u4, hops, u2, u32, lane, iso, gen) \ argument 440 | ((hops) & 0x3) << 6 | (u2) << 5 | (u32) << 4 | (lane) << 3 \
|
/linux-6.12.1/drivers/firmware/arm_scmi/vendors/imx/ |
D | imx-sm-misc.c | 208 int max_msg_size = ph->hops->get_max_msg_size(ph); in scmi_imx_misc_ctrl_get() 247 int max_msg_size = ph->hops->get_max_msg_size(ph); in scmi_imx_misc_ctrl_set()
|