/linux-6.12.1/fs/reiserfs/ |
D | fix_node.c | 54 struct virtual_node *vn = tb->tb_vn; in create_virtual_node() local 61 vn->vn_size = in create_virtual_node() 66 vn->vn_nr_item = (vn->vn_size - DC_SIZE) / (DC_SIZE + KEY_SIZE); in create_virtual_node() 71 vn->vn_nr_item = in create_virtual_node() 72 B_NR_ITEMS(Sh) + ((vn->vn_mode == M_INSERT) ? 1 : 0) - in create_virtual_node() 73 ((vn->vn_mode == M_DELETE) ? 1 : 0); in create_virtual_node() 76 vn->vn_vi = (struct virtual_item *)(tb->tb_vn + 1); in create_virtual_node() 77 memset(vn->vn_vi, 0, vn->vn_nr_item * sizeof(struct virtual_item)); in create_virtual_node() 78 vn->vn_free_ptr += vn->vn_nr_item * sizeof(struct virtual_item); in create_virtual_node() 85 && (vn->vn_mode != M_DELETE || vn->vn_affected_item_num)) in create_virtual_node() [all …]
|
D | item_ops.c | 60 static int sd_create_vi(struct virtual_node *vn, in sd_create_vi() argument 150 static int direct_create_vi(struct virtual_node *vn, in direct_create_vi() argument 288 static int indirect_create_vi(struct virtual_node *vn, in indirect_create_vi() argument 451 static int direntry_create_vi(struct virtual_node *vn, in direntry_create_vi() argument 472 ((is_affected) ? ((vn->vn_mode == M_CUT) ? -1 : in direntry_create_vi() 473 (vn->vn_mode == M_PASTE ? 1 : 0)) : 0); in direntry_create_vi() 476 j = old_entry_num(is_affected, i, vn->vn_pos_in_item, in direntry_create_vi() 477 vn->vn_mode); in direntry_create_vi() 486 if (is_affected && vn->vn_mode == M_PASTE) in direntry_create_vi() 487 dir_u->entry_sizes[vn->vn_pos_in_item] = insert_size; in direntry_create_vi() [all …]
|
/linux-6.12.1/net/8021q/ |
D | vlanproc.c | 94 struct vlan_net *vn = net_generic(net, vlan_net_id); in vlan_proc_cleanup() local 96 if (vn->proc_vlan_conf) in vlan_proc_cleanup() 97 remove_proc_entry(name_conf, vn->proc_vlan_dir); in vlan_proc_cleanup() 99 if (vn->proc_vlan_dir) in vlan_proc_cleanup() 113 struct vlan_net *vn = net_generic(net, vlan_net_id); in vlan_proc_init() local 115 vn->proc_vlan_dir = proc_net_mkdir(net, name_root, net->proc_net); in vlan_proc_init() 116 if (!vn->proc_vlan_dir) in vlan_proc_init() 119 vn->proc_vlan_conf = proc_create_net(name_conf, S_IFREG | 0600, in vlan_proc_init() 120 vn->proc_vlan_dir, &vlan_seq_ops, in vlan_proc_init() 122 if (!vn->proc_vlan_conf) in vlan_proc_init() [all …]
|
D | vlan.c | 219 struct vlan_net *vn = net_generic(net, vlan_net_id); in register_vlan_device() local 232 switch (vn->name_type) { in register_vlan_device() 607 struct vlan_net *vn; in vlan_ioctl_handler() local 609 vn = net_generic(net, vlan_net_id); in vlan_ioctl_handler() 610 vn->name_type = args.u.name_type; in vlan_ioctl_handler() 660 struct vlan_net *vn = net_generic(net, vlan_net_id); in vlan_init_net() local 663 vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD; in vlan_init_net()
|
/linux-6.12.1/mm/ |
D | vmalloc.c | 1058 struct vmap_node *vn; in find_vmap_area_exceed_addr_lock() local 1063 vn = &vmap_nodes[i]; in find_vmap_area_exceed_addr_lock() 1065 spin_lock(&vn->busy.lock); in find_vmap_area_exceed_addr_lock() 1066 *va = __find_vmap_area_exceed_addr(addr, &vn->busy.root); in find_vmap_area_exceed_addr_lock() 1071 spin_unlock(&vn->busy.lock); in find_vmap_area_exceed_addr_lock() 1080 vn = addr_to_node(va_start_lowest); in find_vmap_area_exceed_addr_lock() 1082 spin_lock(&vn->busy.lock); in find_vmap_area_exceed_addr_lock() 1083 *va = __find_vmap_area(va_start_lowest, &vn->busy.root); in find_vmap_area_exceed_addr_lock() 1086 return vn; in find_vmap_area_exceed_addr_lock() 1088 spin_unlock(&vn->busy.lock); in find_vmap_area_exceed_addr_lock() [all …]
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
D | memnv04.c | 35 struct nv04_mem_map_vn vn; in nv04_mem_map() member 41 if ((ret = nvif_unvers(ret, &argv, &argc, args->vn))) in nv04_mem_map() 55 struct nv04_mem_vn vn; in nv04_mem_new() member 59 if ((ret = nvif_unvers(ret, &argv, &argc, args->vn))) in nv04_mem_new()
|
D | memgf100.c | 39 struct gf100_mem_map_vn vn; in gf100_mem_map() member 50 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in gf100_mem_map() 73 struct gf100_mem_vn vn; in gf100_mem_new() member 82 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in gf100_mem_new()
|
D | memnv50.c | 39 struct nv50_mem_map_vn vn; in nv50_mem_map() member 52 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in nv50_mem_map() 70 struct nv50_mem_vn vn; in nv50_mem_new() member 80 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in nv50_mem_new()
|
D | vmmnv04.c | 84 struct nv04_vmm_map_vn vn; in nv04_vmm_valid() member 87 if ((ret = nvif_unvers(ret, &argv, &argc, args->vn))) in nv04_vmm_valid() 108 struct nv04_vmm_vn vn; in nv04_vmm_new_() member 117 return nvif_unvers(-ENOSYS, &argv, &argc, args->vn); in nv04_vmm_new_()
|
D | vmmgp100.c | 411 struct gp100_vmm_map_vn vn; in gp100_vmm_valid() member 429 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in gp100_vmm_valid() 524 struct gp100_vmm_fault_replay_vn vn; in gp100_vmm_fault_replay() member 528 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in gp100_vmm_fault_replay() 606 struct gp100_vmm_vn vn; in gp100_vmm_new_() member 615 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in gp100_vmm_new_()
|
D | vmmgm200.c | 150 struct gm200_vmm_vn vn; in gm200_vmm_new_() member 163 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in gm200_vmm_new_()
|
/linux-6.12.1/drivers/net/vxlan/ |
D | vxlan_multicast.c | 125 bool vxlan_group_used(struct vxlan_net *vn, struct vxlan_dev *dev, in vxlan_group_used() argument 151 list_for_each_entry(vxlan, &vn->vxlan_list, next) { in vxlan_group_used() 218 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_multicast_leave_vnigrp() local 225 !vxlan_group_used(vn, vxlan, v->vni, &v->remote_ip, in vxlan_multicast_leave_vnigrp() 257 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_multicast_leave() local 261 !vxlan_group_used(vn, vxlan, 0, NULL, 0)) { in vxlan_multicast_leave()
|
D | vxlan_vnifilter.c | 43 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_vs_add_del_vninode() local 47 spin_lock(&vn->sock_lock); in vxlan_vs_add_del_vninode() 71 spin_unlock(&vn->sock_lock); in vxlan_vs_add_del_vninode() 78 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_vs_add_vnigrp() local 86 spin_lock(&vn->sock_lock); in vxlan_vs_add_vnigrp() 97 spin_unlock(&vn->sock_lock); in vxlan_vs_add_vnigrp() 103 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_vs_del_vnigrp() local 109 spin_lock(&vn->sock_lock); in vxlan_vs_del_vnigrp() 116 spin_unlock(&vn->sock_lock); in vxlan_vs_del_vnigrp() 521 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_vni_update_group() local [all …]
|
D | vxlan_core.c | 1505 struct vxlan_net *vn; in __vxlan_sock_release_prep() local 1512 vn = net_generic(sock_net(vs->sock->sk), vxlan_net_id); in __vxlan_sock_release_prep() 1513 spin_lock(&vn->sock_lock); in __vxlan_sock_release_prep() 1519 spin_unlock(&vn->sock_lock); in __vxlan_sock_release_prep() 2846 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_vs_del_dev() local 2848 spin_lock(&vn->sock_lock); in vxlan_vs_del_dev() 2853 spin_unlock(&vn->sock_lock); in vxlan_vs_del_dev() 2859 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_vs_add_dev() local 2863 spin_lock(&vn->sock_lock); in vxlan_vs_add_dev() 2865 spin_unlock(&vn->sock_lock); in vxlan_vs_add_dev() [all …]
|
D | vxlan_private.h | 54 struct vxlan_net *vn = net_generic(net, vxlan_net_id); in vs_head() local 56 return &vn->sock_list[hash_32(ntohs(port), PORT_HASH_BITS)]; in vs_head() 224 bool vxlan_group_used(struct vxlan_net *vn, struct vxlan_dev *dev,
|
/linux-6.12.1/drivers/gpu/drm/nouveau/include/nvif/ |
D | if000e.h | 19 } vn; member 24 } vn; member
|
D | if0004.h | 7 } vn; member
|
D | clb069.h | 13 } vn; member
|
D | if0013.h | 15 } vn; member
|
/linux-6.12.1/scripts/gdb/linux/ |
D | vmalloc.py | 34 vn = gdb.parse_and_eval('&vmap_nodes[%d]' % i) 35 … for vmap_area in lists.list_for_each_entry(vn['busy']['head'], vmap_area_ptr_type, "list"):
|
/linux-6.12.1/drivers/scsi/fcoe/ |
D | fcoe_ctlr.c | 539 struct fip_vn_desc *vn; in fcoe_ctlr_send_keep_alive() local 549 len = sizeof(*kal) + ports * sizeof(*vn); in fcoe_ctlr_send_keep_alive() 564 ports * sizeof(*vn)) / FIP_BPW); in fcoe_ctlr_send_keep_alive() 573 vn = (struct fip_vn_desc *)(kal + 1); in fcoe_ctlr_send_keep_alive() 574 vn->fd_desc.fip_dtype = FIP_DT_VN_ID; in fcoe_ctlr_send_keep_alive() 575 vn->fd_desc.fip_dlen = sizeof(*vn) / FIP_BPW; in fcoe_ctlr_send_keep_alive() 576 memcpy(vn->fd_mac, fip->get_src_addr(lport), ETH_ALEN); in fcoe_ctlr_send_keep_alive() 577 hton24(vn->fd_fc_id, lport->port_id); in fcoe_ctlr_send_keep_alive() 578 put_unaligned_be64(lport->wwpn, &vn->fd_wwpn); in fcoe_ctlr_send_keep_alive() 2023 struct fip_vn_desc vn; in fcoe_ctlr_vn_send() member [all …]
|
/linux-6.12.1/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_cmn.h | 904 static inline int func_by_vn(struct bnx2x *bp, int vn) in func_by_vn() argument 906 return 2 * vn + BP_PORT(bp); in func_by_vn() 1327 int vn; in bnx2x_link_sync_notify() local 1330 for (vn = VN_0; vn < BP_MAX_VN_NUM(bp); vn++) { in bnx2x_link_sync_notify() 1331 if (vn == BP_VN(bp)) in bnx2x_link_sync_notify() 1334 func = func_by_vn(bp, vn); in bnx2x_link_sync_notify()
|
/linux-6.12.1/drivers/gpu/drm/sun4i/ |
D | sun8i_vi_layer.c | 62 u32 vn = 0, vm = 0; in sun8i_vi_layer_update_coord() local 139 vn = (u32)ability * dst_h / 100; in sun8i_vi_layer_update_coord() 140 src_h = vn; in sun8i_vi_layer_update_coord() 175 SUN8I_MIXER_CHAN_VI_DS_N(vn) | in sun8i_vi_layer_update_coord() 179 SUN8I_MIXER_CHAN_VI_DS_N(vn) | in sun8i_vi_layer_update_coord()
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/core/ |
D | uevent.c | 41 if (argc != sizeof(args->vn)) in nvkm_uevent_mthd_block() 52 if (argc != sizeof(args->vn)) in nvkm_uevent_mthd_allow()
|
/linux-6.12.1/drivers/gpu/drm/i915/ |
D | i915_gem_evict.c | 478 struct i915_vma *vma, *vn; in i915_gem_evict_vm() local 515 list_for_each_entry_safe(vma, vn, &locked_eviction_list, evict_link) { in i915_gem_evict_vm() 527 list_for_each_entry_safe(vma, vn, &eviction_list, evict_link) { in i915_gem_evict_vm()
|