Home
last modified time | relevance | path

Searched refs:vn (Results 1 – 25 of 52) sorted by relevance

123

/linux-6.12.1/fs/reiserfs/
Dfix_node.c54 struct virtual_node *vn = tb->tb_vn; in create_virtual_node() local
61 vn->vn_size = in create_virtual_node()
66 vn->vn_nr_item = (vn->vn_size - DC_SIZE) / (DC_SIZE + KEY_SIZE); in create_virtual_node()
71 vn->vn_nr_item = in create_virtual_node()
72 B_NR_ITEMS(Sh) + ((vn->vn_mode == M_INSERT) ? 1 : 0) - in create_virtual_node()
73 ((vn->vn_mode == M_DELETE) ? 1 : 0); in create_virtual_node()
76 vn->vn_vi = (struct virtual_item *)(tb->tb_vn + 1); in create_virtual_node()
77 memset(vn->vn_vi, 0, vn->vn_nr_item * sizeof(struct virtual_item)); in create_virtual_node()
78 vn->vn_free_ptr += vn->vn_nr_item * sizeof(struct virtual_item); in create_virtual_node()
85 && (vn->vn_mode != M_DELETE || vn->vn_affected_item_num)) in create_virtual_node()
[all …]
Ditem_ops.c60 static int sd_create_vi(struct virtual_node *vn, in sd_create_vi() argument
150 static int direct_create_vi(struct virtual_node *vn, in direct_create_vi() argument
288 static int indirect_create_vi(struct virtual_node *vn, in indirect_create_vi() argument
451 static int direntry_create_vi(struct virtual_node *vn, in direntry_create_vi() argument
472 ((is_affected) ? ((vn->vn_mode == M_CUT) ? -1 : in direntry_create_vi()
473 (vn->vn_mode == M_PASTE ? 1 : 0)) : 0); in direntry_create_vi()
476 j = old_entry_num(is_affected, i, vn->vn_pos_in_item, in direntry_create_vi()
477 vn->vn_mode); in direntry_create_vi()
486 if (is_affected && vn->vn_mode == M_PASTE) in direntry_create_vi()
487 dir_u->entry_sizes[vn->vn_pos_in_item] = insert_size; in direntry_create_vi()
[all …]
/linux-6.12.1/net/8021q/
Dvlanproc.c94 struct vlan_net *vn = net_generic(net, vlan_net_id); in vlan_proc_cleanup() local
96 if (vn->proc_vlan_conf) in vlan_proc_cleanup()
97 remove_proc_entry(name_conf, vn->proc_vlan_dir); in vlan_proc_cleanup()
99 if (vn->proc_vlan_dir) in vlan_proc_cleanup()
113 struct vlan_net *vn = net_generic(net, vlan_net_id); in vlan_proc_init() local
115 vn->proc_vlan_dir = proc_net_mkdir(net, name_root, net->proc_net); in vlan_proc_init()
116 if (!vn->proc_vlan_dir) in vlan_proc_init()
119 vn->proc_vlan_conf = proc_create_net(name_conf, S_IFREG | 0600, in vlan_proc_init()
120 vn->proc_vlan_dir, &vlan_seq_ops, in vlan_proc_init()
122 if (!vn->proc_vlan_conf) in vlan_proc_init()
[all …]
Dvlan.c219 struct vlan_net *vn = net_generic(net, vlan_net_id); in register_vlan_device() local
232 switch (vn->name_type) { in register_vlan_device()
607 struct vlan_net *vn; in vlan_ioctl_handler() local
609 vn = net_generic(net, vlan_net_id); in vlan_ioctl_handler()
610 vn->name_type = args.u.name_type; in vlan_ioctl_handler()
660 struct vlan_net *vn = net_generic(net, vlan_net_id); in vlan_init_net() local
663 vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD; in vlan_init_net()
/linux-6.12.1/mm/
Dvmalloc.c1058 struct vmap_node *vn; in find_vmap_area_exceed_addr_lock() local
1063 vn = &vmap_nodes[i]; in find_vmap_area_exceed_addr_lock()
1065 spin_lock(&vn->busy.lock); in find_vmap_area_exceed_addr_lock()
1066 *va = __find_vmap_area_exceed_addr(addr, &vn->busy.root); in find_vmap_area_exceed_addr_lock()
1071 spin_unlock(&vn->busy.lock); in find_vmap_area_exceed_addr_lock()
1080 vn = addr_to_node(va_start_lowest); in find_vmap_area_exceed_addr_lock()
1082 spin_lock(&vn->busy.lock); in find_vmap_area_exceed_addr_lock()
1083 *va = __find_vmap_area(va_start_lowest, &vn->busy.root); in find_vmap_area_exceed_addr_lock()
1086 return vn; in find_vmap_area_exceed_addr_lock()
1088 spin_unlock(&vn->busy.lock); in find_vmap_area_exceed_addr_lock()
[all …]
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
Dmemnv04.c35 struct nv04_mem_map_vn vn; in nv04_mem_map() member
41 if ((ret = nvif_unvers(ret, &argv, &argc, args->vn))) in nv04_mem_map()
55 struct nv04_mem_vn vn; in nv04_mem_new() member
59 if ((ret = nvif_unvers(ret, &argv, &argc, args->vn))) in nv04_mem_new()
Dmemgf100.c39 struct gf100_mem_map_vn vn; in gf100_mem_map() member
50 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in gf100_mem_map()
73 struct gf100_mem_vn vn; in gf100_mem_new() member
82 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in gf100_mem_new()
Dmemnv50.c39 struct nv50_mem_map_vn vn; in nv50_mem_map() member
52 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in nv50_mem_map()
70 struct nv50_mem_vn vn; in nv50_mem_new() member
80 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in nv50_mem_new()
Dvmmnv04.c84 struct nv04_vmm_map_vn vn; in nv04_vmm_valid() member
87 if ((ret = nvif_unvers(ret, &argv, &argc, args->vn))) in nv04_vmm_valid()
108 struct nv04_vmm_vn vn; in nv04_vmm_new_() member
117 return nvif_unvers(-ENOSYS, &argv, &argc, args->vn); in nv04_vmm_new_()
Dvmmgp100.c411 struct gp100_vmm_map_vn vn; in gp100_vmm_valid() member
429 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in gp100_vmm_valid()
524 struct gp100_vmm_fault_replay_vn vn; in gp100_vmm_fault_replay() member
528 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in gp100_vmm_fault_replay()
606 struct gp100_vmm_vn vn; in gp100_vmm_new_() member
615 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in gp100_vmm_new_()
Dvmmgm200.c150 struct gm200_vmm_vn vn; in gm200_vmm_new_() member
163 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) { in gm200_vmm_new_()
/linux-6.12.1/drivers/net/vxlan/
Dvxlan_multicast.c125 bool vxlan_group_used(struct vxlan_net *vn, struct vxlan_dev *dev, in vxlan_group_used() argument
151 list_for_each_entry(vxlan, &vn->vxlan_list, next) { in vxlan_group_used()
218 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_multicast_leave_vnigrp() local
225 !vxlan_group_used(vn, vxlan, v->vni, &v->remote_ip, in vxlan_multicast_leave_vnigrp()
257 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_multicast_leave() local
261 !vxlan_group_used(vn, vxlan, 0, NULL, 0)) { in vxlan_multicast_leave()
Dvxlan_vnifilter.c43 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_vs_add_del_vninode() local
47 spin_lock(&vn->sock_lock); in vxlan_vs_add_del_vninode()
71 spin_unlock(&vn->sock_lock); in vxlan_vs_add_del_vninode()
78 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_vs_add_vnigrp() local
86 spin_lock(&vn->sock_lock); in vxlan_vs_add_vnigrp()
97 spin_unlock(&vn->sock_lock); in vxlan_vs_add_vnigrp()
103 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_vs_del_vnigrp() local
109 spin_lock(&vn->sock_lock); in vxlan_vs_del_vnigrp()
116 spin_unlock(&vn->sock_lock); in vxlan_vs_del_vnigrp()
521 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_vni_update_group() local
[all …]
Dvxlan_core.c1505 struct vxlan_net *vn; in __vxlan_sock_release_prep() local
1512 vn = net_generic(sock_net(vs->sock->sk), vxlan_net_id); in __vxlan_sock_release_prep()
1513 spin_lock(&vn->sock_lock); in __vxlan_sock_release_prep()
1519 spin_unlock(&vn->sock_lock); in __vxlan_sock_release_prep()
2846 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_vs_del_dev() local
2848 spin_lock(&vn->sock_lock); in vxlan_vs_del_dev()
2853 spin_unlock(&vn->sock_lock); in vxlan_vs_del_dev()
2859 struct vxlan_net *vn = net_generic(vxlan->net, vxlan_net_id); in vxlan_vs_add_dev() local
2863 spin_lock(&vn->sock_lock); in vxlan_vs_add_dev()
2865 spin_unlock(&vn->sock_lock); in vxlan_vs_add_dev()
[all …]
Dvxlan_private.h54 struct vxlan_net *vn = net_generic(net, vxlan_net_id); in vs_head() local
56 return &vn->sock_list[hash_32(ntohs(port), PORT_HASH_BITS)]; in vs_head()
224 bool vxlan_group_used(struct vxlan_net *vn, struct vxlan_dev *dev,
/linux-6.12.1/drivers/gpu/drm/nouveau/include/nvif/
Dif000e.h19 } vn; member
24 } vn; member
Dif0004.h7 } vn; member
Dclb069.h13 } vn; member
Dif0013.h15 } vn; member
/linux-6.12.1/scripts/gdb/linux/
Dvmalloc.py34 vn = gdb.parse_and_eval('&vmap_nodes[%d]' % i)
35 … for vmap_area in lists.list_for_each_entry(vn['busy']['head'], vmap_area_ptr_type, "list"):
/linux-6.12.1/drivers/scsi/fcoe/
Dfcoe_ctlr.c539 struct fip_vn_desc *vn; in fcoe_ctlr_send_keep_alive() local
549 len = sizeof(*kal) + ports * sizeof(*vn); in fcoe_ctlr_send_keep_alive()
564 ports * sizeof(*vn)) / FIP_BPW); in fcoe_ctlr_send_keep_alive()
573 vn = (struct fip_vn_desc *)(kal + 1); in fcoe_ctlr_send_keep_alive()
574 vn->fd_desc.fip_dtype = FIP_DT_VN_ID; in fcoe_ctlr_send_keep_alive()
575 vn->fd_desc.fip_dlen = sizeof(*vn) / FIP_BPW; in fcoe_ctlr_send_keep_alive()
576 memcpy(vn->fd_mac, fip->get_src_addr(lport), ETH_ALEN); in fcoe_ctlr_send_keep_alive()
577 hton24(vn->fd_fc_id, lport->port_id); in fcoe_ctlr_send_keep_alive()
578 put_unaligned_be64(lport->wwpn, &vn->fd_wwpn); in fcoe_ctlr_send_keep_alive()
2023 struct fip_vn_desc vn; in fcoe_ctlr_vn_send() member
[all …]
/linux-6.12.1/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_cmn.h904 static inline int func_by_vn(struct bnx2x *bp, int vn) in func_by_vn() argument
906 return 2 * vn + BP_PORT(bp); in func_by_vn()
1327 int vn; in bnx2x_link_sync_notify() local
1330 for (vn = VN_0; vn < BP_MAX_VN_NUM(bp); vn++) { in bnx2x_link_sync_notify()
1331 if (vn == BP_VN(bp)) in bnx2x_link_sync_notify()
1334 func = func_by_vn(bp, vn); in bnx2x_link_sync_notify()
/linux-6.12.1/drivers/gpu/drm/sun4i/
Dsun8i_vi_layer.c62 u32 vn = 0, vm = 0; in sun8i_vi_layer_update_coord() local
139 vn = (u32)ability * dst_h / 100; in sun8i_vi_layer_update_coord()
140 src_h = vn; in sun8i_vi_layer_update_coord()
175 SUN8I_MIXER_CHAN_VI_DS_N(vn) | in sun8i_vi_layer_update_coord()
179 SUN8I_MIXER_CHAN_VI_DS_N(vn) | in sun8i_vi_layer_update_coord()
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/core/
Duevent.c41 if (argc != sizeof(args->vn)) in nvkm_uevent_mthd_block()
52 if (argc != sizeof(args->vn)) in nvkm_uevent_mthd_allow()
/linux-6.12.1/drivers/gpu/drm/i915/
Di915_gem_evict.c478 struct i915_vma *vma, *vn; in i915_gem_evict_vm() local
515 list_for_each_entry_safe(vma, vn, &locked_eviction_list, evict_link) { in i915_gem_evict_vm()
527 list_for_each_entry_safe(vma, vn, &eviction_list, evict_link) { in i915_gem_evict_vm()

123