Lines Matching refs:rnet

927 			      struct rdma_dev_net *rnet)  in add_one_compat_dev()  argument
940 if (net_eq(read_pnet(&rnet->net), in add_one_compat_dev()
950 cdev = xa_load(&device->compat_devs, rnet->id); in add_one_compat_dev()
955 ret = xa_reserve(&device->compat_devs, rnet->id, GFP_KERNEL); in add_one_compat_dev()
966 rdma_init_coredev(cdev, device, read_pnet(&rnet->net)); in add_one_compat_dev()
979 ret = xa_err(xa_store(&device->compat_devs, rnet->id, in add_one_compat_dev()
994 xa_release(&device->compat_devs, rnet->id); in add_one_compat_dev()
1025 struct rdma_dev_net *rnet; in add_compat_devs() local
1032 xa_for_each (&rdma_nets, index, rnet) { in add_compat_devs()
1033 ret = add_one_compat_dev(device, rnet); in add_compat_devs()
1064 struct rdma_dev_net *rnet; in add_all_compat_devs() local
1077 xa_for_each (&rdma_nets, net_index, rnet) { in add_all_compat_devs()
1078 ret = add_one_compat_dev(dev, rnet); in add_all_compat_devs()
1092 struct rdma_dev_net *rnet; in rdma_compatdev_set() local
1105 xa_for_each (&rdma_nets, index, rnet) { in rdma_compatdev_set()
1124 struct rdma_dev_net *rnet = rdma_net_to_dev_net(net); in rdma_dev_exit_net() local
1133 ret = xa_err(xa_store(&rdma_nets, rnet->id, NULL, GFP_KERNEL)); in rdma_dev_exit_net()
1146 remove_one_compat_dev(dev, rnet->id); in rdma_dev_exit_net()
1158 rdma_nl_net_exit(rnet); in rdma_dev_exit_net()
1159 xa_erase(&rdma_nets, rnet->id); in rdma_dev_exit_net()
1164 struct rdma_dev_net *rnet = rdma_net_to_dev_net(net); in rdma_dev_init_net() local
1169 write_pnet(&rnet->net, net); in rdma_dev_init_net()
1171 ret = rdma_nl_net_init(rnet); in rdma_dev_init_net()
1179 ret = xa_alloc(&rdma_nets, &rnet->id, rnet, xa_limit_32b, GFP_KERNEL); in rdma_dev_init_net()
1181 rdma_nl_net_exit(rnet); in rdma_dev_init_net()
1191 ret = add_one_compat_dev(dev, rnet); in rdma_dev_init_net()