1  // SPDX-License-Identifier: GPL-2.0-or-later
2  /*
3   * Copyright (c) 2016 Mellanox Technologies. All rights reserved.
4   * Copyright (c) 2016 Jiri Pirko <jiri@mellanox.com>
5   */
6  
7  #include <net/genetlink.h>
8  #define CREATE_TRACE_POINTS
9  #include <trace/events/devlink.h>
10  
11  #include "devl_internal.h"
12  
13  EXPORT_TRACEPOINT_SYMBOL_GPL(devlink_hwmsg);
14  EXPORT_TRACEPOINT_SYMBOL_GPL(devlink_hwerr);
15  EXPORT_TRACEPOINT_SYMBOL_GPL(devlink_trap_report);
16  
17  DEFINE_XARRAY_FLAGS(devlinks, XA_FLAGS_ALLOC);
18  
devlinks_xa_get(unsigned long index)19  static struct devlink *devlinks_xa_get(unsigned long index)
20  {
21  	struct devlink *devlink;
22  
23  	rcu_read_lock();
24  	devlink = xa_find(&devlinks, &index, index, DEVLINK_REGISTERED);
25  	if (!devlink || !devlink_try_get(devlink))
26  		devlink = NULL;
27  	rcu_read_unlock();
28  	return devlink;
29  }
30  
31  /* devlink_rels xarray contains 1:1 relationships between
32   * devlink object and related nested devlink instance.
33   * The xarray index is used to get the nested object from
34   * the nested-in object code.
35   */
36  static DEFINE_XARRAY_FLAGS(devlink_rels, XA_FLAGS_ALLOC1);
37  
38  #define DEVLINK_REL_IN_USE XA_MARK_0
39  
40  struct devlink_rel {
41  	u32 index;
42  	refcount_t refcount;
43  	u32 devlink_index;
44  	struct {
45  		u32 devlink_index;
46  		u32 obj_index;
47  		devlink_rel_notify_cb_t *notify_cb;
48  		devlink_rel_cleanup_cb_t *cleanup_cb;
49  		struct delayed_work notify_work;
50  	} nested_in;
51  };
52  
devlink_rel_free(struct devlink_rel * rel)53  static void devlink_rel_free(struct devlink_rel *rel)
54  {
55  	xa_erase(&devlink_rels, rel->index);
56  	kfree(rel);
57  }
58  
__devlink_rel_get(struct devlink_rel * rel)59  static void __devlink_rel_get(struct devlink_rel *rel)
60  {
61  	refcount_inc(&rel->refcount);
62  }
63  
__devlink_rel_put(struct devlink_rel * rel)64  static void __devlink_rel_put(struct devlink_rel *rel)
65  {
66  	if (refcount_dec_and_test(&rel->refcount))
67  		devlink_rel_free(rel);
68  }
69  
devlink_rel_nested_in_notify_work(struct work_struct * work)70  static void devlink_rel_nested_in_notify_work(struct work_struct *work)
71  {
72  	struct devlink_rel *rel = container_of(work, struct devlink_rel,
73  					       nested_in.notify_work.work);
74  	struct devlink *devlink;
75  
76  	devlink = devlinks_xa_get(rel->nested_in.devlink_index);
77  	if (!devlink)
78  		goto rel_put;
79  	if (!devl_trylock(devlink)) {
80  		devlink_put(devlink);
81  		goto reschedule_work;
82  	}
83  	if (!devl_is_registered(devlink)) {
84  		devl_unlock(devlink);
85  		devlink_put(devlink);
86  		goto rel_put;
87  	}
88  	if (!xa_get_mark(&devlink_rels, rel->index, DEVLINK_REL_IN_USE))
89  		rel->nested_in.cleanup_cb(devlink, rel->nested_in.obj_index, rel->index);
90  	rel->nested_in.notify_cb(devlink, rel->nested_in.obj_index);
91  	devl_unlock(devlink);
92  	devlink_put(devlink);
93  
94  rel_put:
95  	__devlink_rel_put(rel);
96  	return;
97  
98  reschedule_work:
99  	schedule_delayed_work(&rel->nested_in.notify_work, 1);
100  }
101  
devlink_rel_nested_in_notify_work_schedule(struct devlink_rel * rel)102  static void devlink_rel_nested_in_notify_work_schedule(struct devlink_rel *rel)
103  {
104  	__devlink_rel_get(rel);
105  	schedule_delayed_work(&rel->nested_in.notify_work, 0);
106  }
107  
devlink_rel_alloc(void)108  static struct devlink_rel *devlink_rel_alloc(void)
109  {
110  	struct devlink_rel *rel;
111  	static u32 next;
112  	int err;
113  
114  	rel = kzalloc(sizeof(*rel), GFP_KERNEL);
115  	if (!rel)
116  		return ERR_PTR(-ENOMEM);
117  
118  	err = xa_alloc_cyclic(&devlink_rels, &rel->index, rel,
119  			      xa_limit_32b, &next, GFP_KERNEL);
120  	if (err) {
121  		kfree(rel);
122  		return ERR_PTR(err);
123  	}
124  
125  	refcount_set(&rel->refcount, 1);
126  	INIT_DELAYED_WORK(&rel->nested_in.notify_work,
127  			  &devlink_rel_nested_in_notify_work);
128  	return rel;
129  }
130  
devlink_rel_put(struct devlink * devlink)131  static void devlink_rel_put(struct devlink *devlink)
132  {
133  	struct devlink_rel *rel = devlink->rel;
134  
135  	if (!rel)
136  		return;
137  	xa_clear_mark(&devlink_rels, rel->index, DEVLINK_REL_IN_USE);
138  	devlink_rel_nested_in_notify_work_schedule(rel);
139  	__devlink_rel_put(rel);
140  	devlink->rel = NULL;
141  }
142  
devlink_rel_nested_in_clear(u32 rel_index)143  void devlink_rel_nested_in_clear(u32 rel_index)
144  {
145  	xa_clear_mark(&devlink_rels, rel_index, DEVLINK_REL_IN_USE);
146  }
147  
devlink_rel_nested_in_add(u32 * rel_index,u32 devlink_index,u32 obj_index,devlink_rel_notify_cb_t * notify_cb,devlink_rel_cleanup_cb_t * cleanup_cb,struct devlink * devlink)148  int devlink_rel_nested_in_add(u32 *rel_index, u32 devlink_index,
149  			      u32 obj_index, devlink_rel_notify_cb_t *notify_cb,
150  			      devlink_rel_cleanup_cb_t *cleanup_cb,
151  			      struct devlink *devlink)
152  {
153  	struct devlink_rel *rel = devlink_rel_alloc();
154  
155  	ASSERT_DEVLINK_NOT_REGISTERED(devlink);
156  
157  	if (IS_ERR(rel))
158  		return PTR_ERR(rel);
159  
160  	rel->devlink_index = devlink->index;
161  	rel->nested_in.devlink_index = devlink_index;
162  	rel->nested_in.obj_index = obj_index;
163  	rel->nested_in.notify_cb = notify_cb;
164  	rel->nested_in.cleanup_cb = cleanup_cb;
165  	*rel_index = rel->index;
166  	xa_set_mark(&devlink_rels, rel->index, DEVLINK_REL_IN_USE);
167  	devlink->rel = rel;
168  	return 0;
169  }
170  
171  /**
172   * devlink_rel_nested_in_notify - Notify the object this devlink
173   *				  instance is nested in.
174   * @devlink: devlink
175   *
176   * This is called upon network namespace change of devlink instance.
177   * In case this devlink instance is nested in another devlink object,
178   * a notification of a change of this object should be sent
179   * over netlink. The parent devlink instance lock needs to be
180   * taken during the notification preparation.
181   * However, since the devlink lock of nested instance is held here,
182   * we would end with wrong devlink instance lock ordering and
183   * deadlock. Therefore the work is utilized to avoid that.
184   */
devlink_rel_nested_in_notify(struct devlink * devlink)185  void devlink_rel_nested_in_notify(struct devlink *devlink)
186  {
187  	struct devlink_rel *rel = devlink->rel;
188  
189  	if (!rel)
190  		return;
191  	devlink_rel_nested_in_notify_work_schedule(rel);
192  }
193  
devlink_rel_find(unsigned long rel_index)194  static struct devlink_rel *devlink_rel_find(unsigned long rel_index)
195  {
196  	return xa_find(&devlink_rels, &rel_index, rel_index,
197  		       DEVLINK_REL_IN_USE);
198  }
199  
devlink_rel_devlink_get(u32 rel_index)200  static struct devlink *devlink_rel_devlink_get(u32 rel_index)
201  {
202  	struct devlink_rel *rel;
203  	u32 devlink_index;
204  
205  	if (!rel_index)
206  		return NULL;
207  	xa_lock(&devlink_rels);
208  	rel = devlink_rel_find(rel_index);
209  	if (rel)
210  		devlink_index = rel->devlink_index;
211  	xa_unlock(&devlink_rels);
212  	if (!rel)
213  		return NULL;
214  	return devlinks_xa_get(devlink_index);
215  }
216  
devlink_rel_devlink_handle_put(struct sk_buff * msg,struct devlink * devlink,u32 rel_index,int attrtype,bool * msg_updated)217  int devlink_rel_devlink_handle_put(struct sk_buff *msg, struct devlink *devlink,
218  				   u32 rel_index, int attrtype,
219  				   bool *msg_updated)
220  {
221  	struct net *net = devlink_net(devlink);
222  	struct devlink *rel_devlink;
223  	int err;
224  
225  	rel_devlink = devlink_rel_devlink_get(rel_index);
226  	if (!rel_devlink)
227  		return 0;
228  	err = devlink_nl_put_nested_handle(msg, net, rel_devlink, attrtype);
229  	devlink_put(rel_devlink);
230  	if (!err && msg_updated)
231  		*msg_updated = true;
232  	return err;
233  }
234  
devlink_priv(struct devlink * devlink)235  void *devlink_priv(struct devlink *devlink)
236  {
237  	return &devlink->priv;
238  }
239  EXPORT_SYMBOL_GPL(devlink_priv);
240  
priv_to_devlink(void * priv)241  struct devlink *priv_to_devlink(void *priv)
242  {
243  	return container_of(priv, struct devlink, priv);
244  }
245  EXPORT_SYMBOL_GPL(priv_to_devlink);
246  
devlink_to_dev(const struct devlink * devlink)247  struct device *devlink_to_dev(const struct devlink *devlink)
248  {
249  	return devlink->dev;
250  }
251  EXPORT_SYMBOL_GPL(devlink_to_dev);
252  
devlink_net(const struct devlink * devlink)253  struct net *devlink_net(const struct devlink *devlink)
254  {
255  	return read_pnet(&devlink->_net);
256  }
257  EXPORT_SYMBOL_GPL(devlink_net);
258  
devl_assert_locked(struct devlink * devlink)259  void devl_assert_locked(struct devlink *devlink)
260  {
261  	lockdep_assert_held(&devlink->lock);
262  }
263  EXPORT_SYMBOL_GPL(devl_assert_locked);
264  
265  #ifdef CONFIG_LOCKDEP
266  /* For use in conjunction with LOCKDEP only e.g. rcu_dereference_protected() */
devl_lock_is_held(struct devlink * devlink)267  bool devl_lock_is_held(struct devlink *devlink)
268  {
269  	return lockdep_is_held(&devlink->lock);
270  }
271  EXPORT_SYMBOL_GPL(devl_lock_is_held);
272  #endif
273  
devl_lock(struct devlink * devlink)274  void devl_lock(struct devlink *devlink)
275  {
276  	mutex_lock(&devlink->lock);
277  }
278  EXPORT_SYMBOL_GPL(devl_lock);
279  
devl_trylock(struct devlink * devlink)280  int devl_trylock(struct devlink *devlink)
281  {
282  	return mutex_trylock(&devlink->lock);
283  }
284  EXPORT_SYMBOL_GPL(devl_trylock);
285  
devl_unlock(struct devlink * devlink)286  void devl_unlock(struct devlink *devlink)
287  {
288  	mutex_unlock(&devlink->lock);
289  }
290  EXPORT_SYMBOL_GPL(devl_unlock);
291  
292  /**
293   * devlink_try_get() - try to obtain a reference on a devlink instance
294   * @devlink: instance to reference
295   *
296   * Obtain a reference on a devlink instance. A reference on a devlink instance
297   * only implies that it's safe to take the instance lock. It does not imply
298   * that the instance is registered, use devl_is_registered() after taking
299   * the instance lock to check registration status.
300   */
devlink_try_get(struct devlink * devlink)301  struct devlink *__must_check devlink_try_get(struct devlink *devlink)
302  {
303  	if (refcount_inc_not_zero(&devlink->refcount))
304  		return devlink;
305  	return NULL;
306  }
307  
devlink_release(struct work_struct * work)308  static void devlink_release(struct work_struct *work)
309  {
310  	struct devlink *devlink;
311  
312  	devlink = container_of(to_rcu_work(work), struct devlink, rwork);
313  
314  	mutex_destroy(&devlink->lock);
315  	lockdep_unregister_key(&devlink->lock_key);
316  	put_device(devlink->dev);
317  	kvfree(devlink);
318  }
319  
devlink_put(struct devlink * devlink)320  void devlink_put(struct devlink *devlink)
321  {
322  	if (refcount_dec_and_test(&devlink->refcount))
323  		queue_rcu_work(system_wq, &devlink->rwork);
324  }
325  
devlinks_xa_find_get(struct net * net,unsigned long * indexp)326  struct devlink *devlinks_xa_find_get(struct net *net, unsigned long *indexp)
327  {
328  	struct devlink *devlink = NULL;
329  
330  	rcu_read_lock();
331  retry:
332  	devlink = xa_find(&devlinks, indexp, ULONG_MAX, DEVLINK_REGISTERED);
333  	if (!devlink)
334  		goto unlock;
335  
336  	if (!devlink_try_get(devlink))
337  		goto next;
338  	if (!net_eq(devlink_net(devlink), net)) {
339  		devlink_put(devlink);
340  		goto next;
341  	}
342  unlock:
343  	rcu_read_unlock();
344  	return devlink;
345  
346  next:
347  	(*indexp)++;
348  	goto retry;
349  }
350  
351  /**
352   * devl_register - Register devlink instance
353   * @devlink: devlink
354   */
devl_register(struct devlink * devlink)355  int devl_register(struct devlink *devlink)
356  {
357  	ASSERT_DEVLINK_NOT_REGISTERED(devlink);
358  	devl_assert_locked(devlink);
359  
360  	xa_set_mark(&devlinks, devlink->index, DEVLINK_REGISTERED);
361  	devlink_notify_register(devlink);
362  	devlink_rel_nested_in_notify(devlink);
363  
364  	return 0;
365  }
366  EXPORT_SYMBOL_GPL(devl_register);
367  
devlink_register(struct devlink * devlink)368  void devlink_register(struct devlink *devlink)
369  {
370  	devl_lock(devlink);
371  	devl_register(devlink);
372  	devl_unlock(devlink);
373  }
374  EXPORT_SYMBOL_GPL(devlink_register);
375  
376  /**
377   * devl_unregister - Unregister devlink instance
378   * @devlink: devlink
379   */
devl_unregister(struct devlink * devlink)380  void devl_unregister(struct devlink *devlink)
381  {
382  	ASSERT_DEVLINK_REGISTERED(devlink);
383  	devl_assert_locked(devlink);
384  
385  	devlink_notify_unregister(devlink);
386  	xa_clear_mark(&devlinks, devlink->index, DEVLINK_REGISTERED);
387  	devlink_rel_put(devlink);
388  }
389  EXPORT_SYMBOL_GPL(devl_unregister);
390  
devlink_unregister(struct devlink * devlink)391  void devlink_unregister(struct devlink *devlink)
392  {
393  	devl_lock(devlink);
394  	devl_unregister(devlink);
395  	devl_unlock(devlink);
396  }
397  EXPORT_SYMBOL_GPL(devlink_unregister);
398  
399  /**
400   *	devlink_alloc_ns - Allocate new devlink instance resources
401   *	in specific namespace
402   *
403   *	@ops: ops
404   *	@priv_size: size of user private data
405   *	@net: net namespace
406   *	@dev: parent device
407   *
408   *	Allocate new devlink instance resources, including devlink index
409   *	and name.
410   */
devlink_alloc_ns(const struct devlink_ops * ops,size_t priv_size,struct net * net,struct device * dev)411  struct devlink *devlink_alloc_ns(const struct devlink_ops *ops,
412  				 size_t priv_size, struct net *net,
413  				 struct device *dev)
414  {
415  	struct devlink *devlink;
416  	static u32 last_id;
417  	int ret;
418  
419  	WARN_ON(!ops || !dev);
420  	if (!devlink_reload_actions_valid(ops))
421  		return NULL;
422  
423  	devlink = kvzalloc(struct_size(devlink, priv, priv_size), GFP_KERNEL);
424  	if (!devlink)
425  		return NULL;
426  
427  	ret = xa_alloc_cyclic(&devlinks, &devlink->index, devlink, xa_limit_31b,
428  			      &last_id, GFP_KERNEL);
429  	if (ret < 0)
430  		goto err_xa_alloc;
431  
432  	devlink->dev = get_device(dev);
433  	devlink->ops = ops;
434  	xa_init_flags(&devlink->ports, XA_FLAGS_ALLOC);
435  	xa_init_flags(&devlink->params, XA_FLAGS_ALLOC);
436  	xa_init_flags(&devlink->snapshot_ids, XA_FLAGS_ALLOC);
437  	xa_init_flags(&devlink->nested_rels, XA_FLAGS_ALLOC);
438  	write_pnet(&devlink->_net, net);
439  	INIT_LIST_HEAD(&devlink->rate_list);
440  	INIT_LIST_HEAD(&devlink->linecard_list);
441  	INIT_LIST_HEAD(&devlink->sb_list);
442  	INIT_LIST_HEAD_RCU(&devlink->dpipe_table_list);
443  	INIT_LIST_HEAD(&devlink->resource_list);
444  	INIT_LIST_HEAD(&devlink->region_list);
445  	INIT_LIST_HEAD(&devlink->reporter_list);
446  	INIT_LIST_HEAD(&devlink->trap_list);
447  	INIT_LIST_HEAD(&devlink->trap_group_list);
448  	INIT_LIST_HEAD(&devlink->trap_policer_list);
449  	INIT_RCU_WORK(&devlink->rwork, devlink_release);
450  	lockdep_register_key(&devlink->lock_key);
451  	mutex_init(&devlink->lock);
452  	lockdep_set_class(&devlink->lock, &devlink->lock_key);
453  	refcount_set(&devlink->refcount, 1);
454  
455  	return devlink;
456  
457  err_xa_alloc:
458  	kvfree(devlink);
459  	return NULL;
460  }
461  EXPORT_SYMBOL_GPL(devlink_alloc_ns);
462  
463  /**
464   *	devlink_free - Free devlink instance resources
465   *
466   *	@devlink: devlink
467   */
devlink_free(struct devlink * devlink)468  void devlink_free(struct devlink *devlink)
469  {
470  	ASSERT_DEVLINK_NOT_REGISTERED(devlink);
471  
472  	WARN_ON(!list_empty(&devlink->trap_policer_list));
473  	WARN_ON(!list_empty(&devlink->trap_group_list));
474  	WARN_ON(!list_empty(&devlink->trap_list));
475  	WARN_ON(!list_empty(&devlink->reporter_list));
476  	WARN_ON(!list_empty(&devlink->region_list));
477  	WARN_ON(!list_empty(&devlink->resource_list));
478  	WARN_ON(!list_empty(&devlink->dpipe_table_list));
479  	WARN_ON(!list_empty(&devlink->sb_list));
480  	WARN_ON(!list_empty(&devlink->rate_list));
481  	WARN_ON(!list_empty(&devlink->linecard_list));
482  	WARN_ON(!xa_empty(&devlink->ports));
483  
484  	xa_destroy(&devlink->nested_rels);
485  	xa_destroy(&devlink->snapshot_ids);
486  	xa_destroy(&devlink->params);
487  	xa_destroy(&devlink->ports);
488  
489  	xa_erase(&devlinks, devlink->index);
490  
491  	devlink_put(devlink);
492  }
493  EXPORT_SYMBOL_GPL(devlink_free);
494  
devlink_pernet_pre_exit(struct net * net)495  static void __net_exit devlink_pernet_pre_exit(struct net *net)
496  {
497  	struct devlink *devlink;
498  	u32 actions_performed;
499  	unsigned long index;
500  	int err;
501  
502  	/* In case network namespace is getting destroyed, reload
503  	 * all devlink instances from this namespace into init_net.
504  	 */
505  	devlinks_xa_for_each_registered_get(net, index, devlink) {
506  		devl_dev_lock(devlink, true);
507  		err = 0;
508  		if (devl_is_registered(devlink))
509  			err = devlink_reload(devlink, &init_net,
510  					     DEVLINK_RELOAD_ACTION_DRIVER_REINIT,
511  					     DEVLINK_RELOAD_LIMIT_UNSPEC,
512  					     &actions_performed, NULL);
513  		devl_dev_unlock(devlink, true);
514  		devlink_put(devlink);
515  		if (err && err != -EOPNOTSUPP)
516  			pr_warn("Failed to reload devlink instance into init_net\n");
517  	}
518  }
519  
520  static struct pernet_operations devlink_pernet_ops __net_initdata = {
521  	.pre_exit = devlink_pernet_pre_exit,
522  };
523  
524  static struct notifier_block devlink_port_netdevice_nb = {
525  	.notifier_call = devlink_port_netdevice_event,
526  };
527  
devlink_init(void)528  static int __init devlink_init(void)
529  {
530  	int err;
531  
532  	err = register_pernet_subsys(&devlink_pernet_ops);
533  	if (err)
534  		goto out;
535  	err = genl_register_family(&devlink_nl_family);
536  	if (err)
537  		goto out_unreg_pernet_subsys;
538  	err = register_netdevice_notifier(&devlink_port_netdevice_nb);
539  	if (!err)
540  		return 0;
541  
542  	genl_unregister_family(&devlink_nl_family);
543  
544  out_unreg_pernet_subsys:
545  	unregister_pernet_subsys(&devlink_pernet_ops);
546  out:
547  	WARN_ON(err);
548  	return err;
549  }
550  
551  subsys_initcall(devlink_init);
552