Home
last modified time | relevance | path

Searched refs:dev_queue (Results 1 – 25 of 34) sorted by relevance

12

/linux-6.12.1/net/sched/
Dsch_mq.c72 struct netdev_queue *dev_queue; in mq_init() local
89 dev_queue = netdev_get_tx_queue(dev, ntx); in mq_init()
90 qdisc = qdisc_create_dflt(dev_queue, get_default_qdisc_ops(dev, ntx), in mq_init()
115 old = dev_graft_qdisc(qdisc->dev_queue, qdisc); in mq_attach()
178 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); in mq_graft() local
185 *old = dev_graft_qdisc(dev_queue, new); in mq_graft()
203 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); in mq_leaf() local
205 return rtnl_dereference(dev_queue->qdisc_sleeping); in mq_leaf()
220 struct netdev_queue *dev_queue = mq_queue_get(sch, cl); in mq_dump_class() local
224 tcm->tcm_info = rtnl_dereference(dev_queue->qdisc_sleeping)->handle; in mq_dump_class()
[all …]
Dsch_generic.c75 const struct netdev_queue *txq = q->dev_queue; in __skb_dequeue_bad_txq()
232 const struct netdev_queue *txq = q->dev_queue; in dequeue_skb()
672 .dev_queue = &noop_netdev_queue,
936 struct Qdisc *qdisc_alloc(struct netdev_queue *dev_queue, in qdisc_alloc() argument
945 if (!dev_queue) { in qdisc_alloc()
951 dev = dev_queue->dev; in qdisc_alloc()
952 sch = kzalloc_node(size, GFP_KERNEL, netdev_queue_numa_node_read(dev_queue)); in qdisc_alloc()
989 sch->dev_queue = dev_queue; in qdisc_alloc()
1002 struct Qdisc *qdisc_create_dflt(struct netdev_queue *dev_queue, in qdisc_create_dflt() argument
1014 sch = qdisc_alloc(dev_queue, ops, extack); in qdisc_create_dflt()
[all …]
Dsch_htb.c1119 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, ntx); in htb_init() local
1122 qdisc = qdisc_create_dflt(dev_queue, &pfifo_qdisc_ops, in htb_init()
1161 old = dev_graft_qdisc(qdisc->dev_queue, qdisc); in htb_attach_offload()
1166 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, ntx); in htb_attach_offload() local
1167 struct Qdisc *old = dev_graft_qdisc(dev_queue, NULL); in htb_attach_offload()
1183 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, ntx); in htb_attach_software() local
1184 struct Qdisc *old = dev_graft_qdisc(dev_queue, sch); in htb_attach_software()
1369 return sch->dev_queue; in htb_select_queue()
1382 htb_graft_helper(struct netdev_queue *dev_queue, struct Qdisc *new_q) in htb_graft_helper() argument
1384 struct net_device *dev = dev_queue->dev; in htb_graft_helper()
[all …]
Dsch_mqprio.c350 struct netdev_queue *dev_queue; in mqprio_init() local
397 dev_queue = netdev_get_tx_queue(dev, i); in mqprio_init()
398 qdisc = qdisc_create_dflt(dev_queue, in mqprio_init()
442 old = dev_graft_qdisc(qdisc->dev_queue, qdisc); in mqprio_attach()
467 struct netdev_queue *dev_queue = mqprio_queue_get(sch, cl); in mqprio_graft() local
469 if (!dev_queue) in mqprio_graft()
475 *old = dev_graft_qdisc(dev_queue, new); in mqprio_graft()
614 struct netdev_queue *dev_queue = mqprio_queue_get(sch, cl); in mqprio_leaf() local
616 if (!dev_queue) in mqprio_leaf()
619 return rtnl_dereference(dev_queue->qdisc_sleeping); in mqprio_leaf()
[all …]
Dsch_api.c1110 struct netdev_queue *dev_queue; in qdisc_graft() local
1117 dev_queue = dev_ingress_queue(dev); in qdisc_graft()
1118 if (!dev_queue) { in qdisc_graft()
1123 q = rtnl_dereference(dev_queue->qdisc_sleeping); in qdisc_graft()
1145 dev_queue = netdev_get_tx_queue(dev, i); in qdisc_graft()
1146 old = dev_graft_qdisc(dev_queue, new); in qdisc_graft()
1153 old = dev_graft_qdisc(dev_queue, NULL); in qdisc_graft()
1162 dev_graft_qdisc(dev_queue, new); in qdisc_graft()
1258 struct netdev_queue *dev_queue, in qdisc_create() argument
1304 sch = qdisc_alloc(dev_queue, ops, extack); in qdisc_create()
[all …]
Dsch_plug.c184 netif_schedule_queue(sch->dev_queue); in plug_change()
192 netif_schedule_queue(sch->dev_queue); in plug_change()
Dsch_cbs.c412 q->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in cbs_init()
423 q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0); in cbs_init()
498 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in cbs_graft()
Dsch_prio.c198 queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in prio_tune()
294 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in prio_graft()
Dsch_taprio.c2091 struct netdev_queue *dev_queue; in taprio_init() local
2094 dev_queue = netdev_get_tx_queue(dev, i); in taprio_init()
2095 qdisc = qdisc_create_dflt(dev_queue, in taprio_init()
2125 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, ntx); in taprio_attach() local
2143 old = dev_graft_qdisc(dev_queue, dev_queue_qdisc); in taprio_attach()
2171 struct netdev_queue *dev_queue = taprio_queue_get(sch, cl); in taprio_graft() local
2173 if (!dev_queue) in taprio_graft()
2188 WARN_ON_ONCE(dev_graft_qdisc(dev_queue, new) != *old); in taprio_graft()
Dsch_drr.c110 cl->qdisc = qdisc_create_dflt(sch->dev_queue, in drr_change_class()
211 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in drr_graft_class()
Dsch_fifo.c260 q = qdisc_create_dflt(sch->dev_queue, ops, TC_H_MAKE(sch->handle, 1), in fifo_create_dflt()
Dsch_ets.c254 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in ets_class_graft()
637 queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in ets_qdisc_change()
Dsch_multiq.c214 child = qdisc_create_dflt(sch->dev_queue, in multiq_tune()
Dsch_hfsc.c1061 cl->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in hfsc_change_class()
1194 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in hfsc_graft_class()
1420 q->root.qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, in hfsc_init_qdisc()
Dsch_etf.c379 q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0); in etf_init()
/linux-6.12.1/include/linux/
Dnetdevice.h3276 static __always_inline void netif_tx_start_queue(struct netdev_queue *dev_queue) in netif_tx_start_queue() argument
3278 clear_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_start_queue()
3302 void netif_tx_wake_queue(struct netdev_queue *dev_queue);
3326 static __always_inline void netif_tx_stop_queue(struct netdev_queue *dev_queue) in netif_tx_stop_queue() argument
3329 WRITE_ONCE(dev_queue->trans_start, jiffies); in netif_tx_stop_queue()
3335 set_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_stop_queue()
3352 static inline bool netif_tx_queue_stopped(const struct netdev_queue *dev_queue) in netif_tx_queue_stopped() argument
3354 return test_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_queue_stopped()
3368 static inline bool netif_xmit_stopped(const struct netdev_queue *dev_queue) in netif_xmit_stopped() argument
3370 return dev_queue->state & QUEUE_STATE_ANY_XOFF; in netif_xmit_stopped()
[all …]
/linux-6.12.1/include/net/
Dnetdev_queues.h208 netdev_txq_completed_mb(struct netdev_queue *dev_queue, in netdev_txq_completed_mb() argument
212 netdev_tx_completed_queue(dev_queue, pkts, bytes); in netdev_txq_completed_mb()
Dsch_generic.h105 struct netdev_queue *dev_queue; member
550 struct Qdisc *q = rcu_dereference_rtnl(qdisc->dev_queue->qdisc); in qdisc_root()
557 return rcu_dereference_bh(qdisc->dev_queue->qdisc); in qdisc_root_bh()
562 return rcu_dereference_rtnl(qdisc->dev_queue->qdisc_sleeping); in qdisc_root_sleeping()
575 return qdisc->dev_queue->dev; in qdisc_dev()
693 struct Qdisc *dev_graft_qdisc(struct netdev_queue *dev_queue,
727 struct Qdisc *qdisc_alloc(struct netdev_queue *dev_queue,
731 struct Qdisc *qdisc_create_dflt(struct netdev_queue *dev_queue,
Dpkt_sched.h144 return dev_net(q->dev_queue->dev); in qdisc_net()
/linux-6.12.1/drivers/net/ethernet/hisilicon/hns/
Dhns_enet.c333 struct netdev_queue *dev_queue; in hns_nic_net_xmit_hw() local
386 dev_queue = netdev_get_tx_queue(ndev, skb->queue_mapping); in hns_nic_net_xmit_hw()
387 netdev_tx_sent_queue(dev_queue, skb->len); in hns_nic_net_xmit_hw()
976 struct netdev_queue *dev_queue; in hns_nic_tx_poll_one() local
1005 dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index); in hns_nic_tx_poll_one()
1006 netdev_tx_completed_queue(dev_queue, pkts, bytes); in hns_nic_tx_poll_one()
1017 if (netif_tx_queue_stopped(dev_queue) && in hns_nic_tx_poll_one()
1019 netif_tx_wake_queue(dev_queue); in hns_nic_tx_poll_one()
1060 struct netdev_queue *dev_queue; in hns_nic_tx_clr_all_bufs() local
1070 dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index); in hns_nic_tx_clr_all_bufs()
[all …]
Dhns_ethtool.c376 struct netdev_queue *dev_queue; in __lb_other_process() local
407 dev_queue = netdev_get_tx_queue(ndev, ring_data->queue_index); in __lb_other_process()
408 netdev_tx_reset_queue(dev_queue); in __lb_other_process()
/linux-6.12.1/drivers/net/ethernet/mellanox/mlx5/core/en/
Dqos.c376 struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, qid); in mlx5e_reset_qdisc() local
377 struct Qdisc *qdisc = dev_queue->qdisc_sleeping; in mlx5e_reset_qdisc()
/linux-6.12.1/drivers/net/xen-netback/
Dxenbus.c31 struct netdev_queue *dev_queue; in xenvif_read_io_ring() local
92 dev_queue = netdev_get_tx_queue(queue->vif->dev, queue->id); in xenvif_read_io_ring()
97 netif_tx_queue_stopped(dev_queue) ? "stopped" : "running"); in xenvif_read_io_ring()
/linux-6.12.1/drivers/net/ethernet/hisilicon/hns3/
Dhns3_enet.c813 struct netdev_queue *dev_queue; in hns3_reset_tx_queue() local
817 dev_queue = netdev_get_tx_queue(ndev, in hns3_reset_tx_queue()
819 netdev_tx_reset_queue(dev_queue); in hns3_reset_tx_queue()
2299 struct netdev_queue *dev_queue; in hns3_nic_net_xmit() local
2343 dev_queue = netdev_get_tx_queue(netdev, ring->queue_index); in hns3_nic_net_xmit()
2344 doorbell = __netdev_tx_sent_queue(dev_queue, desc_cb->send_bytes, in hns3_nic_net_xmit()
3633 struct netdev_queue *dev_queue; in hns3_clean_tx_ring() local
3650 dev_queue = netdev_get_tx_queue(netdev, ring->tqp->tqp_index); in hns3_clean_tx_ring()
3651 netdev_tx_completed_queue(dev_queue, pkts, bytes); in hns3_clean_tx_ring()
3659 if (netif_tx_queue_stopped(dev_queue) && in hns3_clean_tx_ring()
[all …]
/linux-6.12.1/tools/perf/Documentation/
Dperf-trace.txt326 next block:*_unplug and the next three net:*dev_queue events, this last one
329 …# perf trace -e sched:*switch/nr=2/,block:*_plug/nr=4/,block:*_unplug/nr=1/,net:*dev_queue/nr=3,ma…

12