Lines Matching refs:dev_queue

3276 static __always_inline void netif_tx_start_queue(struct netdev_queue *dev_queue)  in netif_tx_start_queue()  argument
3278 clear_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_start_queue()
3302 void netif_tx_wake_queue(struct netdev_queue *dev_queue);
3326 static __always_inline void netif_tx_stop_queue(struct netdev_queue *dev_queue) in netif_tx_stop_queue() argument
3329 WRITE_ONCE(dev_queue->trans_start, jiffies); in netif_tx_stop_queue()
3335 set_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_stop_queue()
3352 static inline bool netif_tx_queue_stopped(const struct netdev_queue *dev_queue) in netif_tx_queue_stopped() argument
3354 return test_bit(__QUEUE_STATE_DRV_XOFF, &dev_queue->state); in netif_tx_queue_stopped()
3368 static inline bool netif_xmit_stopped(const struct netdev_queue *dev_queue) in netif_xmit_stopped() argument
3370 return dev_queue->state & QUEUE_STATE_ANY_XOFF; in netif_xmit_stopped()
3374 netif_xmit_frozen_or_stopped(const struct netdev_queue *dev_queue) in netif_xmit_frozen_or_stopped() argument
3376 return dev_queue->state & QUEUE_STATE_ANY_XOFF_OR_FROZEN; in netif_xmit_frozen_or_stopped()
3380 netif_xmit_frozen_or_drv_stopped(const struct netdev_queue *dev_queue) in netif_xmit_frozen_or_drv_stopped() argument
3382 return dev_queue->state & QUEUE_STATE_DRV_XOFF_OR_FROZEN; in netif_xmit_frozen_or_drv_stopped()
3395 static inline void netdev_queue_set_dql_min_limit(struct netdev_queue *dev_queue, in netdev_queue_set_dql_min_limit() argument
3399 dev_queue->dql.min_limit = min_limit; in netdev_queue_set_dql_min_limit()
3420 static inline void netdev_txq_bql_enqueue_prefetchw(struct netdev_queue *dev_queue) in netdev_txq_bql_enqueue_prefetchw() argument
3423 prefetchw(&dev_queue->dql.num_queued); in netdev_txq_bql_enqueue_prefetchw()
3434 static inline void netdev_txq_bql_complete_prefetchw(struct netdev_queue *dev_queue) in netdev_txq_bql_complete_prefetchw() argument
3437 prefetchw(&dev_queue->dql.limit); in netdev_txq_bql_complete_prefetchw()
3451 static inline void netdev_tx_sent_queue(struct netdev_queue *dev_queue, in netdev_tx_sent_queue() argument
3455 dql_queued(&dev_queue->dql, bytes); in netdev_tx_sent_queue()
3457 if (likely(dql_avail(&dev_queue->dql) >= 0)) in netdev_tx_sent_queue()
3461 WRITE_ONCE(dev_queue->trans_start, jiffies); in netdev_tx_sent_queue()
3466 set_bit(__QUEUE_STATE_STACK_XOFF, &dev_queue->state); in netdev_tx_sent_queue()
3476 if (unlikely(dql_avail(&dev_queue->dql) >= 0)) in netdev_tx_sent_queue()
3477 clear_bit(__QUEUE_STATE_STACK_XOFF, &dev_queue->state); in netdev_tx_sent_queue()
3487 static inline bool __netdev_tx_sent_queue(struct netdev_queue *dev_queue, in __netdev_tx_sent_queue() argument
3493 dql_queued(&dev_queue->dql, bytes); in __netdev_tx_sent_queue()
3495 return netif_tx_queue_stopped(dev_queue); in __netdev_tx_sent_queue()
3497 netdev_tx_sent_queue(dev_queue, bytes); in __netdev_tx_sent_queue()
3533 static inline void netdev_tx_completed_queue(struct netdev_queue *dev_queue, in netdev_tx_completed_queue() argument
3540 dql_completed(&dev_queue->dql, bytes); in netdev_tx_completed_queue()
3549 if (unlikely(dql_avail(&dev_queue->dql) < 0)) in netdev_tx_completed_queue()
3552 if (test_and_clear_bit(__QUEUE_STATE_STACK_XOFF, &dev_queue->state)) in netdev_tx_completed_queue()
3553 netif_schedule_queue(dev_queue); in netdev_tx_completed_queue()
3599 static inline void netdev_reset_queue(struct net_device *dev_queue) in netdev_reset_queue() argument
3601 netdev_tx_reset_subqueue(dev_queue, 0); in netdev_reset_queue()