/linux-6.12.1/drivers/infiniband/hw/bnxt_re/ |
D | qplib_fp.c | 163 struct bnxt_qplib_nq *nq = nq_work->nq; in bnxt_qpn_cqn_sched_task() local 165 if (cq && nq) { in bnxt_qpn_cqn_sched_task() 167 if (atomic_read(&cq->arm_state) && nq->cqn_handler) { in bnxt_qpn_cqn_sched_task() 168 dev_dbg(&nq->pdev->dev, in bnxt_qpn_cqn_sched_task() 170 __func__, cq, nq); in bnxt_qpn_cqn_sched_task() 171 nq->cqn_handler(nq, cq); in bnxt_qpn_cqn_sched_task() 239 static void clean_nq(struct bnxt_qplib_nq *nq, struct bnxt_qplib_cq *cq) in clean_nq() argument 241 struct bnxt_qplib_hwq *hwq = &nq->hwq; in clean_nq() 243 int budget = nq->budget; in clean_nq() 252 if (!NQE_CMP_VALID(nqe, nq->nq_db.dbinfo.flags)) in clean_nq() [all …]
|
D | qplib_fp.h | 422 struct bnxt_qplib_nq *nq; member 485 typedef int (*cqn_handler_t)(struct bnxt_qplib_nq *nq, 487 typedef int (*srqn_handler_t)(struct bnxt_qplib_nq *nq, 510 struct bnxt_qplib_nq *nq; member 514 void bnxt_qplib_nq_stop_irq(struct bnxt_qplib_nq *nq, bool kill); 515 void bnxt_qplib_disable_nq(struct bnxt_qplib_nq *nq); 516 int bnxt_qplib_nq_start_irq(struct bnxt_qplib_nq *nq, int nq_indx, 518 int bnxt_qplib_enable_nq(struct pci_dev *pdev, struct bnxt_qplib_nq *nq, 563 void bnxt_qplib_free_nq(struct bnxt_qplib_nq *nq); 564 int bnxt_qplib_alloc_nq(struct bnxt_qplib_res *res, struct bnxt_qplib_nq *nq); [all …]
|
D | main.c | 313 struct bnxt_qplib_nq *nq; in bnxt_re_stop_irq() local 320 nq = &rdev->nq[indx - 1]; in bnxt_re_stop_irq() 321 bnxt_qplib_nq_stop_irq(nq, false); in bnxt_re_stop_irq() 333 struct bnxt_qplib_nq *nq; in bnxt_re_start_irq() local 362 nq = &rdev->nq[indx - 1]; in bnxt_re_start_irq() 363 rc = bnxt_qplib_nq_start_irq(nq, indx - 1, in bnxt_re_start_irq() 1244 static int bnxt_re_srqn_handler(struct bnxt_qplib_nq *nq, in bnxt_re_srqn_handler() argument 1263 static int bnxt_re_cqn_handler(struct bnxt_qplib_nq *nq, in bnxt_re_cqn_handler() argument 1280 bnxt_qplib_disable_nq(&rdev->nq[i - 1]); in bnxt_re_cleanup_res() 1296 rc = bnxt_qplib_enable_nq(rdev->en_dev->pdev, &rdev->nq[i - 1], in bnxt_re_init_res() [all …]
|
D | ib_verbs.c | 948 scq_nq = qplib_qp->scq->nq; in bnxt_re_destroy_qp() 949 rcq_nq = qplib_qp->rcq->nq; in bnxt_re_destroy_qp() 1721 struct bnxt_qplib_nq *nq = NULL; in bnxt_re_destroy_srq() local 1724 nq = qplib_srq->cq->nq; in bnxt_re_destroy_srq() 1732 if (nq) in bnxt_re_destroy_srq() 1733 nq->budget--; in bnxt_re_destroy_srq() 1774 struct bnxt_qplib_nq *nq = NULL; in bnxt_re_create_srq() local 1817 srq->qplib_srq.eventq_hw_ring_id = rdev->nq[0].ring_id; in bnxt_re_create_srq() 1818 nq = &rdev->nq[0]; in bnxt_re_create_srq() 1853 if (nq) in bnxt_re_create_srq() [all …]
|
D | bnxt_re.h | 191 struct bnxt_qplib_nq nq[BNXT_MAX_ROCE_MSIX]; member
|
/linux-6.12.1/drivers/block/null_blk/ |
D | main.c | 820 ktime_t kt = cmd->nq->dev->completion_nsec; in null_cmd_end_timer() 1247 struct nullb *nullb = cmd->nq->dev->nullb; in null_handle_rq() 1271 struct nullb_device *dev = cmd->nq->dev; in null_handle_throttled() 1294 struct badblocks *bb = &cmd->nq->dev->badblocks; in null_handle_badblocks() 1309 struct nullb_device *dev = cmd->nq->dev; in null_handle_memory_backed() 1320 struct nullb_device *dev = cmd->nq->dev; in nullb_zero_read_cmd_buffer() 1344 switch (cmd->nq->dev->irqmode) { in nullb_complete_cmd() 1360 struct nullb_device *dev = cmd->nq->dev; in null_process_cmd() 1378 struct nullb_device *dev = cmd->nq->dev; in null_handle_cmd() 1432 struct nullb_device *dev = cmd->nq->dev; in should_timeout_request() [all …]
|
D | null_blk.h | 21 struct nullb_queue *nq; member
|
D | zoned.c | 353 struct nullb_device *dev = cmd->nq->dev; in null_zone_write() 637 struct nullb_device *dev = cmd->nq->dev; in null_zone_mgmt() 715 dev = cmd->nq->dev; in null_process_zoned_cmd()
|
/linux-6.12.1/lib/crypto/ |
D | curve25519-hacl64.c | 544 ladder_smallloop_cmult_small_loop_step(u64 *nq, u64 *nqpq, u64 *nq2, u64 *nqpq2, in ladder_smallloop_cmult_small_loop_step() argument 549 point_swap_conditional(nq, nqpq, bit0); in ladder_smallloop_cmult_small_loop_step() 550 addanddouble_fmonty(nq2, nqpq2, nq, nqpq, q); in ladder_smallloop_cmult_small_loop_step() 556 ladder_smallloop_cmult_small_loop_double_step(u64 *nq, u64 *nqpq, u64 *nq2, in ladder_smallloop_cmult_small_loop_double_step() argument 560 ladder_smallloop_cmult_small_loop_step(nq, nqpq, nq2, nqpq2, q, byt); in ladder_smallloop_cmult_small_loop_double_step() 562 ladder_smallloop_cmult_small_loop_step(nq2, nqpq2, nq, nqpq, q, byt1); in ladder_smallloop_cmult_small_loop_double_step() 566 ladder_smallloop_cmult_small_loop(u64 *nq, u64 *nqpq, u64 *nq2, u64 *nqpq2, in ladder_smallloop_cmult_small_loop() argument 570 ladder_smallloop_cmult_small_loop_double_step(nq, nqpq, nq2, in ladder_smallloop_cmult_small_loop() 576 static __always_inline void ladder_bigloop_cmult_big_loop(u8 *n1, u64 *nq, in ladder_bigloop_cmult_big_loop() argument 583 ladder_smallloop_cmult_small_loop(nq, nqpq, nq2, nqpq2, q, in ladder_bigloop_cmult_big_loop() [all …]
|
/linux-6.12.1/fs/xfs/ |
D | xfs_trans_dquot.c | 79 struct xfs_dqtrx *oq, *nq; in xfs_trans_dup_dqinfo() local 98 nq = &nqa[i]; in xfs_trans_dup_dqinfo() 103 nq->qt_dquot = oq->qt_dquot; in xfs_trans_dup_dqinfo() 104 nq->qt_bcount_delta = nq->qt_icount_delta = 0; in xfs_trans_dup_dqinfo() 105 nq->qt_rtbcount_delta = 0; in xfs_trans_dup_dqinfo() 110 nq->qt_blk_res = oq->qt_blk_res - blk_res_used; in xfs_trans_dup_dqinfo() 113 nq->qt_rtblk_res = oq->qt_rtblk_res - in xfs_trans_dup_dqinfo() 117 nq->qt_ino_res = oq->qt_ino_res - oq->qt_ino_res_used; in xfs_trans_dup_dqinfo()
|
/linux-6.12.1/arch/x86/crypto/ |
D | curve25519-x86_64.c | 977 u64 *nq = p01_tmp1; in point_add_and_double() local 981 u64 *x2 = nq; in point_add_and_double() 982 u64 *z2 = nq + (u32)4U; in point_add_and_double() 1024 fmul2(nq, dc1, ab1, tmp2); in point_add_and_double() 1028 static void point_double(u64 *nq, u64 *tmp1, u64 *tmp2) in point_double() argument 1030 u64 *x2 = nq; in point_double() 1031 u64 *z2 = nq + (u32)4U; in point_double() 1048 fmul2(nq, dc, ab, tmp2); in point_double()
|
/linux-6.12.1/drivers/net/ethernet/engleder/ |
D | tsnep_main.c | 348 struct netdev_queue *nq; in tsnep_tx_enable() local 350 nq = netdev_get_tx_queue(tx->adapter->netdev, tx->queue_index); in tsnep_tx_enable() 352 __netif_tx_lock_bh(nq); in tsnep_tx_enable() 353 netif_tx_wake_queue(nq); in tsnep_tx_enable() 354 __netif_tx_unlock_bh(nq); in tsnep_tx_enable() 359 struct netdev_queue *nq; in tsnep_tx_disable() local 362 nq = netdev_get_tx_queue(tx->adapter->netdev, tx->queue_index); in tsnep_tx_disable() 364 __netif_tx_lock_bh(nq); in tsnep_tx_disable() 365 netif_tx_stop_queue(nq); in tsnep_tx_disable() 366 __netif_tx_unlock_bh(nq); in tsnep_tx_disable() [all …]
|
/linux-6.12.1/drivers/net/ethernet/pensando/ionic/ |
D | ionic_txrx.c | 398 struct netdev_queue *nq; in ionic_xdp_xmit() local 419 nq = netdev_get_tx_queue(netdev, txq->index); in ionic_xdp_xmit() 420 __netif_tx_lock(nq, cpu); in ionic_xdp_xmit() 421 txq_trans_cond_update(nq); in ionic_xdp_xmit() 423 if (netif_tx_queue_stopped(nq) || in ionic_xdp_xmit() 427 __netif_tx_unlock(nq); in ionic_xdp_xmit() 449 __netif_tx_unlock(nq); in ionic_xdp_xmit() 476 struct netdev_queue *nq; in ionic_run_xdp() local 551 nq = netdev_get_tx_queue(netdev, txq->index); in ionic_run_xdp() 552 __netif_tx_lock(nq, smp_processor_id()); in ionic_run_xdp() [all …]
|
/linux-6.12.1/drivers/net/vmxnet3/ |
D | vmxnet3_xdp.c | 205 struct netdev_queue *nq; in vmxnet3_xdp_xmit_back() local 212 nq = netdev_get_tx_queue(adapter->netdev, tq->qid); in vmxnet3_xdp_xmit_back() 214 __netif_tx_lock(nq, smp_processor_id()); in vmxnet3_xdp_xmit_back() 216 __netif_tx_unlock(nq); in vmxnet3_xdp_xmit_back()
|
/linux-6.12.1/drivers/net/ethernet/chelsio/cxgb4/ |
D | cxgb4_uld.c | 111 unsigned int nq = rxq_info->nrxq + rxq_info->nciq; in alloc_uld_rxqs() local 125 for (i = 0; i < nq; i++, q++) { in alloc_uld_rxqs() 403 int nq = txq_info->ntxq; in free_sge_txq_uld() local 406 for (i = 0; i < nq; i++) { in free_sge_txq_uld() 426 int nq = txq_info->ntxq; in alloc_sge_txq_uld() local 429 j = nq / adap->params.nports; in alloc_sge_txq_uld() 430 for (i = 0; i < nq; i++) { in alloc_sge_txq_uld()
|
/linux-6.12.1/drivers/net/ethernet/intel/idpf/ |
D | idpf_singleq_txrx.c | 203 struct netdev_queue *nq; in idpf_tx_singleq_map() local 295 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx); in idpf_tx_singleq_map() 296 netdev_tx_sent_queue(nq, first->bytes); in idpf_tx_singleq_map() 442 struct netdev_queue *nq; in idpf_tx_singleq_clean() local 518 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx); in idpf_tx_singleq_clean() 522 __netif_txq_completed_wake(nq, ss.packets, ss.bytes, in idpf_tx_singleq_clean()
|
D | idpf_txrx.c | 2065 struct netdev_queue *nq; in idpf_tx_clean_complq() local 2075 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx); in idpf_tx_clean_complq() 2081 __netif_txq_completed_wake(nq, tx_q->cleaned_pkts, tx_q->cleaned_bytes, in idpf_tx_clean_complq() 2188 struct netdev_queue *nq; in idpf_tx_buf_hw_update() local 2190 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx); in idpf_tx_buf_hw_update() 2207 if (netif_xmit_stopped(nq) || !xmit_more) in idpf_tx_buf_hw_update() 2344 struct netdev_queue *nq; in idpf_tx_splitq_map() local 2506 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx); in idpf_tx_splitq_map() 2507 netdev_tx_sent_queue(nq, first->bytes); in idpf_tx_splitq_map()
|
/linux-6.12.1/drivers/net/ethernet/marvell/ |
D | mvneta.c | 1870 struct netdev_queue *nq, bool napi) in mvneta_txq_bufs_free() argument 1909 netdev_tx_completed_queue(nq, pkts_compl, bytes_compl); in mvneta_txq_bufs_free() 1916 struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id); in mvneta_txq_done() local 1923 mvneta_txq_bufs_free(pp, txq, tx_done, nq, true); in mvneta_txq_done() 1927 if (netif_tx_queue_stopped(nq)) { in mvneta_txq_done() 1929 netif_tx_wake_queue(nq); in mvneta_txq_done() 2178 struct netdev_queue *nq; in mvneta_xdp_xmit_back() local 2189 nq = netdev_get_tx_queue(pp->dev, txq->id); in mvneta_xdp_xmit_back() 2191 __netif_tx_lock(nq, cpu); in mvneta_xdp_xmit_back() 2206 __netif_tx_unlock(nq); in mvneta_xdp_xmit_back() [all …]
|
D | mv643xx_eth.c | 496 struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index); in txq_maybe_wake() local 498 if (netif_tx_queue_stopped(nq)) { in txq_maybe_wake() 499 __netif_tx_lock(nq, smp_processor_id()); in txq_maybe_wake() 501 netif_tx_wake_queue(nq); in txq_maybe_wake() 502 __netif_tx_unlock(nq); in txq_maybe_wake() 998 struct netdev_queue *nq; in mv643xx_eth_xmit() local 1002 nq = netdev_get_tx_queue(dev, queue); in mv643xx_eth_xmit() 1021 netif_tx_stop_queue(nq); in mv643xx_eth_xmit() 1035 struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index); in txq_kick() local 1039 __netif_tx_lock(nq, smp_processor_id()); in txq_kick() [all …]
|
/linux-6.12.1/drivers/net/ethernet/freescale/ |
D | fec_main.c | 927 struct netdev_queue *nq; in fec_enet_start_xmit() local 932 nq = netdev_get_tx_queue(ndev, queue); in fec_enet_start_xmit() 943 netif_tx_stop_queue(nq); in fec_enet_start_xmit() 1437 struct netdev_queue *nq; in fec_enet_tx_queue() local 1447 nq = netdev_get_tx_queue(ndev, queue_id); in fec_enet_tx_queue() 1571 if (netif_tx_queue_stopped(nq)) { in fec_enet_tx_queue() 1574 netif_tx_wake_queue(nq); in fec_enet_tx_queue() 3905 struct netdev_queue *nq; in fec_enet_xdp_tx_xmit() local 3910 nq = netdev_get_tx_queue(fep->netdev, queue); in fec_enet_xdp_tx_xmit() 3912 __netif_tx_lock(nq, cpu); in fec_enet_xdp_tx_xmit() [all …]
|
/linux-6.12.1/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | t4vf_hw.c | 1272 int nq = min(n, 32); in t4vf_config_rss_range() local 1279 cmd.niqid = cpu_to_be16(nq); in t4vf_config_rss_range() 1285 start += nq; in t4vf_config_rss_range() 1286 n -= nq; in t4vf_config_rss_range() 1293 while (nq > 0) { in t4vf_config_rss_range() 1302 int nqbuf = min(3, nq); in t4vf_config_rss_range() 1304 nq -= nqbuf; in t4vf_config_rss_range()
|
/linux-6.12.1/drivers/net/ |
D | tap.c | 192 struct tap_queue *nq; in tap_disable_queue() local 203 nq = rtnl_dereference(tap->taps[tap->numvtaps - 1]); in tap_disable_queue() 204 nq->queue_index = index; in tap_disable_queue() 206 rcu_assign_pointer(tap->taps[index], nq); in tap_disable_queue()
|
/linux-6.12.1/arch/s390/mm/ |
D | pgtable.c | 803 unsigned char key, bool nq) in set_guest_storage_key() argument 860 page_set_storage_key(paddr, skey, !nq); in set_guest_storage_key() 884 bool nq, bool mr, bool mc) in cond_set_guest_storage_key() argument 903 rc = set_guest_storage_key(current->mm, addr, key, nq); in cond_set_guest_storage_key()
|
/linux-6.12.1/drivers/net/ethernet/intel/igc/ |
D | igc_main.c | 2469 struct netdev_queue *nq; in igc_xdp_xmit_back() local 2477 nq = txring_txq(ring); in igc_xdp_xmit_back() 2479 __netif_tx_lock(nq, cpu); in igc_xdp_xmit_back() 2481 txq_trans_cond_update(nq); in igc_xdp_xmit_back() 2483 __netif_tx_unlock(nq); in igc_xdp_xmit_back() 2550 struct netdev_queue *nq; in igc_finalize_xdp() local 2555 nq = txring_txq(ring); in igc_finalize_xdp() 2557 __netif_tx_lock(nq, cpu); in igc_finalize_xdp() 2559 __netif_tx_unlock(nq); in igc_finalize_xdp() 2966 struct netdev_queue *nq = txring_txq(ring); in igc_xdp_xmit_zc() local [all …]
|
/linux-6.12.1/drivers/net/ethernet/aquantia/atlantic/ |
D | aq_nic.c | 837 struct netdev_queue *nq; in aq_nic_xmit_xdpf() local 848 nq = netdev_get_tx_queue(ndev, tx_ring->idx); in aq_nic_xmit_xdpf() 849 __netif_tx_lock(nq, cpu); in aq_nic_xmit_xdpf() 862 __netif_tx_unlock(nq); in aq_nic_xmit_xdpf()
|