Lines Matching refs:txreq
575 tx = list_entry(l, struct qib_verbs_txreq, txreq.list); in __get_txreq()
604 tx = list_entry(l, struct qib_verbs_txreq, txreq.list); in get_txreq()
627 if (tx->txreq.flags & QIB_SDMA_TXREQ_F_FREEBUF) { in qib_put_txreq()
628 tx->txreq.flags &= ~QIB_SDMA_TXREQ_F_FREEBUF; in qib_put_txreq()
630 tx->txreq.addr, tx->hdr_dwords << 2, in qib_put_txreq()
638 list_add(&tx->txreq.list, &dev->txreq_free); in qib_put_txreq()
686 if (qpp->s_tx->txreq.sg_count > avail) in qib_verbs_sdma_desc_avail()
688 avail -= qpp->s_tx->txreq.sg_count; in qib_verbs_sdma_desc_avail()
714 container_of(cookie, struct qib_verbs_txreq, txreq); in sdma_complete()
724 if (tx->txreq.flags & QIB_SDMA_TXREQ_F_FREEBUF) in sdma_complete()
804 tx->txreq.callback = sdma_complete; in qib_verbs_send_dma()
806 tx->txreq.flags = QIB_SDMA_TXREQ_F_HEADTOHOST; in qib_verbs_send_dma()
808 tx->txreq.flags = QIB_SDMA_TXREQ_F_INTREQ; in qib_verbs_send_dma()
810 tx->txreq.flags |= QIB_SDMA_TXREQ_F_USELARGEBUF; in qib_verbs_send_dma()
827 tx->txreq.flags |= QIB_SDMA_TXREQ_F_FREEDESC; in qib_verbs_send_dma()
828 tx->txreq.sg_count = ndesc; in qib_verbs_send_dma()
829 tx->txreq.addr = dev->pio_hdrs_phys + in qib_verbs_send_dma()
846 tx->txreq.addr = dma_map_single(&dd->pcidev->dev, phdr, in qib_verbs_send_dma()
848 if (dma_mapping_error(&dd->pcidev->dev, tx->txreq.addr)) in qib_verbs_send_dma()
851 tx->txreq.flags |= QIB_SDMA_TXREQ_F_FREEBUF; in qib_verbs_send_dma()
852 tx->txreq.sg_count = 1; in qib_verbs_send_dma()
1538 list_add(&tx->txreq.list, &dev->txreq_free); in qib_register_ib_device()
1630 tx = list_entry(l, struct qib_verbs_txreq, txreq.list); in qib_register_ib_device()
1664 tx = list_entry(l, struct qib_verbs_txreq, txreq.list); in qib_unregister_ib_device()