/linux-6.12.1/drivers/hsi/ |
D | hsi_core.c | 62 kfree(cl->tx_cfg.channels); in hsi_client_release() 77 cl->tx_cfg = info->tx_cfg; in hsi_new_client() 78 if (cl->tx_cfg.channels) { in hsi_new_client() 79 size = cl->tx_cfg.num_channels * sizeof(*cl->tx_cfg.channels); in hsi_new_client() 80 cl->tx_cfg.channels = kmemdup(info->tx_cfg.channels, size, in hsi_new_client() 82 if (!cl->tx_cfg.channels) in hsi_new_client() 110 kfree(cl->tx_cfg.channels); in hsi_new_client() 222 &cl->tx_cfg.mode); in hsi_add_client_from_dt() 227 cl->tx_cfg.mode = mode; in hsi_add_client_from_dt() 231 &cl->tx_cfg.speed); in hsi_add_client_from_dt() [all …]
|
/linux-6.12.1/drivers/net/ethernet/google/gve/ |
D | gve.h | 749 struct gve_queue_config tx_cfg; member 1021 static inline u32 gve_num_tx_qpls(const struct gve_queue_config *tx_cfg, in gve_num_tx_qpls() argument 1027 return tx_cfg->num_queues + num_xdp_queues; in gve_num_tx_qpls() 1056 return priv->tx_cfg.max_queues + rx_qid; in gve_rx_qpl_id() 1059 static inline u32 gve_get_rx_qpl_id(const struct gve_queue_config *tx_cfg, int rx_qid) in gve_get_rx_qpl_id() argument 1061 return tx_cfg->max_queues + rx_qid; in gve_get_rx_qpl_id() 1069 static inline u32 gve_rx_start_qpl_id(const struct gve_queue_config *tx_cfg) in gve_rx_start_qpl_id() argument 1071 return gve_get_rx_qpl_id(tx_cfg, 0); in gve_rx_start_qpl_id() 1086 if (id < gve_rx_start_qpl_id(&priv->tx_cfg)) in gve_qpl_dma_dir() 1100 return priv->tx_cfg.num_queues + priv->num_xdp_queues; in gve_num_tx_queues() [all …]
|
D | gve_main.c | 325 if (block->tx->q_num < priv->tx_cfg.num_queues) in gve_napi_poll() 443 priv->tx_cfg.max_queues = min_t(int, priv->tx_cfg.max_queues, in gve_alloc_notify_blocks() 449 vecs_enabled, priv->tx_cfg.max_queues, in gve_alloc_notify_blocks() 451 if (priv->tx_cfg.num_queues > priv->tx_cfg.max_queues) in gve_alloc_notify_blocks() 452 priv->tx_cfg.num_queues = priv->tx_cfg.max_queues; in gve_alloc_notify_blocks() 763 num_tx_qpls = gve_num_tx_qpls(&priv->tx_cfg, gve_num_xdp_qpls(priv), in gve_register_qpls() 804 num_tx_qpls = gve_num_tx_qpls(&priv->tx_cfg, gve_num_xdp_qpls(priv), in gve_unregister_qpls() 912 for (i = 0; i < priv->tx_cfg.num_queues; i++) in gve_init_sync_stats() 925 cfg->qcfg = &priv->tx_cfg; in gve_tx_get_curr_alloc_cfg() 1072 struct gve_tx_alloc_rings_cfg *tx_cfg, in gve_queues_mem_free() argument [all …]
|
D | gve_ethtool.c | 467 cmd->max_tx = priv->tx_cfg.max_queues; in gve_get_channels() 471 cmd->tx_count = priv->tx_cfg.num_queues; in gve_get_channels() 480 struct gve_queue_config new_tx_cfg = priv->tx_cfg; in gve_set_channels() 496 (new_tx != new_rx || (2 * new_tx > priv->tx_cfg.max_queues))) { in gve_set_channels() 502 priv->tx_cfg.num_queues = new_tx; in gve_set_channels() 759 for (idx = 0; idx < priv->tx_cfg.num_queues; idx++) { in gve_set_coalesce()
|
/linux-6.12.1/drivers/hsi/clients/ |
D | hsi_char.c | 390 tmp = cl->tx_cfg; in hsc_tx_set() 391 cl->tx_cfg.mode = txc->mode; in hsc_tx_set() 392 cl->tx_cfg.num_hw_channels = txc->channels; in hsc_tx_set() 393 cl->tx_cfg.speed = txc->speed; in hsc_tx_set() 394 cl->tx_cfg.arb_mode = txc->arb_mode; in hsc_tx_set() 397 cl->tx_cfg = tmp; in hsc_tx_set() 406 txc->mode = cl->tx_cfg.mode; in hsc_tx_get() 407 txc->channels = cl->tx_cfg.num_hw_channels; in hsc_tx_get() 408 txc->speed = cl->tx_cfg.speed; in hsc_tx_get() 409 txc->arb_mode = cl->tx_cfg.arb_mode; in hsc_tx_get() [all …]
|
D | nokia-modem.c | 176 ssip.tx_cfg = cl->tx_cfg; in nokia_modem_probe() 199 cmtspeech.tx_cfg = cl->tx_cfg; in nokia_modem_probe()
|
/linux-6.12.1/drivers/net/ethernet/samsung/sxgbe/ |
D | sxgbe_core.c | 156 u32 tx_cfg = readl(ioaddr + SXGBE_CORE_TX_CONFIG_REG); in sxgbe_core_set_speed() local 159 tx_cfg &= ~0x60000000; in sxgbe_core_set_speed() 160 tx_cfg |= (speed << SXGBE_SPEED_LSHIFT); in sxgbe_core_set_speed() 163 writel(tx_cfg, ioaddr + SXGBE_CORE_TX_CONFIG_REG); in sxgbe_core_set_speed()
|
/linux-6.12.1/include/linux/hsi/ |
D | hsi.h | 103 struct hsi_config tx_cfg; member 128 struct hsi_config tx_cfg; member 223 struct hsi_config tx_cfg; member
|
/linux-6.12.1/drivers/net/ethernet/ti/icssg/ |
D | icssg_common.c | 281 struct k3_udma_glue_tx_channel_cfg tx_cfg; in prueth_init_tx_chns() local 295 memset(&tx_cfg, 0, sizeof(tx_cfg)); in prueth_init_tx_chns() 296 tx_cfg.swdata_size = PRUETH_NAV_SW_DATA_SIZE; in prueth_init_tx_chns() 297 tx_cfg.tx_cfg = ring_cfg; in prueth_init_tx_chns() 298 tx_cfg.txcq_cfg = ring_cfg; in prueth_init_tx_chns() 313 &tx_cfg); in prueth_init_tx_chns()
|
/linux-6.12.1/drivers/hsi/controllers/ |
D | omap_ssi_port.c | 462 if (cl->tx_cfg.speed) in ssi_setup() 463 omap_ssi->max_speed = cl->tx_cfg.speed; in ssi_setup() 467 cl->tx_cfg.speed, div); in ssi_setup() 479 writel_relaxed(cl->tx_cfg.num_hw_channels, sst + SSI_SST_CHANNELS_REG); in ssi_setup() 480 writel_relaxed(cl->tx_cfg.arb_mode, sst + SSI_SST_ARBMODE_REG); in ssi_setup() 481 writel_relaxed(cl->tx_cfg.mode, sst + SSI_SST_MODE_REG); in ssi_setup() 492 cl->tx_cfg.num_hw_channels); in ssi_setup() 497 omap_port->sst.channels = cl->tx_cfg.num_hw_channels; in ssi_setup() 498 omap_port->sst.arb_mode = cl->tx_cfg.arb_mode; in ssi_setup() 499 omap_port->sst.mode = cl->tx_cfg.mode; in ssi_setup()
|
/linux-6.12.1/drivers/net/ethernet/neterion/ |
D | s2io.c | 582 struct tx_fifo_config *tx_cfg = &config->tx_cfg[i]; in init_shared_mem() local 584 size += tx_cfg->fifo_len; in init_shared_mem() 595 struct tx_fifo_config *tx_cfg = &config->tx_cfg[i]; in init_shared_mem() local 597 size = tx_cfg->fifo_len; in init_shared_mem() 614 struct tx_fifo_config *tx_cfg = &config->tx_cfg[i]; in init_shared_mem() local 615 int fifo_len = tx_cfg->fifo_len; in init_shared_mem() 626 int page_num = TXD_MEM_PAGE_CNT(config->tx_cfg[i].fifo_len, in init_shared_mem() 629 struct tx_fifo_config *tx_cfg = &config->tx_cfg[i]; in init_shared_mem() local 632 fifo->tx_curr_put_info.fifo_len = tx_cfg->fifo_len - 1; in init_shared_mem() 634 fifo->tx_curr_get_info.fifo_len = tx_cfg->fifo_len - 1; in init_shared_mem() [all …]
|
/linux-6.12.1/include/linux/dma/ |
D | k3-udma-glue.h | 14 struct k3_ring_cfg tx_cfg; member
|
/linux-6.12.1/drivers/net/ethernet/micrel/ |
D | ksz884x.c | 1206 u32 tx_cfg; member 2838 u32 tx_cfg; in set_flow_ctrl() local 2841 tx_cfg = hw->tx_cfg; in set_flow_ctrl() 2847 hw->tx_cfg |= DMA_TX_FLOW_ENABLE; in set_flow_ctrl() 2849 hw->tx_cfg &= ~DMA_TX_FLOW_ENABLE; in set_flow_ctrl() 2853 if (tx_cfg != hw->tx_cfg) in set_flow_ctrl() 2854 writel(hw->tx_cfg, hw->io + KS_DMA_TX_CTRL); in set_flow_ctrl() 2893 u32 cfg = hw->tx_cfg; in port_cfg_change() 2897 hw->tx_cfg &= ~DMA_TX_FLOW_ENABLE; in port_cfg_change() 2898 if (hw->enabled && cfg != hw->tx_cfg) in port_cfg_change() [all …]
|
/linux-6.12.1/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
D | hclge_debugfs.c | 2759 u8 tx_cfg; in hclge_get_vlan_tx_offload_cfg() local 2776 tx_cfg = req->vport_vlan_cfg; in hclge_get_vlan_tx_offload_cfg() 2779 vlan_cfg->accept_tag1 = hnae3_get_bit(tx_cfg, HCLGE_ACCEPT_TAG1_B); in hclge_get_vlan_tx_offload_cfg() 2780 vlan_cfg->accept_tag2 = hnae3_get_bit(tx_cfg, HCLGE_ACCEPT_TAG2_B); in hclge_get_vlan_tx_offload_cfg() 2781 vlan_cfg->accept_untag1 = hnae3_get_bit(tx_cfg, HCLGE_ACCEPT_UNTAG1_B); in hclge_get_vlan_tx_offload_cfg() 2782 vlan_cfg->accept_untag2 = hnae3_get_bit(tx_cfg, HCLGE_ACCEPT_UNTAG2_B); in hclge_get_vlan_tx_offload_cfg() 2783 vlan_cfg->insert_tag1 = hnae3_get_bit(tx_cfg, HCLGE_PORT_INS_TAG1_EN_B); in hclge_get_vlan_tx_offload_cfg() 2784 vlan_cfg->insert_tag2 = hnae3_get_bit(tx_cfg, HCLGE_PORT_INS_TAG2_EN_B); in hclge_get_vlan_tx_offload_cfg() 2785 vlan_cfg->shift_tag = hnae3_get_bit(tx_cfg, HCLGE_TAG_SHIFT_MODE_EN_B); in hclge_get_vlan_tx_offload_cfg()
|
/linux-6.12.1/drivers/net/ethernet/brocade/bna/ |
D | bna_tx_rx.c | 3124 cfg_req->tx_cfg.vlan_mode = BFI_ENET_TX_VLAN_WI; in bna_bfi_tx_enet_start() 3125 cfg_req->tx_cfg.vlan_id = htons((u16)tx->txf_vlan_id); in bna_bfi_tx_enet_start() 3126 cfg_req->tx_cfg.admit_tagged_frame = BNA_STATUS_T_ENABLED; in bna_bfi_tx_enet_start() 3127 cfg_req->tx_cfg.apply_vlan_filter = BNA_STATUS_T_DISABLED; in bna_bfi_tx_enet_start() 3364 struct bna_tx_config *tx_cfg, in bna_tx_create() argument 3383 if ((intr_info->num != 1) && (intr_info->num != tx_cfg->num_txq)) in bna_tx_create() 3388 tx = bna_tx_get(tx_mod, tx_cfg->tx_type); in bna_tx_create() 3397 for (i = 0; i < tx_cfg->num_txq; i++) { in bna_tx_create() 3421 tx->num_txq = tx_cfg->num_txq; in bna_tx_create() 3460 txq->ib.coalescing_timeo = tx_cfg->coalescing_timeo; in bna_tx_create() [all …]
|
D | bna.h | 309 struct bna_tx_config *tx_cfg,
|
D | bfi_enet.h | 428 struct bfi_enet_tx_cfg tx_cfg; member
|
/linux-6.12.1/drivers/usb/gadget/udc/ |
D | bcm63xx_udc.c | 468 const struct iudma_ch_cfg *tx_cfg = &iudma_defaults[i + 1]; in bcm63xx_fifo_setup() local 482 ((tx_fifo_slot + tx_cfg->n_fifo_slots - 1) << in bcm63xx_fifo_setup() 484 tx_fifo_slot += tx_cfg->n_fifo_slots; in bcm63xx_fifo_setup() 487 is_hs ? tx_cfg->max_pkt_hs : tx_cfg->max_pkt_fs, in bcm63xx_fifo_setup()
|
/linux-6.12.1/drivers/dma/ti/ |
D | k3-udma-glue.c | 337 cfg->tx_cfg.dma_dev = k3_udma_glue_tx_get_dma_device(tx_chn); in k3_udma_glue_request_tx_chn_common() 338 cfg->txcq_cfg.dma_dev = cfg->tx_cfg.dma_dev; in k3_udma_glue_request_tx_chn_common() 342 cfg->tx_cfg.asel = tx_chn->common.atype_asel; in k3_udma_glue_request_tx_chn_common() 346 ret = k3_ringacc_ring_cfg(tx_chn->ringtx, &cfg->tx_cfg); in k3_udma_glue_request_tx_chn_common()
|
/linux-6.12.1/drivers/soc/qcom/ |
D | qcom-geni-se.c | 431 bool msb_to_lsb, bool tx_cfg, bool rx_cfg) in geni_se_config_packing() argument 464 if (tx_cfg) { in geni_se_config_packing()
|
/linux-6.12.1/drivers/net/ethernet/ti/ |
D | am65-cpsw-nuss.c | 2250 struct k3_udma_glue_tx_channel_cfg tx_cfg = { 0 }; in am65_cpsw_nuss_init_tx_chns() local 2263 tx_cfg.swdata_size = AM65_CPSW_NAV_SW_DATA_SIZE; in am65_cpsw_nuss_init_tx_chns() 2264 tx_cfg.tx_cfg = ring_cfg; in am65_cpsw_nuss_init_tx_chns() 2265 tx_cfg.txcq_cfg = ring_cfg; in am65_cpsw_nuss_init_tx_chns() 2266 tx_cfg.tx_cfg.size = max_desc_num; in am65_cpsw_nuss_init_tx_chns() 2267 tx_cfg.txcq_cfg.size = max_desc_num; in am65_cpsw_nuss_init_tx_chns() 2283 &tx_cfg); in am65_cpsw_nuss_init_tx_chns()
|
/linux-6.12.1/include/linux/soc/qcom/ |
D | geni-se.h | 500 bool msb_to_lsb, bool tx_cfg, bool rx_cfg);
|
/linux-6.12.1/drivers/net/ethernet/sfc/falcon/ |
D | farch.c | 1943 ef4_oword_t tx_cfg; in ef4_farch_filter_push_tx_limits() local 1945 ef4_reado(efx, &tx_cfg, FR_AZ_TX_CFG); in ef4_farch_filter_push_tx_limits() 1950 tx_cfg, FRF_CZ_TX_ETH_FILTER_FULL_SEARCH_RANGE, in ef4_farch_filter_push_tx_limits() 1954 tx_cfg, FRF_CZ_TX_ETH_FILTER_WILD_SEARCH_RANGE, in ef4_farch_filter_push_tx_limits() 1959 ef4_writeo(efx, &tx_cfg, FR_AZ_TX_CFG); in ef4_farch_filter_push_tx_limits()
|
/linux-6.12.1/drivers/net/ethernet/sfc/siena/ |
D | farch.c | 2013 efx_oword_t tx_cfg; in efx_farch_filter_push_tx_limits() local 2015 efx_reado(efx, &tx_cfg, FR_AZ_TX_CFG); in efx_farch_filter_push_tx_limits() 2020 tx_cfg, FRF_CZ_TX_ETH_FILTER_FULL_SEARCH_RANGE, in efx_farch_filter_push_tx_limits() 2024 tx_cfg, FRF_CZ_TX_ETH_FILTER_WILD_SEARCH_RANGE, in efx_farch_filter_push_tx_limits() 2029 efx_writeo(efx, &tx_cfg, FR_AZ_TX_CFG); in efx_farch_filter_push_tx_limits()
|
/linux-6.12.1/drivers/net/ethernet/sun/ |
D | niu.c | 412 u32 tx_cfg, rx_cfg; in serdes_init_niu_10g_fiber() local 415 tx_cfg = (PLL_TX_CFG_ENTX | PLL_TX_CFG_SWING_1375MV); in serdes_init_niu_10g_fiber() 426 tx_cfg |= PLL_TX_CFG_ENTEST; in serdes_init_niu_10g_fiber() 432 int err = esr2_set_tx_cfg(np, i, tx_cfg); in serdes_init_niu_10g_fiber() 452 u32 tx_cfg, rx_cfg; in serdes_init_niu_1g_serdes() local 456 tx_cfg = (PLL_TX_CFG_ENTX | PLL_TX_CFG_SWING_1375MV | in serdes_init_niu_1g_serdes() 471 tx_cfg |= PLL_TX_CFG_ENTEST; in serdes_init_niu_1g_serdes() 500 err = esr2_set_tx_cfg(np, i, tx_cfg); in serdes_init_niu_1g_serdes() 546 u32 tx_cfg, rx_cfg, pll_cfg, pll_sts; in serdes_init_niu_10g_serdes() local 552 tx_cfg = (PLL_TX_CFG_ENTX | PLL_TX_CFG_SWING_1375MV); in serdes_init_niu_10g_serdes() [all …]
|