Lines Matching refs:txr
365 static void bnxt_txr_db_kick(struct bnxt *bp, struct bnxt_tx_ring_info *txr, in bnxt_txr_db_kick() argument
368 bnxt_db_write(bp, &txr->tx_db, prod); in bnxt_txr_db_kick()
369 txr->kick_pending = 0; in bnxt_txr_db_kick()
373 struct bnxt_tx_ring_info *txr, in bnxt_txr_netif_try_stop_queue() argument
384 if (bnxt_tx_avail(bp, txr) >= bp->tx_wake_thresh) { in bnxt_txr_netif_try_stop_queue()
404 struct bnxt_tx_ring_info *txr; in bnxt_start_xmit() local
415 txr = &bp->tx_ring[bp->tx_ring_map[i]]; in bnxt_start_xmit()
416 prod = txr->tx_prod; in bnxt_start_xmit()
418 free_size = bnxt_tx_avail(bp, txr); in bnxt_start_xmit()
421 if (net_ratelimit() && txr->kick_pending) in bnxt_start_xmit()
424 if (bnxt_txr_netif_try_stop_queue(bp, txr, txq)) in bnxt_start_xmit()
432 txbd = &txr->tx_desc_ring[TX_RING(prod)][TX_IDX(prod)]; in bnxt_start_xmit()
436 tx_buf = &txr->tx_buf_ring[prod]; in bnxt_start_xmit()
453 struct tx_push_buffer *tx_push_buf = txr->tx_push; in bnxt_start_xmit()
456 void __iomem *db = txr->tx_db.doorbell; in bnxt_start_xmit()
499 txbd->tx_bd_haddr = txr->data_mapping; in bnxt_start_xmit()
501 txbd = &txr->tx_desc_ring[TX_RING(prod)][TX_IDX(prod)]; in bnxt_start_xmit()
506 txr->tx_prod = prod; in bnxt_start_xmit()
546 &txr->tx_desc_ring[TX_RING(prod)][TX_IDX(prod)]; in bnxt_start_xmit()
589 txbd = &txr->tx_desc_ring[TX_RING(prod)][TX_IDX(prod)]; in bnxt_start_xmit()
598 tx_buf = &txr->tx_buf_ring[prod]; in bnxt_start_xmit()
618 txr->tx_prod = prod; in bnxt_start_xmit()
621 bnxt_txr_db_kick(bp, txr, prod); in bnxt_start_xmit()
623 txr->kick_pending = 1; in bnxt_start_xmit()
627 if (unlikely(bnxt_tx_avail(bp, txr) <= MAX_SKB_FRAGS + 1)) { in bnxt_start_xmit()
629 bnxt_txr_db_kick(bp, txr, prod); in bnxt_start_xmit()
631 bnxt_txr_netif_try_stop_queue(bp, txr, txq); in bnxt_start_xmit()
639 prod = txr->tx_prod; in bnxt_start_xmit()
640 tx_buf = &txr->tx_buf_ring[prod]; in bnxt_start_xmit()
648 tx_buf = &txr->tx_buf_ring[prod]; in bnxt_start_xmit()
657 if (txr->kick_pending) in bnxt_start_xmit()
658 bnxt_txr_db_kick(bp, txr, txr->tx_prod); in bnxt_start_xmit()
659 txr->tx_buf_ring[txr->tx_prod].skb = NULL; in bnxt_start_xmit()
666 struct bnxt_tx_ring_info *txr = bnapi->tx_ring; in bnxt_tx_int() local
667 struct netdev_queue *txq = netdev_get_tx_queue(bp->dev, txr->txq_index); in bnxt_tx_int()
668 u16 cons = txr->tx_cons; in bnxt_tx_int()
678 tx_buf = &txr->tx_buf_ring[cons]; in bnxt_tx_int()
694 tx_buf = &txr->tx_buf_ring[cons]; in bnxt_tx_int()
710 txr->tx_cons = cons; in bnxt_tx_int()
720 bnxt_tx_avail(bp, txr) >= bp->tx_wake_thresh && in bnxt_tx_int()
721 READ_ONCE(txr->dev_state) != BNXT_DEV_STATE_CLOSING) in bnxt_tx_int()
2347 struct bnxt_tx_ring_info *txr = bnapi->tx_ring; in __bnxt_poll_work() local
2348 u16 prod = txr->tx_prod; in __bnxt_poll_work()
2353 bnxt_db_write_relaxed(bp, &txr->tx_db, prod); in __bnxt_poll_work()
2609 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i]; in bnxt_free_tx_skbs() local
2612 if (!txr->tx_buf_ring) in bnxt_free_tx_skbs()
2616 struct bnxt_sw_tx_bd *tx_buf = &txr->tx_buf_ring[j]; in bnxt_free_tx_skbs()
2658 tx_buf = &txr->tx_buf_ring[ring_idx]; in bnxt_free_tx_skbs()
3029 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i]; in bnxt_free_tx_rings() local
3032 if (txr->tx_push) { in bnxt_free_tx_rings()
3034 txr->tx_push, txr->tx_push_mapping); in bnxt_free_tx_rings()
3035 txr->tx_push = NULL; in bnxt_free_tx_rings()
3038 ring = &txr->tx_ring_struct; in bnxt_free_tx_rings()
3065 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i]; in bnxt_alloc_tx_rings() local
3069 ring = &txr->tx_ring_struct; in bnxt_alloc_tx_rings()
3075 ring->grp_idx = txr->bnapi->index; in bnxt_alloc_tx_rings()
3082 txr->tx_push = dma_alloc_coherent(&pdev->dev, in bnxt_alloc_tx_rings()
3084 &txr->tx_push_mapping, in bnxt_alloc_tx_rings()
3087 if (!txr->tx_push) in bnxt_alloc_tx_rings()
3090 mapping = txr->tx_push_mapping + in bnxt_alloc_tx_rings()
3092 txr->data_mapping = cpu_to_le64(mapping); in bnxt_alloc_tx_rings()
3228 struct bnxt_tx_ring_info *txr; in bnxt_init_ring_struct() local
3266 txr = bnapi->tx_ring; in bnxt_init_ring_struct()
3267 if (!txr) in bnxt_init_ring_struct()
3270 ring = &txr->tx_ring_struct; in bnxt_init_ring_struct()
3274 rmem->pg_arr = (void **)txr->tx_desc_ring; in bnxt_init_ring_struct()
3275 rmem->dma_arr = txr->tx_desc_mapping; in bnxt_init_ring_struct()
3277 rmem->vmem = (void **)&txr->tx_buf_ring; in bnxt_init_ring_struct()
3441 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i]; in bnxt_init_tx_rings() local
3442 struct bnxt_ring_struct *ring = &txr->tx_ring_struct; in bnxt_init_tx_rings()
4112 struct bnxt_tx_ring_info *txr; in bnxt_clear_ring_indices() local
4120 txr = bnapi->tx_ring; in bnxt_clear_ring_indices()
4121 if (txr) { in bnxt_clear_ring_indices()
4122 txr->tx_prod = 0; in bnxt_clear_ring_indices()
4123 txr->tx_cons = 0; in bnxt_clear_ring_indices()
4283 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i]; in bnxt_alloc_mem() local
4286 txr->tx_ring_struct.ring_mem.flags = in bnxt_alloc_mem()
4288 txr->bnapi = bp->bnapi[j]; in bnxt_alloc_mem()
4289 bp->bnapi[j]->tx_ring = txr; in bnxt_alloc_mem()
4292 txr->txq_index = i - bp->tx_nr_rings_xdp; in bnxt_alloc_mem()
5083 static u16 bnxt_cp_ring_for_tx(struct bnxt *bp, struct bnxt_tx_ring_info *txr) in bnxt_cp_ring_for_tx() argument
5086 struct bnxt_napi *bnapi = txr->bnapi; in bnxt_cp_ring_for_tx()
5092 return bnxt_cp_ring_from_grp(bp, &txr->tx_ring_struct); in bnxt_cp_ring_for_tx()
5589 struct bnxt_tx_ring_info *txr; in hwrm_ring_alloc_send_msg() local
5591 txr = container_of(ring, struct bnxt_tx_ring_info, in hwrm_ring_alloc_send_msg()
5596 req.cmpl_ring_id = cpu_to_le16(bnxt_cp_ring_for_tx(bp, txr)); in hwrm_ring_alloc_send_msg()
5779 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i]; in bnxt_hwrm_ring_alloc() local
5784 struct bnxt_napi *bnapi = txr->bnapi; in bnxt_hwrm_ring_alloc()
5800 ring = &txr->tx_ring_struct; in bnxt_hwrm_ring_alloc()
5805 bnxt_set_db(bp, &txr->tx_db, type, map_idx, ring->fw_ring_id); in bnxt_hwrm_ring_alloc()
5902 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i]; in bnxt_hwrm_ring_free() local
5903 struct bnxt_ring_struct *ring = &txr->tx_ring_struct; in bnxt_hwrm_ring_free()
5906 u32 cmpl_ring_id = bnxt_cp_ring_for_tx(bp, txr); in bnxt_hwrm_ring_free()
8946 struct bnxt_tx_ring_info *txr; in bnxt_tx_disable() local
8950 txr = &bp->tx_ring[i]; in bnxt_tx_disable()
8951 WRITE_ONCE(txr->dev_state, BNXT_DEV_STATE_CLOSING); in bnxt_tx_disable()
8965 struct bnxt_tx_ring_info *txr; in bnxt_tx_enable() local
8968 txr = &bp->tx_ring[i]; in bnxt_tx_enable()
8969 WRITE_ONCE(txr->dev_state, 0); in bnxt_tx_enable()
10651 struct bnxt_tx_ring_info *txr = bnapi->tx_ring; in bnxt_dump_tx_sw_state() local
10654 if (!txr) in bnxt_dump_tx_sw_state()
10658 i, txr->tx_ring_struct.fw_ring_id, txr->tx_prod, in bnxt_dump_tx_sw_state()
10659 txr->tx_cons); in bnxt_dump_tx_sw_state()