Searched refs:netdev_txq (Results 1 – 5 of 5) sorted by relevance
| /OK3568_Linux_fs/kernel/drivers/net/ethernet/google/gve/ |
| H A D | gve_tx.c | 155 netdev_tx_reset_queue(tx->netdev_txq); in gve_tx_free_ring() 227 tx->netdev_txq = netdev_get_tx_queue(priv->dev, idx); in gve_tx_alloc_ring() 334 netif_tx_stop_queue(tx->netdev_txq); in gve_maybe_stop_tx() 352 netif_tx_start_queue(tx->netdev_txq); in gve_maybe_stop_tx() 499 netdev_tx_sent_queue(tx->netdev_txq, skb->len); in gve_tx() 505 if (!netif_xmit_stopped(tx->netdev_txq) && netdev_xmit_more()) in gve_tx() 554 netdev_tx_completed_queue(tx->netdev_txq, pkts, bytes); in gve_clean_tx_done() 561 if (try_to_wake && netif_tx_queue_stopped(tx->netdev_txq) && in gve_clean_tx_done() 564 netif_tx_wake_queue(tx->netdev_txq); in gve_clean_tx_done()
|
| H A D | gve.h | 142 struct netdev_queue *netdev_txq; member
|
| /OK3568_Linux_fs/kernel/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_tx.c | 493 struct netdev_queue *netdev_txq; in hinic_lb_xmit_frame() local 535 netdev_txq = netdev_get_tx_queue(netdev, q_id); in hinic_lb_xmit_frame() 536 if ((!netdev_xmit_more()) || (netif_xmit_stopped(netdev_txq))) in hinic_lb_xmit_frame() 554 struct netdev_queue *netdev_txq; in hinic_xmit_frame() local 624 netdev_txq = netdev_get_tx_queue(netdev, q_id); in hinic_xmit_frame() 625 if ((!netdev_xmit_more()) || (netif_xmit_stopped(netdev_txq))) in hinic_xmit_frame() 700 struct netdev_queue *netdev_txq; in free_tx_poll() local 744 netdev_txq = netdev_get_tx_queue(txq->netdev, qp->q_id); in free_tx_poll() 746 __netif_tx_lock(netdev_txq, smp_processor_id()); in free_tx_poll() 750 __netif_tx_unlock(netdev_txq); in free_tx_poll()
|
| /OK3568_Linux_fs/kernel/drivers/net/ethernet/qlogic/qede/ |
| H A D | qede_fp.c | 441 struct netdev_queue *netdev_txq; in qede_tx_int() local 445 netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); in qede_tx_int() 467 netdev_tx_completed_queue(netdev_txq, pkts_compl, bytes_compl); in qede_tx_int() 480 if (unlikely(netif_tx_queue_stopped(netdev_txq))) { in qede_tx_int() 491 __netif_tx_lock(netdev_txq, smp_processor_id()); in qede_tx_int() 493 if ((netif_tx_queue_stopped(netdev_txq)) && in qede_tx_int() 497 netif_tx_wake_queue(netdev_txq); in qede_tx_int() 502 __netif_tx_unlock(netdev_txq); in qede_tx_int() 1497 struct netdev_queue *netdev_txq; in qede_start_xmit() local 1516 netdev_txq = netdev_get_tx_queue(ndev, txq_index); in qede_start_xmit() [all …]
|
| H A D | qede_main.c | 1691 struct netdev_queue *netdev_txq; in qede_empty_tx_queue() local 1694 netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); in qede_empty_tx_queue() 1718 netdev_tx_completed_queue(netdev_txq, pkts_compl, bytes_compl); in qede_empty_tx_queue() 2679 struct netdev_queue *netdev_txq; in qede_is_txq_full() local 2681 netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); in qede_is_txq_full() 2682 if (netif_xmit_stopped(netdev_txq)) in qede_is_txq_full()
|