Home
last modified time | relevance | path

Searched refs:netdev_txq (Results 1 – 5 of 5) sorted by relevance

/OK3568_Linux_fs/kernel/drivers/net/ethernet/google/gve/
H A Dgve_tx.c155 netdev_tx_reset_queue(tx->netdev_txq); in gve_tx_free_ring()
227 tx->netdev_txq = netdev_get_tx_queue(priv->dev, idx); in gve_tx_alloc_ring()
334 netif_tx_stop_queue(tx->netdev_txq); in gve_maybe_stop_tx()
352 netif_tx_start_queue(tx->netdev_txq); in gve_maybe_stop_tx()
499 netdev_tx_sent_queue(tx->netdev_txq, skb->len); in gve_tx()
505 if (!netif_xmit_stopped(tx->netdev_txq) && netdev_xmit_more()) in gve_tx()
554 netdev_tx_completed_queue(tx->netdev_txq, pkts, bytes); in gve_clean_tx_done()
561 if (try_to_wake && netif_tx_queue_stopped(tx->netdev_txq) && in gve_clean_tx_done()
564 netif_tx_wake_queue(tx->netdev_txq); in gve_clean_tx_done()
H A Dgve.h142 struct netdev_queue *netdev_txq; member
/OK3568_Linux_fs/kernel/drivers/net/ethernet/huawei/hinic/
H A Dhinic_tx.c493 struct netdev_queue *netdev_txq; in hinic_lb_xmit_frame() local
535 netdev_txq = netdev_get_tx_queue(netdev, q_id); in hinic_lb_xmit_frame()
536 if ((!netdev_xmit_more()) || (netif_xmit_stopped(netdev_txq))) in hinic_lb_xmit_frame()
554 struct netdev_queue *netdev_txq; in hinic_xmit_frame() local
624 netdev_txq = netdev_get_tx_queue(netdev, q_id); in hinic_xmit_frame()
625 if ((!netdev_xmit_more()) || (netif_xmit_stopped(netdev_txq))) in hinic_xmit_frame()
700 struct netdev_queue *netdev_txq; in free_tx_poll() local
744 netdev_txq = netdev_get_tx_queue(txq->netdev, qp->q_id); in free_tx_poll()
746 __netif_tx_lock(netdev_txq, smp_processor_id()); in free_tx_poll()
750 __netif_tx_unlock(netdev_txq); in free_tx_poll()
/OK3568_Linux_fs/kernel/drivers/net/ethernet/qlogic/qede/
H A Dqede_fp.c441 struct netdev_queue *netdev_txq; in qede_tx_int() local
445 netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); in qede_tx_int()
467 netdev_tx_completed_queue(netdev_txq, pkts_compl, bytes_compl); in qede_tx_int()
480 if (unlikely(netif_tx_queue_stopped(netdev_txq))) { in qede_tx_int()
491 __netif_tx_lock(netdev_txq, smp_processor_id()); in qede_tx_int()
493 if ((netif_tx_queue_stopped(netdev_txq)) && in qede_tx_int()
497 netif_tx_wake_queue(netdev_txq); in qede_tx_int()
502 __netif_tx_unlock(netdev_txq); in qede_tx_int()
1497 struct netdev_queue *netdev_txq; in qede_start_xmit() local
1516 netdev_txq = netdev_get_tx_queue(ndev, txq_index); in qede_start_xmit()
[all …]
H A Dqede_main.c1691 struct netdev_queue *netdev_txq; in qede_empty_tx_queue() local
1694 netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); in qede_empty_tx_queue()
1718 netdev_tx_completed_queue(netdev_txq, pkts_compl, bytes_compl); in qede_empty_tx_queue()
2679 struct netdev_queue *netdev_txq; in qede_is_txq_full() local
2681 netdev_txq = netdev_get_tx_queue(edev->ndev, txq->ndev_txq_id); in qede_is_txq_full()
2682 if (netif_xmit_stopped(netdev_txq)) in qede_is_txq_full()