Home
last modified time | relevance | path

Searched full:queue (Results 1 – 25 of 4424) sorted by relevance

12345678910>>...177

/OK3568_Linux_fs/kernel/drivers/net/wireless/st/cw1200/
H A Dqueue.c3 * O(1) TX queue with built-in allocator for ST-Ericsson CW1200 drivers
11 #include "queue.h"
26 static inline void __cw1200_queue_lock(struct cw1200_queue *queue) in __cw1200_queue_lock() argument
28 struct cw1200_queue_stats *stats = queue->stats; in __cw1200_queue_lock()
29 if (queue->tx_locked_cnt++ == 0) { in __cw1200_queue_lock()
30 pr_debug("[TX] Queue %d is locked.\n", in __cw1200_queue_lock()
31 queue->queue_id); in __cw1200_queue_lock()
32 ieee80211_stop_queue(stats->priv->hw, queue->queue_id); in __cw1200_queue_lock()
36 static inline void __cw1200_queue_unlock(struct cw1200_queue *queue) in __cw1200_queue_unlock() argument
38 struct cw1200_queue_stats *stats = queue->stats; in __cw1200_queue_unlock()
[all …]
/OK3568_Linux_fs/kernel/drivers/media/usb/uvc/
H A Duvc_queue.c24 * Video buffers queue management.
30 * the videobuf2 queue operations by serializing calls to videobuf2 and a
31 * spinlock to protect the IRQ queue that holds the buffers to be processed by
36 uvc_queue_to_stream(struct uvc_video_queue *queue) in uvc_queue_to_stream() argument
38 return container_of(queue, struct uvc_streaming, queue); in uvc_queue_to_stream()
49 * This function must be called with the queue spinlock held.
51 static void uvc_queue_return_buffers(struct uvc_video_queue *queue, in uvc_queue_return_buffers() argument
58 while (!list_empty(&queue->irqqueue)) { in uvc_queue_return_buffers()
59 struct uvc_buffer *buf = list_first_entry(&queue->irqqueue, in uvc_queue_return_buffers()
61 queue); in uvc_queue_return_buffers()
[all …]
/OK3568_Linux_fs/kernel/drivers/nvme/target/
H A Dtcp.c56 struct nvmet_tcp_queue *queue; member
158 static inline u16 nvmet_tcp_cmd_tag(struct nvmet_tcp_queue *queue, in nvmet_tcp_cmd_tag() argument
161 if (unlikely(!queue->nr_cmds)) { in nvmet_tcp_cmd_tag()
166 return cmd - queue->cmds; in nvmet_tcp_cmd_tag()
194 nvmet_tcp_get_cmd(struct nvmet_tcp_queue *queue) in nvmet_tcp_get_cmd() argument
198 cmd = list_first_entry_or_null(&queue->free_list, in nvmet_tcp_get_cmd()
214 if (unlikely(cmd == &cmd->queue->connect)) in nvmet_tcp_put_cmd()
217 list_add_tail(&cmd->entry, &cmd->queue->free_list); in nvmet_tcp_put_cmd()
220 static inline int queue_cpu(struct nvmet_tcp_queue *queue) in queue_cpu() argument
222 return queue->sock->sk->sk_incoming_cpu; in queue_cpu()
[all …]
H A Drdma.c47 struct nvmet_rdma_queue *queue; member
62 struct nvmet_rdma_queue *queue; member
144 MODULE_PARM_DESC(use_srq, "Use shared receive queue.");
154 MODULE_PARM_DESC(srq_size, "set Shared Receive Queue (SRQ) size, should >= 256 (default: 1024)");
169 static void nvmet_rdma_queue_disconnect(struct nvmet_rdma_queue *queue);
209 nvmet_rdma_get_rsp(struct nvmet_rdma_queue *queue) in nvmet_rdma_get_rsp() argument
214 spin_lock_irqsave(&queue->rsps_lock, flags); in nvmet_rdma_get_rsp()
215 rsp = list_first_entry_or_null(&queue->free_rsps, in nvmet_rdma_get_rsp()
219 spin_unlock_irqrestore(&queue->rsps_lock, flags); in nvmet_rdma_get_rsp()
227 ret = nvmet_rdma_alloc_rsp(queue->dev, rsp); in nvmet_rdma_get_rsp()
[all …]
/OK3568_Linux_fs/kernel/drivers/net/wireless/broadcom/b43legacy/
H A Dpio.c22 static void tx_start(struct b43legacy_pioqueue *queue) in tx_start() argument
24 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL, in tx_start()
28 static void tx_octet(struct b43legacy_pioqueue *queue, in tx_octet() argument
31 if (queue->need_workarounds) { in tx_octet()
32 b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, octet); in tx_octet()
33 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL, in tx_octet()
36 b43legacy_pio_write(queue, B43legacy_PIO_TXCTL, in tx_octet()
38 b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, octet); in tx_octet()
63 static void tx_data(struct b43legacy_pioqueue *queue, in tx_data() argument
71 if (queue->need_workarounds) { in tx_data()
[all …]
/OK3568_Linux_fs/kernel/drivers/usb/gadget/function/
H A Duvc_queue.c26 * Video buffers queue management.
32 * the videobuf2 queue operations by serializing calls to videobuf2 and a
33 * spinlock to protect the IRQ queue that holds the buffers to be processed by
38 * videobuf2 queue operations
45 struct uvc_video_queue *queue = vb2_get_drv_priv(vq); in uvc_queue_setup() local
46 struct uvc_video *video = container_of(queue, struct uvc_video, queue); in uvc_queue_setup()
130 struct uvc_video_queue *queue) in uvc_buffer_mem_prepare() argument
132 struct uvc_video *video = container_of(queue, struct uvc_video, queue); in uvc_buffer_mem_prepare()
150 struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue); in uvc_buffer_prepare() local
160 if (unlikely(queue->flags & UVC_QUEUE_DISCONNECTED)) in uvc_buffer_prepare()
[all …]
/OK3568_Linux_fs/kernel/drivers/scsi/arm/
H A Dqueue.c3 * linux/drivers/acorn/scsi/queue.c: queue handling primitives
46 #include "queue.h"
51 * Function: void queue_initialise (Queue_t *queue)
52 * Purpose : initialise a queue
53 * Params : queue - queue to initialise
55 int queue_initialise (Queue_t *queue) in queue_initialise() argument
60 spin_lock_init(&queue->queue_lock); in queue_initialise()
61 INIT_LIST_HEAD(&queue->head); in queue_initialise()
62 INIT_LIST_HEAD(&queue->free); in queue_initialise()
70 queue->alloc = q = kmalloc_array(nqueues, sizeof(QE_t), GFP_KERNEL); in queue_initialise()
[all …]
H A Dqueue.h3 * linux/drivers/acorn/scsi/queue.h: queue handling
18 * Function: void queue_initialise (Queue_t *queue)
19 * Purpose : initialise a queue
20 * Params : queue - queue to initialise
22 extern int queue_initialise (Queue_t *queue);
25 * Function: void queue_free (Queue_t *queue)
26 * Purpose : free a queue
27 * Params : queue - queue to free
29 extern void queue_free (Queue_t *queue);
32 * Function: struct scsi_cmnd *queue_remove (queue)
[all …]
/OK3568_Linux_fs/kernel/drivers/net/xen-netback/
H A Drx.c42 static void xenvif_update_needed_slots(struct xenvif_queue *queue, in xenvif_update_needed_slots() argument
55 WRITE_ONCE(queue->rx_slots_needed, needed); in xenvif_update_needed_slots()
58 static bool xenvif_rx_ring_slots_available(struct xenvif_queue *queue) in xenvif_rx_ring_slots_available() argument
63 needed = READ_ONCE(queue->rx_slots_needed); in xenvif_rx_ring_slots_available()
68 prod = queue->rx.sring->req_prod; in xenvif_rx_ring_slots_available()
69 cons = queue->rx.req_cons; in xenvif_rx_ring_slots_available()
74 queue->rx.sring->req_event = prod + 1; in xenvif_rx_ring_slots_available()
80 } while (queue->rx.sring->req_prod != prod); in xenvif_rx_ring_slots_available()
85 bool xenvif_rx_queue_tail(struct xenvif_queue *queue, struct sk_buff *skb) in xenvif_rx_queue_tail() argument
90 spin_lock_irqsave(&queue->rx_queue.lock, flags); in xenvif_rx_queue_tail()
[all …]
H A Dinterface.c47 /* Number of bytes allowed on the internal guest Rx queue. */
55 void xenvif_skb_zerocopy_prepare(struct xenvif_queue *queue, in xenvif_skb_zerocopy_prepare() argument
59 atomic_inc(&queue->inflight_packets); in xenvif_skb_zerocopy_prepare()
62 void xenvif_skb_zerocopy_complete(struct xenvif_queue *queue) in xenvif_skb_zerocopy_complete() argument
64 atomic_dec(&queue->inflight_packets); in xenvif_skb_zerocopy_complete()
70 wake_up(&queue->dealloc_wq); in xenvif_skb_zerocopy_complete()
80 static bool xenvif_handle_tx_interrupt(struct xenvif_queue *queue) in xenvif_handle_tx_interrupt() argument
84 rc = RING_HAS_UNCONSUMED_REQUESTS(&queue->tx); in xenvif_handle_tx_interrupt()
86 napi_schedule(&queue->napi); in xenvif_handle_tx_interrupt()
92 struct xenvif_queue *queue = dev_id; in xenvif_tx_interrupt() local
[all …]
H A Dnetback.c58 /* The time that packets can stay on the guest Rx internal queue
106 static void xenvif_idx_release(struct xenvif_queue *queue, u16 pending_idx,
109 static void make_tx_response(struct xenvif_queue *queue,
113 static void push_tx_responses(struct xenvif_queue *queue);
115 static void xenvif_idx_unmap(struct xenvif_queue *queue, u16 pending_idx);
117 static inline int tx_work_todo(struct xenvif_queue *queue);
119 static inline unsigned long idx_to_pfn(struct xenvif_queue *queue, in idx_to_pfn() argument
122 return page_to_pfn(queue->mmap_pages[idx]); in idx_to_pfn()
125 static inline unsigned long idx_to_kaddr(struct xenvif_queue *queue, in idx_to_kaddr() argument
128 return (unsigned long)pfn_to_kaddr(idx_to_pfn(queue, idx)); in idx_to_kaddr()
[all …]
/OK3568_Linux_fs/kernel/drivers/nvme/host/
H A Dtcp.c81 struct nvme_tcp_queue *queue; member
176 static int nvme_tcp_try_send(struct nvme_tcp_queue *queue);
183 static inline int nvme_tcp_queue_id(struct nvme_tcp_queue *queue) in nvme_tcp_queue_id() argument
185 return queue - queue->ctrl->queues; in nvme_tcp_queue_id()
188 static inline struct blk_mq_tags *nvme_tcp_tagset(struct nvme_tcp_queue *queue) in nvme_tcp_tagset() argument
190 u32 queue_idx = nvme_tcp_queue_id(queue); in nvme_tcp_tagset()
193 return queue->ctrl->admin_tag_set.tags[queue_idx]; in nvme_tcp_tagset()
194 return queue->ctrl->tag_set.tags[queue_idx - 1]; in nvme_tcp_tagset()
197 static inline u8 nvme_tcp_hdgst_len(struct nvme_tcp_queue *queue) in nvme_tcp_hdgst_len() argument
199 return queue->hdr_digest ? NVME_TCP_DIGEST_LENGTH : 0; in nvme_tcp_hdgst_len()
[all …]
/OK3568_Linux_fs/kernel/drivers/net/
H A Dxen-netfront.c93 /* Queue name is interface name with "-qNNN" appended */
96 /* IRQ name is queue name with "-tx" or "-rx" appended */
110 unsigned int id; /* Queue ID, 0-based */
168 /* Multi-queue support */
220 static struct sk_buff *xennet_get_rx_skb(struct netfront_queue *queue, in xennet_get_rx_skb() argument
224 struct sk_buff *skb = queue->rx_skbs[i]; in xennet_get_rx_skb()
225 queue->rx_skbs[i] = NULL; in xennet_get_rx_skb()
229 static grant_ref_t xennet_get_rx_ref(struct netfront_queue *queue, in xennet_get_rx_ref() argument
233 grant_ref_t ref = queue->grant_rx_ref[i]; in xennet_get_rx_ref()
234 queue->grant_rx_ref[i] = GRANT_INVALID_REF; in xennet_get_rx_ref()
[all …]
/OK3568_Linux_fs/kernel/drivers/iio/buffer/
H A Dindustrialio-buffer-dma.c33 * means of two queues. The incoming queue and the outgoing queue. Blocks on the
34 * incoming queue are waiting for the DMA controller to pick them up and fill
35 * them with data. Block on the outgoing queue have been filled with data and
51 * incoming or outgoing queue the block will be freed.
100 dma_free_coherent(block->queue->dev, PAGE_ALIGN(block->size), in iio_buffer_block_release()
103 iio_buffer_put(&block->queue->buffer); in iio_buffer_block_release()
166 struct iio_dma_buffer_queue *queue, size_t size) in iio_dma_buffer_alloc_block() argument
174 block->vaddr = dma_alloc_coherent(queue->dev, PAGE_ALIGN(size), in iio_dma_buffer_alloc_block()
183 block->queue = queue; in iio_dma_buffer_alloc_block()
187 iio_buffer_get(&queue->buffer); in iio_dma_buffer_alloc_block()
[all …]
/OK3568_Linux_fs/kernel/drivers/misc/genwqe/
H A Dcard_ddcb.c14 * Device Driver Control Block (DDCB) queue support. Definition of
15 * interrupt handlers for queue support as well as triggering the
40 * Situation (1): Empty queue
56 * Situation (3): Queue wrapped, A > N
64 * Situation (4a): Queue full N > A
73 * Situation (4a): Queue full A > N
82 static int queue_empty(struct ddcb_queue *queue) in queue_empty() argument
84 return queue->ddcb_next == queue->ddcb_act; in queue_empty()
87 static int queue_enqueued_ddcbs(struct ddcb_queue *queue) in queue_enqueued_ddcbs() argument
89 if (queue->ddcb_next >= queue->ddcb_act) in queue_enqueued_ddcbs()
[all …]
/OK3568_Linux_fs/kernel/drivers/net/wireless/ralink/rt2x00/
H A Drt2x00queue.c12 Abstract: rt2x00 queue specific routines.
25 struct data_queue *queue = entry->queue; in rt2x00queue_alloc_rxskb() local
26 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev; in rt2x00queue_alloc_rxskb()
37 frame_size = queue->data_size + queue->desc_size + queue->winfo_size; in rt2x00queue_alloc_rxskb()
95 struct device *dev = entry->queue->rt2x00dev->dev; in rt2x00queue_map_txskb()
112 struct device *dev = entry->queue->rt2x00dev->dev; in rt2x00queue_unmap_skb()
490 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev; in rt2x00queue_write_tx_data()
495 * a queue corruption! in rt2x00queue_write_tx_data()
500 "Corrupt queue %d, accessing entry which is not ours\n" in rt2x00queue_write_tx_data()
502 entry->queue->qid, DRV_PROJECT); in rt2x00queue_write_tx_data()
[all …]
/OK3568_Linux_fs/kernel/drivers/crypto/hisilicon/sec/
H A Dsec_drv.c227 static int sec_queue_map_io(struct sec_queue *queue) in sec_queue_map_io() argument
229 struct device *dev = queue->dev_info->dev; in sec_queue_map_io()
234 2 + queue->queue_id); in sec_queue_map_io()
236 dev_err(dev, "Failed to get queue %d memory resource\n", in sec_queue_map_io()
237 queue->queue_id); in sec_queue_map_io()
240 queue->regs = ioremap(res->start, resource_size(res)); in sec_queue_map_io()
241 if (!queue->regs) in sec_queue_map_io()
247 static void sec_queue_unmap_io(struct sec_queue *queue) in sec_queue_unmap_io() argument
249 iounmap(queue->regs); in sec_queue_unmap_io()
252 static int sec_queue_ar_pkgattr(struct sec_queue *queue, u32 ar_pkg) in sec_queue_ar_pkgattr() argument
[all …]
/OK3568_Linux_fs/kernel/drivers/crypto/cavium/zip/
H A Dzip_device.c50 * zip_cmd_queue_consumed - Calculates the space consumed in the command queue.
53 * @queue: Queue number
55 * Return: Bytes consumed in the command queue buffer.
57 static inline u32 zip_cmd_queue_consumed(struct zip_device *zip_dev, int queue) in zip_cmd_queue_consumed() argument
59 return ((zip_dev->iq[queue].sw_head - zip_dev->iq[queue].sw_tail) * in zip_cmd_queue_consumed()
64 * zip_load_instr - Submits the instruction into the ZIP command queue
69 * This function copies the ZIP instruction to the command queue and rings the
71 * queue is maintained in a circular fashion. When there is space for exactly
72 * one instruction in the queue, next chunk pointer of the queue is made to
73 * point to the head of the queue, thus maintaining a circular queue.
[all …]
/OK3568_Linux_fs/kernel/drivers/soc/ixp4xx/
H A Dixp4xx-qmgr.c3 * Intel IXP4xx Queue Manager driver for Linux
28 void qmgr_put_entry(unsigned int queue, u32 val) in qmgr_put_entry() argument
31 BUG_ON(!qmgr_queue_descs[queue]); /* not yet requested */ in qmgr_put_entry()
33 printk(KERN_DEBUG "Queue %s(%i) put %X\n", in qmgr_put_entry()
34 qmgr_queue_descs[queue], queue, val); in qmgr_put_entry()
36 __raw_writel(val, &qmgr_regs->acc[queue][0]); in qmgr_put_entry()
39 u32 qmgr_get_entry(unsigned int queue) in qmgr_get_entry() argument
42 val = __raw_readl(&qmgr_regs->acc[queue][0]); in qmgr_get_entry()
44 BUG_ON(!qmgr_queue_descs[queue]); /* not yet requested */ in qmgr_get_entry()
46 printk(KERN_DEBUG "Queue %s(%i) get %X\n", in qmgr_get_entry()
[all …]
/OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_kcpu.c180 * @queue: The queue containing this JIT allocation
181 * @cmd: The JIT allocation that is blocking this queue
184 struct kbase_kcpu_command_queue *queue, in kbase_jit_add_to_pending_alloc_list() argument
187 struct kbase_context *const kctx = queue->kctx; in kbase_jit_add_to_pending_alloc_list()
192 lockdep_assert_held(&queue->lock); in kbase_jit_add_to_pending_alloc_list()
208 list_add_tail(&queue->jit_blocked, target_list_head); in kbase_jit_add_to_pending_alloc_list()
214 * @queue: The queue containing this JIT allocation
224 struct kbase_kcpu_command_queue *queue, in kbase_kcpu_jit_allocate_process() argument
227 struct kbase_context *const kctx = queue->kctx; in kbase_kcpu_jit_allocate_process()
238 lockdep_assert_held(&queue->lock); in kbase_kcpu_jit_allocate_process()
[all …]
/OK3568_Linux_fs/kernel/drivers/net/ethernet/ibm/ehea/
H A Dehea_qmr.h40 * WQE - Work Queue Entry
41 * SWQE - Send Work Queue Entry
42 * RWQE - Receive Work Queue Entry
43 * CQE - Completion Queue Entry
44 * EQE - Event Queue Entry
196 static inline void *hw_qeit_calc(struct hw_queue *queue, u64 q_offset) in hw_qeit_calc() argument
200 if (q_offset >= queue->queue_length) in hw_qeit_calc()
201 q_offset -= queue->queue_length; in hw_qeit_calc()
202 current_page = (queue->queue_pages)[q_offset >> EHEA_PAGESHIFT]; in hw_qeit_calc()
206 static inline void *hw_qeit_get(struct hw_queue *queue) in hw_qeit_get() argument
[all …]
/OK3568_Linux_fs/kernel/net/sunrpc/
H A Dsched.c76 * queue->lock and bh_disabled in order to avoid races within
80 __rpc_disable_timer(struct rpc_wait_queue *queue, struct rpc_task *task) in __rpc_disable_timer() argument
86 if (list_empty(&queue->timer_list.list)) in __rpc_disable_timer()
87 cancel_delayed_work(&queue->timer_list.dwork); in __rpc_disable_timer()
91 rpc_set_queue_timer(struct rpc_wait_queue *queue, unsigned long expires) in rpc_set_queue_timer() argument
94 queue->timer_list.expires = expires; in rpc_set_queue_timer()
99 mod_delayed_work(rpciod_workqueue, &queue->timer_list.dwork, expires); in rpc_set_queue_timer()
106 __rpc_add_timer(struct rpc_wait_queue *queue, struct rpc_task *task, in __rpc_add_timer() argument
110 if (list_empty(&queue->timer_list.list) || time_before(timeout, queue->timer_list.expires)) in __rpc_add_timer()
111 rpc_set_queue_timer(queue, timeout); in __rpc_add_timer()
[all …]
/OK3568_Linux_fs/kernel/drivers/net/wireguard/
H A Dqueueing.c25 int wg_packet_queue_init(struct crypt_queue *queue, work_func_t function, in wg_packet_queue_init() argument
30 memset(queue, 0, sizeof(*queue)); in wg_packet_queue_init()
31 ret = ptr_ring_init(&queue->ring, len, GFP_KERNEL); in wg_packet_queue_init()
34 queue->worker = wg_packet_percpu_multicore_worker_alloc(function, queue); in wg_packet_queue_init()
35 if (!queue->worker) { in wg_packet_queue_init()
36 ptr_ring_cleanup(&queue->ring, NULL); in wg_packet_queue_init()
42 void wg_packet_queue_free(struct crypt_queue *queue, bool purge) in wg_packet_queue_free() argument
44 free_percpu(queue->worker); in wg_packet_queue_free()
45 WARN_ON(!purge && !__ptr_ring_empty(&queue->ring)); in wg_packet_queue_free()
46 ptr_ring_cleanup(&queue->ring, purge ? __skb_array_destroy_skb : NULL); in wg_packet_queue_free()
[all …]
/OK3568_Linux_fs/kernel/drivers/net/wireless/rockchip_wlan/rkwifi/bcmdhd_indep_power/
H A Ddhd_flowring.h44 /* Max pkts held in a flow ring's backup queue */
92 /* Flow Ring Queue Enqueue overflow callback */
93 typedef int (*flow_queue_cb_t)(struct flow_queue * queue, void * pkt);
96 * Each flow ring has an associated (tx flow controlled) queue. 802.3 packets are transferred
97 * between queue and ring. A packet from the host stack is first added to the queue, and in a later
98 * stage transferred to the flow ring. Packets in the queue are dhd owned, whereas packets in the
102 dll_t list; /* manage a flowring queue in a double linked list */
103 void * head; /* first packet in the queue */
104 void * tail; /* last packet in the queue */
105 uint16 len; /* number of packets in the queue */
[all …]
/OK3568_Linux_fs/kernel/drivers/net/wireless/ath/ath5k/
H A Dqcu.c20 Queue Control Unit, DCF Control Unit Functions
31 * DOC: Queue Control Unit (QCU)/DCF Control Unit (DCU) functions
35 * basically we have 10 queues to play with. Each queue has a matching
36 * QCU that controls when the queue will get triggered and multiple QCUs
39 * and DCUs allowing us to have different DFS settings for each queue.
41 * When a frame goes into a TX queue, QCU decides when it'll trigger a
43 * it's buffer or -if it's a beacon queue- if it's time to fire up the queue
58 * ath5k_hw_num_tx_pending() - Get number of pending frames for a given queue
60 * @queue: One of enum ath5k_tx_queue_id
63 ath5k_hw_num_tx_pending(struct ath5k_hw *ah, unsigned int queue) in ath5k_hw_num_tx_pending() argument
[all …]

12345678910>>...177