Lines Matching refs:ib_cq
91 struct ib_cq *ib_cq; member
153 static void nvme_rdma_recv_done(struct ib_cq *cq, struct ib_wc *wc);
276 init_attr.send_cq = queue->ib_cq; in nvme_rdma_create_qp()
277 init_attr.recv_cq = queue->ib_cq; in nvme_rdma_create_qp()
418 ib_free_cq(queue->ib_cq); in nvme_rdma_free_cq()
420 ib_cq_pool_put(queue->ib_cq, queue->cq_size); in nvme_rdma_free_cq()
479 queue->ib_cq = ib_alloc_cq(ibdev, queue, queue->cq_size, in nvme_rdma_create_cq()
483 queue->ib_cq = ib_cq_pool_get(ibdev, queue->cq_size, in nvme_rdma_create_cq()
487 if (IS_ERR(queue->ib_cq)) { in nvme_rdma_create_cq()
488 ret = PTR_ERR(queue->ib_cq); in nvme_rdma_create_cq()
1246 static void nvme_rdma_wr_error(struct ib_cq *cq, struct ib_wc *wc, in nvme_rdma_wr_error()
1260 static void nvme_rdma_memreg_done(struct ib_cq *cq, struct ib_wc *wc) in nvme_rdma_memreg_done()
1266 static void nvme_rdma_inv_rkey_done(struct ib_cq *cq, struct ib_wc *wc) in nvme_rdma_inv_rkey_done()
1472 static void nvme_rdma_sig_done(struct ib_cq *cq, struct ib_wc *wc) in nvme_rdma_sig_done()
1634 static void nvme_rdma_send_done(struct ib_cq *cq, struct ib_wc *wc) in nvme_rdma_send_done()
1713 static void nvme_rdma_async_done(struct ib_cq *cq, struct ib_wc *wc) in nvme_rdma_async_done()
1790 static void nvme_rdma_recv_done(struct ib_cq *cq, struct ib_wc *wc) in nvme_rdma_recv_done()
2129 return ib_process_cq_direct(queue->ib_cq, -1); in nvme_rdma_poll()