Lines Matching refs:phba

57 lpfc_get_nvme_buf(struct lpfc_hba *phba, struct lpfc_nodelist *ndlp,
376 spin_lock_irq(&vport->phba->hbalock); in lpfc_nvme_remoteport_delete()
384 spin_unlock_irq(&vport->phba->hbalock); in lpfc_nvme_remoteport_delete()
391 spin_unlock_irq(&vport->phba->hbalock); in lpfc_nvme_remoteport_delete()
415 lpfc_nvme_handle_lsreq(struct lpfc_hba *phba, in lpfc_nvme_handle_lsreq() argument
442 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_DISC, in lpfc_nvme_handle_lsreq()
469 __lpfc_nvme_ls_req_cmp(struct lpfc_hba *phba, struct lpfc_vport *vport, in __lpfc_nvme_ls_req_cmp() argument
491 lpfc_nvmeio_data(phba, "NVMEx LS CMPL: xri x%x stat x%x parm x%x\n", in __lpfc_nvme_ls_req_cmp()
496 lpfc_mbuf_free(phba, buf_ptr->virt, buf_ptr->phys); in __lpfc_nvme_ls_req_cmp()
512 lpfc_sli_release_iocbq(phba, cmdwqe); in __lpfc_nvme_ls_req_cmp()
516 lpfc_nvme_ls_req_cmp(struct lpfc_hba *phba, struct lpfc_iocbq *cmdwqe, in lpfc_nvme_ls_req_cmp() argument
537 __lpfc_nvme_ls_req_cmp(phba, vport, cmdwqe, wcqe); in lpfc_nvme_ls_req_cmp()
549 struct lpfc_hba *phba = vport->phba; in lpfc_nvme_gen_req() local
557 genwqe = lpfc_sli_get_iocbq(phba); in lpfc_nvme_gen_req()
575 tmo = (3 * phba->fc_ratov); in lpfc_nvme_gen_req()
613 phba->sli4_hba.rpi_ids[ndlp->nlp_rpi]); in lpfc_nvme_gen_req()
617 bf_set(wqe_tmo, &wqe->gen_req.wqe_com, (vport->phba->fc_ratov-1)); in lpfc_nvme_gen_req()
647 lpfc_nvmeio_data(phba, "NVME LS XMIT: xri x%x iotag x%x to x%06x\n", in lpfc_nvme_gen_req()
650 rc = lpfc_sli4_issue_wqe(phba, &phba->sli4_hba.hdwq[0], genwqe); in lpfc_nvme_gen_req()
657 lpfc_sli_release_iocbq(phba, genwqe); in lpfc_nvme_gen_req()
689 void (*gen_req_cmp)(struct lpfc_hba *phba, in __lpfc_nvme_ls_req() argument
717 if (!vport->phba->sli4_hba.nvmels_wq) in __lpfc_nvme_ls_req()
741 bmp->virt = lpfc_mbuf_alloc(vport->phba, MEM_PRI, &(bmp->phys)); in __lpfc_nvme_ls_req()
782 lpfc_mbuf_free(vport->phba, bmp->virt, bmp->phys); in __lpfc_nvme_ls_req()
850 struct lpfc_hba *phba = vport->phba; in __lpfc_nvme_ls_abort() local
874 pring = phba->sli4_hba.nvmels_wq->pring; in __lpfc_nvme_ls_abort()
875 spin_lock_irq(&phba->hbalock); in __lpfc_nvme_ls_abort()
887 lpfc_sli_issue_abort_iotag(phba, pring, wqe); in __lpfc_nvme_ls_abort()
888 spin_unlock_irq(&phba->hbalock); in __lpfc_nvme_ls_abort()
909 if (axchg->phba->pport->load_flag & FC_UNLOADING) in lpfc_nvme_xmt_ls_rsp()
946 struct lpfc_hba *phba; in lpfc_nvme_ls_abort() local
954 phba = vport->phba; in lpfc_nvme_ls_abort()
972 struct lpfc_hba *phba = vport->phba; in lpfc_nvme_adj_fcp_sgls() local
994 if (phba->cfg_nvme_embed_cmd) { in lpfc_nvme_adj_fcp_sgls()
1076 lpfc_nvme_io_cmd_wqe_cmpl(struct lpfc_hba *phba, struct lpfc_iocbq *pwqeIn, in lpfc_nvme_io_cmd_wqe_cmpl() argument
1114 lpfc_release_nvme_buf(phba, lpfc_ncmd); in lpfc_nvme_io_cmd_wqe_cmpl()
1121 phba->sli4_hba.hdwq[idx].nvme_cstat.io_cmpls++; in lpfc_nvme_io_cmd_wqe_cmpl()
1132 lpfc_nvmeio_data(phba, "NVME FCP CMPL: xri x%x stat x%x parm x%x\n", in lpfc_nvme_io_cmd_wqe_cmpl()
1262 phba->ktime_last_cmd = lpfc_ncmd->ts_data_io; in lpfc_nvme_io_cmd_wqe_cmpl()
1263 lpfc_io_ktime(phba, lpfc_ncmd); in lpfc_nvme_io_cmd_wqe_cmpl()
1265 if (unlikely(phba->hdwqstat_on & LPFC_CHECK_NVME_IO)) { in lpfc_nvme_io_cmd_wqe_cmpl()
1267 this_cpu_inc(phba->sli4_hba.c_stat->cmpl_io); in lpfc_nvme_io_cmd_wqe_cmpl()
1291 lpfc_release_nvme_buf(phba, lpfc_ncmd); in lpfc_nvme_io_cmd_wqe_cmpl()
1316 struct lpfc_hba *phba = vport->phba; in lpfc_nvme_prep_io_cmd() local
1341 if ((phba->cfg_nvme_enable_fb) && in lpfc_nvme_prep_io_cmd()
1394 phba->sli4_hba.rpi_ids[pnode->nlp_rpi]); in lpfc_nvme_prep_io_cmd()
1427 struct lpfc_hba *phba = vport->phba; in lpfc_nvme_prep_io_dma() local
1459 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT, in lpfc_nvme_prep_io_dma()
1463 phba->cfg_nvme_seg_cnt + 1, in lpfc_nvme_prep_io_dma()
1482 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT, in lpfc_nvme_prep_io_dma()
1499 !((j + 1) % phba->border_sge_num) && in lpfc_nvme_prep_io_dma()
1506 phba, lpfc_ncmd); in lpfc_nvme_prep_io_dma()
1549 phba->cfg_sg_dma_buf_size); in lpfc_nvme_prep_io_dma()
1559 if (phba->cfg_enable_pbde) { in lpfc_nvme_prep_io_dma()
1583 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT, in lpfc_nvme_prep_io_dma()
1622 struct lpfc_hba *phba; in lpfc_nvme_fcp_io_submit() local
1652 phba = vport->phba; in lpfc_nvme_fcp_io_submit()
1672 if (phba->ktime_on) in lpfc_nvme_fcp_io_submit()
1722 if (lpfc_ndlp_check_qdepth(phba, ndlp)) { in lpfc_nvme_fcp_io_submit()
1738 if (phba->cfg_fcp_io_sched == LPFC_FCP_SCHED_BY_HDWQ) { in lpfc_nvme_fcp_io_submit()
1742 idx = phba->sli4_hba.cpu_map[cpu].hdwq; in lpfc_nvme_fcp_io_submit()
1745 lpfc_ncmd = lpfc_get_nvme_buf(phba, ndlp, idx, expedite); in lpfc_nvme_fcp_io_submit()
1758 lpfc_ncmd->ts_last_cmd = phba->ktime_last_cmd; in lpfc_nvme_fcp_io_submit()
1784 cstat = &phba->sli4_hba.hdwq[idx].nvme_cstat; in lpfc_nvme_fcp_io_submit()
1798 lpfc_nvmeio_data(phba, "NVME FCP XMIT: xri x%x idx %d to %06x\n", in lpfc_nvme_fcp_io_submit()
1802 ret = lpfc_sli4_issue_wqe(phba, lpfc_ncmd->hdwq, &lpfc_ncmd->cur_iocbq); in lpfc_nvme_fcp_io_submit()
1813 if (phba->cfg_xri_rebalancing) in lpfc_nvme_fcp_io_submit()
1814 lpfc_keep_pvt_pool_above_lowwm(phba, lpfc_ncmd->hdwq_no); in lpfc_nvme_fcp_io_submit()
1820 if (phba->hdwqstat_on & LPFC_CHECK_NVME_IO) { in lpfc_nvme_fcp_io_submit()
1822 this_cpu_inc(phba->sli4_hba.c_stat->xmt_io); in lpfc_nvme_fcp_io_submit()
1843 lpfc_release_nvme_buf(phba, lpfc_ncmd); in lpfc_nvme_fcp_io_submit()
1860 lpfc_nvme_abort_fcreq_cmpl(struct lpfc_hba *phba, struct lpfc_iocbq *cmdiocb, in lpfc_nvme_abort_fcreq_cmpl() argument
1863 lpfc_printf_log(phba, KERN_INFO, LOG_NVME, in lpfc_nvme_abort_fcreq_cmpl()
1873 lpfc_sli_release_iocbq(phba, cmdiocb); in lpfc_nvme_abort_fcreq_cmpl()
1901 struct lpfc_hba *phba; in lpfc_nvme_fcp_abort() local
1924 phba = vport->phba; in lpfc_nvme_fcp_abort()
1942 spin_lock_irqsave(&phba->hbalock, flags); in lpfc_nvme_fcp_abort()
1944 if (phba->hba_flag & HBA_IOQ_FLUSH) { in lpfc_nvme_fcp_abort()
1945 spin_unlock_irqrestore(&phba->hbalock, flags); in lpfc_nvme_fcp_abort()
1949 phba->hba_flag); in lpfc_nvme_fcp_abort()
1955 spin_unlock_irqrestore(&phba->hbalock, flags); in lpfc_nvme_fcp_abort()
1961 spin_unlock_irqrestore(&phba->hbalock, flags); in lpfc_nvme_fcp_abort()
1999 lpfc_nvmeio_data(phba, "NVME FCP ABORT: xri x%x idx %d to %06x\n", in lpfc_nvme_fcp_abort()
2014 abts_buf = __lpfc_sli_get_iocbq(phba); in lpfc_nvme_fcp_abort()
2033 ret_val = lpfc_sli4_issue_wqe(phba, lpfc_nbuf->hdwq, abts_buf); in lpfc_nvme_fcp_abort()
2035 spin_unlock_irqrestore(&phba->hbalock, flags); in lpfc_nvme_fcp_abort()
2041 lpfc_sli_release_iocbq(phba, abts_buf); in lpfc_nvme_fcp_abort()
2054 spin_unlock_irqrestore(&phba->hbalock, flags); in lpfc_nvme_fcp_abort()
2097 lpfc_get_nvme_buf(struct lpfc_hba *phba, struct lpfc_nodelist *ndlp, in lpfc_get_nvme_buf() argument
2106 lpfc_ncmd = lpfc_get_io_buf(phba, NULL, idx, expedite); in lpfc_get_nvme_buf()
2133 if (lpfc_ndlp_check_qdepth(phba, ndlp)) { in lpfc_get_nvme_buf()
2139 qp = &phba->sli4_hba.hdwq[idx]; in lpfc_get_nvme_buf()
2157 lpfc_release_nvme_buf(struct lpfc_hba *phba, struct lpfc_io_buf *lpfc_ncmd) in lpfc_release_nvme_buf() argument
2170 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_release_nvme_buf()
2182 lpfc_release_io_buf(phba, (struct lpfc_io_buf *)lpfc_ncmd, qp); in lpfc_release_nvme_buf()
2205 struct lpfc_hba *phba = vport->phba; in lpfc_nvme_create_localport() local
2222 lpfc_nvme_template.max_sgl_segments = phba->cfg_nvme_seg_cnt + 1; in lpfc_nvme_create_localport()
2227 lpfc_nvme_template.max_hw_queues = phba->cfg_hdw_queue; in lpfc_nvme_create_localport()
2237 &vport->phba->pcidev->dev, &localport); in lpfc_nvme_create_localport()
2292 struct lpfc_hba *phba = vport->phba; in lpfc_nvme_lport_unreg_wait() local
2307 for (i = 0; i < phba->cfg_hdw_queue; i++) { in lpfc_nvme_lport_unreg_wait()
2308 qp = &phba->sli4_hba.hdwq[i]; in lpfc_nvme_lport_unreg_wait()
2475 spin_lock_irq(&vport->phba->hbalock); in lpfc_nvme_register_port()
2479 spin_unlock_irq(&vport->phba->hbalock); in lpfc_nvme_register_port()
2481 spin_unlock_irq(&vport->phba->hbalock); in lpfc_nvme_register_port()
2494 spin_lock_irq(&vport->phba->hbalock); in lpfc_nvme_register_port()
2496 spin_unlock_irq(&vport->phba->hbalock); in lpfc_nvme_register_port()
2504 spin_lock_irq(&vport->phba->hbalock); in lpfc_nvme_register_port()
2507 spin_unlock_irq(&vport->phba->hbalock); in lpfc_nvme_register_port()
2526 spin_lock_irq(&vport->phba->hbalock); in lpfc_nvme_register_port()
2528 spin_unlock_irq(&vport->phba->hbalock); in lpfc_nvme_register_port()
2567 spin_lock_irq(&vport->phba->hbalock); in lpfc_nvme_rescan_port()
2571 spin_unlock_irq(&vport->phba->hbalock); in lpfc_nvme_rescan_port()
2634 spin_lock_irq(&vport->phba->hbalock); in lpfc_nvme_unregister_port()
2638 spin_unlock_irq(&vport->phba->hbalock); in lpfc_nvme_unregister_port()
2696 lpfc_sli4_nvme_xri_aborted(struct lpfc_hba *phba, in lpfc_sli4_nvme_xri_aborted() argument
2706 lpfc_sli4_abts_err_handler(phba, ndlp, axri); in lpfc_sli4_nvme_xri_aborted()
2708 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_ABTS, in lpfc_sli4_nvme_xri_aborted()
2723 lpfc_release_nvme_buf(phba, lpfc_ncmd); in lpfc_sli4_nvme_xri_aborted()
2737 lpfc_nvme_wait_for_io_drain(struct lpfc_hba *phba) in lpfc_nvme_wait_for_io_drain() argument
2742 if (phba->sli_rev < LPFC_SLI_REV4 || !phba->sli4_hba.hdwq) in lpfc_nvme_wait_for_io_drain()
2748 for (i = 0; i < phba->cfg_hdw_queue; i++) { in lpfc_nvme_wait_for_io_drain()
2749 if (!phba->sli4_hba.hdwq[i].io_wq) in lpfc_nvme_wait_for_io_drain()
2751 pring = phba->sli4_hba.hdwq[i].io_wq->pring; in lpfc_nvme_wait_for_io_drain()
2765 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT, in lpfc_nvme_wait_for_io_drain()
2774 lpfc_nvme_cancel_iocb(struct lpfc_hba *phba, struct lpfc_iocbq *pwqeIn) in lpfc_nvme_cancel_iocb() argument
2782 lpfc_sli_release_iocbq(phba, pwqeIn); in lpfc_nvme_cancel_iocb()
2788 lpfc_sli_release_iocbq(phba, pwqeIn); in lpfc_nvme_cancel_iocb()
2796 lpfc_release_nvme_buf(phba, lpfc_ncmd); in lpfc_nvme_cancel_iocb()
2801 lpfc_printf_log(phba, KERN_INFO, LOG_NVME_IOERR, in lpfc_nvme_cancel_iocb()
2816 lpfc_release_nvme_buf(phba, lpfc_ncmd); in lpfc_nvme_cancel_iocb()