Lines Matching refs:hrrq
697 struct ipr_cmnd *__ipr_get_free_ipr_cmnd(struct ipr_hrr_queue *hrrq) in __ipr_get_free_ipr_cmnd() argument
701 if (likely(!list_empty(&hrrq->hrrq_free_q))) { in __ipr_get_free_ipr_cmnd()
702 ipr_cmd = list_entry(hrrq->hrrq_free_q.next, in __ipr_get_free_ipr_cmnd()
722 __ipr_get_free_ipr_cmnd(&ioa_cfg->hrrq[IPR_INIT_HRRQ]); in ipr_get_free_ipr_cmnd()
745 spin_lock(&ioa_cfg->hrrq[i]._lock); in ipr_mask_and_clear_interrupts()
746 ioa_cfg->hrrq[i].allow_interrupts = 0; in ipr_mask_and_clear_interrupts()
747 spin_unlock(&ioa_cfg->hrrq[i]._lock); in ipr_mask_and_clear_interrupts()
829 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in __ipr_sata_eh_done()
844 struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq; in ipr_sata_eh_done() local
847 spin_lock_irqsave(&hrrq->_lock, hrrq_flags); in ipr_sata_eh_done()
849 spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags); in ipr_sata_eh_done()
872 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in __ipr_scsi_eh_done()
888 struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq; in ipr_scsi_eh_done() local
890 spin_lock_irqsave(&hrrq->_lock, hrrq_flags); in ipr_scsi_eh_done()
892 spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags); in ipr_scsi_eh_done()
907 struct ipr_hrr_queue *hrrq; in ipr_fail_all_ops() local
910 for_each_hrrq(hrrq, ioa_cfg) { in ipr_fail_all_ops()
911 spin_lock(&hrrq->_lock); in ipr_fail_all_ops()
913 temp, &hrrq->hrrq_pending_q, queue) { in ipr_fail_all_ops()
931 spin_unlock(&hrrq->_lock); in ipr_fail_all_ops()
982 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); in ipr_do_req()
1084 unsigned int hrrq; in ipr_get_hrrq_index() local
1087 hrrq = 0; in ipr_get_hrrq_index()
1089 hrrq = atomic_add_return(1, &ioa_cfg->hrrq_index); in ipr_get_hrrq_index()
1090 hrrq = (hrrq % (ioa_cfg->hrrq_num - 1)) + 1; in ipr_get_hrrq_index()
1092 return hrrq; in ipr_get_hrrq_index()
1114 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds) { in ipr_send_hcam()
1116 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); in ipr_send_hcam()
1504 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_process_ccn()
2648 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_process_error()
3339 if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds) { in ipr_add_remove_thread()
3654 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) in ipr_show_adapter_state()
3687 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead && in ipr_store_adapter_state()
3690 spin_lock(&ioa_cfg->hrrq[i]._lock); in ipr_store_adapter_state()
3691 ioa_cfg->hrrq[i].ioa_is_dead = 0; in ipr_store_adapter_state()
3692 spin_unlock(&ioa_cfg->hrrq[i]._lock); in ipr_store_adapter_state()
3819 irq_poll_disable(&ioa_cfg->hrrq[i].iopoll); in ipr_store_iopoll_weight()
3826 irq_poll_init(&ioa_cfg->hrrq[i].iopoll, in ipr_store_iopoll_weight()
4402 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead && !ioa_cfg->dump_taken) { in ipr_alloc_dump()
5086 list_for_each_entry(loop_cmd, &ipr_cmd->hrrq->hrrq_free_q, queue) { in ipr_cmnd_is_free()
5126 struct ipr_hrr_queue *hrrq; in ipr_wait_for_ops() local
5134 for_each_hrrq(hrrq, ioa_cfg) { in ipr_wait_for_ops()
5135 spin_lock_irqsave(hrrq->lock, flags); in ipr_wait_for_ops()
5136 for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) { in ipr_wait_for_ops()
5145 spin_unlock_irqrestore(hrrq->lock, flags); in ipr_wait_for_ops()
5154 for_each_hrrq(hrrq, ioa_cfg) { in ipr_wait_for_ops()
5155 spin_lock_irqsave(hrrq->lock, flags); in ipr_wait_for_ops()
5156 for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) { in ipr_wait_for_ops()
5165 spin_unlock_irqrestore(hrrq->lock, flags); in ipr_wait_for_ops()
5190 if (!ioa_cfg->in_reset_reload && !ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) { in ipr_eh_host_reset()
5205 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) { in ipr_eh_host_reset()
5260 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_device_reset()
5341 struct ipr_hrr_queue *hrrq; in __ipr_eh_dev_reset() local
5354 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) in __ipr_eh_dev_reset()
5357 for_each_hrrq(hrrq, ioa_cfg) { in __ipr_eh_dev_reset()
5358 spin_lock(&hrrq->_lock); in __ipr_eh_dev_reset()
5359 for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) { in __ipr_eh_dev_reset()
5375 spin_unlock(&hrrq->_lock); in __ipr_eh_dev_reset()
5452 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_bus_reset_done()
5514 struct ipr_hrr_queue *hrrq; in ipr_cancel_op() local
5525 ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) in ipr_cancel_op()
5540 for_each_hrrq(hrrq, ioa_cfg) { in ipr_cancel_op()
5541 spin_lock(&hrrq->_lock); in ipr_cancel_op()
5542 for (i = hrrq->min_cmd_id; i <= hrrq->max_cmd_id; i++) { in ipr_cancel_op()
5550 spin_unlock(&hrrq->_lock); in ipr_cancel_op()
5577 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_cancel_op()
5600 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead || ioa_cfg->scan_done) in ipr_scan_finished()
5782 struct ipr_hrr_queue *hrrq; in ipr_iopoll() local
5788 hrrq = container_of(iop, struct ipr_hrr_queue, iopoll); in ipr_iopoll()
5790 spin_lock_irqsave(hrrq->lock, hrrq_flags); in ipr_iopoll()
5791 completed_ops = ipr_process_hrrq(hrrq, budget, &doneq); in ipr_iopoll()
5795 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_iopoll()
5816 struct ipr_hrr_queue *hrrq = (struct ipr_hrr_queue *)devp; in ipr_isr() local
5817 struct ipr_ioa_cfg *ioa_cfg = hrrq->ioa_cfg; in ipr_isr()
5826 spin_lock_irqsave(hrrq->lock, hrrq_flags); in ipr_isr()
5828 if (!hrrq->allow_interrupts) { in ipr_isr()
5829 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_isr()
5834 if (ipr_process_hrrq(hrrq, -1, &doneq)) { in ipr_isr()
5865 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_isr()
5884 struct ipr_hrr_queue *hrrq = (struct ipr_hrr_queue *)devp; in ipr_isr_mhrrq() local
5885 struct ipr_ioa_cfg *ioa_cfg = hrrq->ioa_cfg; in ipr_isr_mhrrq()
5891 spin_lock_irqsave(hrrq->lock, hrrq_flags); in ipr_isr_mhrrq()
5894 if (!hrrq->allow_interrupts) { in ipr_isr_mhrrq()
5895 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_isr_mhrrq()
5900 if ((be32_to_cpu(*hrrq->hrrq_curr) & IPR_HRRQ_TOGGLE_BIT) == in ipr_isr_mhrrq()
5901 hrrq->toggle_bit) { in ipr_isr_mhrrq()
5902 irq_poll_sched(&hrrq->iopoll); in ipr_isr_mhrrq()
5903 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_isr_mhrrq()
5907 if ((be32_to_cpu(*hrrq->hrrq_curr) & IPR_HRRQ_TOGGLE_BIT) == in ipr_isr_mhrrq()
5908 hrrq->toggle_bit) in ipr_isr_mhrrq()
5910 if (ipr_process_hrrq(hrrq, -1, &doneq)) in ipr_isr_mhrrq()
5914 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_isr_mhrrq()
6071 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in __ipr_erp_done()
6086 struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq; in ipr_erp_done() local
6089 spin_lock_irqsave(&hrrq->_lock, hrrq_flags); in ipr_erp_done()
6091 spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags); in ipr_erp_done()
6173 struct ipr_hrr_queue *hrrq = ipr_cmd->hrrq; in ipr_erp_request_sense() local
6176 spin_lock_irqsave(&hrrq->_lock, hrrq_flags); in ipr_erp_request_sense()
6178 spin_unlock_irqrestore(&hrrq->_lock, hrrq_flags); in ipr_erp_request_sense()
6508 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_erp_start()
6533 spin_lock_irqsave(ipr_cmd->hrrq->lock, lock_flags); in ipr_scsi_done()
6537 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_scsi_done()
6538 spin_unlock_irqrestore(ipr_cmd->hrrq->lock, lock_flags); in ipr_scsi_done()
6541 spin_lock(&ipr_cmd->hrrq->_lock); in ipr_scsi_done()
6543 spin_unlock(&ipr_cmd->hrrq->_lock); in ipr_scsi_done()
6569 struct ipr_hrr_queue *hrrq; in ipr_queuecommand() local
6585 hrrq = &ioa_cfg->hrrq[hrrq_id]; in ipr_queuecommand()
6587 spin_lock_irqsave(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6593 if (unlikely(!hrrq->allow_cmds && !hrrq->ioa_is_dead && !hrrq->removing_ioa)) { in ipr_queuecommand()
6594 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6602 if (unlikely(hrrq->ioa_is_dead || hrrq->removing_ioa || !res)) { in ipr_queuecommand()
6603 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6607 ipr_cmd = __ipr_get_free_ipr_cmnd(hrrq); in ipr_queuecommand()
6609 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6612 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6657 spin_lock_irqsave(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6658 if (unlikely(rc || (!hrrq->allow_cmds && !hrrq->ioa_is_dead))) { in ipr_queuecommand()
6659 list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_free_q); in ipr_queuecommand()
6660 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6666 if (unlikely(hrrq->ioa_is_dead)) { in ipr_queuecommand()
6667 list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_free_q); in ipr_queuecommand()
6668 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6678 list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_pending_q); in ipr_queuecommand()
6681 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6685 spin_lock_irqsave(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6689 spin_unlock_irqrestore(hrrq->lock, hrrq_flags); in ipr_queuecommand()
6791 if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds) in ipr_ata_phy_reset()
6822 struct ipr_hrr_queue *hrrq; in ipr_ata_post_internal() local
6832 for_each_hrrq(hrrq, ioa_cfg) { in ipr_ata_post_internal()
6833 spin_lock(&hrrq->_lock); in ipr_ata_post_internal()
6834 list_for_each_entry(ipr_cmd, &hrrq->hrrq_pending_q, queue) { in ipr_ata_post_internal()
6840 spin_unlock(&hrrq->_lock); in ipr_ata_post_internal()
6889 spin_lock(&ipr_cmd->hrrq->_lock); in ipr_sata_done()
6905 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_sata_done()
6906 spin_unlock(&ipr_cmd->hrrq->_lock); in ipr_sata_done()
7014 struct ipr_hrr_queue *hrrq; in ipr_qc_defer() local
7018 hrrq = &ioa_cfg->hrrq[hrrq_id]; in ipr_qc_defer()
7021 spin_lock(&hrrq->_lock); in ipr_qc_defer()
7022 if (unlikely(hrrq->ioa_is_dead)) { in ipr_qc_defer()
7023 spin_unlock(&hrrq->_lock); in ipr_qc_defer()
7027 if (unlikely(!hrrq->allow_cmds)) { in ipr_qc_defer()
7028 spin_unlock(&hrrq->_lock); in ipr_qc_defer()
7032 ipr_cmd = __ipr_get_free_ipr_cmnd(hrrq); in ipr_qc_defer()
7034 spin_unlock(&hrrq->_lock); in ipr_qc_defer()
7039 spin_unlock(&hrrq->_lock); in ipr_qc_defer()
7068 spin_lock(&ipr_cmd->hrrq->_lock); in ipr_qc_issue()
7069 if (unlikely(!ipr_cmd->hrrq->allow_cmds || in ipr_qc_issue()
7070 ipr_cmd->hrrq->ioa_is_dead)) { in ipr_qc_issue()
7071 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_qc_issue()
7072 spin_unlock(&ipr_cmd->hrrq->_lock); in ipr_qc_issue()
7088 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); in ipr_qc_issue()
7128 spin_unlock(&ipr_cmd->hrrq->_lock); in ipr_qc_issue()
7133 spin_unlock(&ipr_cmd->hrrq->_lock); in ipr_qc_issue()
7242 if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].removing_ioa) { in ipr_ioa_bringdown_done()
7251 spin_lock(&ioa_cfg->hrrq[i]._lock); in ipr_ioa_bringdown_done()
7252 ioa_cfg->hrrq[i].ioa_is_dead = 1; in ipr_ioa_bringdown_done()
7253 spin_unlock(&ioa_cfg->hrrq[i]._lock); in ipr_ioa_bringdown_done()
7257 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_ioa_bringdown_done()
7284 spin_lock(&ioa_cfg->hrrq[j]._lock); in ipr_ioa_reset_done()
7285 ioa_cfg->hrrq[j].allow_cmds = 1; in ipr_ioa_reset_done()
7286 spin_unlock(&ioa_cfg->hrrq[j]._lock); in ipr_ioa_reset_done()
7316 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_ioa_reset_done()
7651 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_reset_cmd_failed()
8164 &ioa_cfg->hrrq->hrrq_free_q); in ipr_ioafp_page0_inquiry()
8217 struct ipr_hrr_queue *hrrq; in ipr_ioafp_identify_hrrq() local
8225 hrrq = &ioa_cfg->hrrq[ioa_cfg->identify_hrrq_index]; in ipr_ioafp_identify_hrrq()
8240 ((u64) hrrq->host_rrq_dma >> 24) & 0xff; in ipr_ioafp_identify_hrrq()
8242 ((u64) hrrq->host_rrq_dma >> 16) & 0xff; in ipr_ioafp_identify_hrrq()
8244 ((u64) hrrq->host_rrq_dma >> 8) & 0xff; in ipr_ioafp_identify_hrrq()
8246 ((u64) hrrq->host_rrq_dma) & 0xff; in ipr_ioafp_identify_hrrq()
8248 ((sizeof(u32) * hrrq->size) >> 8) & 0xff; in ipr_ioafp_identify_hrrq()
8250 (sizeof(u32) * hrrq->size) & 0xff; in ipr_ioafp_identify_hrrq()
8258 ((u64) hrrq->host_rrq_dma >> 56) & 0xff; in ipr_ioafp_identify_hrrq()
8260 ((u64) hrrq->host_rrq_dma >> 48) & 0xff; in ipr_ioafp_identify_hrrq()
8262 ((u64) hrrq->host_rrq_dma >> 40) & 0xff; in ipr_ioafp_identify_hrrq()
8264 ((u64) hrrq->host_rrq_dma >> 32) & 0xff; in ipr_ioafp_identify_hrrq()
8333 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); in ipr_reset_start_timer()
8350 struct ipr_hrr_queue *hrrq; in ipr_init_ioa_mem() local
8352 for_each_hrrq(hrrq, ioa_cfg) { in ipr_init_ioa_mem()
8353 spin_lock(&hrrq->_lock); in ipr_init_ioa_mem()
8354 memset(hrrq->host_rrq, 0, sizeof(u32) * hrrq->size); in ipr_init_ioa_mem()
8357 hrrq->hrrq_start = hrrq->host_rrq; in ipr_init_ioa_mem()
8358 hrrq->hrrq_end = &hrrq->host_rrq[hrrq->size - 1]; in ipr_init_ioa_mem()
8359 hrrq->hrrq_curr = hrrq->hrrq_start; in ipr_init_ioa_mem()
8360 hrrq->toggle_bit = 1; in ipr_init_ioa_mem()
8361 spin_unlock(&hrrq->_lock); in ipr_init_ioa_mem()
8426 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); in ipr_reset_next_stage()
8453 spin_lock(&ioa_cfg->hrrq[i]._lock); in ipr_reset_enable_ioa()
8454 ioa_cfg->hrrq[i].allow_interrupts = 1; in ipr_reset_enable_ioa()
8455 spin_unlock(&ioa_cfg->hrrq[i]._lock); in ipr_reset_enable_ioa()
8495 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); in ipr_reset_enable_ioa()
9016 struct ipr_hrr_queue *hrrq; in ipr_reset_cancel_hcam_done() local
9023 for_each_hrrq(hrrq, ioa_cfg) { in ipr_reset_cancel_hcam_done()
9024 spin_lock(&hrrq->_lock); in ipr_reset_cancel_hcam_done()
9025 list_for_each_entry(loop_cmd, &hrrq->hrrq_pending_q, queue) { in ipr_reset_cancel_hcam_done()
9028 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_reset_cancel_hcam_done()
9032 spin_unlock(&hrrq->_lock); in ipr_reset_cancel_hcam_done()
9057 struct ipr_hrr_queue *hrrq = &ioa_cfg->hrrq[IPR_INIT_HRRQ]; in ipr_reset_cancel_hcam() local
9062 if (!hrrq->ioa_is_dead) { in ipr_reset_cancel_hcam()
9064 list_for_each_entry(hcam_cmd, &hrrq->hrrq_pending_q, queue) { in ipr_reset_cancel_hcam()
9183 !ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) { in ipr_reset_shutdown_ioa()
9232 &ipr_cmd->hrrq->hrrq_free_q); in ipr_reset_ioa_job()
9271 spin_lock(&ioa_cfg->hrrq[i]._lock); in _ipr_initiate_ioa_reset()
9272 ioa_cfg->hrrq[i].allow_cmds = 0; in _ipr_initiate_ioa_reset()
9273 spin_unlock(&ioa_cfg->hrrq[i]._lock); in _ipr_initiate_ioa_reset()
9276 if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].removing_ioa) { in _ipr_initiate_ioa_reset()
9307 if (ioa_cfg->hrrq[IPR_INIT_HRRQ].ioa_is_dead) in ipr_initiate_ioa_reset()
9323 spin_lock(&ioa_cfg->hrrq[i]._lock); in ipr_initiate_ioa_reset()
9324 ioa_cfg->hrrq[i].ioa_is_dead = 1; in ipr_initiate_ioa_reset()
9325 spin_unlock(&ioa_cfg->hrrq[i]._lock); in ipr_initiate_ioa_reset()
9335 if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].removing_ioa) { in ipr_initiate_ioa_reset()
9365 spin_lock(&ioa_cfg->hrrq[i]._lock); in ipr_reset_freeze()
9366 ioa_cfg->hrrq[i].allow_interrupts = 0; in ipr_reset_freeze()
9367 spin_unlock(&ioa_cfg->hrrq[i]._lock); in ipr_reset_freeze()
9370 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q); in ipr_reset_freeze()
9459 spin_lock(&ioa_cfg->hrrq[i]._lock); in ipr_pci_perm_failure()
9460 ioa_cfg->hrrq[i].allow_cmds = 0; in ipr_pci_perm_failure()
9461 spin_unlock(&ioa_cfg->hrrq[i]._lock); in ipr_pci_perm_failure()
9577 sizeof(u32) * ioa_cfg->hrrq[i].size, in ipr_free_mem()
9578 ioa_cfg->hrrq[i].host_rrq, in ipr_free_mem()
9579 ioa_cfg->hrrq[i].host_rrq_dma); in ipr_free_mem()
9611 free_irq(pci_irq_vector(pdev, i), &ioa_cfg->hrrq[i]); in ipr_free_irqs()
9673 ioa_cfg->hrrq[i].min_cmd_id = 0; in ipr_alloc_cmd_blks()
9674 ioa_cfg->hrrq[i].max_cmd_id = in ipr_alloc_cmd_blks()
9680 ioa_cfg->hrrq[i].min_cmd_id = in ipr_alloc_cmd_blks()
9683 ioa_cfg->hrrq[i].max_cmd_id = in ipr_alloc_cmd_blks()
9689 ioa_cfg->hrrq[i].min_cmd_id = 0; in ipr_alloc_cmd_blks()
9690 ioa_cfg->hrrq[i].max_cmd_id = (entries_each_hrrq - 1); in ipr_alloc_cmd_blks()
9692 ioa_cfg->hrrq[i].size = entries_each_hrrq; in ipr_alloc_cmd_blks()
9698 ioa_cfg->hrrq[ioa_cfg->hrrq_num - 1].max_cmd_id - 1; in ipr_alloc_cmd_blks()
9700 ioa_cfg->hrrq[ioa_cfg->hrrq_num - 1].size += i; in ipr_alloc_cmd_blks()
9701 ioa_cfg->hrrq[ioa_cfg->hrrq_num - 1].max_cmd_id += i; in ipr_alloc_cmd_blks()
9743 ipr_cmd->hrrq = &ioa_cfg->hrrq[hrrq_id]; in ipr_alloc_cmd_blks()
9744 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_alloc_cmd_blks()
9745 if (i >= ioa_cfg->hrrq[hrrq_id].max_cmd_id) in ipr_alloc_cmd_blks()
9789 ioa_cfg->hrrq[i].host_rrq = dma_alloc_coherent(&pdev->dev, in ipr_alloc_mem()
9790 sizeof(u32) * ioa_cfg->hrrq[i].size, in ipr_alloc_mem()
9791 &ioa_cfg->hrrq[i].host_rrq_dma, in ipr_alloc_mem()
9794 if (!ioa_cfg->hrrq[i].host_rrq) { in ipr_alloc_mem()
9797 sizeof(u32) * ioa_cfg->hrrq[i].size, in ipr_alloc_mem()
9798 ioa_cfg->hrrq[i].host_rrq, in ipr_alloc_mem()
9799 ioa_cfg->hrrq[i].host_rrq_dma); in ipr_alloc_mem()
9802 ioa_cfg->hrrq[i].ioa_cfg = ioa_cfg; in ipr_alloc_mem()
9851 sizeof(u32) * ioa_cfg->hrrq[i].size, in ipr_alloc_mem()
9852 ioa_cfg->hrrq[i].host_rrq, in ipr_alloc_mem()
9853 ioa_cfg->hrrq[i].host_rrq_dma); in ipr_alloc_mem()
9994 for (i = 0; i < ARRAY_SIZE(ioa_cfg->hrrq); i++) { in ipr_init_ioa_cfg()
9995 INIT_LIST_HEAD(&ioa_cfg->hrrq[i].hrrq_free_q); in ipr_init_ioa_cfg()
9996 INIT_LIST_HEAD(&ioa_cfg->hrrq[i].hrrq_pending_q); in ipr_init_ioa_cfg()
9997 spin_lock_init(&ioa_cfg->hrrq[i]._lock); in ipr_init_ioa_cfg()
9999 ioa_cfg->hrrq[i].lock = ioa_cfg->host->host_lock; in ipr_init_ioa_cfg()
10001 ioa_cfg->hrrq[i].lock = &ioa_cfg->hrrq[i]._lock; in ipr_init_ioa_cfg()
10066 &ioa_cfg->hrrq[i]); in ipr_request_other_msi_irqs()
10070 &ioa_cfg->hrrq[i]); in ipr_request_other_msi_irqs()
10386 &ioa_cfg->hrrq[0]); in ipr_probe_ioa()
10392 IPR_NAME, &ioa_cfg->hrrq[0]); in ipr_probe_ioa()
10493 spin_lock(&ioa_cfg->hrrq[i]._lock); in __ipr_remove()
10494 ioa_cfg->hrrq[i].removing_ioa = 1; in __ipr_remove()
10495 spin_unlock(&ioa_cfg->hrrq[i]._lock); in __ipr_remove()
10626 irq_poll_init(&ioa_cfg->hrrq[i].iopoll, in ipr_probe()
10657 irq_poll_disable(&ioa_cfg->hrrq[i].iopoll); in ipr_shutdown()
10816 list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q); in ipr_halt_done()
10841 if (!ioa_cfg->hrrq[IPR_INIT_HRRQ].allow_cmds || in ipr_halt()