xref: /OK3568_Linux_fs/kernel/drivers/scsi/qla4xxx/ql4_83xx.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * QLogic iSCSI HBA Driver
4*4882a593Smuzhiyun  * Copyright (c)   2003-2013 QLogic Corporation
5*4882a593Smuzhiyun  */
6*4882a593Smuzhiyun 
7*4882a593Smuzhiyun #include <linux/ratelimit.h>
8*4882a593Smuzhiyun 
9*4882a593Smuzhiyun #include "ql4_def.h"
10*4882a593Smuzhiyun #include "ql4_version.h"
11*4882a593Smuzhiyun #include "ql4_glbl.h"
12*4882a593Smuzhiyun #include "ql4_dbg.h"
13*4882a593Smuzhiyun #include "ql4_inline.h"
14*4882a593Smuzhiyun 
qla4_83xx_rd_reg(struct scsi_qla_host * ha,ulong addr)15*4882a593Smuzhiyun uint32_t qla4_83xx_rd_reg(struct scsi_qla_host *ha, ulong addr)
16*4882a593Smuzhiyun {
17*4882a593Smuzhiyun 	return readl((void __iomem *)(ha->nx_pcibase + addr));
18*4882a593Smuzhiyun }
19*4882a593Smuzhiyun 
qla4_83xx_wr_reg(struct scsi_qla_host * ha,ulong addr,uint32_t val)20*4882a593Smuzhiyun void qla4_83xx_wr_reg(struct scsi_qla_host *ha, ulong addr, uint32_t val)
21*4882a593Smuzhiyun {
22*4882a593Smuzhiyun 	writel(val, (void __iomem *)(ha->nx_pcibase + addr));
23*4882a593Smuzhiyun }
24*4882a593Smuzhiyun 
qla4_83xx_set_win_base(struct scsi_qla_host * ha,uint32_t addr)25*4882a593Smuzhiyun static int qla4_83xx_set_win_base(struct scsi_qla_host *ha, uint32_t addr)
26*4882a593Smuzhiyun {
27*4882a593Smuzhiyun 	uint32_t val;
28*4882a593Smuzhiyun 	int ret_val = QLA_SUCCESS;
29*4882a593Smuzhiyun 
30*4882a593Smuzhiyun 	qla4_83xx_wr_reg(ha, QLA83XX_CRB_WIN_FUNC(ha->func_num), addr);
31*4882a593Smuzhiyun 	val = qla4_83xx_rd_reg(ha, QLA83XX_CRB_WIN_FUNC(ha->func_num));
32*4882a593Smuzhiyun 	if (val != addr) {
33*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Failed to set register window : addr written 0x%x, read 0x%x!\n",
34*4882a593Smuzhiyun 			   __func__, addr, val);
35*4882a593Smuzhiyun 		ret_val = QLA_ERROR;
36*4882a593Smuzhiyun 	}
37*4882a593Smuzhiyun 
38*4882a593Smuzhiyun 	return ret_val;
39*4882a593Smuzhiyun }
40*4882a593Smuzhiyun 
qla4_83xx_rd_reg_indirect(struct scsi_qla_host * ha,uint32_t addr,uint32_t * data)41*4882a593Smuzhiyun int qla4_83xx_rd_reg_indirect(struct scsi_qla_host *ha, uint32_t addr,
42*4882a593Smuzhiyun 			      uint32_t *data)
43*4882a593Smuzhiyun {
44*4882a593Smuzhiyun 	int ret_val;
45*4882a593Smuzhiyun 
46*4882a593Smuzhiyun 	ret_val = qla4_83xx_set_win_base(ha, addr);
47*4882a593Smuzhiyun 
48*4882a593Smuzhiyun 	if (ret_val == QLA_SUCCESS) {
49*4882a593Smuzhiyun 		*data = qla4_83xx_rd_reg(ha, QLA83XX_WILDCARD);
50*4882a593Smuzhiyun 	} else {
51*4882a593Smuzhiyun 		*data = 0xffffffff;
52*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: failed read of addr 0x%x!\n",
53*4882a593Smuzhiyun 			   __func__, addr);
54*4882a593Smuzhiyun 	}
55*4882a593Smuzhiyun 
56*4882a593Smuzhiyun 	return ret_val;
57*4882a593Smuzhiyun }
58*4882a593Smuzhiyun 
qla4_83xx_wr_reg_indirect(struct scsi_qla_host * ha,uint32_t addr,uint32_t data)59*4882a593Smuzhiyun int qla4_83xx_wr_reg_indirect(struct scsi_qla_host *ha, uint32_t addr,
60*4882a593Smuzhiyun 			      uint32_t data)
61*4882a593Smuzhiyun {
62*4882a593Smuzhiyun 	int ret_val;
63*4882a593Smuzhiyun 
64*4882a593Smuzhiyun 	ret_val = qla4_83xx_set_win_base(ha, addr);
65*4882a593Smuzhiyun 
66*4882a593Smuzhiyun 	if (ret_val == QLA_SUCCESS)
67*4882a593Smuzhiyun 		qla4_83xx_wr_reg(ha, QLA83XX_WILDCARD, data);
68*4882a593Smuzhiyun 	else
69*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: failed wrt to addr 0x%x, data 0x%x\n",
70*4882a593Smuzhiyun 			   __func__, addr, data);
71*4882a593Smuzhiyun 
72*4882a593Smuzhiyun 	return ret_val;
73*4882a593Smuzhiyun }
74*4882a593Smuzhiyun 
qla4_83xx_flash_lock(struct scsi_qla_host * ha)75*4882a593Smuzhiyun static int qla4_83xx_flash_lock(struct scsi_qla_host *ha)
76*4882a593Smuzhiyun {
77*4882a593Smuzhiyun 	int lock_owner;
78*4882a593Smuzhiyun 	int timeout = 0;
79*4882a593Smuzhiyun 	uint32_t lock_status = 0;
80*4882a593Smuzhiyun 	int ret_val = QLA_SUCCESS;
81*4882a593Smuzhiyun 
82*4882a593Smuzhiyun 	while (lock_status == 0) {
83*4882a593Smuzhiyun 		lock_status = qla4_83xx_rd_reg(ha, QLA83XX_FLASH_LOCK);
84*4882a593Smuzhiyun 		if (lock_status)
85*4882a593Smuzhiyun 			break;
86*4882a593Smuzhiyun 
87*4882a593Smuzhiyun 		if (++timeout >= QLA83XX_FLASH_LOCK_TIMEOUT / 20) {
88*4882a593Smuzhiyun 			lock_owner = qla4_83xx_rd_reg(ha,
89*4882a593Smuzhiyun 						      QLA83XX_FLASH_LOCK_ID);
90*4882a593Smuzhiyun 			ql4_printk(KERN_ERR, ha, "%s: flash lock by func %d failed, held by func %d\n",
91*4882a593Smuzhiyun 				   __func__, ha->func_num, lock_owner);
92*4882a593Smuzhiyun 			ret_val = QLA_ERROR;
93*4882a593Smuzhiyun 			break;
94*4882a593Smuzhiyun 		}
95*4882a593Smuzhiyun 		msleep(20);
96*4882a593Smuzhiyun 	}
97*4882a593Smuzhiyun 
98*4882a593Smuzhiyun 	qla4_83xx_wr_reg(ha, QLA83XX_FLASH_LOCK_ID, ha->func_num);
99*4882a593Smuzhiyun 	return ret_val;
100*4882a593Smuzhiyun }
101*4882a593Smuzhiyun 
qla4_83xx_flash_unlock(struct scsi_qla_host * ha)102*4882a593Smuzhiyun static void qla4_83xx_flash_unlock(struct scsi_qla_host *ha)
103*4882a593Smuzhiyun {
104*4882a593Smuzhiyun 	/* Reading FLASH_UNLOCK register unlocks the Flash */
105*4882a593Smuzhiyun 	qla4_83xx_wr_reg(ha, QLA83XX_FLASH_LOCK_ID, 0xFF);
106*4882a593Smuzhiyun 	qla4_83xx_rd_reg(ha, QLA83XX_FLASH_UNLOCK);
107*4882a593Smuzhiyun }
108*4882a593Smuzhiyun 
qla4_83xx_flash_read_u32(struct scsi_qla_host * ha,uint32_t flash_addr,uint8_t * p_data,int u32_word_count)109*4882a593Smuzhiyun int qla4_83xx_flash_read_u32(struct scsi_qla_host *ha, uint32_t flash_addr,
110*4882a593Smuzhiyun 			     uint8_t *p_data, int u32_word_count)
111*4882a593Smuzhiyun {
112*4882a593Smuzhiyun 	int i;
113*4882a593Smuzhiyun 	uint32_t u32_word;
114*4882a593Smuzhiyun 	uint32_t addr = flash_addr;
115*4882a593Smuzhiyun 	int ret_val = QLA_SUCCESS;
116*4882a593Smuzhiyun 
117*4882a593Smuzhiyun 	ret_val = qla4_83xx_flash_lock(ha);
118*4882a593Smuzhiyun 	if (ret_val == QLA_ERROR)
119*4882a593Smuzhiyun 		goto exit_lock_error;
120*4882a593Smuzhiyun 
121*4882a593Smuzhiyun 	if (addr & 0x03) {
122*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Illegal addr = 0x%x\n",
123*4882a593Smuzhiyun 			   __func__, addr);
124*4882a593Smuzhiyun 		ret_val = QLA_ERROR;
125*4882a593Smuzhiyun 		goto exit_flash_read;
126*4882a593Smuzhiyun 	}
127*4882a593Smuzhiyun 
128*4882a593Smuzhiyun 	for (i = 0; i < u32_word_count; i++) {
129*4882a593Smuzhiyun 		ret_val = qla4_83xx_wr_reg_indirect(ha,
130*4882a593Smuzhiyun 						    QLA83XX_FLASH_DIRECT_WINDOW,
131*4882a593Smuzhiyun 						    (addr & 0xFFFF0000));
132*4882a593Smuzhiyun 		if (ret_val == QLA_ERROR) {
133*4882a593Smuzhiyun 			ql4_printk(KERN_ERR, ha, "%s: failed to write addr 0x%x to FLASH_DIRECT_WINDOW\n!",
134*4882a593Smuzhiyun 				   __func__, addr);
135*4882a593Smuzhiyun 			goto exit_flash_read;
136*4882a593Smuzhiyun 		}
137*4882a593Smuzhiyun 
138*4882a593Smuzhiyun 		ret_val = qla4_83xx_rd_reg_indirect(ha,
139*4882a593Smuzhiyun 						QLA83XX_FLASH_DIRECT_DATA(addr),
140*4882a593Smuzhiyun 						&u32_word);
141*4882a593Smuzhiyun 		if (ret_val == QLA_ERROR) {
142*4882a593Smuzhiyun 			ql4_printk(KERN_ERR, ha, "%s: failed to read addr 0x%x!\n",
143*4882a593Smuzhiyun 				   __func__, addr);
144*4882a593Smuzhiyun 			goto exit_flash_read;
145*4882a593Smuzhiyun 		}
146*4882a593Smuzhiyun 
147*4882a593Smuzhiyun 		*(__le32 *)p_data = le32_to_cpu(u32_word);
148*4882a593Smuzhiyun 		p_data = p_data + 4;
149*4882a593Smuzhiyun 		addr = addr + 4;
150*4882a593Smuzhiyun 	}
151*4882a593Smuzhiyun 
152*4882a593Smuzhiyun exit_flash_read:
153*4882a593Smuzhiyun 	qla4_83xx_flash_unlock(ha);
154*4882a593Smuzhiyun 
155*4882a593Smuzhiyun exit_lock_error:
156*4882a593Smuzhiyun 	return ret_val;
157*4882a593Smuzhiyun }
158*4882a593Smuzhiyun 
qla4_83xx_lockless_flash_read_u32(struct scsi_qla_host * ha,uint32_t flash_addr,uint8_t * p_data,int u32_word_count)159*4882a593Smuzhiyun int qla4_83xx_lockless_flash_read_u32(struct scsi_qla_host *ha,
160*4882a593Smuzhiyun 				      uint32_t flash_addr, uint8_t *p_data,
161*4882a593Smuzhiyun 				      int u32_word_count)
162*4882a593Smuzhiyun {
163*4882a593Smuzhiyun 	uint32_t i;
164*4882a593Smuzhiyun 	uint32_t u32_word;
165*4882a593Smuzhiyun 	uint32_t flash_offset;
166*4882a593Smuzhiyun 	uint32_t addr = flash_addr;
167*4882a593Smuzhiyun 	int ret_val = QLA_SUCCESS;
168*4882a593Smuzhiyun 
169*4882a593Smuzhiyun 	flash_offset = addr & (QLA83XX_FLASH_SECTOR_SIZE - 1);
170*4882a593Smuzhiyun 
171*4882a593Smuzhiyun 	if (addr & 0x3) {
172*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Illegal addr = 0x%x\n",
173*4882a593Smuzhiyun 			   __func__, addr);
174*4882a593Smuzhiyun 		ret_val = QLA_ERROR;
175*4882a593Smuzhiyun 		goto exit_lockless_read;
176*4882a593Smuzhiyun 	}
177*4882a593Smuzhiyun 
178*4882a593Smuzhiyun 	ret_val = qla4_83xx_wr_reg_indirect(ha, QLA83XX_FLASH_DIRECT_WINDOW,
179*4882a593Smuzhiyun 					    addr);
180*4882a593Smuzhiyun 	if (ret_val == QLA_ERROR) {
181*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: failed to write addr 0x%x to FLASH_DIRECT_WINDOW!\n",
182*4882a593Smuzhiyun 			   __func__, addr);
183*4882a593Smuzhiyun 		goto exit_lockless_read;
184*4882a593Smuzhiyun 	}
185*4882a593Smuzhiyun 
186*4882a593Smuzhiyun 	/* Check if data is spread across multiple sectors  */
187*4882a593Smuzhiyun 	if ((flash_offset + (u32_word_count * sizeof(uint32_t))) >
188*4882a593Smuzhiyun 	    (QLA83XX_FLASH_SECTOR_SIZE - 1)) {
189*4882a593Smuzhiyun 
190*4882a593Smuzhiyun 		/* Multi sector read */
191*4882a593Smuzhiyun 		for (i = 0; i < u32_word_count; i++) {
192*4882a593Smuzhiyun 			ret_val = qla4_83xx_rd_reg_indirect(ha,
193*4882a593Smuzhiyun 						QLA83XX_FLASH_DIRECT_DATA(addr),
194*4882a593Smuzhiyun 						&u32_word);
195*4882a593Smuzhiyun 			if (ret_val == QLA_ERROR) {
196*4882a593Smuzhiyun 				ql4_printk(KERN_ERR, ha, "%s: failed to read addr 0x%x!\n",
197*4882a593Smuzhiyun 					   __func__, addr);
198*4882a593Smuzhiyun 				goto exit_lockless_read;
199*4882a593Smuzhiyun 			}
200*4882a593Smuzhiyun 
201*4882a593Smuzhiyun 			*(__le32 *)p_data  = le32_to_cpu(u32_word);
202*4882a593Smuzhiyun 			p_data = p_data + 4;
203*4882a593Smuzhiyun 			addr = addr + 4;
204*4882a593Smuzhiyun 			flash_offset = flash_offset + 4;
205*4882a593Smuzhiyun 
206*4882a593Smuzhiyun 			if (flash_offset > (QLA83XX_FLASH_SECTOR_SIZE - 1)) {
207*4882a593Smuzhiyun 				/* This write is needed once for each sector */
208*4882a593Smuzhiyun 				ret_val = qla4_83xx_wr_reg_indirect(ha,
209*4882a593Smuzhiyun 						   QLA83XX_FLASH_DIRECT_WINDOW,
210*4882a593Smuzhiyun 						   addr);
211*4882a593Smuzhiyun 				if (ret_val == QLA_ERROR) {
212*4882a593Smuzhiyun 					ql4_printk(KERN_ERR, ha, "%s: failed to write addr 0x%x to FLASH_DIRECT_WINDOW!\n",
213*4882a593Smuzhiyun 						   __func__, addr);
214*4882a593Smuzhiyun 					goto exit_lockless_read;
215*4882a593Smuzhiyun 				}
216*4882a593Smuzhiyun 				flash_offset = 0;
217*4882a593Smuzhiyun 			}
218*4882a593Smuzhiyun 		}
219*4882a593Smuzhiyun 	} else {
220*4882a593Smuzhiyun 		/* Single sector read */
221*4882a593Smuzhiyun 		for (i = 0; i < u32_word_count; i++) {
222*4882a593Smuzhiyun 			ret_val = qla4_83xx_rd_reg_indirect(ha,
223*4882a593Smuzhiyun 						QLA83XX_FLASH_DIRECT_DATA(addr),
224*4882a593Smuzhiyun 						&u32_word);
225*4882a593Smuzhiyun 			if (ret_val == QLA_ERROR) {
226*4882a593Smuzhiyun 				ql4_printk(KERN_ERR, ha, "%s: failed to read addr 0x%x!\n",
227*4882a593Smuzhiyun 					   __func__, addr);
228*4882a593Smuzhiyun 				goto exit_lockless_read;
229*4882a593Smuzhiyun 			}
230*4882a593Smuzhiyun 
231*4882a593Smuzhiyun 			*(__le32 *)p_data = le32_to_cpu(u32_word);
232*4882a593Smuzhiyun 			p_data = p_data + 4;
233*4882a593Smuzhiyun 			addr = addr + 4;
234*4882a593Smuzhiyun 		}
235*4882a593Smuzhiyun 	}
236*4882a593Smuzhiyun 
237*4882a593Smuzhiyun exit_lockless_read:
238*4882a593Smuzhiyun 	return ret_val;
239*4882a593Smuzhiyun }
240*4882a593Smuzhiyun 
qla4_83xx_rom_lock_recovery(struct scsi_qla_host * ha)241*4882a593Smuzhiyun void qla4_83xx_rom_lock_recovery(struct scsi_qla_host *ha)
242*4882a593Smuzhiyun {
243*4882a593Smuzhiyun 	if (qla4_83xx_flash_lock(ha))
244*4882a593Smuzhiyun 		ql4_printk(KERN_INFO, ha, "%s: Resetting rom lock\n", __func__);
245*4882a593Smuzhiyun 
246*4882a593Smuzhiyun 	/*
247*4882a593Smuzhiyun 	 * We got the lock, or someone else is holding the lock
248*4882a593Smuzhiyun 	 * since we are restting, forcefully unlock
249*4882a593Smuzhiyun 	 */
250*4882a593Smuzhiyun 	qla4_83xx_flash_unlock(ha);
251*4882a593Smuzhiyun }
252*4882a593Smuzhiyun 
253*4882a593Smuzhiyun #define INTENT_TO_RECOVER	0x01
254*4882a593Smuzhiyun #define PROCEED_TO_RECOVER	0x02
255*4882a593Smuzhiyun 
qla4_83xx_lock_recovery(struct scsi_qla_host * ha)256*4882a593Smuzhiyun static int qla4_83xx_lock_recovery(struct scsi_qla_host *ha)
257*4882a593Smuzhiyun {
258*4882a593Smuzhiyun 
259*4882a593Smuzhiyun 	uint32_t lock = 0, lockid;
260*4882a593Smuzhiyun 	int ret_val = QLA_ERROR;
261*4882a593Smuzhiyun 
262*4882a593Smuzhiyun 	lockid = ha->isp_ops->rd_reg_direct(ha, QLA83XX_DRV_LOCKRECOVERY);
263*4882a593Smuzhiyun 
264*4882a593Smuzhiyun 	/* Check for other Recovery in progress, go wait */
265*4882a593Smuzhiyun 	if ((lockid & 0x3) != 0)
266*4882a593Smuzhiyun 		goto exit_lock_recovery;
267*4882a593Smuzhiyun 
268*4882a593Smuzhiyun 	/* Intent to Recover */
269*4882a593Smuzhiyun 	ha->isp_ops->wr_reg_direct(ha, QLA83XX_DRV_LOCKRECOVERY,
270*4882a593Smuzhiyun 				   (ha->func_num << 2) | INTENT_TO_RECOVER);
271*4882a593Smuzhiyun 
272*4882a593Smuzhiyun 	msleep(200);
273*4882a593Smuzhiyun 
274*4882a593Smuzhiyun 	/* Check Intent to Recover is advertised */
275*4882a593Smuzhiyun 	lockid = ha->isp_ops->rd_reg_direct(ha, QLA83XX_DRV_LOCKRECOVERY);
276*4882a593Smuzhiyun 	if ((lockid & 0x3C) != (ha->func_num << 2))
277*4882a593Smuzhiyun 		goto exit_lock_recovery;
278*4882a593Smuzhiyun 
279*4882a593Smuzhiyun 	ql4_printk(KERN_INFO, ha, "%s: IDC Lock recovery initiated for func %d\n",
280*4882a593Smuzhiyun 		   __func__, ha->func_num);
281*4882a593Smuzhiyun 
282*4882a593Smuzhiyun 	/* Proceed to Recover */
283*4882a593Smuzhiyun 	ha->isp_ops->wr_reg_direct(ha, QLA83XX_DRV_LOCKRECOVERY,
284*4882a593Smuzhiyun 				   (ha->func_num << 2) | PROCEED_TO_RECOVER);
285*4882a593Smuzhiyun 
286*4882a593Smuzhiyun 	/* Force Unlock */
287*4882a593Smuzhiyun 	ha->isp_ops->wr_reg_direct(ha, QLA83XX_DRV_LOCK_ID, 0xFF);
288*4882a593Smuzhiyun 	ha->isp_ops->rd_reg_direct(ha, QLA83XX_DRV_UNLOCK);
289*4882a593Smuzhiyun 
290*4882a593Smuzhiyun 	/* Clear bits 0-5 in IDC_RECOVERY register*/
291*4882a593Smuzhiyun 	ha->isp_ops->wr_reg_direct(ha, QLA83XX_DRV_LOCKRECOVERY, 0);
292*4882a593Smuzhiyun 
293*4882a593Smuzhiyun 	/* Get lock */
294*4882a593Smuzhiyun 	lock = ha->isp_ops->rd_reg_direct(ha, QLA83XX_DRV_LOCK);
295*4882a593Smuzhiyun 	if (lock) {
296*4882a593Smuzhiyun 		lockid = ha->isp_ops->rd_reg_direct(ha, QLA83XX_DRV_LOCK_ID);
297*4882a593Smuzhiyun 		lockid = ((lockid + (1 << 8)) & ~0xFF) | ha->func_num;
298*4882a593Smuzhiyun 		ha->isp_ops->wr_reg_direct(ha, QLA83XX_DRV_LOCK_ID, lockid);
299*4882a593Smuzhiyun 		ret_val = QLA_SUCCESS;
300*4882a593Smuzhiyun 	}
301*4882a593Smuzhiyun 
302*4882a593Smuzhiyun exit_lock_recovery:
303*4882a593Smuzhiyun 	return ret_val;
304*4882a593Smuzhiyun }
305*4882a593Smuzhiyun 
306*4882a593Smuzhiyun #define	QLA83XX_DRV_LOCK_MSLEEP		200
307*4882a593Smuzhiyun 
qla4_83xx_drv_lock(struct scsi_qla_host * ha)308*4882a593Smuzhiyun int qla4_83xx_drv_lock(struct scsi_qla_host *ha)
309*4882a593Smuzhiyun {
310*4882a593Smuzhiyun 	int timeout = 0;
311*4882a593Smuzhiyun 	uint32_t status = 0;
312*4882a593Smuzhiyun 	int ret_val = QLA_SUCCESS;
313*4882a593Smuzhiyun 	uint32_t first_owner = 0;
314*4882a593Smuzhiyun 	uint32_t tmo_owner = 0;
315*4882a593Smuzhiyun 	uint32_t lock_id;
316*4882a593Smuzhiyun 	uint32_t func_num;
317*4882a593Smuzhiyun 	uint32_t lock_cnt;
318*4882a593Smuzhiyun 
319*4882a593Smuzhiyun 	while (status == 0) {
320*4882a593Smuzhiyun 		status = qla4_83xx_rd_reg(ha, QLA83XX_DRV_LOCK);
321*4882a593Smuzhiyun 		if (status) {
322*4882a593Smuzhiyun 			/* Increment Counter (8-31) and update func_num (0-7) on
323*4882a593Smuzhiyun 			 * getting a successful lock  */
324*4882a593Smuzhiyun 			lock_id = qla4_83xx_rd_reg(ha, QLA83XX_DRV_LOCK_ID);
325*4882a593Smuzhiyun 			lock_id = ((lock_id + (1 << 8)) & ~0xFF) | ha->func_num;
326*4882a593Smuzhiyun 			qla4_83xx_wr_reg(ha, QLA83XX_DRV_LOCK_ID, lock_id);
327*4882a593Smuzhiyun 			break;
328*4882a593Smuzhiyun 		}
329*4882a593Smuzhiyun 
330*4882a593Smuzhiyun 		if (timeout == 0)
331*4882a593Smuzhiyun 			/* Save counter + ID of function holding the lock for
332*4882a593Smuzhiyun 			 * first failure */
333*4882a593Smuzhiyun 			first_owner = ha->isp_ops->rd_reg_direct(ha,
334*4882a593Smuzhiyun 							  QLA83XX_DRV_LOCK_ID);
335*4882a593Smuzhiyun 
336*4882a593Smuzhiyun 		if (++timeout >=
337*4882a593Smuzhiyun 		    (QLA83XX_DRV_LOCK_TIMEOUT / QLA83XX_DRV_LOCK_MSLEEP)) {
338*4882a593Smuzhiyun 			tmo_owner = qla4_83xx_rd_reg(ha, QLA83XX_DRV_LOCK_ID);
339*4882a593Smuzhiyun 			func_num = tmo_owner & 0xFF;
340*4882a593Smuzhiyun 			lock_cnt = tmo_owner >> 8;
341*4882a593Smuzhiyun 			ql4_printk(KERN_INFO, ha, "%s: Lock by func %d failed after 2s, lock held by func %d, lock count %d, first_owner %d\n",
342*4882a593Smuzhiyun 				   __func__, ha->func_num, func_num, lock_cnt,
343*4882a593Smuzhiyun 				   (first_owner & 0xFF));
344*4882a593Smuzhiyun 
345*4882a593Smuzhiyun 			if (first_owner != tmo_owner) {
346*4882a593Smuzhiyun 				/* Some other driver got lock, OR same driver
347*4882a593Smuzhiyun 				 * got lock again (counter value changed), when
348*4882a593Smuzhiyun 				 * we were waiting for lock.
349*4882a593Smuzhiyun 				 * Retry for another 2 sec */
350*4882a593Smuzhiyun 				ql4_printk(KERN_INFO, ha, "%s: IDC lock failed for func %d\n",
351*4882a593Smuzhiyun 					   __func__, ha->func_num);
352*4882a593Smuzhiyun 				timeout = 0;
353*4882a593Smuzhiyun 			} else {
354*4882a593Smuzhiyun 				/* Same driver holding lock > 2sec.
355*4882a593Smuzhiyun 				 * Force Recovery */
356*4882a593Smuzhiyun 				ret_val = qla4_83xx_lock_recovery(ha);
357*4882a593Smuzhiyun 				if (ret_val == QLA_SUCCESS) {
358*4882a593Smuzhiyun 					/* Recovered and got lock */
359*4882a593Smuzhiyun 					ql4_printk(KERN_INFO, ha, "%s: IDC lock Recovery by %d successful\n",
360*4882a593Smuzhiyun 						   __func__, ha->func_num);
361*4882a593Smuzhiyun 					break;
362*4882a593Smuzhiyun 				}
363*4882a593Smuzhiyun 				/* Recovery Failed, some other function
364*4882a593Smuzhiyun 				 * has the lock, wait for 2secs and retry */
365*4882a593Smuzhiyun 				ql4_printk(KERN_INFO, ha, "%s: IDC lock Recovery by %d failed, Retrying timeout\n",
366*4882a593Smuzhiyun 					   __func__, ha->func_num);
367*4882a593Smuzhiyun 				timeout = 0;
368*4882a593Smuzhiyun 			}
369*4882a593Smuzhiyun 		}
370*4882a593Smuzhiyun 		msleep(QLA83XX_DRV_LOCK_MSLEEP);
371*4882a593Smuzhiyun 	}
372*4882a593Smuzhiyun 
373*4882a593Smuzhiyun 	return ret_val;
374*4882a593Smuzhiyun }
375*4882a593Smuzhiyun 
qla4_83xx_drv_unlock(struct scsi_qla_host * ha)376*4882a593Smuzhiyun void qla4_83xx_drv_unlock(struct scsi_qla_host *ha)
377*4882a593Smuzhiyun {
378*4882a593Smuzhiyun 	int id;
379*4882a593Smuzhiyun 
380*4882a593Smuzhiyun 	id = qla4_83xx_rd_reg(ha, QLA83XX_DRV_LOCK_ID);
381*4882a593Smuzhiyun 
382*4882a593Smuzhiyun 	if ((id & 0xFF) != ha->func_num) {
383*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: IDC Unlock by %d failed, lock owner is %d\n",
384*4882a593Smuzhiyun 			   __func__, ha->func_num, (id & 0xFF));
385*4882a593Smuzhiyun 		return;
386*4882a593Smuzhiyun 	}
387*4882a593Smuzhiyun 
388*4882a593Smuzhiyun 	/* Keep lock counter value, update the ha->func_num to 0xFF */
389*4882a593Smuzhiyun 	qla4_83xx_wr_reg(ha, QLA83XX_DRV_LOCK_ID, (id | 0xFF));
390*4882a593Smuzhiyun 	qla4_83xx_rd_reg(ha, QLA83XX_DRV_UNLOCK);
391*4882a593Smuzhiyun }
392*4882a593Smuzhiyun 
qla4_83xx_set_idc_dontreset(struct scsi_qla_host * ha)393*4882a593Smuzhiyun void qla4_83xx_set_idc_dontreset(struct scsi_qla_host *ha)
394*4882a593Smuzhiyun {
395*4882a593Smuzhiyun 	uint32_t idc_ctrl;
396*4882a593Smuzhiyun 
397*4882a593Smuzhiyun 	idc_ctrl = qla4_83xx_rd_reg(ha, QLA83XX_IDC_DRV_CTRL);
398*4882a593Smuzhiyun 	idc_ctrl |= DONTRESET_BIT0;
399*4882a593Smuzhiyun 	qla4_83xx_wr_reg(ha, QLA83XX_IDC_DRV_CTRL, idc_ctrl);
400*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha, "%s: idc_ctrl = %d\n", __func__,
401*4882a593Smuzhiyun 			  idc_ctrl));
402*4882a593Smuzhiyun }
403*4882a593Smuzhiyun 
qla4_83xx_clear_idc_dontreset(struct scsi_qla_host * ha)404*4882a593Smuzhiyun void qla4_83xx_clear_idc_dontreset(struct scsi_qla_host *ha)
405*4882a593Smuzhiyun {
406*4882a593Smuzhiyun 	uint32_t idc_ctrl;
407*4882a593Smuzhiyun 
408*4882a593Smuzhiyun 	idc_ctrl = qla4_83xx_rd_reg(ha, QLA83XX_IDC_DRV_CTRL);
409*4882a593Smuzhiyun 	idc_ctrl &= ~DONTRESET_BIT0;
410*4882a593Smuzhiyun 	qla4_83xx_wr_reg(ha, QLA83XX_IDC_DRV_CTRL, idc_ctrl);
411*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha, "%s: idc_ctrl = %d\n", __func__,
412*4882a593Smuzhiyun 			  idc_ctrl));
413*4882a593Smuzhiyun }
414*4882a593Smuzhiyun 
qla4_83xx_idc_dontreset(struct scsi_qla_host * ha)415*4882a593Smuzhiyun int qla4_83xx_idc_dontreset(struct scsi_qla_host *ha)
416*4882a593Smuzhiyun {
417*4882a593Smuzhiyun 	uint32_t idc_ctrl;
418*4882a593Smuzhiyun 
419*4882a593Smuzhiyun 	idc_ctrl = qla4_83xx_rd_reg(ha, QLA83XX_IDC_DRV_CTRL);
420*4882a593Smuzhiyun 	return idc_ctrl & DONTRESET_BIT0;
421*4882a593Smuzhiyun }
422*4882a593Smuzhiyun 
423*4882a593Smuzhiyun /*-------------------------IDC State Machine ---------------------*/
424*4882a593Smuzhiyun 
425*4882a593Smuzhiyun enum {
426*4882a593Smuzhiyun 	UNKNOWN_CLASS = 0,
427*4882a593Smuzhiyun 	NIC_CLASS,
428*4882a593Smuzhiyun 	FCOE_CLASS,
429*4882a593Smuzhiyun 	ISCSI_CLASS
430*4882a593Smuzhiyun };
431*4882a593Smuzhiyun 
432*4882a593Smuzhiyun struct device_info {
433*4882a593Smuzhiyun 	int func_num;
434*4882a593Smuzhiyun 	int device_type;
435*4882a593Smuzhiyun 	int port_num;
436*4882a593Smuzhiyun };
437*4882a593Smuzhiyun 
qla4_83xx_can_perform_reset(struct scsi_qla_host * ha)438*4882a593Smuzhiyun int qla4_83xx_can_perform_reset(struct scsi_qla_host *ha)
439*4882a593Smuzhiyun {
440*4882a593Smuzhiyun 	uint32_t drv_active;
441*4882a593Smuzhiyun 	uint32_t dev_part, dev_part1, dev_part2;
442*4882a593Smuzhiyun 	int i;
443*4882a593Smuzhiyun 	struct device_info device_map[16];
444*4882a593Smuzhiyun 	int func_nibble;
445*4882a593Smuzhiyun 	int nibble;
446*4882a593Smuzhiyun 	int nic_present = 0;
447*4882a593Smuzhiyun 	int iscsi_present = 0;
448*4882a593Smuzhiyun 	int iscsi_func_low = 0;
449*4882a593Smuzhiyun 
450*4882a593Smuzhiyun 	/* Use the dev_partition register to determine the PCI function number
451*4882a593Smuzhiyun 	 * and then check drv_active register to see which driver is loaded */
452*4882a593Smuzhiyun 	dev_part1 = qla4_83xx_rd_reg(ha,
453*4882a593Smuzhiyun 				     ha->reg_tbl[QLA8XXX_CRB_DEV_PART_INFO]);
454*4882a593Smuzhiyun 	dev_part2 = qla4_83xx_rd_reg(ha, QLA83XX_CRB_DEV_PART_INFO2);
455*4882a593Smuzhiyun 	drv_active = qla4_83xx_rd_reg(ha, ha->reg_tbl[QLA8XXX_CRB_DRV_ACTIVE]);
456*4882a593Smuzhiyun 
457*4882a593Smuzhiyun 	/* Each function has 4 bits in dev_partition Info register,
458*4882a593Smuzhiyun 	 * Lower 2 bits - device type, Upper 2 bits - physical port number */
459*4882a593Smuzhiyun 	dev_part = dev_part1;
460*4882a593Smuzhiyun 	for (i = nibble = 0; i <= 15; i++, nibble++) {
461*4882a593Smuzhiyun 		func_nibble = dev_part & (0xF << (nibble * 4));
462*4882a593Smuzhiyun 		func_nibble >>= (nibble * 4);
463*4882a593Smuzhiyun 		device_map[i].func_num = i;
464*4882a593Smuzhiyun 		device_map[i].device_type = func_nibble & 0x3;
465*4882a593Smuzhiyun 		device_map[i].port_num = func_nibble & 0xC;
466*4882a593Smuzhiyun 
467*4882a593Smuzhiyun 		if (device_map[i].device_type == NIC_CLASS) {
468*4882a593Smuzhiyun 			if (drv_active & (1 << device_map[i].func_num)) {
469*4882a593Smuzhiyun 				nic_present++;
470*4882a593Smuzhiyun 				break;
471*4882a593Smuzhiyun 			}
472*4882a593Smuzhiyun 		} else if (device_map[i].device_type == ISCSI_CLASS) {
473*4882a593Smuzhiyun 			if (drv_active & (1 << device_map[i].func_num)) {
474*4882a593Smuzhiyun 				if (!iscsi_present ||
475*4882a593Smuzhiyun 				    (iscsi_present &&
476*4882a593Smuzhiyun 				     (iscsi_func_low > device_map[i].func_num)))
477*4882a593Smuzhiyun 					iscsi_func_low = device_map[i].func_num;
478*4882a593Smuzhiyun 
479*4882a593Smuzhiyun 				iscsi_present++;
480*4882a593Smuzhiyun 			}
481*4882a593Smuzhiyun 		}
482*4882a593Smuzhiyun 
483*4882a593Smuzhiyun 		/* For function_num[8..15] get info from dev_part2 register */
484*4882a593Smuzhiyun 		if (nibble == 7) {
485*4882a593Smuzhiyun 			nibble = 0;
486*4882a593Smuzhiyun 			dev_part = dev_part2;
487*4882a593Smuzhiyun 		}
488*4882a593Smuzhiyun 	}
489*4882a593Smuzhiyun 
490*4882a593Smuzhiyun 	/* NIC, iSCSI and FCOE are the Reset owners based on order, NIC gets
491*4882a593Smuzhiyun 	 * precedence over iSCSI and FCOE and iSCSI over FCOE, based on drivers
492*4882a593Smuzhiyun 	 * present. */
493*4882a593Smuzhiyun 	if (!nic_present && (ha->func_num == iscsi_func_low)) {
494*4882a593Smuzhiyun 		DEBUG2(ql4_printk(KERN_INFO, ha,
495*4882a593Smuzhiyun 				  "%s: can reset - NIC not present and lower iSCSI function is %d\n",
496*4882a593Smuzhiyun 				  __func__, ha->func_num));
497*4882a593Smuzhiyun 		return 1;
498*4882a593Smuzhiyun 	}
499*4882a593Smuzhiyun 
500*4882a593Smuzhiyun 	return 0;
501*4882a593Smuzhiyun }
502*4882a593Smuzhiyun 
503*4882a593Smuzhiyun /**
504*4882a593Smuzhiyun  * qla4_83xx_need_reset_handler - Code to start reset sequence
505*4882a593Smuzhiyun  * @ha: pointer to adapter structure
506*4882a593Smuzhiyun  *
507*4882a593Smuzhiyun  * Note: IDC lock must be held upon entry
508*4882a593Smuzhiyun  **/
qla4_83xx_need_reset_handler(struct scsi_qla_host * ha)509*4882a593Smuzhiyun void qla4_83xx_need_reset_handler(struct scsi_qla_host *ha)
510*4882a593Smuzhiyun {
511*4882a593Smuzhiyun 	uint32_t dev_state, drv_state, drv_active;
512*4882a593Smuzhiyun 	unsigned long reset_timeout, dev_init_timeout;
513*4882a593Smuzhiyun 
514*4882a593Smuzhiyun 	ql4_printk(KERN_INFO, ha, "%s: Performing ISP error recovery\n",
515*4882a593Smuzhiyun 		   __func__);
516*4882a593Smuzhiyun 
517*4882a593Smuzhiyun 	if (!test_bit(AF_8XXX_RST_OWNER, &ha->flags)) {
518*4882a593Smuzhiyun 		DEBUG2(ql4_printk(KERN_INFO, ha, "%s: reset acknowledged\n",
519*4882a593Smuzhiyun 				  __func__));
520*4882a593Smuzhiyun 		qla4_8xxx_set_rst_ready(ha);
521*4882a593Smuzhiyun 
522*4882a593Smuzhiyun 		/* Non-reset owners ACK Reset and wait for device INIT state
523*4882a593Smuzhiyun 		 * as part of Reset Recovery by Reset Owner */
524*4882a593Smuzhiyun 		dev_init_timeout = jiffies + (ha->nx_dev_init_timeout * HZ);
525*4882a593Smuzhiyun 
526*4882a593Smuzhiyun 		do {
527*4882a593Smuzhiyun 			if (time_after_eq(jiffies, dev_init_timeout)) {
528*4882a593Smuzhiyun 				ql4_printk(KERN_INFO, ha, "%s: Non Reset owner dev init timeout\n",
529*4882a593Smuzhiyun 					   __func__);
530*4882a593Smuzhiyun 				break;
531*4882a593Smuzhiyun 			}
532*4882a593Smuzhiyun 
533*4882a593Smuzhiyun 			ha->isp_ops->idc_unlock(ha);
534*4882a593Smuzhiyun 			msleep(1000);
535*4882a593Smuzhiyun 			ha->isp_ops->idc_lock(ha);
536*4882a593Smuzhiyun 
537*4882a593Smuzhiyun 			dev_state = qla4_8xxx_rd_direct(ha,
538*4882a593Smuzhiyun 							QLA8XXX_CRB_DEV_STATE);
539*4882a593Smuzhiyun 		} while (dev_state == QLA8XXX_DEV_NEED_RESET);
540*4882a593Smuzhiyun 	} else {
541*4882a593Smuzhiyun 		qla4_8xxx_set_rst_ready(ha);
542*4882a593Smuzhiyun 		reset_timeout = jiffies + (ha->nx_reset_timeout * HZ);
543*4882a593Smuzhiyun 		drv_state = qla4_8xxx_rd_direct(ha, QLA8XXX_CRB_DRV_STATE);
544*4882a593Smuzhiyun 		drv_active = qla4_8xxx_rd_direct(ha, QLA8XXX_CRB_DRV_ACTIVE);
545*4882a593Smuzhiyun 
546*4882a593Smuzhiyun 		ql4_printk(KERN_INFO, ha, "%s: drv_state = 0x%x, drv_active = 0x%x\n",
547*4882a593Smuzhiyun 			   __func__, drv_state, drv_active);
548*4882a593Smuzhiyun 
549*4882a593Smuzhiyun 		while (drv_state != drv_active) {
550*4882a593Smuzhiyun 			if (time_after_eq(jiffies, reset_timeout)) {
551*4882a593Smuzhiyun 				ql4_printk(KERN_INFO, ha, "%s: %s: RESET TIMEOUT! drv_state: 0x%08x, drv_active: 0x%08x\n",
552*4882a593Smuzhiyun 					   __func__, DRIVER_NAME, drv_state,
553*4882a593Smuzhiyun 					   drv_active);
554*4882a593Smuzhiyun 				break;
555*4882a593Smuzhiyun 			}
556*4882a593Smuzhiyun 
557*4882a593Smuzhiyun 			ha->isp_ops->idc_unlock(ha);
558*4882a593Smuzhiyun 			msleep(1000);
559*4882a593Smuzhiyun 			ha->isp_ops->idc_lock(ha);
560*4882a593Smuzhiyun 
561*4882a593Smuzhiyun 			drv_state = qla4_8xxx_rd_direct(ha,
562*4882a593Smuzhiyun 							QLA8XXX_CRB_DRV_STATE);
563*4882a593Smuzhiyun 			drv_active = qla4_8xxx_rd_direct(ha,
564*4882a593Smuzhiyun 							QLA8XXX_CRB_DRV_ACTIVE);
565*4882a593Smuzhiyun 		}
566*4882a593Smuzhiyun 
567*4882a593Smuzhiyun 		if (drv_state != drv_active) {
568*4882a593Smuzhiyun 			ql4_printk(KERN_INFO, ha, "%s: Reset_owner turning off drv_active of non-acking function 0x%x\n",
569*4882a593Smuzhiyun 				   __func__, (drv_active ^ drv_state));
570*4882a593Smuzhiyun 			drv_active = drv_active & drv_state;
571*4882a593Smuzhiyun 			qla4_8xxx_wr_direct(ha, QLA8XXX_CRB_DRV_ACTIVE,
572*4882a593Smuzhiyun 					    drv_active);
573*4882a593Smuzhiyun 		}
574*4882a593Smuzhiyun 
575*4882a593Smuzhiyun 		clear_bit(AF_8XXX_RST_OWNER, &ha->flags);
576*4882a593Smuzhiyun 		/* Start Reset Recovery */
577*4882a593Smuzhiyun 		qla4_8xxx_device_bootstrap(ha);
578*4882a593Smuzhiyun 	}
579*4882a593Smuzhiyun }
580*4882a593Smuzhiyun 
qla4_83xx_get_idc_param(struct scsi_qla_host * ha)581*4882a593Smuzhiyun void qla4_83xx_get_idc_param(struct scsi_qla_host *ha)
582*4882a593Smuzhiyun {
583*4882a593Smuzhiyun 	uint32_t idc_params, ret_val;
584*4882a593Smuzhiyun 
585*4882a593Smuzhiyun 	ret_val = qla4_83xx_flash_read_u32(ha, QLA83XX_IDC_PARAM_ADDR,
586*4882a593Smuzhiyun 					   (uint8_t *)&idc_params, 1);
587*4882a593Smuzhiyun 	if (ret_val == QLA_SUCCESS) {
588*4882a593Smuzhiyun 		ha->nx_dev_init_timeout = idc_params & 0xFFFF;
589*4882a593Smuzhiyun 		ha->nx_reset_timeout = (idc_params >> 16) & 0xFFFF;
590*4882a593Smuzhiyun 	} else {
591*4882a593Smuzhiyun 		ha->nx_dev_init_timeout = ROM_DEV_INIT_TIMEOUT;
592*4882a593Smuzhiyun 		ha->nx_reset_timeout = ROM_DRV_RESET_ACK_TIMEOUT;
593*4882a593Smuzhiyun 	}
594*4882a593Smuzhiyun 
595*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_DEBUG, ha,
596*4882a593Smuzhiyun 			  "%s: ha->nx_dev_init_timeout = %d, ha->nx_reset_timeout = %d\n",
597*4882a593Smuzhiyun 			  __func__, ha->nx_dev_init_timeout,
598*4882a593Smuzhiyun 			  ha->nx_reset_timeout));
599*4882a593Smuzhiyun }
600*4882a593Smuzhiyun 
601*4882a593Smuzhiyun /*-------------------------Reset Sequence Functions-----------------------*/
602*4882a593Smuzhiyun 
qla4_83xx_dump_reset_seq_hdr(struct scsi_qla_host * ha)603*4882a593Smuzhiyun static void qla4_83xx_dump_reset_seq_hdr(struct scsi_qla_host *ha)
604*4882a593Smuzhiyun {
605*4882a593Smuzhiyun 	uint8_t *phdr;
606*4882a593Smuzhiyun 
607*4882a593Smuzhiyun 	if (!ha->reset_tmplt.buff) {
608*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Error: Invalid reset_seq_template\n",
609*4882a593Smuzhiyun 			   __func__);
610*4882a593Smuzhiyun 		return;
611*4882a593Smuzhiyun 	}
612*4882a593Smuzhiyun 
613*4882a593Smuzhiyun 	phdr = ha->reset_tmplt.buff;
614*4882a593Smuzhiyun 
615*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha,
616*4882a593Smuzhiyun 			  "Reset Template: 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X 0x%X\n",
617*4882a593Smuzhiyun 			  *phdr, *(phdr+1), *(phdr+2), *(phdr+3), *(phdr+4),
618*4882a593Smuzhiyun 			  *(phdr+5), *(phdr+6), *(phdr+7), *(phdr + 8),
619*4882a593Smuzhiyun 			  *(phdr+9), *(phdr+10), *(phdr+11), *(phdr+12),
620*4882a593Smuzhiyun 			  *(phdr+13), *(phdr+14), *(phdr+15)));
621*4882a593Smuzhiyun }
622*4882a593Smuzhiyun 
qla4_83xx_copy_bootloader(struct scsi_qla_host * ha)623*4882a593Smuzhiyun static int qla4_83xx_copy_bootloader(struct scsi_qla_host *ha)
624*4882a593Smuzhiyun {
625*4882a593Smuzhiyun 	uint8_t *p_cache;
626*4882a593Smuzhiyun 	uint32_t src, count, size;
627*4882a593Smuzhiyun 	uint64_t dest;
628*4882a593Smuzhiyun 	int ret_val = QLA_SUCCESS;
629*4882a593Smuzhiyun 
630*4882a593Smuzhiyun 	src = QLA83XX_BOOTLOADER_FLASH_ADDR;
631*4882a593Smuzhiyun 	dest = qla4_83xx_rd_reg(ha, QLA83XX_BOOTLOADER_ADDR);
632*4882a593Smuzhiyun 	size = qla4_83xx_rd_reg(ha, QLA83XX_BOOTLOADER_SIZE);
633*4882a593Smuzhiyun 
634*4882a593Smuzhiyun 	/* 128 bit alignment check */
635*4882a593Smuzhiyun 	if (size & 0xF)
636*4882a593Smuzhiyun 		size = (size + 16) & ~0xF;
637*4882a593Smuzhiyun 
638*4882a593Smuzhiyun 	/* 16 byte count */
639*4882a593Smuzhiyun 	count = size/16;
640*4882a593Smuzhiyun 
641*4882a593Smuzhiyun 	p_cache = vmalloc(size);
642*4882a593Smuzhiyun 	if (p_cache == NULL) {
643*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Failed to allocate memory for boot loader cache\n",
644*4882a593Smuzhiyun 			   __func__);
645*4882a593Smuzhiyun 		ret_val = QLA_ERROR;
646*4882a593Smuzhiyun 		goto exit_copy_bootloader;
647*4882a593Smuzhiyun 	}
648*4882a593Smuzhiyun 
649*4882a593Smuzhiyun 	ret_val = qla4_83xx_lockless_flash_read_u32(ha, src, p_cache,
650*4882a593Smuzhiyun 						    size / sizeof(uint32_t));
651*4882a593Smuzhiyun 	if (ret_val == QLA_ERROR) {
652*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Error reading firmware from flash\n",
653*4882a593Smuzhiyun 			   __func__);
654*4882a593Smuzhiyun 		goto exit_copy_error;
655*4882a593Smuzhiyun 	}
656*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha, "%s: Read firmware from flash\n",
657*4882a593Smuzhiyun 			  __func__));
658*4882a593Smuzhiyun 
659*4882a593Smuzhiyun 	/* 128 bit/16 byte write to MS memory */
660*4882a593Smuzhiyun 	ret_val = qla4_8xxx_ms_mem_write_128b(ha, dest, (uint32_t *)p_cache,
661*4882a593Smuzhiyun 					      count);
662*4882a593Smuzhiyun 	if (ret_val == QLA_ERROR) {
663*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Error writing firmware to MS\n",
664*4882a593Smuzhiyun 			   __func__);
665*4882a593Smuzhiyun 		goto exit_copy_error;
666*4882a593Smuzhiyun 	}
667*4882a593Smuzhiyun 
668*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha, "%s: Wrote firmware size %d to MS\n",
669*4882a593Smuzhiyun 			  __func__, size));
670*4882a593Smuzhiyun 
671*4882a593Smuzhiyun exit_copy_error:
672*4882a593Smuzhiyun 	vfree(p_cache);
673*4882a593Smuzhiyun 
674*4882a593Smuzhiyun exit_copy_bootloader:
675*4882a593Smuzhiyun 	return ret_val;
676*4882a593Smuzhiyun }
677*4882a593Smuzhiyun 
qla4_83xx_check_cmd_peg_status(struct scsi_qla_host * ha)678*4882a593Smuzhiyun static int qla4_83xx_check_cmd_peg_status(struct scsi_qla_host *ha)
679*4882a593Smuzhiyun {
680*4882a593Smuzhiyun 	uint32_t val, ret_val = QLA_ERROR;
681*4882a593Smuzhiyun 	int retries = CRB_CMDPEG_CHECK_RETRY_COUNT;
682*4882a593Smuzhiyun 
683*4882a593Smuzhiyun 	do {
684*4882a593Smuzhiyun 		val = qla4_83xx_rd_reg(ha, QLA83XX_CMDPEG_STATE);
685*4882a593Smuzhiyun 		if (val == PHAN_INITIALIZE_COMPLETE) {
686*4882a593Smuzhiyun 			DEBUG2(ql4_printk(KERN_INFO, ha,
687*4882a593Smuzhiyun 					  "%s: Command Peg initialization complete. State=0x%x\n",
688*4882a593Smuzhiyun 					  __func__, val));
689*4882a593Smuzhiyun 			ret_val = QLA_SUCCESS;
690*4882a593Smuzhiyun 			break;
691*4882a593Smuzhiyun 		}
692*4882a593Smuzhiyun 		msleep(CRB_CMDPEG_CHECK_DELAY);
693*4882a593Smuzhiyun 	} while (--retries);
694*4882a593Smuzhiyun 
695*4882a593Smuzhiyun 	return ret_val;
696*4882a593Smuzhiyun }
697*4882a593Smuzhiyun 
698*4882a593Smuzhiyun /**
699*4882a593Smuzhiyun  * qla4_83xx_poll_reg - Poll the given CRB addr for duration msecs till
700*4882a593Smuzhiyun  * value read ANDed with test_mask is equal to test_result.
701*4882a593Smuzhiyun  *
702*4882a593Smuzhiyun  * @ha : Pointer to adapter structure
703*4882a593Smuzhiyun  * @addr : CRB register address
704*4882a593Smuzhiyun  * @duration : Poll for total of "duration" msecs
705*4882a593Smuzhiyun  * @test_mask : Mask value read with "test_mask"
706*4882a593Smuzhiyun  * @test_result : Compare (value&test_mask) with test_result.
707*4882a593Smuzhiyun  **/
qla4_83xx_poll_reg(struct scsi_qla_host * ha,uint32_t addr,int duration,uint32_t test_mask,uint32_t test_result)708*4882a593Smuzhiyun static int qla4_83xx_poll_reg(struct scsi_qla_host *ha, uint32_t addr,
709*4882a593Smuzhiyun 			      int duration, uint32_t test_mask,
710*4882a593Smuzhiyun 			      uint32_t test_result)
711*4882a593Smuzhiyun {
712*4882a593Smuzhiyun 	uint32_t value;
713*4882a593Smuzhiyun 	uint8_t retries;
714*4882a593Smuzhiyun 	int ret_val = QLA_SUCCESS;
715*4882a593Smuzhiyun 
716*4882a593Smuzhiyun 	ret_val = qla4_83xx_rd_reg_indirect(ha, addr, &value);
717*4882a593Smuzhiyun 	if (ret_val == QLA_ERROR)
718*4882a593Smuzhiyun 		goto exit_poll_reg;
719*4882a593Smuzhiyun 
720*4882a593Smuzhiyun 	retries = duration / 10;
721*4882a593Smuzhiyun 	do {
722*4882a593Smuzhiyun 		if ((value & test_mask) != test_result) {
723*4882a593Smuzhiyun 			msleep(duration / 10);
724*4882a593Smuzhiyun 			ret_val = qla4_83xx_rd_reg_indirect(ha, addr, &value);
725*4882a593Smuzhiyun 			if (ret_val == QLA_ERROR)
726*4882a593Smuzhiyun 				goto exit_poll_reg;
727*4882a593Smuzhiyun 
728*4882a593Smuzhiyun 			ret_val = QLA_ERROR;
729*4882a593Smuzhiyun 		} else {
730*4882a593Smuzhiyun 			ret_val = QLA_SUCCESS;
731*4882a593Smuzhiyun 			break;
732*4882a593Smuzhiyun 		}
733*4882a593Smuzhiyun 	} while (retries--);
734*4882a593Smuzhiyun 
735*4882a593Smuzhiyun exit_poll_reg:
736*4882a593Smuzhiyun 	if (ret_val == QLA_ERROR) {
737*4882a593Smuzhiyun 		ha->reset_tmplt.seq_error++;
738*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Poll Failed:  0x%08x 0x%08x 0x%08x\n",
739*4882a593Smuzhiyun 			   __func__, value, test_mask, test_result);
740*4882a593Smuzhiyun 	}
741*4882a593Smuzhiyun 
742*4882a593Smuzhiyun 	return ret_val;
743*4882a593Smuzhiyun }
744*4882a593Smuzhiyun 
qla4_83xx_reset_seq_checksum_test(struct scsi_qla_host * ha)745*4882a593Smuzhiyun static int qla4_83xx_reset_seq_checksum_test(struct scsi_qla_host *ha)
746*4882a593Smuzhiyun {
747*4882a593Smuzhiyun 	uint32_t sum =  0;
748*4882a593Smuzhiyun 	uint16_t *buff = (uint16_t *)ha->reset_tmplt.buff;
749*4882a593Smuzhiyun 	int u16_count =  ha->reset_tmplt.hdr->size / sizeof(uint16_t);
750*4882a593Smuzhiyun 	int ret_val;
751*4882a593Smuzhiyun 
752*4882a593Smuzhiyun 	while (u16_count-- > 0)
753*4882a593Smuzhiyun 		sum += *buff++;
754*4882a593Smuzhiyun 
755*4882a593Smuzhiyun 	while (sum >> 16)
756*4882a593Smuzhiyun 		sum = (sum & 0xFFFF) +  (sum >> 16);
757*4882a593Smuzhiyun 
758*4882a593Smuzhiyun 	/* checksum of 0 indicates a valid template */
759*4882a593Smuzhiyun 	if (~sum) {
760*4882a593Smuzhiyun 		ret_val = QLA_SUCCESS;
761*4882a593Smuzhiyun 	} else {
762*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Reset seq checksum failed\n",
763*4882a593Smuzhiyun 			   __func__);
764*4882a593Smuzhiyun 		ret_val = QLA_ERROR;
765*4882a593Smuzhiyun 	}
766*4882a593Smuzhiyun 
767*4882a593Smuzhiyun 	return ret_val;
768*4882a593Smuzhiyun }
769*4882a593Smuzhiyun 
770*4882a593Smuzhiyun /**
771*4882a593Smuzhiyun  * qla4_83xx_read_reset_template - Read Reset Template from Flash
772*4882a593Smuzhiyun  * @ha: Pointer to adapter structure
773*4882a593Smuzhiyun  **/
qla4_83xx_read_reset_template(struct scsi_qla_host * ha)774*4882a593Smuzhiyun void qla4_83xx_read_reset_template(struct scsi_qla_host *ha)
775*4882a593Smuzhiyun {
776*4882a593Smuzhiyun 	uint8_t *p_buff;
777*4882a593Smuzhiyun 	uint32_t addr, tmplt_hdr_def_size, tmplt_hdr_size;
778*4882a593Smuzhiyun 	uint32_t ret_val;
779*4882a593Smuzhiyun 
780*4882a593Smuzhiyun 	ha->reset_tmplt.seq_error = 0;
781*4882a593Smuzhiyun 	ha->reset_tmplt.buff = vmalloc(QLA83XX_RESTART_TEMPLATE_SIZE);
782*4882a593Smuzhiyun 	if (ha->reset_tmplt.buff == NULL) {
783*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Failed to allocate reset template resources\n",
784*4882a593Smuzhiyun 			   __func__);
785*4882a593Smuzhiyun 		goto exit_read_reset_template;
786*4882a593Smuzhiyun 	}
787*4882a593Smuzhiyun 
788*4882a593Smuzhiyun 	p_buff = ha->reset_tmplt.buff;
789*4882a593Smuzhiyun 	addr = QLA83XX_RESET_TEMPLATE_ADDR;
790*4882a593Smuzhiyun 
791*4882a593Smuzhiyun 	tmplt_hdr_def_size = sizeof(struct qla4_83xx_reset_template_hdr) /
792*4882a593Smuzhiyun 				    sizeof(uint32_t);
793*4882a593Smuzhiyun 
794*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha,
795*4882a593Smuzhiyun 			  "%s: Read template hdr size %d from Flash\n",
796*4882a593Smuzhiyun 			  __func__, tmplt_hdr_def_size));
797*4882a593Smuzhiyun 
798*4882a593Smuzhiyun 	/* Copy template header from flash */
799*4882a593Smuzhiyun 	ret_val = qla4_83xx_flash_read_u32(ha, addr, p_buff,
800*4882a593Smuzhiyun 					   tmplt_hdr_def_size);
801*4882a593Smuzhiyun 	if (ret_val != QLA_SUCCESS) {
802*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Failed to read reset template\n",
803*4882a593Smuzhiyun 			   __func__);
804*4882a593Smuzhiyun 		goto exit_read_template_error;
805*4882a593Smuzhiyun 	}
806*4882a593Smuzhiyun 
807*4882a593Smuzhiyun 	ha->reset_tmplt.hdr =
808*4882a593Smuzhiyun 		(struct qla4_83xx_reset_template_hdr *)ha->reset_tmplt.buff;
809*4882a593Smuzhiyun 
810*4882a593Smuzhiyun 	/* Validate the template header size and signature */
811*4882a593Smuzhiyun 	tmplt_hdr_size = ha->reset_tmplt.hdr->hdr_size/sizeof(uint32_t);
812*4882a593Smuzhiyun 	if ((tmplt_hdr_size != tmplt_hdr_def_size) ||
813*4882a593Smuzhiyun 	    (ha->reset_tmplt.hdr->signature != RESET_TMPLT_HDR_SIGNATURE)) {
814*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Template Header size %d is invalid, tmplt_hdr_def_size %d\n",
815*4882a593Smuzhiyun 			   __func__, tmplt_hdr_size, tmplt_hdr_def_size);
816*4882a593Smuzhiyun 		goto exit_read_template_error;
817*4882a593Smuzhiyun 	}
818*4882a593Smuzhiyun 
819*4882a593Smuzhiyun 	addr = QLA83XX_RESET_TEMPLATE_ADDR + ha->reset_tmplt.hdr->hdr_size;
820*4882a593Smuzhiyun 	p_buff = ha->reset_tmplt.buff + ha->reset_tmplt.hdr->hdr_size;
821*4882a593Smuzhiyun 	tmplt_hdr_def_size = (ha->reset_tmplt.hdr->size -
822*4882a593Smuzhiyun 			      ha->reset_tmplt.hdr->hdr_size) / sizeof(uint32_t);
823*4882a593Smuzhiyun 
824*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha,
825*4882a593Smuzhiyun 			  "%s: Read rest of the template size %d\n",
826*4882a593Smuzhiyun 			  __func__, ha->reset_tmplt.hdr->size));
827*4882a593Smuzhiyun 
828*4882a593Smuzhiyun 	/* Copy rest of the template */
829*4882a593Smuzhiyun 	ret_val = qla4_83xx_flash_read_u32(ha, addr, p_buff,
830*4882a593Smuzhiyun 					   tmplt_hdr_def_size);
831*4882a593Smuzhiyun 	if (ret_val != QLA_SUCCESS) {
832*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Failed to read reset template\n",
833*4882a593Smuzhiyun 			   __func__);
834*4882a593Smuzhiyun 		goto exit_read_template_error;
835*4882a593Smuzhiyun 	}
836*4882a593Smuzhiyun 
837*4882a593Smuzhiyun 	/* Integrity check */
838*4882a593Smuzhiyun 	if (qla4_83xx_reset_seq_checksum_test(ha)) {
839*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Reset Seq checksum failed!\n",
840*4882a593Smuzhiyun 			   __func__);
841*4882a593Smuzhiyun 		goto exit_read_template_error;
842*4882a593Smuzhiyun 	}
843*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha,
844*4882a593Smuzhiyun 			  "%s: Reset Seq checksum passed, Get stop, start and init seq offsets\n",
845*4882a593Smuzhiyun 			  __func__));
846*4882a593Smuzhiyun 
847*4882a593Smuzhiyun 	/* Get STOP, START, INIT sequence offsets */
848*4882a593Smuzhiyun 	ha->reset_tmplt.init_offset = ha->reset_tmplt.buff +
849*4882a593Smuzhiyun 				      ha->reset_tmplt.hdr->init_seq_offset;
850*4882a593Smuzhiyun 	ha->reset_tmplt.start_offset = ha->reset_tmplt.buff +
851*4882a593Smuzhiyun 				       ha->reset_tmplt.hdr->start_seq_offset;
852*4882a593Smuzhiyun 	ha->reset_tmplt.stop_offset = ha->reset_tmplt.buff +
853*4882a593Smuzhiyun 				      ha->reset_tmplt.hdr->hdr_size;
854*4882a593Smuzhiyun 	qla4_83xx_dump_reset_seq_hdr(ha);
855*4882a593Smuzhiyun 
856*4882a593Smuzhiyun 	goto exit_read_reset_template;
857*4882a593Smuzhiyun 
858*4882a593Smuzhiyun exit_read_template_error:
859*4882a593Smuzhiyun 	vfree(ha->reset_tmplt.buff);
860*4882a593Smuzhiyun 
861*4882a593Smuzhiyun exit_read_reset_template:
862*4882a593Smuzhiyun 	return;
863*4882a593Smuzhiyun }
864*4882a593Smuzhiyun 
865*4882a593Smuzhiyun /**
866*4882a593Smuzhiyun  * qla4_83xx_read_write_crb_reg - Read from raddr and write value to waddr.
867*4882a593Smuzhiyun  *
868*4882a593Smuzhiyun  * @ha : Pointer to adapter structure
869*4882a593Smuzhiyun  * @raddr : CRB address to read from
870*4882a593Smuzhiyun  * @waddr : CRB address to write to
871*4882a593Smuzhiyun  **/
qla4_83xx_read_write_crb_reg(struct scsi_qla_host * ha,uint32_t raddr,uint32_t waddr)872*4882a593Smuzhiyun static void qla4_83xx_read_write_crb_reg(struct scsi_qla_host *ha,
873*4882a593Smuzhiyun 					 uint32_t raddr, uint32_t waddr)
874*4882a593Smuzhiyun {
875*4882a593Smuzhiyun 	uint32_t value;
876*4882a593Smuzhiyun 
877*4882a593Smuzhiyun 	qla4_83xx_rd_reg_indirect(ha, raddr, &value);
878*4882a593Smuzhiyun 	qla4_83xx_wr_reg_indirect(ha, waddr, value);
879*4882a593Smuzhiyun }
880*4882a593Smuzhiyun 
881*4882a593Smuzhiyun /**
882*4882a593Smuzhiyun  * qla4_83xx_rmw_crb_reg - Read Modify Write crb register
883*4882a593Smuzhiyun  *
884*4882a593Smuzhiyun  * This function read value from raddr, AND with test_mask,
885*4882a593Smuzhiyun  * Shift Left,Right/OR/XOR with values RMW header and write value to waddr.
886*4882a593Smuzhiyun  *
887*4882a593Smuzhiyun  * @ha : Pointer to adapter structure
888*4882a593Smuzhiyun  * @raddr : CRB address to read from
889*4882a593Smuzhiyun  * @waddr : CRB address to write to
890*4882a593Smuzhiyun  * @p_rmw_hdr : header with shift/or/xor values.
891*4882a593Smuzhiyun  **/
qla4_83xx_rmw_crb_reg(struct scsi_qla_host * ha,uint32_t raddr,uint32_t waddr,struct qla4_83xx_rmw * p_rmw_hdr)892*4882a593Smuzhiyun static void qla4_83xx_rmw_crb_reg(struct scsi_qla_host *ha, uint32_t raddr,
893*4882a593Smuzhiyun 				  uint32_t waddr,
894*4882a593Smuzhiyun 				  struct qla4_83xx_rmw *p_rmw_hdr)
895*4882a593Smuzhiyun {
896*4882a593Smuzhiyun 	uint32_t value;
897*4882a593Smuzhiyun 
898*4882a593Smuzhiyun 	if (p_rmw_hdr->index_a)
899*4882a593Smuzhiyun 		value = ha->reset_tmplt.array[p_rmw_hdr->index_a];
900*4882a593Smuzhiyun 	else
901*4882a593Smuzhiyun 		qla4_83xx_rd_reg_indirect(ha, raddr, &value);
902*4882a593Smuzhiyun 
903*4882a593Smuzhiyun 	value &= p_rmw_hdr->test_mask;
904*4882a593Smuzhiyun 	value <<= p_rmw_hdr->shl;
905*4882a593Smuzhiyun 	value >>= p_rmw_hdr->shr;
906*4882a593Smuzhiyun 	value |= p_rmw_hdr->or_value;
907*4882a593Smuzhiyun 	value ^= p_rmw_hdr->xor_value;
908*4882a593Smuzhiyun 
909*4882a593Smuzhiyun 	qla4_83xx_wr_reg_indirect(ha, waddr, value);
910*4882a593Smuzhiyun 
911*4882a593Smuzhiyun 	return;
912*4882a593Smuzhiyun }
913*4882a593Smuzhiyun 
qla4_83xx_write_list(struct scsi_qla_host * ha,struct qla4_83xx_reset_entry_hdr * p_hdr)914*4882a593Smuzhiyun static void qla4_83xx_write_list(struct scsi_qla_host *ha,
915*4882a593Smuzhiyun 				 struct qla4_83xx_reset_entry_hdr *p_hdr)
916*4882a593Smuzhiyun {
917*4882a593Smuzhiyun 	struct qla4_83xx_entry *p_entry;
918*4882a593Smuzhiyun 	uint32_t i;
919*4882a593Smuzhiyun 
920*4882a593Smuzhiyun 	p_entry = (struct qla4_83xx_entry *)
921*4882a593Smuzhiyun 		  ((char *)p_hdr + sizeof(struct qla4_83xx_reset_entry_hdr));
922*4882a593Smuzhiyun 
923*4882a593Smuzhiyun 	for (i = 0; i < p_hdr->count; i++, p_entry++) {
924*4882a593Smuzhiyun 		qla4_83xx_wr_reg_indirect(ha, p_entry->arg1, p_entry->arg2);
925*4882a593Smuzhiyun 		if (p_hdr->delay)
926*4882a593Smuzhiyun 			udelay((uint32_t)(p_hdr->delay));
927*4882a593Smuzhiyun 	}
928*4882a593Smuzhiyun }
929*4882a593Smuzhiyun 
qla4_83xx_read_write_list(struct scsi_qla_host * ha,struct qla4_83xx_reset_entry_hdr * p_hdr)930*4882a593Smuzhiyun static void qla4_83xx_read_write_list(struct scsi_qla_host *ha,
931*4882a593Smuzhiyun 				      struct qla4_83xx_reset_entry_hdr *p_hdr)
932*4882a593Smuzhiyun {
933*4882a593Smuzhiyun 	struct qla4_83xx_entry *p_entry;
934*4882a593Smuzhiyun 	uint32_t i;
935*4882a593Smuzhiyun 
936*4882a593Smuzhiyun 	p_entry = (struct qla4_83xx_entry *)
937*4882a593Smuzhiyun 		  ((char *)p_hdr + sizeof(struct qla4_83xx_reset_entry_hdr));
938*4882a593Smuzhiyun 
939*4882a593Smuzhiyun 	for (i = 0; i < p_hdr->count; i++, p_entry++) {
940*4882a593Smuzhiyun 		qla4_83xx_read_write_crb_reg(ha, p_entry->arg1, p_entry->arg2);
941*4882a593Smuzhiyun 		if (p_hdr->delay)
942*4882a593Smuzhiyun 			udelay((uint32_t)(p_hdr->delay));
943*4882a593Smuzhiyun 	}
944*4882a593Smuzhiyun }
945*4882a593Smuzhiyun 
qla4_83xx_poll_list(struct scsi_qla_host * ha,struct qla4_83xx_reset_entry_hdr * p_hdr)946*4882a593Smuzhiyun static void qla4_83xx_poll_list(struct scsi_qla_host *ha,
947*4882a593Smuzhiyun 				struct qla4_83xx_reset_entry_hdr *p_hdr)
948*4882a593Smuzhiyun {
949*4882a593Smuzhiyun 	long delay;
950*4882a593Smuzhiyun 	struct qla4_83xx_entry *p_entry;
951*4882a593Smuzhiyun 	struct qla4_83xx_poll *p_poll;
952*4882a593Smuzhiyun 	uint32_t i;
953*4882a593Smuzhiyun 	uint32_t value;
954*4882a593Smuzhiyun 
955*4882a593Smuzhiyun 	p_poll = (struct qla4_83xx_poll *)
956*4882a593Smuzhiyun 		 ((char *)p_hdr + sizeof(struct qla4_83xx_reset_entry_hdr));
957*4882a593Smuzhiyun 
958*4882a593Smuzhiyun 	/* Entries start after 8 byte qla4_83xx_poll, poll header contains
959*4882a593Smuzhiyun 	 * the test_mask, test_value. */
960*4882a593Smuzhiyun 	p_entry = (struct qla4_83xx_entry *)((char *)p_poll +
961*4882a593Smuzhiyun 					     sizeof(struct qla4_83xx_poll));
962*4882a593Smuzhiyun 
963*4882a593Smuzhiyun 	delay = (long)p_hdr->delay;
964*4882a593Smuzhiyun 	if (!delay) {
965*4882a593Smuzhiyun 		for (i = 0; i < p_hdr->count; i++, p_entry++) {
966*4882a593Smuzhiyun 			qla4_83xx_poll_reg(ha, p_entry->arg1, delay,
967*4882a593Smuzhiyun 					   p_poll->test_mask,
968*4882a593Smuzhiyun 					   p_poll->test_value);
969*4882a593Smuzhiyun 		}
970*4882a593Smuzhiyun 	} else {
971*4882a593Smuzhiyun 		for (i = 0; i < p_hdr->count; i++, p_entry++) {
972*4882a593Smuzhiyun 			if (qla4_83xx_poll_reg(ha, p_entry->arg1, delay,
973*4882a593Smuzhiyun 					       p_poll->test_mask,
974*4882a593Smuzhiyun 					       p_poll->test_value)) {
975*4882a593Smuzhiyun 				qla4_83xx_rd_reg_indirect(ha, p_entry->arg1,
976*4882a593Smuzhiyun 							  &value);
977*4882a593Smuzhiyun 				qla4_83xx_rd_reg_indirect(ha, p_entry->arg2,
978*4882a593Smuzhiyun 							  &value);
979*4882a593Smuzhiyun 			}
980*4882a593Smuzhiyun 		}
981*4882a593Smuzhiyun 	}
982*4882a593Smuzhiyun }
983*4882a593Smuzhiyun 
qla4_83xx_poll_write_list(struct scsi_qla_host * ha,struct qla4_83xx_reset_entry_hdr * p_hdr)984*4882a593Smuzhiyun static void qla4_83xx_poll_write_list(struct scsi_qla_host *ha,
985*4882a593Smuzhiyun 				      struct qla4_83xx_reset_entry_hdr *p_hdr)
986*4882a593Smuzhiyun {
987*4882a593Smuzhiyun 	long delay;
988*4882a593Smuzhiyun 	struct qla4_83xx_quad_entry *p_entry;
989*4882a593Smuzhiyun 	struct qla4_83xx_poll *p_poll;
990*4882a593Smuzhiyun 	uint32_t i;
991*4882a593Smuzhiyun 
992*4882a593Smuzhiyun 	p_poll = (struct qla4_83xx_poll *)
993*4882a593Smuzhiyun 		 ((char *)p_hdr + sizeof(struct qla4_83xx_reset_entry_hdr));
994*4882a593Smuzhiyun 	p_entry = (struct qla4_83xx_quad_entry *)
995*4882a593Smuzhiyun 		  ((char *)p_poll + sizeof(struct qla4_83xx_poll));
996*4882a593Smuzhiyun 	delay = (long)p_hdr->delay;
997*4882a593Smuzhiyun 
998*4882a593Smuzhiyun 	for (i = 0; i < p_hdr->count; i++, p_entry++) {
999*4882a593Smuzhiyun 		qla4_83xx_wr_reg_indirect(ha, p_entry->dr_addr,
1000*4882a593Smuzhiyun 					  p_entry->dr_value);
1001*4882a593Smuzhiyun 		qla4_83xx_wr_reg_indirect(ha, p_entry->ar_addr,
1002*4882a593Smuzhiyun 					  p_entry->ar_value);
1003*4882a593Smuzhiyun 		if (delay) {
1004*4882a593Smuzhiyun 			if (qla4_83xx_poll_reg(ha, p_entry->ar_addr, delay,
1005*4882a593Smuzhiyun 					       p_poll->test_mask,
1006*4882a593Smuzhiyun 					       p_poll->test_value)) {
1007*4882a593Smuzhiyun 				DEBUG2(ql4_printk(KERN_INFO, ha,
1008*4882a593Smuzhiyun 						  "%s: Timeout Error: poll list, item_num %d, entry_num %d\n",
1009*4882a593Smuzhiyun 						  __func__, i,
1010*4882a593Smuzhiyun 						  ha->reset_tmplt.seq_index));
1011*4882a593Smuzhiyun 			}
1012*4882a593Smuzhiyun 		}
1013*4882a593Smuzhiyun 	}
1014*4882a593Smuzhiyun }
1015*4882a593Smuzhiyun 
qla4_83xx_read_modify_write(struct scsi_qla_host * ha,struct qla4_83xx_reset_entry_hdr * p_hdr)1016*4882a593Smuzhiyun static void qla4_83xx_read_modify_write(struct scsi_qla_host *ha,
1017*4882a593Smuzhiyun 					struct qla4_83xx_reset_entry_hdr *p_hdr)
1018*4882a593Smuzhiyun {
1019*4882a593Smuzhiyun 	struct qla4_83xx_entry *p_entry;
1020*4882a593Smuzhiyun 	struct qla4_83xx_rmw *p_rmw_hdr;
1021*4882a593Smuzhiyun 	uint32_t i;
1022*4882a593Smuzhiyun 
1023*4882a593Smuzhiyun 	p_rmw_hdr = (struct qla4_83xx_rmw *)
1024*4882a593Smuzhiyun 		    ((char *)p_hdr + sizeof(struct qla4_83xx_reset_entry_hdr));
1025*4882a593Smuzhiyun 	p_entry = (struct qla4_83xx_entry *)
1026*4882a593Smuzhiyun 		  ((char *)p_rmw_hdr + sizeof(struct qla4_83xx_rmw));
1027*4882a593Smuzhiyun 
1028*4882a593Smuzhiyun 	for (i = 0; i < p_hdr->count; i++, p_entry++) {
1029*4882a593Smuzhiyun 		qla4_83xx_rmw_crb_reg(ha, p_entry->arg1, p_entry->arg2,
1030*4882a593Smuzhiyun 				      p_rmw_hdr);
1031*4882a593Smuzhiyun 		if (p_hdr->delay)
1032*4882a593Smuzhiyun 			udelay((uint32_t)(p_hdr->delay));
1033*4882a593Smuzhiyun 	}
1034*4882a593Smuzhiyun }
1035*4882a593Smuzhiyun 
qla4_83xx_pause(struct scsi_qla_host * ha,struct qla4_83xx_reset_entry_hdr * p_hdr)1036*4882a593Smuzhiyun static void qla4_83xx_pause(struct scsi_qla_host *ha,
1037*4882a593Smuzhiyun 			    struct qla4_83xx_reset_entry_hdr *p_hdr)
1038*4882a593Smuzhiyun {
1039*4882a593Smuzhiyun 	if (p_hdr->delay)
1040*4882a593Smuzhiyun 		mdelay((uint32_t)((long)p_hdr->delay));
1041*4882a593Smuzhiyun }
1042*4882a593Smuzhiyun 
qla4_83xx_poll_read_list(struct scsi_qla_host * ha,struct qla4_83xx_reset_entry_hdr * p_hdr)1043*4882a593Smuzhiyun static void qla4_83xx_poll_read_list(struct scsi_qla_host *ha,
1044*4882a593Smuzhiyun 				     struct qla4_83xx_reset_entry_hdr *p_hdr)
1045*4882a593Smuzhiyun {
1046*4882a593Smuzhiyun 	long delay;
1047*4882a593Smuzhiyun 	int index;
1048*4882a593Smuzhiyun 	struct qla4_83xx_quad_entry *p_entry;
1049*4882a593Smuzhiyun 	struct qla4_83xx_poll *p_poll;
1050*4882a593Smuzhiyun 	uint32_t i;
1051*4882a593Smuzhiyun 	uint32_t value;
1052*4882a593Smuzhiyun 
1053*4882a593Smuzhiyun 	p_poll = (struct qla4_83xx_poll *)
1054*4882a593Smuzhiyun 		 ((char *)p_hdr + sizeof(struct qla4_83xx_reset_entry_hdr));
1055*4882a593Smuzhiyun 	p_entry = (struct qla4_83xx_quad_entry *)
1056*4882a593Smuzhiyun 		  ((char *)p_poll + sizeof(struct qla4_83xx_poll));
1057*4882a593Smuzhiyun 	delay = (long)p_hdr->delay;
1058*4882a593Smuzhiyun 
1059*4882a593Smuzhiyun 	for (i = 0; i < p_hdr->count; i++, p_entry++) {
1060*4882a593Smuzhiyun 		qla4_83xx_wr_reg_indirect(ha, p_entry->ar_addr,
1061*4882a593Smuzhiyun 					  p_entry->ar_value);
1062*4882a593Smuzhiyun 		if (delay) {
1063*4882a593Smuzhiyun 			if (qla4_83xx_poll_reg(ha, p_entry->ar_addr, delay,
1064*4882a593Smuzhiyun 					       p_poll->test_mask,
1065*4882a593Smuzhiyun 					       p_poll->test_value)) {
1066*4882a593Smuzhiyun 				DEBUG2(ql4_printk(KERN_INFO, ha,
1067*4882a593Smuzhiyun 						  "%s: Timeout Error: poll list, Item_num %d, entry_num %d\n",
1068*4882a593Smuzhiyun 						  __func__, i,
1069*4882a593Smuzhiyun 						  ha->reset_tmplt.seq_index));
1070*4882a593Smuzhiyun 			} else {
1071*4882a593Smuzhiyun 				index = ha->reset_tmplt.array_index;
1072*4882a593Smuzhiyun 				qla4_83xx_rd_reg_indirect(ha, p_entry->dr_addr,
1073*4882a593Smuzhiyun 							  &value);
1074*4882a593Smuzhiyun 				ha->reset_tmplt.array[index++] = value;
1075*4882a593Smuzhiyun 
1076*4882a593Smuzhiyun 				if (index == QLA83XX_MAX_RESET_SEQ_ENTRIES)
1077*4882a593Smuzhiyun 					ha->reset_tmplt.array_index = 1;
1078*4882a593Smuzhiyun 			}
1079*4882a593Smuzhiyun 		}
1080*4882a593Smuzhiyun 	}
1081*4882a593Smuzhiyun }
1082*4882a593Smuzhiyun 
qla4_83xx_seq_end(struct scsi_qla_host * ha,struct qla4_83xx_reset_entry_hdr * p_hdr)1083*4882a593Smuzhiyun static void qla4_83xx_seq_end(struct scsi_qla_host *ha,
1084*4882a593Smuzhiyun 			      struct qla4_83xx_reset_entry_hdr *p_hdr)
1085*4882a593Smuzhiyun {
1086*4882a593Smuzhiyun 	ha->reset_tmplt.seq_end = 1;
1087*4882a593Smuzhiyun }
1088*4882a593Smuzhiyun 
qla4_83xx_template_end(struct scsi_qla_host * ha,struct qla4_83xx_reset_entry_hdr * p_hdr)1089*4882a593Smuzhiyun static void qla4_83xx_template_end(struct scsi_qla_host *ha,
1090*4882a593Smuzhiyun 				   struct qla4_83xx_reset_entry_hdr *p_hdr)
1091*4882a593Smuzhiyun {
1092*4882a593Smuzhiyun 	ha->reset_tmplt.template_end = 1;
1093*4882a593Smuzhiyun 
1094*4882a593Smuzhiyun 	if (ha->reset_tmplt.seq_error == 0) {
1095*4882a593Smuzhiyun 		DEBUG2(ql4_printk(KERN_INFO, ha,
1096*4882a593Smuzhiyun 				  "%s: Reset sequence completed SUCCESSFULLY.\n",
1097*4882a593Smuzhiyun 				  __func__));
1098*4882a593Smuzhiyun 	} else {
1099*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Reset sequence completed with some timeout errors.\n",
1100*4882a593Smuzhiyun 			   __func__);
1101*4882a593Smuzhiyun 	}
1102*4882a593Smuzhiyun }
1103*4882a593Smuzhiyun 
1104*4882a593Smuzhiyun /**
1105*4882a593Smuzhiyun  * qla4_83xx_process_reset_template - Process reset template.
1106*4882a593Smuzhiyun  *
1107*4882a593Smuzhiyun  * Process all entries in reset template till entry with SEQ_END opcode,
1108*4882a593Smuzhiyun  * which indicates end of the reset template processing. Each entry has a
1109*4882a593Smuzhiyun  * Reset Entry header, entry opcode/command, with size of the entry, number
1110*4882a593Smuzhiyun  * of entries in sub-sequence and delay in microsecs or timeout in millisecs.
1111*4882a593Smuzhiyun  *
1112*4882a593Smuzhiyun  * @ha : Pointer to adapter structure
1113*4882a593Smuzhiyun  * @p_buff : Common reset entry header.
1114*4882a593Smuzhiyun  **/
qla4_83xx_process_reset_template(struct scsi_qla_host * ha,char * p_buff)1115*4882a593Smuzhiyun static void qla4_83xx_process_reset_template(struct scsi_qla_host *ha,
1116*4882a593Smuzhiyun 					     char *p_buff)
1117*4882a593Smuzhiyun {
1118*4882a593Smuzhiyun 	int index, entries;
1119*4882a593Smuzhiyun 	struct qla4_83xx_reset_entry_hdr *p_hdr;
1120*4882a593Smuzhiyun 	char *p_entry = p_buff;
1121*4882a593Smuzhiyun 
1122*4882a593Smuzhiyun 	ha->reset_tmplt.seq_end = 0;
1123*4882a593Smuzhiyun 	ha->reset_tmplt.template_end = 0;
1124*4882a593Smuzhiyun 	entries = ha->reset_tmplt.hdr->entries;
1125*4882a593Smuzhiyun 	index = ha->reset_tmplt.seq_index;
1126*4882a593Smuzhiyun 
1127*4882a593Smuzhiyun 	for (; (!ha->reset_tmplt.seq_end) && (index  < entries); index++) {
1128*4882a593Smuzhiyun 
1129*4882a593Smuzhiyun 		p_hdr = (struct qla4_83xx_reset_entry_hdr *)p_entry;
1130*4882a593Smuzhiyun 		switch (p_hdr->cmd) {
1131*4882a593Smuzhiyun 		case OPCODE_NOP:
1132*4882a593Smuzhiyun 			break;
1133*4882a593Smuzhiyun 		case OPCODE_WRITE_LIST:
1134*4882a593Smuzhiyun 			qla4_83xx_write_list(ha, p_hdr);
1135*4882a593Smuzhiyun 			break;
1136*4882a593Smuzhiyun 		case OPCODE_READ_WRITE_LIST:
1137*4882a593Smuzhiyun 			qla4_83xx_read_write_list(ha, p_hdr);
1138*4882a593Smuzhiyun 			break;
1139*4882a593Smuzhiyun 		case OPCODE_POLL_LIST:
1140*4882a593Smuzhiyun 			qla4_83xx_poll_list(ha, p_hdr);
1141*4882a593Smuzhiyun 			break;
1142*4882a593Smuzhiyun 		case OPCODE_POLL_WRITE_LIST:
1143*4882a593Smuzhiyun 			qla4_83xx_poll_write_list(ha, p_hdr);
1144*4882a593Smuzhiyun 			break;
1145*4882a593Smuzhiyun 		case OPCODE_READ_MODIFY_WRITE:
1146*4882a593Smuzhiyun 			qla4_83xx_read_modify_write(ha, p_hdr);
1147*4882a593Smuzhiyun 			break;
1148*4882a593Smuzhiyun 		case OPCODE_SEQ_PAUSE:
1149*4882a593Smuzhiyun 			qla4_83xx_pause(ha, p_hdr);
1150*4882a593Smuzhiyun 			break;
1151*4882a593Smuzhiyun 		case OPCODE_SEQ_END:
1152*4882a593Smuzhiyun 			qla4_83xx_seq_end(ha, p_hdr);
1153*4882a593Smuzhiyun 			break;
1154*4882a593Smuzhiyun 		case OPCODE_TMPL_END:
1155*4882a593Smuzhiyun 			qla4_83xx_template_end(ha, p_hdr);
1156*4882a593Smuzhiyun 			break;
1157*4882a593Smuzhiyun 		case OPCODE_POLL_READ_LIST:
1158*4882a593Smuzhiyun 			qla4_83xx_poll_read_list(ha, p_hdr);
1159*4882a593Smuzhiyun 			break;
1160*4882a593Smuzhiyun 		default:
1161*4882a593Smuzhiyun 			ql4_printk(KERN_ERR, ha, "%s: Unknown command ==> 0x%04x on entry = %d\n",
1162*4882a593Smuzhiyun 				   __func__, p_hdr->cmd, index);
1163*4882a593Smuzhiyun 			break;
1164*4882a593Smuzhiyun 		}
1165*4882a593Smuzhiyun 
1166*4882a593Smuzhiyun 		/* Set pointer to next entry in the sequence. */
1167*4882a593Smuzhiyun 		p_entry += p_hdr->size;
1168*4882a593Smuzhiyun 	}
1169*4882a593Smuzhiyun 
1170*4882a593Smuzhiyun 	ha->reset_tmplt.seq_index = index;
1171*4882a593Smuzhiyun }
1172*4882a593Smuzhiyun 
qla4_83xx_process_stop_seq(struct scsi_qla_host * ha)1173*4882a593Smuzhiyun static void qla4_83xx_process_stop_seq(struct scsi_qla_host *ha)
1174*4882a593Smuzhiyun {
1175*4882a593Smuzhiyun 	ha->reset_tmplt.seq_index = 0;
1176*4882a593Smuzhiyun 	qla4_83xx_process_reset_template(ha, ha->reset_tmplt.stop_offset);
1177*4882a593Smuzhiyun 
1178*4882a593Smuzhiyun 	if (ha->reset_tmplt.seq_end != 1)
1179*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Abrupt STOP Sub-Sequence end.\n",
1180*4882a593Smuzhiyun 			   __func__);
1181*4882a593Smuzhiyun }
1182*4882a593Smuzhiyun 
qla4_83xx_process_start_seq(struct scsi_qla_host * ha)1183*4882a593Smuzhiyun static void qla4_83xx_process_start_seq(struct scsi_qla_host *ha)
1184*4882a593Smuzhiyun {
1185*4882a593Smuzhiyun 	qla4_83xx_process_reset_template(ha, ha->reset_tmplt.start_offset);
1186*4882a593Smuzhiyun 
1187*4882a593Smuzhiyun 	if (ha->reset_tmplt.template_end != 1)
1188*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Abrupt START Sub-Sequence end.\n",
1189*4882a593Smuzhiyun 			   __func__);
1190*4882a593Smuzhiyun }
1191*4882a593Smuzhiyun 
qla4_83xx_process_init_seq(struct scsi_qla_host * ha)1192*4882a593Smuzhiyun static void qla4_83xx_process_init_seq(struct scsi_qla_host *ha)
1193*4882a593Smuzhiyun {
1194*4882a593Smuzhiyun 	qla4_83xx_process_reset_template(ha, ha->reset_tmplt.init_offset);
1195*4882a593Smuzhiyun 
1196*4882a593Smuzhiyun 	if (ha->reset_tmplt.seq_end != 1)
1197*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Abrupt INIT Sub-Sequence end.\n",
1198*4882a593Smuzhiyun 			   __func__);
1199*4882a593Smuzhiyun }
1200*4882a593Smuzhiyun 
qla4_83xx_restart(struct scsi_qla_host * ha)1201*4882a593Smuzhiyun static int qla4_83xx_restart(struct scsi_qla_host *ha)
1202*4882a593Smuzhiyun {
1203*4882a593Smuzhiyun 	int ret_val = QLA_SUCCESS;
1204*4882a593Smuzhiyun 	uint32_t idc_ctrl;
1205*4882a593Smuzhiyun 
1206*4882a593Smuzhiyun 	qla4_83xx_process_stop_seq(ha);
1207*4882a593Smuzhiyun 
1208*4882a593Smuzhiyun 	/*
1209*4882a593Smuzhiyun 	 * Collect minidump.
1210*4882a593Smuzhiyun 	 * If IDC_CTRL BIT1 is set, clear it on going to INIT state and
1211*4882a593Smuzhiyun 	 * don't collect minidump
1212*4882a593Smuzhiyun 	 */
1213*4882a593Smuzhiyun 	idc_ctrl = qla4_83xx_rd_reg(ha, QLA83XX_IDC_DRV_CTRL);
1214*4882a593Smuzhiyun 	if (idc_ctrl & GRACEFUL_RESET_BIT1) {
1215*4882a593Smuzhiyun 		qla4_83xx_wr_reg(ha, QLA83XX_IDC_DRV_CTRL,
1216*4882a593Smuzhiyun 				 (idc_ctrl & ~GRACEFUL_RESET_BIT1));
1217*4882a593Smuzhiyun 		ql4_printk(KERN_INFO, ha, "%s: Graceful RESET: Not collecting minidump\n",
1218*4882a593Smuzhiyun 			   __func__);
1219*4882a593Smuzhiyun 	} else {
1220*4882a593Smuzhiyun 		qla4_8xxx_get_minidump(ha);
1221*4882a593Smuzhiyun 	}
1222*4882a593Smuzhiyun 
1223*4882a593Smuzhiyun 	qla4_83xx_process_init_seq(ha);
1224*4882a593Smuzhiyun 
1225*4882a593Smuzhiyun 	if (qla4_83xx_copy_bootloader(ha)) {
1226*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Copy bootloader, firmware restart failed!\n",
1227*4882a593Smuzhiyun 			   __func__);
1228*4882a593Smuzhiyun 		ret_val = QLA_ERROR;
1229*4882a593Smuzhiyun 		goto exit_restart;
1230*4882a593Smuzhiyun 	}
1231*4882a593Smuzhiyun 
1232*4882a593Smuzhiyun 	qla4_83xx_wr_reg(ha, QLA83XX_FW_IMAGE_VALID, QLA83XX_BOOT_FROM_FLASH);
1233*4882a593Smuzhiyun 	qla4_83xx_process_start_seq(ha);
1234*4882a593Smuzhiyun 
1235*4882a593Smuzhiyun exit_restart:
1236*4882a593Smuzhiyun 	return ret_val;
1237*4882a593Smuzhiyun }
1238*4882a593Smuzhiyun 
qla4_83xx_start_firmware(struct scsi_qla_host * ha)1239*4882a593Smuzhiyun int qla4_83xx_start_firmware(struct scsi_qla_host *ha)
1240*4882a593Smuzhiyun {
1241*4882a593Smuzhiyun 	int ret_val = QLA_SUCCESS;
1242*4882a593Smuzhiyun 
1243*4882a593Smuzhiyun 	ret_val = qla4_83xx_restart(ha);
1244*4882a593Smuzhiyun 	if (ret_val == QLA_ERROR) {
1245*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Restart error\n", __func__);
1246*4882a593Smuzhiyun 		goto exit_start_fw;
1247*4882a593Smuzhiyun 	} else {
1248*4882a593Smuzhiyun 		DEBUG2(ql4_printk(KERN_INFO, ha, "%s: Restart done\n",
1249*4882a593Smuzhiyun 				  __func__));
1250*4882a593Smuzhiyun 	}
1251*4882a593Smuzhiyun 
1252*4882a593Smuzhiyun 	ret_val = qla4_83xx_check_cmd_peg_status(ha);
1253*4882a593Smuzhiyun 	if (ret_val == QLA_ERROR)
1254*4882a593Smuzhiyun 		ql4_printk(KERN_ERR, ha, "%s: Peg not initialized\n",
1255*4882a593Smuzhiyun 			   __func__);
1256*4882a593Smuzhiyun 
1257*4882a593Smuzhiyun exit_start_fw:
1258*4882a593Smuzhiyun 	return ret_val;
1259*4882a593Smuzhiyun }
1260*4882a593Smuzhiyun 
1261*4882a593Smuzhiyun /*----------------------Interrupt Related functions ---------------------*/
1262*4882a593Smuzhiyun 
qla4_83xx_disable_iocb_intrs(struct scsi_qla_host * ha)1263*4882a593Smuzhiyun static void qla4_83xx_disable_iocb_intrs(struct scsi_qla_host *ha)
1264*4882a593Smuzhiyun {
1265*4882a593Smuzhiyun 	if (test_and_clear_bit(AF_83XX_IOCB_INTR_ON, &ha->flags))
1266*4882a593Smuzhiyun 		qla4_8xxx_intr_disable(ha);
1267*4882a593Smuzhiyun }
1268*4882a593Smuzhiyun 
qla4_83xx_disable_mbox_intrs(struct scsi_qla_host * ha)1269*4882a593Smuzhiyun static void qla4_83xx_disable_mbox_intrs(struct scsi_qla_host *ha)
1270*4882a593Smuzhiyun {
1271*4882a593Smuzhiyun 	uint32_t mb_int, ret;
1272*4882a593Smuzhiyun 
1273*4882a593Smuzhiyun 	if (test_and_clear_bit(AF_83XX_MBOX_INTR_ON, &ha->flags)) {
1274*4882a593Smuzhiyun 		ret = readl(&ha->qla4_83xx_reg->mbox_int);
1275*4882a593Smuzhiyun 		mb_int = ret & ~INT_ENABLE_FW_MB;
1276*4882a593Smuzhiyun 		writel(mb_int, &ha->qla4_83xx_reg->mbox_int);
1277*4882a593Smuzhiyun 		writel(1, &ha->qla4_83xx_reg->leg_int_mask);
1278*4882a593Smuzhiyun 	}
1279*4882a593Smuzhiyun }
1280*4882a593Smuzhiyun 
qla4_83xx_disable_intrs(struct scsi_qla_host * ha)1281*4882a593Smuzhiyun void qla4_83xx_disable_intrs(struct scsi_qla_host *ha)
1282*4882a593Smuzhiyun {
1283*4882a593Smuzhiyun 	qla4_83xx_disable_mbox_intrs(ha);
1284*4882a593Smuzhiyun 	qla4_83xx_disable_iocb_intrs(ha);
1285*4882a593Smuzhiyun }
1286*4882a593Smuzhiyun 
qla4_83xx_enable_iocb_intrs(struct scsi_qla_host * ha)1287*4882a593Smuzhiyun static void qla4_83xx_enable_iocb_intrs(struct scsi_qla_host *ha)
1288*4882a593Smuzhiyun {
1289*4882a593Smuzhiyun 	if (!test_bit(AF_83XX_IOCB_INTR_ON, &ha->flags)) {
1290*4882a593Smuzhiyun 		qla4_8xxx_intr_enable(ha);
1291*4882a593Smuzhiyun 		set_bit(AF_83XX_IOCB_INTR_ON, &ha->flags);
1292*4882a593Smuzhiyun 	}
1293*4882a593Smuzhiyun }
1294*4882a593Smuzhiyun 
qla4_83xx_enable_mbox_intrs(struct scsi_qla_host * ha)1295*4882a593Smuzhiyun void qla4_83xx_enable_mbox_intrs(struct scsi_qla_host *ha)
1296*4882a593Smuzhiyun {
1297*4882a593Smuzhiyun 	uint32_t mb_int;
1298*4882a593Smuzhiyun 
1299*4882a593Smuzhiyun 	if (!test_bit(AF_83XX_MBOX_INTR_ON, &ha->flags)) {
1300*4882a593Smuzhiyun 		mb_int = INT_ENABLE_FW_MB;
1301*4882a593Smuzhiyun 		writel(mb_int, &ha->qla4_83xx_reg->mbox_int);
1302*4882a593Smuzhiyun 		writel(0, &ha->qla4_83xx_reg->leg_int_mask);
1303*4882a593Smuzhiyun 		set_bit(AF_83XX_MBOX_INTR_ON, &ha->flags);
1304*4882a593Smuzhiyun 	}
1305*4882a593Smuzhiyun }
1306*4882a593Smuzhiyun 
1307*4882a593Smuzhiyun 
qla4_83xx_enable_intrs(struct scsi_qla_host * ha)1308*4882a593Smuzhiyun void qla4_83xx_enable_intrs(struct scsi_qla_host *ha)
1309*4882a593Smuzhiyun {
1310*4882a593Smuzhiyun 	qla4_83xx_enable_mbox_intrs(ha);
1311*4882a593Smuzhiyun 	qla4_83xx_enable_iocb_intrs(ha);
1312*4882a593Smuzhiyun }
1313*4882a593Smuzhiyun 
1314*4882a593Smuzhiyun 
qla4_83xx_queue_mbox_cmd(struct scsi_qla_host * ha,uint32_t * mbx_cmd,int incount)1315*4882a593Smuzhiyun void qla4_83xx_queue_mbox_cmd(struct scsi_qla_host *ha, uint32_t *mbx_cmd,
1316*4882a593Smuzhiyun 			      int incount)
1317*4882a593Smuzhiyun {
1318*4882a593Smuzhiyun 	int i;
1319*4882a593Smuzhiyun 
1320*4882a593Smuzhiyun 	/* Load all mailbox registers, except mailbox 0. */
1321*4882a593Smuzhiyun 	for (i = 1; i < incount; i++)
1322*4882a593Smuzhiyun 		writel(mbx_cmd[i], &ha->qla4_83xx_reg->mailbox_in[i]);
1323*4882a593Smuzhiyun 
1324*4882a593Smuzhiyun 	writel(mbx_cmd[0], &ha->qla4_83xx_reg->mailbox_in[0]);
1325*4882a593Smuzhiyun 
1326*4882a593Smuzhiyun 	/* Set Host Interrupt register to 1, to tell the firmware that
1327*4882a593Smuzhiyun 	 * a mailbox command is pending. Firmware after reading the
1328*4882a593Smuzhiyun 	 * mailbox command, clears the host interrupt register */
1329*4882a593Smuzhiyun 	writel(HINT_MBX_INT_PENDING, &ha->qla4_83xx_reg->host_intr);
1330*4882a593Smuzhiyun }
1331*4882a593Smuzhiyun 
qla4_83xx_process_mbox_intr(struct scsi_qla_host * ha,int outcount)1332*4882a593Smuzhiyun void qla4_83xx_process_mbox_intr(struct scsi_qla_host *ha, int outcount)
1333*4882a593Smuzhiyun {
1334*4882a593Smuzhiyun 	int intr_status;
1335*4882a593Smuzhiyun 
1336*4882a593Smuzhiyun 	intr_status = readl(&ha->qla4_83xx_reg->risc_intr);
1337*4882a593Smuzhiyun 	if (intr_status) {
1338*4882a593Smuzhiyun 		ha->mbox_status_count = outcount;
1339*4882a593Smuzhiyun 		ha->isp_ops->interrupt_service_routine(ha, intr_status);
1340*4882a593Smuzhiyun 	}
1341*4882a593Smuzhiyun }
1342*4882a593Smuzhiyun 
1343*4882a593Smuzhiyun /**
1344*4882a593Smuzhiyun  * qla4_83xx_isp_reset - Resets ISP and aborts all outstanding commands.
1345*4882a593Smuzhiyun  * @ha: pointer to host adapter structure.
1346*4882a593Smuzhiyun  **/
qla4_83xx_isp_reset(struct scsi_qla_host * ha)1347*4882a593Smuzhiyun int qla4_83xx_isp_reset(struct scsi_qla_host *ha)
1348*4882a593Smuzhiyun {
1349*4882a593Smuzhiyun 	int rval;
1350*4882a593Smuzhiyun 	uint32_t dev_state;
1351*4882a593Smuzhiyun 
1352*4882a593Smuzhiyun 	ha->isp_ops->idc_lock(ha);
1353*4882a593Smuzhiyun 	dev_state = qla4_8xxx_rd_direct(ha, QLA8XXX_CRB_DEV_STATE);
1354*4882a593Smuzhiyun 
1355*4882a593Smuzhiyun 	if (ql4xdontresethba)
1356*4882a593Smuzhiyun 		qla4_83xx_set_idc_dontreset(ha);
1357*4882a593Smuzhiyun 
1358*4882a593Smuzhiyun 	if (dev_state == QLA8XXX_DEV_READY) {
1359*4882a593Smuzhiyun 		/* If IDC_CTRL DONTRESETHBA_BIT0 is set dont do reset
1360*4882a593Smuzhiyun 		 * recovery */
1361*4882a593Smuzhiyun 		if (qla4_83xx_idc_dontreset(ha) == DONTRESET_BIT0) {
1362*4882a593Smuzhiyun 			ql4_printk(KERN_ERR, ha, "%s: Reset recovery disabled\n",
1363*4882a593Smuzhiyun 				   __func__);
1364*4882a593Smuzhiyun 			rval = QLA_ERROR;
1365*4882a593Smuzhiyun 			goto exit_isp_reset;
1366*4882a593Smuzhiyun 		}
1367*4882a593Smuzhiyun 
1368*4882a593Smuzhiyun 		DEBUG2(ql4_printk(KERN_INFO, ha, "%s: HW State: NEED RESET\n",
1369*4882a593Smuzhiyun 				  __func__));
1370*4882a593Smuzhiyun 		qla4_8xxx_wr_direct(ha, QLA8XXX_CRB_DEV_STATE,
1371*4882a593Smuzhiyun 				    QLA8XXX_DEV_NEED_RESET);
1372*4882a593Smuzhiyun 
1373*4882a593Smuzhiyun 	} else {
1374*4882a593Smuzhiyun 		/* If device_state is NEED_RESET, go ahead with
1375*4882a593Smuzhiyun 		 * Reset,irrespective of ql4xdontresethba. This is to allow a
1376*4882a593Smuzhiyun 		 * non-reset-owner to force a reset. Non-reset-owner sets
1377*4882a593Smuzhiyun 		 * the IDC_CTRL BIT0 to prevent Reset-owner from doing a Reset
1378*4882a593Smuzhiyun 		 * and then forces a Reset by setting device_state to
1379*4882a593Smuzhiyun 		 * NEED_RESET. */
1380*4882a593Smuzhiyun 		DEBUG2(ql4_printk(KERN_INFO, ha,
1381*4882a593Smuzhiyun 				  "%s: HW state already set to NEED_RESET\n",
1382*4882a593Smuzhiyun 				  __func__));
1383*4882a593Smuzhiyun 	}
1384*4882a593Smuzhiyun 
1385*4882a593Smuzhiyun 	/* For ISP8324 and ISP8042, Reset owner is NIC, iSCSI or FCOE based on
1386*4882a593Smuzhiyun 	 * priority and which drivers are present. Unlike ISP8022, the function
1387*4882a593Smuzhiyun 	 * setting NEED_RESET, may not be the Reset owner. */
1388*4882a593Smuzhiyun 	if (qla4_83xx_can_perform_reset(ha))
1389*4882a593Smuzhiyun 		set_bit(AF_8XXX_RST_OWNER, &ha->flags);
1390*4882a593Smuzhiyun 
1391*4882a593Smuzhiyun 	ha->isp_ops->idc_unlock(ha);
1392*4882a593Smuzhiyun 	rval = qla4_8xxx_device_state_handler(ha);
1393*4882a593Smuzhiyun 
1394*4882a593Smuzhiyun 	ha->isp_ops->idc_lock(ha);
1395*4882a593Smuzhiyun 	qla4_8xxx_clear_rst_ready(ha);
1396*4882a593Smuzhiyun exit_isp_reset:
1397*4882a593Smuzhiyun 	ha->isp_ops->idc_unlock(ha);
1398*4882a593Smuzhiyun 
1399*4882a593Smuzhiyun 	if (rval == QLA_SUCCESS)
1400*4882a593Smuzhiyun 		clear_bit(AF_FW_RECOVERY, &ha->flags);
1401*4882a593Smuzhiyun 
1402*4882a593Smuzhiyun 	return rval;
1403*4882a593Smuzhiyun }
1404*4882a593Smuzhiyun 
qla4_83xx_dump_pause_control_regs(struct scsi_qla_host * ha)1405*4882a593Smuzhiyun static void qla4_83xx_dump_pause_control_regs(struct scsi_qla_host *ha)
1406*4882a593Smuzhiyun {
1407*4882a593Smuzhiyun 	u32 val = 0, val1 = 0;
1408*4882a593Smuzhiyun 	int i;
1409*4882a593Smuzhiyun 
1410*4882a593Smuzhiyun 	qla4_83xx_rd_reg_indirect(ha, QLA83XX_SRE_SHIM_CONTROL, &val);
1411*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha, "SRE-Shim Ctrl:0x%x\n", val));
1412*4882a593Smuzhiyun 
1413*4882a593Smuzhiyun 	/* Port 0 Rx Buffer Pause Threshold Registers. */
1414*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha,
1415*4882a593Smuzhiyun 		"Port 0 Rx Buffer Pause Threshold Registers[TC7..TC0]:"));
1416*4882a593Smuzhiyun 	for (i = 0; i < 8; i++) {
1417*4882a593Smuzhiyun 		qla4_83xx_rd_reg_indirect(ha,
1418*4882a593Smuzhiyun 				QLA83XX_PORT0_RXB_PAUSE_THRS + (i * 0x4), &val);
1419*4882a593Smuzhiyun 		DEBUG2(pr_info("0x%x ", val));
1420*4882a593Smuzhiyun 	}
1421*4882a593Smuzhiyun 
1422*4882a593Smuzhiyun 	DEBUG2(pr_info("\n"));
1423*4882a593Smuzhiyun 
1424*4882a593Smuzhiyun 	/* Port 1 Rx Buffer Pause Threshold Registers. */
1425*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha,
1426*4882a593Smuzhiyun 		"Port 1 Rx Buffer Pause Threshold Registers[TC7..TC0]:"));
1427*4882a593Smuzhiyun 	for (i = 0; i < 8; i++) {
1428*4882a593Smuzhiyun 		qla4_83xx_rd_reg_indirect(ha,
1429*4882a593Smuzhiyun 				QLA83XX_PORT1_RXB_PAUSE_THRS + (i * 0x4), &val);
1430*4882a593Smuzhiyun 		DEBUG2(pr_info("0x%x  ", val));
1431*4882a593Smuzhiyun 	}
1432*4882a593Smuzhiyun 
1433*4882a593Smuzhiyun 	DEBUG2(pr_info("\n"));
1434*4882a593Smuzhiyun 
1435*4882a593Smuzhiyun 	/* Port 0 RxB Traffic Class Max Cell Registers. */
1436*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha,
1437*4882a593Smuzhiyun 		"Port 0 RxB Traffic Class Max Cell Registers[3..0]:"));
1438*4882a593Smuzhiyun 	for (i = 0; i < 4; i++) {
1439*4882a593Smuzhiyun 		qla4_83xx_rd_reg_indirect(ha,
1440*4882a593Smuzhiyun 			       QLA83XX_PORT0_RXB_TC_MAX_CELL + (i * 0x4), &val);
1441*4882a593Smuzhiyun 		DEBUG2(pr_info("0x%x  ", val));
1442*4882a593Smuzhiyun 	}
1443*4882a593Smuzhiyun 
1444*4882a593Smuzhiyun 	DEBUG2(pr_info("\n"));
1445*4882a593Smuzhiyun 
1446*4882a593Smuzhiyun 	/* Port 1 RxB Traffic Class Max Cell Registers. */
1447*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha,
1448*4882a593Smuzhiyun 		"Port 1 RxB Traffic Class Max Cell Registers[3..0]:"));
1449*4882a593Smuzhiyun 	for (i = 0; i < 4; i++) {
1450*4882a593Smuzhiyun 		qla4_83xx_rd_reg_indirect(ha,
1451*4882a593Smuzhiyun 			       QLA83XX_PORT1_RXB_TC_MAX_CELL + (i * 0x4), &val);
1452*4882a593Smuzhiyun 		DEBUG2(pr_info("0x%x  ", val));
1453*4882a593Smuzhiyun 	}
1454*4882a593Smuzhiyun 
1455*4882a593Smuzhiyun 	DEBUG2(pr_info("\n"));
1456*4882a593Smuzhiyun 
1457*4882a593Smuzhiyun 	/* Port 0 RxB Rx Traffic Class Stats. */
1458*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha,
1459*4882a593Smuzhiyun 			  "Port 0 RxB Rx Traffic Class Stats [TC7..TC0]"));
1460*4882a593Smuzhiyun 	for (i = 7; i >= 0; i--) {
1461*4882a593Smuzhiyun 		qla4_83xx_rd_reg_indirect(ha, QLA83XX_PORT0_RXB_TC_STATS, &val);
1462*4882a593Smuzhiyun 		val &= ~(0x7 << 29);    /* Reset bits 29 to 31 */
1463*4882a593Smuzhiyun 		qla4_83xx_wr_reg_indirect(ha, QLA83XX_PORT0_RXB_TC_STATS,
1464*4882a593Smuzhiyun 					  (val | (i << 29)));
1465*4882a593Smuzhiyun 		qla4_83xx_rd_reg_indirect(ha, QLA83XX_PORT0_RXB_TC_STATS, &val);
1466*4882a593Smuzhiyun 		DEBUG2(pr_info("0x%x  ", val));
1467*4882a593Smuzhiyun 	}
1468*4882a593Smuzhiyun 
1469*4882a593Smuzhiyun 	DEBUG2(pr_info("\n"));
1470*4882a593Smuzhiyun 
1471*4882a593Smuzhiyun 	/* Port 1 RxB Rx Traffic Class Stats. */
1472*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha,
1473*4882a593Smuzhiyun 			  "Port 1 RxB Rx Traffic Class Stats [TC7..TC0]"));
1474*4882a593Smuzhiyun 	for (i = 7; i >= 0; i--) {
1475*4882a593Smuzhiyun 		qla4_83xx_rd_reg_indirect(ha, QLA83XX_PORT1_RXB_TC_STATS, &val);
1476*4882a593Smuzhiyun 		val &= ~(0x7 << 29);    /* Reset bits 29 to 31 */
1477*4882a593Smuzhiyun 		qla4_83xx_wr_reg_indirect(ha, QLA83XX_PORT1_RXB_TC_STATS,
1478*4882a593Smuzhiyun 					  (val | (i << 29)));
1479*4882a593Smuzhiyun 		qla4_83xx_rd_reg_indirect(ha, QLA83XX_PORT1_RXB_TC_STATS, &val);
1480*4882a593Smuzhiyun 		DEBUG2(pr_info("0x%x  ", val));
1481*4882a593Smuzhiyun 	}
1482*4882a593Smuzhiyun 
1483*4882a593Smuzhiyun 	DEBUG2(pr_info("\n"));
1484*4882a593Smuzhiyun 
1485*4882a593Smuzhiyun 	qla4_83xx_rd_reg_indirect(ha, QLA83XX_PORT2_IFB_PAUSE_THRS, &val);
1486*4882a593Smuzhiyun 	qla4_83xx_rd_reg_indirect(ha, QLA83XX_PORT3_IFB_PAUSE_THRS, &val1);
1487*4882a593Smuzhiyun 
1488*4882a593Smuzhiyun 	DEBUG2(ql4_printk(KERN_INFO, ha,
1489*4882a593Smuzhiyun 			  "IFB-Pause Thresholds: Port 2:0x%x, Port 3:0x%x\n",
1490*4882a593Smuzhiyun 			  val, val1));
1491*4882a593Smuzhiyun }
1492*4882a593Smuzhiyun 
__qla4_83xx_disable_pause(struct scsi_qla_host * ha)1493*4882a593Smuzhiyun static void __qla4_83xx_disable_pause(struct scsi_qla_host *ha)
1494*4882a593Smuzhiyun {
1495*4882a593Smuzhiyun 	int i;
1496*4882a593Smuzhiyun 
1497*4882a593Smuzhiyun 	/* set SRE-Shim Control Register */
1498*4882a593Smuzhiyun 	qla4_83xx_wr_reg_indirect(ha, QLA83XX_SRE_SHIM_CONTROL,
1499*4882a593Smuzhiyun 				  QLA83XX_SET_PAUSE_VAL);
1500*4882a593Smuzhiyun 
1501*4882a593Smuzhiyun 	for (i = 0; i < 8; i++) {
1502*4882a593Smuzhiyun 		/* Port 0 Rx Buffer Pause Threshold Registers. */
1503*4882a593Smuzhiyun 		qla4_83xx_wr_reg_indirect(ha,
1504*4882a593Smuzhiyun 				      QLA83XX_PORT0_RXB_PAUSE_THRS + (i * 0x4),
1505*4882a593Smuzhiyun 				      QLA83XX_SET_PAUSE_VAL);
1506*4882a593Smuzhiyun 		/* Port 1 Rx Buffer Pause Threshold Registers. */
1507*4882a593Smuzhiyun 		qla4_83xx_wr_reg_indirect(ha,
1508*4882a593Smuzhiyun 				      QLA83XX_PORT1_RXB_PAUSE_THRS + (i * 0x4),
1509*4882a593Smuzhiyun 				      QLA83XX_SET_PAUSE_VAL);
1510*4882a593Smuzhiyun 	}
1511*4882a593Smuzhiyun 
1512*4882a593Smuzhiyun 	for (i = 0; i < 4; i++) {
1513*4882a593Smuzhiyun 		/* Port 0 RxB Traffic Class Max Cell Registers. */
1514*4882a593Smuzhiyun 		qla4_83xx_wr_reg_indirect(ha,
1515*4882a593Smuzhiyun 				     QLA83XX_PORT0_RXB_TC_MAX_CELL + (i * 0x4),
1516*4882a593Smuzhiyun 				     QLA83XX_SET_TC_MAX_CELL_VAL);
1517*4882a593Smuzhiyun 		/* Port 1 RxB Traffic Class Max Cell Registers. */
1518*4882a593Smuzhiyun 		qla4_83xx_wr_reg_indirect(ha,
1519*4882a593Smuzhiyun 				     QLA83XX_PORT1_RXB_TC_MAX_CELL + (i * 0x4),
1520*4882a593Smuzhiyun 				     QLA83XX_SET_TC_MAX_CELL_VAL);
1521*4882a593Smuzhiyun 	}
1522*4882a593Smuzhiyun 
1523*4882a593Smuzhiyun 	qla4_83xx_wr_reg_indirect(ha, QLA83XX_PORT2_IFB_PAUSE_THRS,
1524*4882a593Smuzhiyun 				  QLA83XX_SET_PAUSE_VAL);
1525*4882a593Smuzhiyun 	qla4_83xx_wr_reg_indirect(ha, QLA83XX_PORT3_IFB_PAUSE_THRS,
1526*4882a593Smuzhiyun 				  QLA83XX_SET_PAUSE_VAL);
1527*4882a593Smuzhiyun 
1528*4882a593Smuzhiyun 	ql4_printk(KERN_INFO, ha, "Disabled pause frames successfully.\n");
1529*4882a593Smuzhiyun }
1530*4882a593Smuzhiyun 
1531*4882a593Smuzhiyun /**
1532*4882a593Smuzhiyun  * qla4_83xx_eport_init - Initialize EPort.
1533*4882a593Smuzhiyun  * @ha: Pointer to host adapter structure.
1534*4882a593Smuzhiyun  *
1535*4882a593Smuzhiyun  * If EPort hardware is in reset state before disabling pause, there would be
1536*4882a593Smuzhiyun  * serious hardware wedging issues. To prevent this perform eport init everytime
1537*4882a593Smuzhiyun  * before disabling pause frames.
1538*4882a593Smuzhiyun  **/
qla4_83xx_eport_init(struct scsi_qla_host * ha)1539*4882a593Smuzhiyun static void qla4_83xx_eport_init(struct scsi_qla_host *ha)
1540*4882a593Smuzhiyun {
1541*4882a593Smuzhiyun 	/* Clear the 8 registers */
1542*4882a593Smuzhiyun 	qla4_83xx_wr_reg_indirect(ha, QLA83XX_RESET_REG, 0x0);
1543*4882a593Smuzhiyun 	qla4_83xx_wr_reg_indirect(ha, QLA83XX_RESET_PORT0, 0x0);
1544*4882a593Smuzhiyun 	qla4_83xx_wr_reg_indirect(ha, QLA83XX_RESET_PORT1, 0x0);
1545*4882a593Smuzhiyun 	qla4_83xx_wr_reg_indirect(ha, QLA83XX_RESET_PORT2, 0x0);
1546*4882a593Smuzhiyun 	qla4_83xx_wr_reg_indirect(ha, QLA83XX_RESET_PORT3, 0x0);
1547*4882a593Smuzhiyun 	qla4_83xx_wr_reg_indirect(ha, QLA83XX_RESET_SRE_SHIM, 0x0);
1548*4882a593Smuzhiyun 	qla4_83xx_wr_reg_indirect(ha, QLA83XX_RESET_EPG_SHIM, 0x0);
1549*4882a593Smuzhiyun 	qla4_83xx_wr_reg_indirect(ha, QLA83XX_RESET_ETHER_PCS, 0x0);
1550*4882a593Smuzhiyun 
1551*4882a593Smuzhiyun 	/* Write any value to Reset Control register */
1552*4882a593Smuzhiyun 	qla4_83xx_wr_reg_indirect(ha, QLA83XX_RESET_CONTROL, 0xFF);
1553*4882a593Smuzhiyun 
1554*4882a593Smuzhiyun 	ql4_printk(KERN_INFO, ha, "EPORT is out of reset.\n");
1555*4882a593Smuzhiyun }
1556*4882a593Smuzhiyun 
qla4_83xx_disable_pause(struct scsi_qla_host * ha)1557*4882a593Smuzhiyun void qla4_83xx_disable_pause(struct scsi_qla_host *ha)
1558*4882a593Smuzhiyun {
1559*4882a593Smuzhiyun 	ha->isp_ops->idc_lock(ha);
1560*4882a593Smuzhiyun 	/* Before disabling pause frames, ensure that eport is not in reset */
1561*4882a593Smuzhiyun 	qla4_83xx_eport_init(ha);
1562*4882a593Smuzhiyun 	qla4_83xx_dump_pause_control_regs(ha);
1563*4882a593Smuzhiyun 	__qla4_83xx_disable_pause(ha);
1564*4882a593Smuzhiyun 	ha->isp_ops->idc_unlock(ha);
1565*4882a593Smuzhiyun }
1566*4882a593Smuzhiyun 
1567*4882a593Smuzhiyun /**
1568*4882a593Smuzhiyun  * qla4_83xx_is_detached - Check if we are marked invisible.
1569*4882a593Smuzhiyun  * @ha: Pointer to host adapter structure.
1570*4882a593Smuzhiyun  **/
qla4_83xx_is_detached(struct scsi_qla_host * ha)1571*4882a593Smuzhiyun int qla4_83xx_is_detached(struct scsi_qla_host *ha)
1572*4882a593Smuzhiyun {
1573*4882a593Smuzhiyun 	uint32_t drv_active;
1574*4882a593Smuzhiyun 
1575*4882a593Smuzhiyun 	drv_active = qla4_8xxx_rd_direct(ha, QLA8XXX_CRB_DRV_ACTIVE);
1576*4882a593Smuzhiyun 
1577*4882a593Smuzhiyun 	if (test_bit(AF_INIT_DONE, &ha->flags) &&
1578*4882a593Smuzhiyun 	    !(drv_active & (1 << ha->func_num))) {
1579*4882a593Smuzhiyun 		DEBUG2(ql4_printk(KERN_INFO, ha, "%s: drv_active = 0x%X\n",
1580*4882a593Smuzhiyun 				  __func__, drv_active));
1581*4882a593Smuzhiyun 		return QLA_SUCCESS;
1582*4882a593Smuzhiyun 	}
1583*4882a593Smuzhiyun 
1584*4882a593Smuzhiyun 	return QLA_ERROR;
1585*4882a593Smuzhiyun }
1586