xref: /OK3568_Linux_fs/external/rkwifibt/drivers/infineon/aiutils.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun  * Misc utility routines for accessing chip-specific features
3*4882a593Smuzhiyun  * of the SiliconBackplane-based Broadcom chips.
4*4882a593Smuzhiyun  *
5*4882a593Smuzhiyun  * Portions of this code are copyright (c) 2021 Cypress Semiconductor Corporation
6*4882a593Smuzhiyun  *
7*4882a593Smuzhiyun  * Copyright (C) 1999-2017, Broadcom Corporation
8*4882a593Smuzhiyun  *
9*4882a593Smuzhiyun  *      Unless you and Broadcom execute a separate written software license
10*4882a593Smuzhiyun  * agreement governing use of this software, this software is licensed to you
11*4882a593Smuzhiyun  * under the terms of the GNU General Public License version 2 (the "GPL"),
12*4882a593Smuzhiyun  * available at http://www.broadcom.com/licenses/GPLv2.php, with the
13*4882a593Smuzhiyun  * following added to such license:
14*4882a593Smuzhiyun  *
15*4882a593Smuzhiyun  *      As a special exception, the copyright holders of this software give you
16*4882a593Smuzhiyun  * permission to link this software with independent modules, and to copy and
17*4882a593Smuzhiyun  * distribute the resulting executable under terms of your choice, provided that
18*4882a593Smuzhiyun  * you also meet, for each linked independent module, the terms and conditions of
19*4882a593Smuzhiyun  * the license of that module.  An independent module is a module which is not
20*4882a593Smuzhiyun  * derived from this software.  The special exception does not apply to any
21*4882a593Smuzhiyun  * modifications of the software.
22*4882a593Smuzhiyun  *
23*4882a593Smuzhiyun  *      Notwithstanding the above, under no circumstances may you combine this
24*4882a593Smuzhiyun  * software in any way with any other Broadcom software provided under a license
25*4882a593Smuzhiyun  * other than the GPL, without Broadcom's express prior written consent.
26*4882a593Smuzhiyun  *
27*4882a593Smuzhiyun  *
28*4882a593Smuzhiyun  * <<Broadcom-WL-IPTag/Open:>>
29*4882a593Smuzhiyun  *
30*4882a593Smuzhiyun  * $Id: aiutils.c 701122 2017-05-23 19:32:45Z $
31*4882a593Smuzhiyun  */
32*4882a593Smuzhiyun #include <bcm_cfg.h>
33*4882a593Smuzhiyun #include <typedefs.h>
34*4882a593Smuzhiyun #include <bcmdefs.h>
35*4882a593Smuzhiyun #include <osl.h>
36*4882a593Smuzhiyun #include <bcmutils.h>
37*4882a593Smuzhiyun #include <siutils.h>
38*4882a593Smuzhiyun #include <hndsoc.h>
39*4882a593Smuzhiyun #include <sbchipc.h>
40*4882a593Smuzhiyun #include <pcicfg.h>
41*4882a593Smuzhiyun 
42*4882a593Smuzhiyun #include "siutils_priv.h"
43*4882a593Smuzhiyun #include <bcmdevs.h>
44*4882a593Smuzhiyun 
45*4882a593Smuzhiyun #define BCM53573_DMP() (0)
46*4882a593Smuzhiyun #define BCM4707_DMP() (0)
47*4882a593Smuzhiyun #define PMU_DMP() (0)
48*4882a593Smuzhiyun #define GCI_DMP() (0)
49*4882a593Smuzhiyun 
50*4882a593Smuzhiyun #if defined(BCM_BACKPLANE_TIMEOUT)
51*4882a593Smuzhiyun static bool ai_get_apb_bridge(si_t *sih, uint32 coreidx, uint32 *apb_id, uint32 *apb_coreuinit);
52*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
53*4882a593Smuzhiyun 
54*4882a593Smuzhiyun #if defined(AXI_TIMEOUTS) || defined(BCM_BACKPLANE_TIMEOUT)
55*4882a593Smuzhiyun static void ai_reset_axi_to(si_info_t *sii, aidmp_t *ai);
56*4882a593Smuzhiyun #endif	/* defined (AXI_TIMEOUTS) || defined (BCM_BACKPLANE_TIMEOUT) */
57*4882a593Smuzhiyun 
58*4882a593Smuzhiyun /* EROM parsing */
59*4882a593Smuzhiyun 
60*4882a593Smuzhiyun #ifdef BCMQT
61*4882a593Smuzhiyun #define SPINWAIT_TIME_US	3000
62*4882a593Smuzhiyun #else
63*4882a593Smuzhiyun #define SPINWAIT_TIME_US	300
64*4882a593Smuzhiyun #endif /* BCMQT */
65*4882a593Smuzhiyun 
66*4882a593Smuzhiyun static uint32
get_erom_ent(si_t * sih,uint32 ** eromptr,uint32 mask,uint32 match)67*4882a593Smuzhiyun get_erom_ent(si_t *sih, uint32 **eromptr, uint32 mask, uint32 match)
68*4882a593Smuzhiyun {
69*4882a593Smuzhiyun 	uint32 ent;
70*4882a593Smuzhiyun 	uint inv = 0, nom = 0;
71*4882a593Smuzhiyun 	uint32 size = 0;
72*4882a593Smuzhiyun 
73*4882a593Smuzhiyun 	while (TRUE) {
74*4882a593Smuzhiyun 		ent = R_REG(si_osh(sih), *eromptr);
75*4882a593Smuzhiyun 		(*eromptr)++;
76*4882a593Smuzhiyun 
77*4882a593Smuzhiyun 		if (mask == 0)
78*4882a593Smuzhiyun 			break;
79*4882a593Smuzhiyun 
80*4882a593Smuzhiyun 		if ((ent & ER_VALID) == 0) {
81*4882a593Smuzhiyun 			inv++;
82*4882a593Smuzhiyun 			continue;
83*4882a593Smuzhiyun 		}
84*4882a593Smuzhiyun 
85*4882a593Smuzhiyun 		if (ent == (ER_END | ER_VALID))
86*4882a593Smuzhiyun 			break;
87*4882a593Smuzhiyun 
88*4882a593Smuzhiyun 		if ((ent & mask) == match)
89*4882a593Smuzhiyun 			break;
90*4882a593Smuzhiyun 
91*4882a593Smuzhiyun 		/* escape condition related EROM size if it has invalid values */
92*4882a593Smuzhiyun 		size += sizeof(*eromptr);
93*4882a593Smuzhiyun 		if (size >= ER_SZ_MAX) {
94*4882a593Smuzhiyun 			SI_ERROR(("Failed to find end of EROM marker\n"));
95*4882a593Smuzhiyun 			break;
96*4882a593Smuzhiyun 		}
97*4882a593Smuzhiyun 
98*4882a593Smuzhiyun 		nom++;
99*4882a593Smuzhiyun 	}
100*4882a593Smuzhiyun 
101*4882a593Smuzhiyun 	SI_VMSG(("%s: Returning ent 0x%08x\n", __FUNCTION__, ent));
102*4882a593Smuzhiyun 	if (inv + nom) {
103*4882a593Smuzhiyun 		SI_VMSG(("  after %d invalid and %d non-matching entries\n", inv, nom));
104*4882a593Smuzhiyun 	}
105*4882a593Smuzhiyun 	return ent;
106*4882a593Smuzhiyun }
107*4882a593Smuzhiyun 
108*4882a593Smuzhiyun static uint32
get_asd(si_t * sih,uint32 ** eromptr,uint sp,uint ad,uint st,uint32 * addrl,uint32 * addrh,uint32 * sizel,uint32 * sizeh)109*4882a593Smuzhiyun get_asd(si_t *sih, uint32 **eromptr, uint sp, uint ad, uint st, uint32 *addrl, uint32 *addrh,
110*4882a593Smuzhiyun 	uint32 *sizel, uint32 *sizeh)
111*4882a593Smuzhiyun {
112*4882a593Smuzhiyun 	uint32 asd, sz, szd;
113*4882a593Smuzhiyun 
114*4882a593Smuzhiyun 	BCM_REFERENCE(ad);
115*4882a593Smuzhiyun 
116*4882a593Smuzhiyun 	asd = get_erom_ent(sih, eromptr, ER_VALID, ER_VALID);
117*4882a593Smuzhiyun 	if (((asd & ER_TAG1) != ER_ADD) ||
118*4882a593Smuzhiyun 	    (((asd & AD_SP_MASK) >> AD_SP_SHIFT) != sp) ||
119*4882a593Smuzhiyun 	    ((asd & AD_ST_MASK) != st)) {
120*4882a593Smuzhiyun 		/* This is not what we want, "push" it back */
121*4882a593Smuzhiyun 		(*eromptr)--;
122*4882a593Smuzhiyun 		return 0;
123*4882a593Smuzhiyun 	}
124*4882a593Smuzhiyun 	*addrl = asd & AD_ADDR_MASK;
125*4882a593Smuzhiyun 	if (asd & AD_AG32)
126*4882a593Smuzhiyun 		*addrh = get_erom_ent(sih, eromptr, 0, 0);
127*4882a593Smuzhiyun 	else
128*4882a593Smuzhiyun 		*addrh = 0;
129*4882a593Smuzhiyun 	*sizeh = 0;
130*4882a593Smuzhiyun 	sz = asd & AD_SZ_MASK;
131*4882a593Smuzhiyun 	if (sz == AD_SZ_SZD) {
132*4882a593Smuzhiyun 		szd = get_erom_ent(sih, eromptr, 0, 0);
133*4882a593Smuzhiyun 		*sizel = szd & SD_SZ_MASK;
134*4882a593Smuzhiyun 		if (szd & SD_SG32)
135*4882a593Smuzhiyun 			*sizeh = get_erom_ent(sih, eromptr, 0, 0);
136*4882a593Smuzhiyun 	} else
137*4882a593Smuzhiyun 		*sizel = AD_SZ_BASE << (sz >> AD_SZ_SHIFT);
138*4882a593Smuzhiyun 
139*4882a593Smuzhiyun 	SI_VMSG(("  SP %d, ad %d: st = %d, 0x%08x_0x%08x @ 0x%08x_0x%08x\n",
140*4882a593Smuzhiyun 	        sp, ad, st, *sizeh, *sizel, *addrh, *addrl));
141*4882a593Smuzhiyun 
142*4882a593Smuzhiyun 	return asd;
143*4882a593Smuzhiyun }
144*4882a593Smuzhiyun 
145*4882a593Smuzhiyun /* Parse the enumeration rom to identify all cores
146*4882a593Smuzhiyun  * Erom content format can be found in:
147*4882a593Smuzhiyun  * http://hwnbu-twiki.broadcom.com/twiki/pub/Mwgroup/ArmDocumentation/SystemDiscovery.pdf
148*4882a593Smuzhiyun  */
149*4882a593Smuzhiyun void
ai_scan(si_t * sih,void * regs,uint32 erombase,uint devid)150*4882a593Smuzhiyun ai_scan(si_t *sih, void *regs, uint32 erombase, uint devid)
151*4882a593Smuzhiyun {
152*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
153*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
154*4882a593Smuzhiyun 	uint32 *eromptr, *eromlim;
155*4882a593Smuzhiyun 	axi_wrapper_t * axi_wrapper = sii->axi_wrapper;
156*4882a593Smuzhiyun 
157*4882a593Smuzhiyun 	BCM_REFERENCE(devid);
158*4882a593Smuzhiyun 
159*4882a593Smuzhiyun 	switch (BUSTYPE(sih->bustype)) {
160*4882a593Smuzhiyun 	case SI_BUS:
161*4882a593Smuzhiyun 		eromptr = (uint32 *)REG_MAP(erombase, SI_CORE_SIZE);
162*4882a593Smuzhiyun 		break;
163*4882a593Smuzhiyun 
164*4882a593Smuzhiyun 	case PCI_BUS:
165*4882a593Smuzhiyun 		/* Set wrappers address */
166*4882a593Smuzhiyun 		sii->curwrap = (void *)((uintptr)regs + SI_CORE_SIZE);
167*4882a593Smuzhiyun 
168*4882a593Smuzhiyun 		/* Now point the window at the erom */
169*4882a593Smuzhiyun 		OSL_PCI_WRITE_CONFIG(sii->osh, PCI_BAR0_WIN, 4, erombase);
170*4882a593Smuzhiyun 		eromptr = regs;
171*4882a593Smuzhiyun 		break;
172*4882a593Smuzhiyun 
173*4882a593Smuzhiyun #ifdef BCMSDIO
174*4882a593Smuzhiyun 	case SPI_BUS:
175*4882a593Smuzhiyun 	case SDIO_BUS:
176*4882a593Smuzhiyun 		eromptr = (uint32 *)(uintptr)erombase;
177*4882a593Smuzhiyun 		break;
178*4882a593Smuzhiyun #endif	/* BCMSDIO */
179*4882a593Smuzhiyun 
180*4882a593Smuzhiyun 	case PCMCIA_BUS:
181*4882a593Smuzhiyun 	default:
182*4882a593Smuzhiyun 		SI_ERROR(("Don't know how to do AXI enumertion on bus %d\n", sih->bustype));
183*4882a593Smuzhiyun 		ASSERT(0);
184*4882a593Smuzhiyun 		return;
185*4882a593Smuzhiyun 	}
186*4882a593Smuzhiyun 	eromlim = eromptr + (ER_REMAPCONTROL / sizeof(uint32));
187*4882a593Smuzhiyun 	sii->axi_num_wrappers = 0;
188*4882a593Smuzhiyun 
189*4882a593Smuzhiyun 	SI_VMSG(("ai_scan: regs = 0x%p, erombase = 0x%08x, eromptr = 0x%p, eromlim = 0x%p\n",
190*4882a593Smuzhiyun 	         OSL_OBFUSCATE_BUF(regs), erombase,
191*4882a593Smuzhiyun 		OSL_OBFUSCATE_BUF(eromptr), OSL_OBFUSATE_BUF(eromlim)));
192*4882a593Smuzhiyun 	while (eromptr < eromlim) {
193*4882a593Smuzhiyun 		uint32 cia, cib, cid, mfg, crev, nmw, nsw, nmp, nsp;
194*4882a593Smuzhiyun 		uint32 mpd, asd, addrl, addrh, sizel, sizeh;
195*4882a593Smuzhiyun 		uint i, j, idx;
196*4882a593Smuzhiyun 		bool br;
197*4882a593Smuzhiyun 
198*4882a593Smuzhiyun 		br = FALSE;
199*4882a593Smuzhiyun 
200*4882a593Smuzhiyun 		/* Grok a component */
201*4882a593Smuzhiyun 		cia = get_erom_ent(sih, &eromptr, ER_TAG, ER_CI);
202*4882a593Smuzhiyun 		if (cia == (ER_END | ER_VALID)) {
203*4882a593Smuzhiyun 			SI_VMSG(("Found END of erom after %d cores\n", sii->numcores));
204*4882a593Smuzhiyun 			return;
205*4882a593Smuzhiyun 		}
206*4882a593Smuzhiyun 
207*4882a593Smuzhiyun 		cib = get_erom_ent(sih, &eromptr, 0, 0);
208*4882a593Smuzhiyun 
209*4882a593Smuzhiyun 		if ((cib & ER_TAG) != ER_CI) {
210*4882a593Smuzhiyun 			SI_ERROR(("CIA not followed by CIB\n"));
211*4882a593Smuzhiyun 			goto error;
212*4882a593Smuzhiyun 		}
213*4882a593Smuzhiyun 
214*4882a593Smuzhiyun 		cid = (cia & CIA_CID_MASK) >> CIA_CID_SHIFT;
215*4882a593Smuzhiyun 		mfg = (cia & CIA_MFG_MASK) >> CIA_MFG_SHIFT;
216*4882a593Smuzhiyun 		crev = (cib & CIB_REV_MASK) >> CIB_REV_SHIFT;
217*4882a593Smuzhiyun 		nmw = (cib & CIB_NMW_MASK) >> CIB_NMW_SHIFT;
218*4882a593Smuzhiyun 		nsw = (cib & CIB_NSW_MASK) >> CIB_NSW_SHIFT;
219*4882a593Smuzhiyun 		nmp = (cib & CIB_NMP_MASK) >> CIB_NMP_SHIFT;
220*4882a593Smuzhiyun 		nsp = (cib & CIB_NSP_MASK) >> CIB_NSP_SHIFT;
221*4882a593Smuzhiyun 
222*4882a593Smuzhiyun #ifdef BCMDBG_SI
223*4882a593Smuzhiyun 		SI_VMSG(("Found component 0x%04x/0x%04x rev %d at erom addr 0x%p, with nmw = %d, "
224*4882a593Smuzhiyun 		         "nsw = %d, nmp = %d & nsp = %d\n",
225*4882a593Smuzhiyun 		         mfg, cid, crev, OSL_OBFUSCATE_BUF(eromptr - 1), nmw, nsw, nmp, nsp));
226*4882a593Smuzhiyun #else
227*4882a593Smuzhiyun 		BCM_REFERENCE(crev);
228*4882a593Smuzhiyun #endif // endif
229*4882a593Smuzhiyun 
230*4882a593Smuzhiyun 		if (BCM4347_CHIP(sih->chip)) {
231*4882a593Smuzhiyun 			/* 4347 has more entries for ARM core
232*4882a593Smuzhiyun 			 * This should apply to all chips but crashes on router
233*4882a593Smuzhiyun 			 * This is a temp fix to be further analyze
234*4882a593Smuzhiyun 			 */
235*4882a593Smuzhiyun 			if (nsp == 0)
236*4882a593Smuzhiyun 				continue;
237*4882a593Smuzhiyun 		} else
238*4882a593Smuzhiyun 		{
239*4882a593Smuzhiyun 			/* Include Default slave wrapper for timeout monitoring */
240*4882a593Smuzhiyun 			if ((nsp == 0) ||
241*4882a593Smuzhiyun #if !defined(AXI_TIMEOUTS) && !defined(BCM_BACKPLANE_TIMEOUT)
242*4882a593Smuzhiyun 				((mfg == MFGID_ARM) && (cid == DEF_AI_COMP)) ||
243*4882a593Smuzhiyun #else
244*4882a593Smuzhiyun 				((CHIPTYPE(sii->pub.socitype) == SOCI_NAI) &&
245*4882a593Smuzhiyun 				(mfg == MFGID_ARM) && (cid == DEF_AI_COMP)) ||
246*4882a593Smuzhiyun #endif /* !defined(AXI_TIMEOUTS) && !defined(BCM_BACKPLANE_TIMEOUT) */
247*4882a593Smuzhiyun 				FALSE) {
248*4882a593Smuzhiyun 				continue;
249*4882a593Smuzhiyun 			}
250*4882a593Smuzhiyun 		}
251*4882a593Smuzhiyun 
252*4882a593Smuzhiyun 		if ((nmw + nsw == 0)) {
253*4882a593Smuzhiyun 			/* A component which is not a core */
254*4882a593Smuzhiyun 			if (cid == OOB_ROUTER_CORE_ID) {
255*4882a593Smuzhiyun 				asd = get_asd(sih, &eromptr, 0, 0, AD_ST_SLAVE,
256*4882a593Smuzhiyun 					&addrl, &addrh, &sizel, &sizeh);
257*4882a593Smuzhiyun 				if (asd != 0) {
258*4882a593Smuzhiyun 					if ((sii->oob_router != 0) && (sii->oob_router != addrl)) {
259*4882a593Smuzhiyun 						sii->oob_router1 = addrl;
260*4882a593Smuzhiyun 					} else {
261*4882a593Smuzhiyun 						sii->oob_router = addrl;
262*4882a593Smuzhiyun 					}
263*4882a593Smuzhiyun 				}
264*4882a593Smuzhiyun 			}
265*4882a593Smuzhiyun 			if (cid != NS_CCB_CORE_ID &&
266*4882a593Smuzhiyun 				cid != PMU_CORE_ID && cid != GCI_CORE_ID && cid != SR_CORE_ID &&
267*4882a593Smuzhiyun 				cid != HUB_CORE_ID && cid != HND_OOBR_CORE_ID)
268*4882a593Smuzhiyun 				continue;
269*4882a593Smuzhiyun 		}
270*4882a593Smuzhiyun 
271*4882a593Smuzhiyun 		idx = sii->numcores;
272*4882a593Smuzhiyun 
273*4882a593Smuzhiyun 		cores_info->cia[idx] = cia;
274*4882a593Smuzhiyun 		cores_info->cib[idx] = cib;
275*4882a593Smuzhiyun 		cores_info->coreid[idx] = cid;
276*4882a593Smuzhiyun 
277*4882a593Smuzhiyun 		for (i = 0; i < nmp; i++) {
278*4882a593Smuzhiyun 			mpd = get_erom_ent(sih, &eromptr, ER_VALID, ER_VALID);
279*4882a593Smuzhiyun 			if ((mpd & ER_TAG) != ER_MP) {
280*4882a593Smuzhiyun 				SI_ERROR(("Not enough MP entries for component 0x%x\n", cid));
281*4882a593Smuzhiyun 				goto error;
282*4882a593Smuzhiyun 			}
283*4882a593Smuzhiyun 			SI_VMSG(("  Master port %d, mp: %d id: %d\n", i,
284*4882a593Smuzhiyun 			         (mpd & MPD_MP_MASK) >> MPD_MP_SHIFT,
285*4882a593Smuzhiyun 			         (mpd & MPD_MUI_MASK) >> MPD_MUI_SHIFT));
286*4882a593Smuzhiyun 		}
287*4882a593Smuzhiyun 
288*4882a593Smuzhiyun 		/* First Slave Address Descriptor should be port 0:
289*4882a593Smuzhiyun 		 * the main register space for the core
290*4882a593Smuzhiyun 		 */
291*4882a593Smuzhiyun 		asd = get_asd(sih, &eromptr, 0, 0, AD_ST_SLAVE, &addrl, &addrh, &sizel, &sizeh);
292*4882a593Smuzhiyun 		if (asd == 0) {
293*4882a593Smuzhiyun 			do {
294*4882a593Smuzhiyun 			/* Try again to see if it is a bridge */
295*4882a593Smuzhiyun 			asd = get_asd(sih, &eromptr, 0, 0, AD_ST_BRIDGE, &addrl, &addrh,
296*4882a593Smuzhiyun 			              &sizel, &sizeh);
297*4882a593Smuzhiyun 			if (asd != 0)
298*4882a593Smuzhiyun 				br = TRUE;
299*4882a593Smuzhiyun 			else {
300*4882a593Smuzhiyun 					if (br == TRUE) {
301*4882a593Smuzhiyun 						break;
302*4882a593Smuzhiyun 					}
303*4882a593Smuzhiyun 					else if ((addrh != 0) || (sizeh != 0) ||
304*4882a593Smuzhiyun 						(sizel != SI_CORE_SIZE)) {
305*4882a593Smuzhiyun 						SI_ERROR(("addrh = 0x%x\t sizeh = 0x%x\t size1 ="
306*4882a593Smuzhiyun 							"0x%x\n", addrh, sizeh, sizel));
307*4882a593Smuzhiyun 						SI_ERROR(("First Slave ASD for"
308*4882a593Smuzhiyun 							"core 0x%04x malformed "
309*4882a593Smuzhiyun 							"(0x%08x)\n", cid, asd));
310*4882a593Smuzhiyun 						goto error;
311*4882a593Smuzhiyun 					}
312*4882a593Smuzhiyun 				}
313*4882a593Smuzhiyun 			} while (1);
314*4882a593Smuzhiyun 		}
315*4882a593Smuzhiyun 		cores_info->coresba[idx] = addrl;
316*4882a593Smuzhiyun 		cores_info->coresba_size[idx] = sizel;
317*4882a593Smuzhiyun 		/* Get any more ASDs in first port */
318*4882a593Smuzhiyun 		j = 1;
319*4882a593Smuzhiyun 		do {
320*4882a593Smuzhiyun 			asd = get_asd(sih, &eromptr, 0, j, AD_ST_SLAVE, &addrl, &addrh,
321*4882a593Smuzhiyun 			              &sizel, &sizeh);
322*4882a593Smuzhiyun 			if ((asd != 0) && (j == 1) && (sizel == SI_CORE_SIZE)) {
323*4882a593Smuzhiyun 				cores_info->coresba2[idx] = addrl;
324*4882a593Smuzhiyun 				cores_info->coresba2_size[idx] = sizel;
325*4882a593Smuzhiyun 			}
326*4882a593Smuzhiyun 			j++;
327*4882a593Smuzhiyun 		} while (asd != 0);
328*4882a593Smuzhiyun 
329*4882a593Smuzhiyun 		/* Go through the ASDs for other slave ports */
330*4882a593Smuzhiyun 		for (i = 1; i < nsp; i++) {
331*4882a593Smuzhiyun 			j = 0;
332*4882a593Smuzhiyun 			do {
333*4882a593Smuzhiyun 				asd = get_asd(sih, &eromptr, i, j, AD_ST_SLAVE, &addrl, &addrh,
334*4882a593Smuzhiyun 				              &sizel, &sizeh);
335*4882a593Smuzhiyun 				/* To get the first base address of second slave port */
336*4882a593Smuzhiyun 				if ((asd != 0) && (i == 1) && (j == 0)) {
337*4882a593Smuzhiyun 					cores_info->csp2ba[idx] = addrl;
338*4882a593Smuzhiyun 					cores_info->csp2ba_size[idx] = sizel;
339*4882a593Smuzhiyun 				}
340*4882a593Smuzhiyun 				if (asd == 0)
341*4882a593Smuzhiyun 					break;
342*4882a593Smuzhiyun 				j++;
343*4882a593Smuzhiyun 			} while (1);
344*4882a593Smuzhiyun 			if (j == 0) {
345*4882a593Smuzhiyun 				SI_ERROR((" SP %d has no address descriptors\n", i));
346*4882a593Smuzhiyun 				goto error;
347*4882a593Smuzhiyun 			}
348*4882a593Smuzhiyun 		}
349*4882a593Smuzhiyun 
350*4882a593Smuzhiyun 		/* Now get master wrappers */
351*4882a593Smuzhiyun 		for (i = 0; i < nmw; i++) {
352*4882a593Smuzhiyun 			asd = get_asd(sih, &eromptr, i, 0, AD_ST_MWRAP, &addrl, &addrh,
353*4882a593Smuzhiyun 			              &sizel, &sizeh);
354*4882a593Smuzhiyun 			if (asd == 0) {
355*4882a593Smuzhiyun 				SI_ERROR(("Missing descriptor for MW %d\n", i));
356*4882a593Smuzhiyun 				goto error;
357*4882a593Smuzhiyun 			}
358*4882a593Smuzhiyun 			if ((sizeh != 0) || (sizel != SI_CORE_SIZE)) {
359*4882a593Smuzhiyun 				SI_ERROR(("Master wrapper %d is not 4KB\n", i));
360*4882a593Smuzhiyun 				goto error;
361*4882a593Smuzhiyun 			}
362*4882a593Smuzhiyun 			if (i == 0) {
363*4882a593Smuzhiyun 				cores_info->wrapba[idx] = addrl;
364*4882a593Smuzhiyun 			} else if (i == 1) {
365*4882a593Smuzhiyun 				cores_info->wrapba2[idx] = addrl;
366*4882a593Smuzhiyun 			} else if (i == 2) {
367*4882a593Smuzhiyun 				cores_info->wrapba3[idx] = addrl;
368*4882a593Smuzhiyun 			}
369*4882a593Smuzhiyun 
370*4882a593Smuzhiyun 			if (axi_wrapper &&
371*4882a593Smuzhiyun 				(sii->axi_num_wrappers < SI_MAX_AXI_WRAPPERS)) {
372*4882a593Smuzhiyun 				axi_wrapper[sii->axi_num_wrappers].mfg = mfg;
373*4882a593Smuzhiyun 				axi_wrapper[sii->axi_num_wrappers].cid = cid;
374*4882a593Smuzhiyun 				axi_wrapper[sii->axi_num_wrappers].rev = crev;
375*4882a593Smuzhiyun 				axi_wrapper[sii->axi_num_wrappers].wrapper_type = AI_MASTER_WRAPPER;
376*4882a593Smuzhiyun 				axi_wrapper[sii->axi_num_wrappers].wrapper_addr = addrl;
377*4882a593Smuzhiyun 				sii->axi_num_wrappers++;
378*4882a593Smuzhiyun 				SI_VMSG(("MASTER WRAPPER: %d, mfg:%x, cid:%x,"
379*4882a593Smuzhiyun 					"rev:%x, addr:%x, size:%x\n",
380*4882a593Smuzhiyun 					sii->axi_num_wrappers, mfg, cid, crev, addrl, sizel));
381*4882a593Smuzhiyun 			}
382*4882a593Smuzhiyun 		}
383*4882a593Smuzhiyun 
384*4882a593Smuzhiyun 		/* And finally slave wrappers */
385*4882a593Smuzhiyun 		for (i = 0; i < nsw; i++) {
386*4882a593Smuzhiyun 			uint fwp = (nsp == 1) ? 0 : 1;
387*4882a593Smuzhiyun 			asd = get_asd(sih, &eromptr, fwp + i, 0, AD_ST_SWRAP, &addrl, &addrh,
388*4882a593Smuzhiyun 			              &sizel, &sizeh);
389*4882a593Smuzhiyun 
390*4882a593Smuzhiyun 			/* cache APB bridge wrapper address for set/clear timeout */
391*4882a593Smuzhiyun 			if ((mfg == MFGID_ARM) && (cid == APB_BRIDGE_ID)) {
392*4882a593Smuzhiyun 				ASSERT(sii->num_br < SI_MAXBR);
393*4882a593Smuzhiyun 				sii->br_wrapba[sii->num_br++] = addrl;
394*4882a593Smuzhiyun 			}
395*4882a593Smuzhiyun 
396*4882a593Smuzhiyun 			if (asd == 0) {
397*4882a593Smuzhiyun 				SI_ERROR(("Missing descriptor for SW %d\n", i));
398*4882a593Smuzhiyun 				goto error;
399*4882a593Smuzhiyun 			}
400*4882a593Smuzhiyun 			if ((sizeh != 0) || (sizel != SI_CORE_SIZE)) {
401*4882a593Smuzhiyun 				SI_ERROR(("Slave wrapper %d is not 4KB\n", i));
402*4882a593Smuzhiyun 				goto error;
403*4882a593Smuzhiyun 			}
404*4882a593Smuzhiyun 			if ((nmw == 0) && (i == 0)) {
405*4882a593Smuzhiyun 				cores_info->wrapba[idx] = addrl;
406*4882a593Smuzhiyun 			} else if ((nmw == 0) && (i == 1)) {
407*4882a593Smuzhiyun 				cores_info->wrapba2[idx] = addrl;
408*4882a593Smuzhiyun 			} else if ((nmw == 0) && (i == 2)) {
409*4882a593Smuzhiyun 				cores_info->wrapba3[idx] = addrl;
410*4882a593Smuzhiyun 			}
411*4882a593Smuzhiyun 
412*4882a593Smuzhiyun 			/* Include all slave wrappers to the list to
413*4882a593Smuzhiyun 			 * enable and monitor watchdog timeouts
414*4882a593Smuzhiyun 			 */
415*4882a593Smuzhiyun 
416*4882a593Smuzhiyun 			if (axi_wrapper &&
417*4882a593Smuzhiyun 				(sii->axi_num_wrappers < SI_MAX_AXI_WRAPPERS)) {
418*4882a593Smuzhiyun 				axi_wrapper[sii->axi_num_wrappers].mfg = mfg;
419*4882a593Smuzhiyun 				axi_wrapper[sii->axi_num_wrappers].cid = cid;
420*4882a593Smuzhiyun 				axi_wrapper[sii->axi_num_wrappers].rev = crev;
421*4882a593Smuzhiyun 				axi_wrapper[sii->axi_num_wrappers].wrapper_type = AI_SLAVE_WRAPPER;
422*4882a593Smuzhiyun 
423*4882a593Smuzhiyun 			/* Software WAR as discussed with hardware team, to ensure proper
424*4882a593Smuzhiyun 			* Slave Wrapper Base address is set for 4364 Chip ID.
425*4882a593Smuzhiyun 			* Current address is 0x1810c000, Corrected the same to 0x1810e000.
426*4882a593Smuzhiyun 			* This ensures AXI default slave wrapper is registered along with
427*4882a593Smuzhiyun 			* other slave wrapper cores and is useful while generating trap info
428*4882a593Smuzhiyun 			* when write operation is tried on Invalid Core / Wrapper register
429*4882a593Smuzhiyun 			*/
430*4882a593Smuzhiyun 
431*4882a593Smuzhiyun 				if ((CHIPID(sih->chip) == BCM4364_CHIP_ID) &&
432*4882a593Smuzhiyun 						(cid == DEF_AI_COMP)) {
433*4882a593Smuzhiyun 					axi_wrapper[sii->axi_num_wrappers].wrapper_addr =
434*4882a593Smuzhiyun 						0x1810e000;
435*4882a593Smuzhiyun 				} else {
436*4882a593Smuzhiyun 					axi_wrapper[sii->axi_num_wrappers].wrapper_addr = addrl;
437*4882a593Smuzhiyun 				}
438*4882a593Smuzhiyun 
439*4882a593Smuzhiyun 				sii->axi_num_wrappers++;
440*4882a593Smuzhiyun 
441*4882a593Smuzhiyun 				SI_VMSG(("SLAVE WRAPPER: %d,  mfg:%x, cid:%x,"
442*4882a593Smuzhiyun 					"rev:%x, addr:%x, size:%x\n",
443*4882a593Smuzhiyun 					sii->axi_num_wrappers,  mfg, cid, crev, addrl, sizel));
444*4882a593Smuzhiyun 			}
445*4882a593Smuzhiyun 		}
446*4882a593Smuzhiyun 
447*4882a593Smuzhiyun #ifndef BCM_BACKPLANE_TIMEOUT
448*4882a593Smuzhiyun 		/* Don't record bridges */
449*4882a593Smuzhiyun 		if (br)
450*4882a593Smuzhiyun 			continue;
451*4882a593Smuzhiyun #endif // endif
452*4882a593Smuzhiyun 
453*4882a593Smuzhiyun 		/* Done with core */
454*4882a593Smuzhiyun 		sii->numcores++;
455*4882a593Smuzhiyun 	}
456*4882a593Smuzhiyun 
457*4882a593Smuzhiyun 	SI_ERROR(("Reached end of erom without finding END"));
458*4882a593Smuzhiyun 
459*4882a593Smuzhiyun error:
460*4882a593Smuzhiyun 	sii->numcores = 0;
461*4882a593Smuzhiyun 	return;
462*4882a593Smuzhiyun }
463*4882a593Smuzhiyun 
464*4882a593Smuzhiyun #define AI_SETCOREIDX_MAPSIZE(coreid) \
465*4882a593Smuzhiyun 	(((coreid) == NS_CCB_CORE_ID) ? 15 * SI_CORE_SIZE : SI_CORE_SIZE)
466*4882a593Smuzhiyun 
467*4882a593Smuzhiyun /* This function changes the logical "focus" to the indicated core.
468*4882a593Smuzhiyun  * Return the current core's virtual address.
469*4882a593Smuzhiyun  */
470*4882a593Smuzhiyun static volatile void *
_ai_setcoreidx(si_t * sih,uint coreidx,uint use_wrapn)471*4882a593Smuzhiyun _ai_setcoreidx(si_t *sih, uint coreidx, uint use_wrapn)
472*4882a593Smuzhiyun {
473*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
474*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
475*4882a593Smuzhiyun 	uint32 addr, wrap, wrap2, wrap3;
476*4882a593Smuzhiyun 	volatile void *regs;
477*4882a593Smuzhiyun 
478*4882a593Smuzhiyun 	if (coreidx >= MIN(sii->numcores, SI_MAXCORES))
479*4882a593Smuzhiyun 		return (NULL);
480*4882a593Smuzhiyun 
481*4882a593Smuzhiyun 	addr = cores_info->coresba[coreidx];
482*4882a593Smuzhiyun 	wrap = cores_info->wrapba[coreidx];
483*4882a593Smuzhiyun 	wrap2 = cores_info->wrapba2[coreidx];
484*4882a593Smuzhiyun 	wrap3 = cores_info->wrapba3[coreidx];
485*4882a593Smuzhiyun 
486*4882a593Smuzhiyun #ifdef BCM_BACKPLANE_TIMEOUT
487*4882a593Smuzhiyun 	/* No need to disable interrupts while entering/exiting APB bridge core */
488*4882a593Smuzhiyun 	if ((cores_info->coreid[coreidx] != APB_BRIDGE_CORE_ID) &&
489*4882a593Smuzhiyun 		(cores_info->coreid[sii->curidx] != APB_BRIDGE_CORE_ID))
490*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
491*4882a593Smuzhiyun 	{
492*4882a593Smuzhiyun 		/*
493*4882a593Smuzhiyun 		 * If the user has provided an interrupt mask enabled function,
494*4882a593Smuzhiyun 		 * then assert interrupts are disabled before switching the core.
495*4882a593Smuzhiyun 		 */
496*4882a593Smuzhiyun 		ASSERT((sii->intrsenabled_fn == NULL) ||
497*4882a593Smuzhiyun 			!(*(sii)->intrsenabled_fn)((sii)->intr_arg));
498*4882a593Smuzhiyun 	}
499*4882a593Smuzhiyun 
500*4882a593Smuzhiyun 	switch (BUSTYPE(sih->bustype)) {
501*4882a593Smuzhiyun 	case SI_BUS:
502*4882a593Smuzhiyun 		/* map new one */
503*4882a593Smuzhiyun 		if (!cores_info->regs[coreidx]) {
504*4882a593Smuzhiyun 			cores_info->regs[coreidx] = REG_MAP(addr,
505*4882a593Smuzhiyun 				AI_SETCOREIDX_MAPSIZE(cores_info->coreid[coreidx]));
506*4882a593Smuzhiyun 			ASSERT(GOODREGS(cores_info->regs[coreidx]));
507*4882a593Smuzhiyun 		}
508*4882a593Smuzhiyun 		sii->curmap = regs = cores_info->regs[coreidx];
509*4882a593Smuzhiyun 		if (!cores_info->wrappers[coreidx] && (wrap != 0)) {
510*4882a593Smuzhiyun 			cores_info->wrappers[coreidx] = REG_MAP(wrap, SI_CORE_SIZE);
511*4882a593Smuzhiyun 			ASSERT(GOODREGS(cores_info->wrappers[coreidx]));
512*4882a593Smuzhiyun 		}
513*4882a593Smuzhiyun 		if (!cores_info->wrappers2[coreidx] && (wrap2 != 0)) {
514*4882a593Smuzhiyun 			cores_info->wrappers2[coreidx] = REG_MAP(wrap2, SI_CORE_SIZE);
515*4882a593Smuzhiyun 			ASSERT(GOODREGS(cores_info->wrappers2[coreidx]));
516*4882a593Smuzhiyun 		}
517*4882a593Smuzhiyun 		if (!cores_info->wrappers3[coreidx] && (wrap3 != 0)) {
518*4882a593Smuzhiyun 			cores_info->wrappers3[coreidx] = REG_MAP(wrap3, SI_CORE_SIZE);
519*4882a593Smuzhiyun 			ASSERT(GOODREGS(cores_info->wrappers3[coreidx]));
520*4882a593Smuzhiyun 		}
521*4882a593Smuzhiyun 
522*4882a593Smuzhiyun 		if (use_wrapn == 2) {
523*4882a593Smuzhiyun 			sii->curwrap = cores_info->wrappers3[coreidx];
524*4882a593Smuzhiyun 		} else if (use_wrapn == 1) {
525*4882a593Smuzhiyun 			sii->curwrap = cores_info->wrappers2[coreidx];
526*4882a593Smuzhiyun 		} else {
527*4882a593Smuzhiyun 			sii->curwrap = cores_info->wrappers[coreidx];
528*4882a593Smuzhiyun 		}
529*4882a593Smuzhiyun 		break;
530*4882a593Smuzhiyun 
531*4882a593Smuzhiyun 	case PCI_BUS:
532*4882a593Smuzhiyun #ifdef BCM_BACKPLANE_TIMEOUT
533*4882a593Smuzhiyun 		/* No need to set the BAR0 if core is APB Bridge.
534*4882a593Smuzhiyun 		 * This is to reduce 2 PCI writes while checkng for errlog
535*4882a593Smuzhiyun 		 */
536*4882a593Smuzhiyun 		if (cores_info->coreid[coreidx] != APB_BRIDGE_CORE_ID)
537*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
538*4882a593Smuzhiyun 		{
539*4882a593Smuzhiyun 			/* point bar0 window */
540*4882a593Smuzhiyun 			OSL_PCI_WRITE_CONFIG(sii->osh, PCI_BAR0_WIN, 4, addr);
541*4882a593Smuzhiyun 		}
542*4882a593Smuzhiyun 
543*4882a593Smuzhiyun 		regs = sii->curmap;
544*4882a593Smuzhiyun 		/* point bar0 2nd 4KB window to the primary wrapper */
545*4882a593Smuzhiyun 		if (use_wrapn)
546*4882a593Smuzhiyun 			wrap = wrap2;
547*4882a593Smuzhiyun 		if (PCIE_GEN2(sii))
548*4882a593Smuzhiyun 			OSL_PCI_WRITE_CONFIG(sii->osh, PCIE2_BAR0_WIN2, 4, wrap);
549*4882a593Smuzhiyun 		else
550*4882a593Smuzhiyun 			OSL_PCI_WRITE_CONFIG(sii->osh, PCI_BAR0_WIN2, 4, wrap);
551*4882a593Smuzhiyun 		break;
552*4882a593Smuzhiyun 
553*4882a593Smuzhiyun #ifdef BCMSDIO
554*4882a593Smuzhiyun 	case SPI_BUS:
555*4882a593Smuzhiyun 	case SDIO_BUS:
556*4882a593Smuzhiyun 		sii->curmap = regs = (void *)((uintptr)addr);
557*4882a593Smuzhiyun 		if (use_wrapn)
558*4882a593Smuzhiyun 			sii->curwrap = (void *)((uintptr)wrap2);
559*4882a593Smuzhiyun 		else
560*4882a593Smuzhiyun 			sii->curwrap = (void *)((uintptr)wrap);
561*4882a593Smuzhiyun 		break;
562*4882a593Smuzhiyun #endif	/* BCMSDIO */
563*4882a593Smuzhiyun 
564*4882a593Smuzhiyun 	case PCMCIA_BUS:
565*4882a593Smuzhiyun 	default:
566*4882a593Smuzhiyun 		ASSERT(0);
567*4882a593Smuzhiyun 		regs = NULL;
568*4882a593Smuzhiyun 		break;
569*4882a593Smuzhiyun 	}
570*4882a593Smuzhiyun 
571*4882a593Smuzhiyun 	sii->curmap = regs;
572*4882a593Smuzhiyun 	sii->curidx = coreidx;
573*4882a593Smuzhiyun 
574*4882a593Smuzhiyun 	return regs;
575*4882a593Smuzhiyun }
576*4882a593Smuzhiyun 
577*4882a593Smuzhiyun volatile void *
ai_setcoreidx(si_t * sih,uint coreidx)578*4882a593Smuzhiyun ai_setcoreidx(si_t *sih, uint coreidx)
579*4882a593Smuzhiyun {
580*4882a593Smuzhiyun 	return _ai_setcoreidx(sih, coreidx, 0);
581*4882a593Smuzhiyun }
582*4882a593Smuzhiyun 
583*4882a593Smuzhiyun volatile void *
ai_setcoreidx_2ndwrap(si_t * sih,uint coreidx)584*4882a593Smuzhiyun ai_setcoreidx_2ndwrap(si_t *sih, uint coreidx)
585*4882a593Smuzhiyun {
586*4882a593Smuzhiyun 	return _ai_setcoreidx(sih, coreidx, 1);
587*4882a593Smuzhiyun }
588*4882a593Smuzhiyun 
589*4882a593Smuzhiyun volatile void *
ai_setcoreidx_3rdwrap(si_t * sih,uint coreidx)590*4882a593Smuzhiyun ai_setcoreidx_3rdwrap(si_t *sih, uint coreidx)
591*4882a593Smuzhiyun {
592*4882a593Smuzhiyun 	return _ai_setcoreidx(sih, coreidx, 2);
593*4882a593Smuzhiyun }
594*4882a593Smuzhiyun 
595*4882a593Smuzhiyun void
ai_coreaddrspaceX(si_t * sih,uint asidx,uint32 * addr,uint32 * size)596*4882a593Smuzhiyun ai_coreaddrspaceX(si_t *sih, uint asidx, uint32 *addr, uint32 *size)
597*4882a593Smuzhiyun {
598*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
599*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
600*4882a593Smuzhiyun 	chipcregs_t *cc = NULL;
601*4882a593Smuzhiyun 	uint32 erombase, *eromptr, *eromlim;
602*4882a593Smuzhiyun 	uint i, j, cidx;
603*4882a593Smuzhiyun 	uint32 cia, cib, nmp, nsp;
604*4882a593Smuzhiyun 	uint32 asd, addrl, addrh, sizel, sizeh;
605*4882a593Smuzhiyun 
606*4882a593Smuzhiyun 	for (i = 0; i < sii->numcores; i++) {
607*4882a593Smuzhiyun 		if (cores_info->coreid[i] == CC_CORE_ID) {
608*4882a593Smuzhiyun 			cc = (chipcregs_t *)cores_info->regs[i];
609*4882a593Smuzhiyun 			break;
610*4882a593Smuzhiyun 		}
611*4882a593Smuzhiyun 	}
612*4882a593Smuzhiyun 	if (cc == NULL)
613*4882a593Smuzhiyun 		goto error;
614*4882a593Smuzhiyun 
615*4882a593Smuzhiyun 	erombase = R_REG(sii->osh, &cc->eromptr);
616*4882a593Smuzhiyun 	eromptr = (uint32 *)REG_MAP(erombase, SI_CORE_SIZE);
617*4882a593Smuzhiyun 	eromlim = eromptr + (ER_REMAPCONTROL / sizeof(uint32));
618*4882a593Smuzhiyun 
619*4882a593Smuzhiyun 	cidx = sii->curidx;
620*4882a593Smuzhiyun 	cia = cores_info->cia[cidx];
621*4882a593Smuzhiyun 	cib = cores_info->cib[cidx];
622*4882a593Smuzhiyun 
623*4882a593Smuzhiyun 	nmp = (cib & CIB_NMP_MASK) >> CIB_NMP_SHIFT;
624*4882a593Smuzhiyun 	nsp = (cib & CIB_NSP_MASK) >> CIB_NSP_SHIFT;
625*4882a593Smuzhiyun 
626*4882a593Smuzhiyun 	/* scan for cores */
627*4882a593Smuzhiyun 	while (eromptr < eromlim) {
628*4882a593Smuzhiyun 		if ((get_erom_ent(sih, &eromptr, ER_TAG, ER_CI) == cia) &&
629*4882a593Smuzhiyun 			(get_erom_ent(sih, &eromptr, 0, 0) == cib)) {
630*4882a593Smuzhiyun 			break;
631*4882a593Smuzhiyun 		}
632*4882a593Smuzhiyun 	}
633*4882a593Smuzhiyun 
634*4882a593Smuzhiyun 	/* skip master ports */
635*4882a593Smuzhiyun 	for (i = 0; i < nmp; i++)
636*4882a593Smuzhiyun 		get_erom_ent(sih, &eromptr, ER_VALID, ER_VALID);
637*4882a593Smuzhiyun 
638*4882a593Smuzhiyun 	/* Skip ASDs in port 0 */
639*4882a593Smuzhiyun 	asd = get_asd(sih, &eromptr, 0, 0, AD_ST_SLAVE, &addrl, &addrh, &sizel, &sizeh);
640*4882a593Smuzhiyun 	if (asd == 0) {
641*4882a593Smuzhiyun 		/* Try again to see if it is a bridge */
642*4882a593Smuzhiyun 		asd = get_asd(sih, &eromptr, 0, 0, AD_ST_BRIDGE, &addrl, &addrh,
643*4882a593Smuzhiyun 		              &sizel, &sizeh);
644*4882a593Smuzhiyun 	}
645*4882a593Smuzhiyun 
646*4882a593Smuzhiyun 	j = 1;
647*4882a593Smuzhiyun 	do {
648*4882a593Smuzhiyun 		asd = get_asd(sih, &eromptr, 0, j, AD_ST_SLAVE, &addrl, &addrh,
649*4882a593Smuzhiyun 		              &sizel, &sizeh);
650*4882a593Smuzhiyun 		j++;
651*4882a593Smuzhiyun 	} while (asd != 0);
652*4882a593Smuzhiyun 
653*4882a593Smuzhiyun 	/* Go through the ASDs for other slave ports */
654*4882a593Smuzhiyun 	for (i = 1; i < nsp; i++) {
655*4882a593Smuzhiyun 		j = 0;
656*4882a593Smuzhiyun 		do {
657*4882a593Smuzhiyun 			asd = get_asd(sih, &eromptr, i, j, AD_ST_SLAVE, &addrl, &addrh,
658*4882a593Smuzhiyun 				&sizel, &sizeh);
659*4882a593Smuzhiyun 			if (asd == 0)
660*4882a593Smuzhiyun 				break;
661*4882a593Smuzhiyun 
662*4882a593Smuzhiyun 			if (!asidx--) {
663*4882a593Smuzhiyun 				*addr = addrl;
664*4882a593Smuzhiyun 				*size = sizel;
665*4882a593Smuzhiyun 				return;
666*4882a593Smuzhiyun 			}
667*4882a593Smuzhiyun 			j++;
668*4882a593Smuzhiyun 		} while (1);
669*4882a593Smuzhiyun 
670*4882a593Smuzhiyun 		if (j == 0) {
671*4882a593Smuzhiyun 			SI_ERROR((" SP %d has no address descriptors\n", i));
672*4882a593Smuzhiyun 			break;
673*4882a593Smuzhiyun 		}
674*4882a593Smuzhiyun 	}
675*4882a593Smuzhiyun 
676*4882a593Smuzhiyun error:
677*4882a593Smuzhiyun 	*size = 0;
678*4882a593Smuzhiyun 	return;
679*4882a593Smuzhiyun }
680*4882a593Smuzhiyun 
681*4882a593Smuzhiyun /* Return the number of address spaces in current core */
682*4882a593Smuzhiyun int
ai_numaddrspaces(si_t * sih)683*4882a593Smuzhiyun ai_numaddrspaces(si_t *sih)
684*4882a593Smuzhiyun {
685*4882a593Smuzhiyun 
686*4882a593Smuzhiyun 	BCM_REFERENCE(sih);
687*4882a593Smuzhiyun 
688*4882a593Smuzhiyun 	return 2;
689*4882a593Smuzhiyun }
690*4882a593Smuzhiyun 
691*4882a593Smuzhiyun /* Return the address of the nth address space in the current core
692*4882a593Smuzhiyun  * Arguments:
693*4882a593Smuzhiyun  * sih : Pointer to struct si_t
694*4882a593Smuzhiyun  * spidx : slave port index
695*4882a593Smuzhiyun  * baidx : base address index
696*4882a593Smuzhiyun  */
697*4882a593Smuzhiyun uint32
ai_addrspace(si_t * sih,uint spidx,uint baidx)698*4882a593Smuzhiyun ai_addrspace(si_t *sih, uint spidx, uint baidx)
699*4882a593Smuzhiyun {
700*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
701*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
702*4882a593Smuzhiyun 	uint cidx;
703*4882a593Smuzhiyun 
704*4882a593Smuzhiyun 	cidx = sii->curidx;
705*4882a593Smuzhiyun 
706*4882a593Smuzhiyun 	if (spidx == CORE_SLAVE_PORT_0) {
707*4882a593Smuzhiyun 		if (baidx == CORE_BASE_ADDR_0)
708*4882a593Smuzhiyun 			return cores_info->coresba[cidx];
709*4882a593Smuzhiyun 		else if (baidx == CORE_BASE_ADDR_1)
710*4882a593Smuzhiyun 			return cores_info->coresba2[cidx];
711*4882a593Smuzhiyun 	}
712*4882a593Smuzhiyun 	else if (spidx == CORE_SLAVE_PORT_1) {
713*4882a593Smuzhiyun 		if (baidx == CORE_BASE_ADDR_0)
714*4882a593Smuzhiyun 			return cores_info->csp2ba[cidx];
715*4882a593Smuzhiyun 	}
716*4882a593Smuzhiyun 
717*4882a593Smuzhiyun 	SI_ERROR(("%s: Need to parse the erom again to find %d base addr in %d slave port\n",
718*4882a593Smuzhiyun 	      __FUNCTION__, baidx, spidx));
719*4882a593Smuzhiyun 
720*4882a593Smuzhiyun 	return 0;
721*4882a593Smuzhiyun 
722*4882a593Smuzhiyun }
723*4882a593Smuzhiyun 
724*4882a593Smuzhiyun /* Return the size of the nth address space in the current core
725*4882a593Smuzhiyun * Arguments:
726*4882a593Smuzhiyun * sih : Pointer to struct si_t
727*4882a593Smuzhiyun * spidx : slave port index
728*4882a593Smuzhiyun * baidx : base address index
729*4882a593Smuzhiyun */
730*4882a593Smuzhiyun uint32
ai_addrspacesize(si_t * sih,uint spidx,uint baidx)731*4882a593Smuzhiyun ai_addrspacesize(si_t *sih, uint spidx, uint baidx)
732*4882a593Smuzhiyun {
733*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
734*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
735*4882a593Smuzhiyun 	uint cidx;
736*4882a593Smuzhiyun 
737*4882a593Smuzhiyun 	cidx = sii->curidx;
738*4882a593Smuzhiyun 	if (spidx == CORE_SLAVE_PORT_0) {
739*4882a593Smuzhiyun 		if (baidx == CORE_BASE_ADDR_0)
740*4882a593Smuzhiyun 			return cores_info->coresba_size[cidx];
741*4882a593Smuzhiyun 		else if (baidx == CORE_BASE_ADDR_1)
742*4882a593Smuzhiyun 			return cores_info->coresba2_size[cidx];
743*4882a593Smuzhiyun 	}
744*4882a593Smuzhiyun 	else if (spidx == CORE_SLAVE_PORT_1) {
745*4882a593Smuzhiyun 		if (baidx == CORE_BASE_ADDR_0)
746*4882a593Smuzhiyun 			return cores_info->csp2ba_size[cidx];
747*4882a593Smuzhiyun 	}
748*4882a593Smuzhiyun 
749*4882a593Smuzhiyun 	SI_ERROR(("%s: Need to parse the erom again to find %d base addr in %d slave port\n",
750*4882a593Smuzhiyun 	      __FUNCTION__, baidx, spidx));
751*4882a593Smuzhiyun 
752*4882a593Smuzhiyun 	return 0;
753*4882a593Smuzhiyun }
754*4882a593Smuzhiyun 
755*4882a593Smuzhiyun uint
ai_flag(si_t * sih)756*4882a593Smuzhiyun ai_flag(si_t *sih)
757*4882a593Smuzhiyun {
758*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
759*4882a593Smuzhiyun 	aidmp_t *ai;
760*4882a593Smuzhiyun 
761*4882a593Smuzhiyun 	if (BCM4707_DMP()) {
762*4882a593Smuzhiyun 		SI_ERROR(("%s: Attempting to read CHIPCOMMONB DMP registers on 4707\n",
763*4882a593Smuzhiyun 			__FUNCTION__));
764*4882a593Smuzhiyun 		return sii->curidx;
765*4882a593Smuzhiyun 	}
766*4882a593Smuzhiyun 	if (BCM53573_DMP()) {
767*4882a593Smuzhiyun 		SI_ERROR(("%s: Attempting to read DMP registers on 53573\n", __FUNCTION__));
768*4882a593Smuzhiyun 		return sii->curidx;
769*4882a593Smuzhiyun 	}
770*4882a593Smuzhiyun 	if (PMU_DMP()) {
771*4882a593Smuzhiyun 		uint idx, flag;
772*4882a593Smuzhiyun 		idx = sii->curidx;
773*4882a593Smuzhiyun 		ai_setcoreidx(sih, SI_CC_IDX);
774*4882a593Smuzhiyun 		flag = ai_flag_alt(sih);
775*4882a593Smuzhiyun 		ai_setcoreidx(sih, idx);
776*4882a593Smuzhiyun 		return flag;
777*4882a593Smuzhiyun 	}
778*4882a593Smuzhiyun 
779*4882a593Smuzhiyun 	ai = sii->curwrap;
780*4882a593Smuzhiyun 	ASSERT(ai != NULL);
781*4882a593Smuzhiyun 
782*4882a593Smuzhiyun 	return (R_REG(sii->osh, &ai->oobselouta30) & 0x1f);
783*4882a593Smuzhiyun }
784*4882a593Smuzhiyun 
785*4882a593Smuzhiyun uint
ai_flag_alt(si_t * sih)786*4882a593Smuzhiyun ai_flag_alt(si_t *sih)
787*4882a593Smuzhiyun {
788*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
789*4882a593Smuzhiyun 	aidmp_t *ai;
790*4882a593Smuzhiyun 
791*4882a593Smuzhiyun 	if (BCM4707_DMP()) {
792*4882a593Smuzhiyun 		SI_ERROR(("%s: Attempting to read CHIPCOMMONB DMP registers on 4707\n",
793*4882a593Smuzhiyun 			__FUNCTION__));
794*4882a593Smuzhiyun 		return sii->curidx;
795*4882a593Smuzhiyun 	}
796*4882a593Smuzhiyun 
797*4882a593Smuzhiyun 	ai = sii->curwrap;
798*4882a593Smuzhiyun 
799*4882a593Smuzhiyun 	return ((R_REG(sii->osh, &ai->oobselouta30) >> AI_OOBSEL_1_SHIFT) & AI_OOBSEL_MASK);
800*4882a593Smuzhiyun }
801*4882a593Smuzhiyun 
802*4882a593Smuzhiyun void
ai_setint(si_t * sih,int siflag)803*4882a593Smuzhiyun ai_setint(si_t *sih, int siflag)
804*4882a593Smuzhiyun {
805*4882a593Smuzhiyun 	BCM_REFERENCE(sih);
806*4882a593Smuzhiyun 	BCM_REFERENCE(siflag);
807*4882a593Smuzhiyun 
808*4882a593Smuzhiyun }
809*4882a593Smuzhiyun 
810*4882a593Smuzhiyun uint
ai_wrap_reg(si_t * sih,uint32 offset,uint32 mask,uint32 val)811*4882a593Smuzhiyun ai_wrap_reg(si_t *sih, uint32 offset, uint32 mask, uint32 val)
812*4882a593Smuzhiyun {
813*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
814*4882a593Smuzhiyun 	uint32 *addr = (uint32 *) ((uchar *)(sii->curwrap) + offset);
815*4882a593Smuzhiyun 
816*4882a593Smuzhiyun 	if (mask || val) {
817*4882a593Smuzhiyun 		uint32 w = R_REG(sii->osh, addr);
818*4882a593Smuzhiyun 		w &= ~mask;
819*4882a593Smuzhiyun 		w |= val;
820*4882a593Smuzhiyun 		W_REG(sii->osh, addr, w);
821*4882a593Smuzhiyun 	}
822*4882a593Smuzhiyun 	return (R_REG(sii->osh, addr));
823*4882a593Smuzhiyun }
824*4882a593Smuzhiyun 
825*4882a593Smuzhiyun uint
ai_corevendor(si_t * sih)826*4882a593Smuzhiyun ai_corevendor(si_t *sih)
827*4882a593Smuzhiyun {
828*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
829*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
830*4882a593Smuzhiyun 	uint32 cia;
831*4882a593Smuzhiyun 
832*4882a593Smuzhiyun 	cia = cores_info->cia[sii->curidx];
833*4882a593Smuzhiyun 	return ((cia & CIA_MFG_MASK) >> CIA_MFG_SHIFT);
834*4882a593Smuzhiyun }
835*4882a593Smuzhiyun 
836*4882a593Smuzhiyun uint
ai_corerev(si_t * sih)837*4882a593Smuzhiyun ai_corerev(si_t *sih)
838*4882a593Smuzhiyun {
839*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
840*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
841*4882a593Smuzhiyun 	uint32 cib;
842*4882a593Smuzhiyun 
843*4882a593Smuzhiyun 	cib = cores_info->cib[sii->curidx];
844*4882a593Smuzhiyun 	return ((cib & CIB_REV_MASK) >> CIB_REV_SHIFT);
845*4882a593Smuzhiyun }
846*4882a593Smuzhiyun 
847*4882a593Smuzhiyun uint
ai_corerev_minor(si_t * sih)848*4882a593Smuzhiyun ai_corerev_minor(si_t *sih)
849*4882a593Smuzhiyun {
850*4882a593Smuzhiyun 	return (ai_core_sflags(sih, 0, 0) >> SISF_MINORREV_D11_SHIFT) &
851*4882a593Smuzhiyun 			SISF_MINORREV_D11_MASK;
852*4882a593Smuzhiyun }
853*4882a593Smuzhiyun 
854*4882a593Smuzhiyun bool
ai_iscoreup(si_t * sih)855*4882a593Smuzhiyun ai_iscoreup(si_t *sih)
856*4882a593Smuzhiyun {
857*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
858*4882a593Smuzhiyun 	aidmp_t *ai;
859*4882a593Smuzhiyun 
860*4882a593Smuzhiyun 	ai = sii->curwrap;
861*4882a593Smuzhiyun 
862*4882a593Smuzhiyun 	return (((R_REG(sii->osh, &ai->ioctrl) & (SICF_FGC | SICF_CLOCK_EN)) == SICF_CLOCK_EN) &&
863*4882a593Smuzhiyun 	        ((R_REG(sii->osh, &ai->resetctrl) & AIRC_RESET) == 0));
864*4882a593Smuzhiyun }
865*4882a593Smuzhiyun 
866*4882a593Smuzhiyun /*
867*4882a593Smuzhiyun  * Switch to 'coreidx', issue a single arbitrary 32bit register mask&set operation,
868*4882a593Smuzhiyun  * switch back to the original core, and return the new value.
869*4882a593Smuzhiyun  *
870*4882a593Smuzhiyun  * When using the silicon backplane, no fiddling with interrupts or core switches is needed.
871*4882a593Smuzhiyun  *
872*4882a593Smuzhiyun  * Also, when using pci/pcie, we can optimize away the core switching for pci registers
873*4882a593Smuzhiyun  * and (on newer pci cores) chipcommon registers.
874*4882a593Smuzhiyun  */
875*4882a593Smuzhiyun uint
ai_corereg(si_t * sih,uint coreidx,uint regoff,uint mask,uint val)876*4882a593Smuzhiyun ai_corereg(si_t *sih, uint coreidx, uint regoff, uint mask, uint val)
877*4882a593Smuzhiyun {
878*4882a593Smuzhiyun 	uint origidx = 0;
879*4882a593Smuzhiyun 	volatile uint32 *r = NULL;
880*4882a593Smuzhiyun 	uint w;
881*4882a593Smuzhiyun 	uint intr_val = 0;
882*4882a593Smuzhiyun 	bool fast = FALSE;
883*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
884*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
885*4882a593Smuzhiyun 
886*4882a593Smuzhiyun 	ASSERT(GOODIDX(coreidx));
887*4882a593Smuzhiyun 	ASSERT(regoff < SI_CORE_SIZE);
888*4882a593Smuzhiyun 	ASSERT((val & ~mask) == 0);
889*4882a593Smuzhiyun 
890*4882a593Smuzhiyun 	if (coreidx >= SI_MAXCORES)
891*4882a593Smuzhiyun 		return 0;
892*4882a593Smuzhiyun 
893*4882a593Smuzhiyun 	if (BUSTYPE(sih->bustype) == SI_BUS) {
894*4882a593Smuzhiyun 		/* If internal bus, we can always get at everything */
895*4882a593Smuzhiyun 		fast = TRUE;
896*4882a593Smuzhiyun 		/* map if does not exist */
897*4882a593Smuzhiyun 		if (!cores_info->regs[coreidx]) {
898*4882a593Smuzhiyun 			cores_info->regs[coreidx] = REG_MAP(cores_info->coresba[coreidx],
899*4882a593Smuzhiyun 			                            SI_CORE_SIZE);
900*4882a593Smuzhiyun 			ASSERT(GOODREGS(cores_info->regs[coreidx]));
901*4882a593Smuzhiyun 		}
902*4882a593Smuzhiyun 		r = (volatile uint32 *)((volatile uchar *)cores_info->regs[coreidx] + regoff);
903*4882a593Smuzhiyun 	} else if (BUSTYPE(sih->bustype) == PCI_BUS) {
904*4882a593Smuzhiyun 		/* If pci/pcie, we can get at pci/pcie regs and on newer cores to chipc */
905*4882a593Smuzhiyun 
906*4882a593Smuzhiyun 		if ((cores_info->coreid[coreidx] == CC_CORE_ID) && SI_FAST(sii)) {
907*4882a593Smuzhiyun 			/* Chipc registers are mapped at 12KB */
908*4882a593Smuzhiyun 
909*4882a593Smuzhiyun 			fast = TRUE;
910*4882a593Smuzhiyun 			r = (volatile uint32 *)((volatile char *)sii->curmap +
911*4882a593Smuzhiyun 			               PCI_16KB0_CCREGS_OFFSET + regoff);
912*4882a593Smuzhiyun 		} else if (sii->pub.buscoreidx == coreidx) {
913*4882a593Smuzhiyun 			/* pci registers are at either in the last 2KB of an 8KB window
914*4882a593Smuzhiyun 			 * or, in pcie and pci rev 13 at 8KB
915*4882a593Smuzhiyun 			 */
916*4882a593Smuzhiyun 			fast = TRUE;
917*4882a593Smuzhiyun 			if (SI_FAST(sii))
918*4882a593Smuzhiyun 				r = (volatile uint32 *)((volatile char *)sii->curmap +
919*4882a593Smuzhiyun 				               PCI_16KB0_PCIREGS_OFFSET + regoff);
920*4882a593Smuzhiyun 			else
921*4882a593Smuzhiyun 				r = (volatile uint32 *)((volatile char *)sii->curmap +
922*4882a593Smuzhiyun 				               ((regoff >= SBCONFIGOFF) ?
923*4882a593Smuzhiyun 				                PCI_BAR0_PCISBR_OFFSET : PCI_BAR0_PCIREGS_OFFSET) +
924*4882a593Smuzhiyun 				               regoff);
925*4882a593Smuzhiyun 		}
926*4882a593Smuzhiyun 	}
927*4882a593Smuzhiyun 
928*4882a593Smuzhiyun 	if (!fast) {
929*4882a593Smuzhiyun 		INTR_OFF(sii, intr_val);
930*4882a593Smuzhiyun 
931*4882a593Smuzhiyun 		/* save current core index */
932*4882a593Smuzhiyun 		origidx = si_coreidx(&sii->pub);
933*4882a593Smuzhiyun 
934*4882a593Smuzhiyun 		/* switch core */
935*4882a593Smuzhiyun 		r = (volatile uint32*) ((volatile uchar*) ai_setcoreidx(&sii->pub, coreidx) +
936*4882a593Smuzhiyun 		               regoff);
937*4882a593Smuzhiyun 	}
938*4882a593Smuzhiyun 	ASSERT(r != NULL);
939*4882a593Smuzhiyun 
940*4882a593Smuzhiyun 	/* mask and set */
941*4882a593Smuzhiyun 	if (mask || val) {
942*4882a593Smuzhiyun 		w = (R_REG(sii->osh, r) & ~mask) | val;
943*4882a593Smuzhiyun 		W_REG(sii->osh, r, w);
944*4882a593Smuzhiyun 	}
945*4882a593Smuzhiyun 
946*4882a593Smuzhiyun 	/* readback */
947*4882a593Smuzhiyun 	w = R_REG(sii->osh, r);
948*4882a593Smuzhiyun 
949*4882a593Smuzhiyun 	if (!fast) {
950*4882a593Smuzhiyun 		/* restore core index */
951*4882a593Smuzhiyun 		if (origidx != coreidx)
952*4882a593Smuzhiyun 			ai_setcoreidx(&sii->pub, origidx);
953*4882a593Smuzhiyun 
954*4882a593Smuzhiyun 		INTR_RESTORE(sii, intr_val);
955*4882a593Smuzhiyun 	}
956*4882a593Smuzhiyun 
957*4882a593Smuzhiyun 	return (w);
958*4882a593Smuzhiyun }
959*4882a593Smuzhiyun 
960*4882a593Smuzhiyun /*
961*4882a593Smuzhiyun  * Switch to 'coreidx', issue a single arbitrary 32bit register mask&set operation,
962*4882a593Smuzhiyun  * switch back to the original core, and return the new value.
963*4882a593Smuzhiyun  *
964*4882a593Smuzhiyun  * When using the silicon backplane, no fiddling with interrupts or core switches is needed.
965*4882a593Smuzhiyun  *
966*4882a593Smuzhiyun  * Also, when using pci/pcie, we can optimize away the core switching for pci registers
967*4882a593Smuzhiyun  * and (on newer pci cores) chipcommon registers.
968*4882a593Smuzhiyun  */
969*4882a593Smuzhiyun uint
ai_corereg_writeonly(si_t * sih,uint coreidx,uint regoff,uint mask,uint val)970*4882a593Smuzhiyun ai_corereg_writeonly(si_t *sih, uint coreidx, uint regoff, uint mask, uint val)
971*4882a593Smuzhiyun {
972*4882a593Smuzhiyun 	uint origidx = 0;
973*4882a593Smuzhiyun 	volatile uint32 *r = NULL;
974*4882a593Smuzhiyun 	uint w = 0;
975*4882a593Smuzhiyun 	uint intr_val = 0;
976*4882a593Smuzhiyun 	bool fast = FALSE;
977*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
978*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
979*4882a593Smuzhiyun 
980*4882a593Smuzhiyun 	ASSERT(GOODIDX(coreidx));
981*4882a593Smuzhiyun 	ASSERT(regoff < SI_CORE_SIZE);
982*4882a593Smuzhiyun 	ASSERT((val & ~mask) == 0);
983*4882a593Smuzhiyun 
984*4882a593Smuzhiyun 	if (coreidx >= SI_MAXCORES)
985*4882a593Smuzhiyun 		return 0;
986*4882a593Smuzhiyun 
987*4882a593Smuzhiyun 	if (BUSTYPE(sih->bustype) == SI_BUS) {
988*4882a593Smuzhiyun 		/* If internal bus, we can always get at everything */
989*4882a593Smuzhiyun 		fast = TRUE;
990*4882a593Smuzhiyun 		/* map if does not exist */
991*4882a593Smuzhiyun 		if (!cores_info->regs[coreidx]) {
992*4882a593Smuzhiyun 			cores_info->regs[coreidx] = REG_MAP(cores_info->coresba[coreidx],
993*4882a593Smuzhiyun 			                            SI_CORE_SIZE);
994*4882a593Smuzhiyun 			ASSERT(GOODREGS(cores_info->regs[coreidx]));
995*4882a593Smuzhiyun 		}
996*4882a593Smuzhiyun 		r = (volatile uint32 *)((volatile uchar *)cores_info->regs[coreidx] + regoff);
997*4882a593Smuzhiyun 	} else if (BUSTYPE(sih->bustype) == PCI_BUS) {
998*4882a593Smuzhiyun 		/* If pci/pcie, we can get at pci/pcie regs and on newer cores to chipc */
999*4882a593Smuzhiyun 
1000*4882a593Smuzhiyun 		if ((cores_info->coreid[coreidx] == CC_CORE_ID) && SI_FAST(sii)) {
1001*4882a593Smuzhiyun 			/* Chipc registers are mapped at 12KB */
1002*4882a593Smuzhiyun 
1003*4882a593Smuzhiyun 			fast = TRUE;
1004*4882a593Smuzhiyun 			r = (volatile uint32 *)((volatile char *)sii->curmap +
1005*4882a593Smuzhiyun 			               PCI_16KB0_CCREGS_OFFSET + regoff);
1006*4882a593Smuzhiyun 		} else if (sii->pub.buscoreidx == coreidx) {
1007*4882a593Smuzhiyun 			/* pci registers are at either in the last 2KB of an 8KB window
1008*4882a593Smuzhiyun 			 * or, in pcie and pci rev 13 at 8KB
1009*4882a593Smuzhiyun 			 */
1010*4882a593Smuzhiyun 			fast = TRUE;
1011*4882a593Smuzhiyun 			if (SI_FAST(sii))
1012*4882a593Smuzhiyun 				r = (volatile uint32 *)((volatile char *)sii->curmap +
1013*4882a593Smuzhiyun 				               PCI_16KB0_PCIREGS_OFFSET + regoff);
1014*4882a593Smuzhiyun 			else
1015*4882a593Smuzhiyun 				r = (volatile uint32 *)((volatile char *)sii->curmap +
1016*4882a593Smuzhiyun 				               ((regoff >= SBCONFIGOFF) ?
1017*4882a593Smuzhiyun 				                PCI_BAR0_PCISBR_OFFSET : PCI_BAR0_PCIREGS_OFFSET) +
1018*4882a593Smuzhiyun 				               regoff);
1019*4882a593Smuzhiyun 		}
1020*4882a593Smuzhiyun 	}
1021*4882a593Smuzhiyun 
1022*4882a593Smuzhiyun 	if (!fast) {
1023*4882a593Smuzhiyun 		INTR_OFF(sii, intr_val);
1024*4882a593Smuzhiyun 
1025*4882a593Smuzhiyun 		/* save current core index */
1026*4882a593Smuzhiyun 		origidx = si_coreidx(&sii->pub);
1027*4882a593Smuzhiyun 
1028*4882a593Smuzhiyun 		/* switch core */
1029*4882a593Smuzhiyun 		r = (volatile uint32*) ((volatile uchar*) ai_setcoreidx(&sii->pub, coreidx) +
1030*4882a593Smuzhiyun 		               regoff);
1031*4882a593Smuzhiyun 	}
1032*4882a593Smuzhiyun 	ASSERT(r != NULL);
1033*4882a593Smuzhiyun 
1034*4882a593Smuzhiyun 	/* mask and set */
1035*4882a593Smuzhiyun 	if (mask || val) {
1036*4882a593Smuzhiyun 		w = (R_REG(sii->osh, r) & ~mask) | val;
1037*4882a593Smuzhiyun 		W_REG(sii->osh, r, w);
1038*4882a593Smuzhiyun 	}
1039*4882a593Smuzhiyun 
1040*4882a593Smuzhiyun 	if (!fast) {
1041*4882a593Smuzhiyun 		/* restore core index */
1042*4882a593Smuzhiyun 		if (origidx != coreidx)
1043*4882a593Smuzhiyun 			ai_setcoreidx(&sii->pub, origidx);
1044*4882a593Smuzhiyun 
1045*4882a593Smuzhiyun 		INTR_RESTORE(sii, intr_val);
1046*4882a593Smuzhiyun 	}
1047*4882a593Smuzhiyun 
1048*4882a593Smuzhiyun 	return (w);
1049*4882a593Smuzhiyun }
1050*4882a593Smuzhiyun 
1051*4882a593Smuzhiyun /*
1052*4882a593Smuzhiyun  * If there is no need for fiddling with interrupts or core switches (typically silicon
1053*4882a593Smuzhiyun  * back plane registers, pci registers and chipcommon registers), this function
1054*4882a593Smuzhiyun  * returns the register offset on this core to a mapped address. This address can
1055*4882a593Smuzhiyun  * be used for W_REG/R_REG directly.
1056*4882a593Smuzhiyun  *
1057*4882a593Smuzhiyun  * For accessing registers that would need a core switch, this function will return
1058*4882a593Smuzhiyun  * NULL.
1059*4882a593Smuzhiyun  */
1060*4882a593Smuzhiyun volatile uint32 *
ai_corereg_addr(si_t * sih,uint coreidx,uint regoff)1061*4882a593Smuzhiyun ai_corereg_addr(si_t *sih, uint coreidx, uint regoff)
1062*4882a593Smuzhiyun {
1063*4882a593Smuzhiyun 	volatile uint32 *r = NULL;
1064*4882a593Smuzhiyun 	bool fast = FALSE;
1065*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
1066*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
1067*4882a593Smuzhiyun 
1068*4882a593Smuzhiyun 	ASSERT(GOODIDX(coreidx));
1069*4882a593Smuzhiyun 	ASSERT(regoff < SI_CORE_SIZE);
1070*4882a593Smuzhiyun 
1071*4882a593Smuzhiyun 	if (coreidx >= SI_MAXCORES)
1072*4882a593Smuzhiyun 		return 0;
1073*4882a593Smuzhiyun 
1074*4882a593Smuzhiyun 	if (BUSTYPE(sih->bustype) == SI_BUS) {
1075*4882a593Smuzhiyun 		/* If internal bus, we can always get at everything */
1076*4882a593Smuzhiyun 		fast = TRUE;
1077*4882a593Smuzhiyun 		/* map if does not exist */
1078*4882a593Smuzhiyun 		if (!cores_info->regs[coreidx]) {
1079*4882a593Smuzhiyun 			cores_info->regs[coreidx] = REG_MAP(cores_info->coresba[coreidx],
1080*4882a593Smuzhiyun 			                            SI_CORE_SIZE);
1081*4882a593Smuzhiyun 			ASSERT(GOODREGS(cores_info->regs[coreidx]));
1082*4882a593Smuzhiyun 		}
1083*4882a593Smuzhiyun 		r = (volatile uint32 *)((volatile uchar *)cores_info->regs[coreidx] + regoff);
1084*4882a593Smuzhiyun 	} else if (BUSTYPE(sih->bustype) == PCI_BUS) {
1085*4882a593Smuzhiyun 		/* If pci/pcie, we can get at pci/pcie regs and on newer cores to chipc */
1086*4882a593Smuzhiyun 
1087*4882a593Smuzhiyun 		if ((cores_info->coreid[coreidx] == CC_CORE_ID) && SI_FAST(sii)) {
1088*4882a593Smuzhiyun 			/* Chipc registers are mapped at 12KB */
1089*4882a593Smuzhiyun 
1090*4882a593Smuzhiyun 			fast = TRUE;
1091*4882a593Smuzhiyun 			r = (volatile uint32 *)((volatile char *)sii->curmap +
1092*4882a593Smuzhiyun 			               PCI_16KB0_CCREGS_OFFSET + regoff);
1093*4882a593Smuzhiyun 		} else if (sii->pub.buscoreidx == coreidx) {
1094*4882a593Smuzhiyun 			/* pci registers are at either in the last 2KB of an 8KB window
1095*4882a593Smuzhiyun 			 * or, in pcie and pci rev 13 at 8KB
1096*4882a593Smuzhiyun 			 */
1097*4882a593Smuzhiyun 			fast = TRUE;
1098*4882a593Smuzhiyun 			if (SI_FAST(sii))
1099*4882a593Smuzhiyun 				r = (volatile uint32 *)((volatile char *)sii->curmap +
1100*4882a593Smuzhiyun 				               PCI_16KB0_PCIREGS_OFFSET + regoff);
1101*4882a593Smuzhiyun 			else
1102*4882a593Smuzhiyun 				r = (volatile uint32 *)((volatile char *)sii->curmap +
1103*4882a593Smuzhiyun 				               ((regoff >= SBCONFIGOFF) ?
1104*4882a593Smuzhiyun 				                PCI_BAR0_PCISBR_OFFSET : PCI_BAR0_PCIREGS_OFFSET) +
1105*4882a593Smuzhiyun 				               regoff);
1106*4882a593Smuzhiyun 		}
1107*4882a593Smuzhiyun 	}
1108*4882a593Smuzhiyun 
1109*4882a593Smuzhiyun 	if (!fast) {
1110*4882a593Smuzhiyun 		ASSERT(sii->curidx == coreidx);
1111*4882a593Smuzhiyun 		r = (volatile uint32*) ((volatile uchar*)sii->curmap + regoff);
1112*4882a593Smuzhiyun 	}
1113*4882a593Smuzhiyun 
1114*4882a593Smuzhiyun 	return (r);
1115*4882a593Smuzhiyun }
1116*4882a593Smuzhiyun 
1117*4882a593Smuzhiyun void
ai_core_disable(si_t * sih,uint32 bits)1118*4882a593Smuzhiyun ai_core_disable(si_t *sih, uint32 bits)
1119*4882a593Smuzhiyun {
1120*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
1121*4882a593Smuzhiyun 	volatile uint32 dummy;
1122*4882a593Smuzhiyun 	uint32 status;
1123*4882a593Smuzhiyun 	aidmp_t *ai;
1124*4882a593Smuzhiyun 
1125*4882a593Smuzhiyun 	ASSERT(GOODREGS(sii->curwrap));
1126*4882a593Smuzhiyun 	ai = sii->curwrap;
1127*4882a593Smuzhiyun 
1128*4882a593Smuzhiyun 	/* if core is already in reset, just return */
1129*4882a593Smuzhiyun 	if (R_REG(sii->osh, &ai->resetctrl) & AIRC_RESET) {
1130*4882a593Smuzhiyun 		return;
1131*4882a593Smuzhiyun 	}
1132*4882a593Smuzhiyun 
1133*4882a593Smuzhiyun 	/* ensure there are no pending backplane operations */
1134*4882a593Smuzhiyun 	SPINWAIT(((status = R_REG(sii->osh, &ai->resetstatus)) != 0), 300);
1135*4882a593Smuzhiyun 
1136*4882a593Smuzhiyun 	/* if pending backplane ops still, try waiting longer */
1137*4882a593Smuzhiyun 	if (status != 0) {
1138*4882a593Smuzhiyun 		/* 300usecs was sufficient to allow backplane ops to clear for big hammer */
1139*4882a593Smuzhiyun 		/* during driver load we may need more time */
1140*4882a593Smuzhiyun 		SPINWAIT(((status = R_REG(sii->osh, &ai->resetstatus)) != 0), 10000);
1141*4882a593Smuzhiyun 		/* if still pending ops, continue on and try disable anyway */
1142*4882a593Smuzhiyun 		/* this is in big hammer path, so don't call wl_reinit in this case... */
1143*4882a593Smuzhiyun 	}
1144*4882a593Smuzhiyun 
1145*4882a593Smuzhiyun 	W_REG(sii->osh, &ai->resetctrl, AIRC_RESET);
1146*4882a593Smuzhiyun 	dummy = R_REG(sii->osh, &ai->resetctrl);
1147*4882a593Smuzhiyun 	BCM_REFERENCE(dummy);
1148*4882a593Smuzhiyun 	OSL_DELAY(1);
1149*4882a593Smuzhiyun 
1150*4882a593Smuzhiyun 	W_REG(sii->osh, &ai->ioctrl, bits);
1151*4882a593Smuzhiyun 	dummy = R_REG(sii->osh, &ai->ioctrl);
1152*4882a593Smuzhiyun 	BCM_REFERENCE(dummy);
1153*4882a593Smuzhiyun 	OSL_DELAY(10);
1154*4882a593Smuzhiyun }
1155*4882a593Smuzhiyun 
1156*4882a593Smuzhiyun /* reset and re-enable a core
1157*4882a593Smuzhiyun  * inputs:
1158*4882a593Smuzhiyun  * bits - core specific bits that are set during and after reset sequence
1159*4882a593Smuzhiyun  * resetbits - core specific bits that are set only during reset sequence
1160*4882a593Smuzhiyun  */
1161*4882a593Smuzhiyun static void
_ai_core_reset(si_t * sih,uint32 bits,uint32 resetbits)1162*4882a593Smuzhiyun _ai_core_reset(si_t *sih, uint32 bits, uint32 resetbits)
1163*4882a593Smuzhiyun {
1164*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
1165*4882a593Smuzhiyun #if defined(UCM_CORRUPTION_WAR)
1166*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
1167*4882a593Smuzhiyun #endif // endif
1168*4882a593Smuzhiyun 	aidmp_t *ai;
1169*4882a593Smuzhiyun 	volatile uint32 dummy;
1170*4882a593Smuzhiyun 	uint loop_counter = 10;
1171*4882a593Smuzhiyun 
1172*4882a593Smuzhiyun 	ASSERT(GOODREGS(sii->curwrap));
1173*4882a593Smuzhiyun 	ai = sii->curwrap;
1174*4882a593Smuzhiyun 
1175*4882a593Smuzhiyun 	/* ensure there are no pending backplane operations */
1176*4882a593Smuzhiyun 	SPINWAIT(((dummy = R_REG(sii->osh, &ai->resetstatus)) != 0), SPINWAIT_TIME_US);
1177*4882a593Smuzhiyun 
1178*4882a593Smuzhiyun 	/* put core into reset state */
1179*4882a593Smuzhiyun 	W_REG(sii->osh, &ai->resetctrl, AIRC_RESET);
1180*4882a593Smuzhiyun 	OSL_DELAY(10);
1181*4882a593Smuzhiyun 
1182*4882a593Smuzhiyun 	/* ensure there are no pending backplane operations */
1183*4882a593Smuzhiyun 	SPINWAIT((R_REG(sii->osh, &ai->resetstatus) != 0), SPINWAIT_TIME_US);
1184*4882a593Smuzhiyun 
1185*4882a593Smuzhiyun 	W_REG(sii->osh, &ai->ioctrl, (bits | resetbits | SICF_FGC | SICF_CLOCK_EN));
1186*4882a593Smuzhiyun 	dummy = R_REG(sii->osh, &ai->ioctrl);
1187*4882a593Smuzhiyun 	BCM_REFERENCE(dummy);
1188*4882a593Smuzhiyun #ifdef UCM_CORRUPTION_WAR
1189*4882a593Smuzhiyun 	if (cores_info->coreid[sii->curidx] == D11_CORE_ID) {
1190*4882a593Smuzhiyun 		/* Reset FGC */
1191*4882a593Smuzhiyun 		OSL_DELAY(1);
1192*4882a593Smuzhiyun 		W_REG(sii->osh, &ai->ioctrl, (dummy & (~SICF_FGC)));
1193*4882a593Smuzhiyun 	}
1194*4882a593Smuzhiyun #endif /* UCM_CORRUPTION_WAR */
1195*4882a593Smuzhiyun 	/* ensure there are no pending backplane operations */
1196*4882a593Smuzhiyun 	SPINWAIT(((dummy = R_REG(sii->osh, &ai->resetstatus)) != 0), SPINWAIT_TIME_US);
1197*4882a593Smuzhiyun 
1198*4882a593Smuzhiyun 	while (R_REG(sii->osh, &ai->resetctrl) != 0 && --loop_counter != 0) {
1199*4882a593Smuzhiyun 		/* ensure there are no pending backplane operations */
1200*4882a593Smuzhiyun 		SPINWAIT(((dummy = R_REG(sii->osh, &ai->resetstatus)) != 0), SPINWAIT_TIME_US);
1201*4882a593Smuzhiyun 
1202*4882a593Smuzhiyun 		/* take core out of reset */
1203*4882a593Smuzhiyun 		W_REG(sii->osh, &ai->resetctrl, 0);
1204*4882a593Smuzhiyun 
1205*4882a593Smuzhiyun 		/* ensure there are no pending backplane operations */
1206*4882a593Smuzhiyun 		SPINWAIT((R_REG(sii->osh, &ai->resetstatus) != 0), SPINWAIT_TIME_US);
1207*4882a593Smuzhiyun 	}
1208*4882a593Smuzhiyun 
1209*4882a593Smuzhiyun #ifdef UCM_CORRUPTION_WAR
1210*4882a593Smuzhiyun 	/* Pulse FGC after lifting Reset */
1211*4882a593Smuzhiyun 	W_REG(sii->osh, &ai->ioctrl, (bits | SICF_FGC | SICF_CLOCK_EN));
1212*4882a593Smuzhiyun #else
1213*4882a593Smuzhiyun 	W_REG(sii->osh, &ai->ioctrl, (bits | SICF_CLOCK_EN));
1214*4882a593Smuzhiyun #endif /* UCM_CORRUPTION_WAR */
1215*4882a593Smuzhiyun 	dummy = R_REG(sii->osh, &ai->ioctrl);
1216*4882a593Smuzhiyun 	BCM_REFERENCE(dummy);
1217*4882a593Smuzhiyun #ifdef UCM_CORRUPTION_WAR
1218*4882a593Smuzhiyun 	if (cores_info->coreid[sii->curidx] == D11_CORE_ID) {
1219*4882a593Smuzhiyun 		/* Reset FGC */
1220*4882a593Smuzhiyun 		OSL_DELAY(1);
1221*4882a593Smuzhiyun 		W_REG(sii->osh, &ai->ioctrl, (dummy & (~SICF_FGC)));
1222*4882a593Smuzhiyun 	}
1223*4882a593Smuzhiyun #endif /* UCM_CORRUPTION_WAR */
1224*4882a593Smuzhiyun 	OSL_DELAY(1);
1225*4882a593Smuzhiyun 
1226*4882a593Smuzhiyun }
1227*4882a593Smuzhiyun 
1228*4882a593Smuzhiyun void
ai_core_reset(si_t * sih,uint32 bits,uint32 resetbits)1229*4882a593Smuzhiyun ai_core_reset(si_t *sih, uint32 bits, uint32 resetbits)
1230*4882a593Smuzhiyun {
1231*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
1232*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
1233*4882a593Smuzhiyun 	uint idx = sii->curidx;
1234*4882a593Smuzhiyun 
1235*4882a593Smuzhiyun 	if (cores_info->wrapba3[idx] != 0) {
1236*4882a593Smuzhiyun 		ai_setcoreidx_3rdwrap(sih, idx);
1237*4882a593Smuzhiyun 		_ai_core_reset(sih, bits, resetbits);
1238*4882a593Smuzhiyun 		ai_setcoreidx(sih, idx);
1239*4882a593Smuzhiyun 	}
1240*4882a593Smuzhiyun 
1241*4882a593Smuzhiyun 	if (cores_info->wrapba2[idx] != 0) {
1242*4882a593Smuzhiyun 		ai_setcoreidx_2ndwrap(sih, idx);
1243*4882a593Smuzhiyun 		_ai_core_reset(sih, bits, resetbits);
1244*4882a593Smuzhiyun 		ai_setcoreidx(sih, idx);
1245*4882a593Smuzhiyun 	}
1246*4882a593Smuzhiyun 
1247*4882a593Smuzhiyun 	_ai_core_reset(sih, bits, resetbits);
1248*4882a593Smuzhiyun }
1249*4882a593Smuzhiyun 
1250*4882a593Smuzhiyun void
ai_core_cflags_wo(si_t * sih,uint32 mask,uint32 val)1251*4882a593Smuzhiyun ai_core_cflags_wo(si_t *sih, uint32 mask, uint32 val)
1252*4882a593Smuzhiyun {
1253*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
1254*4882a593Smuzhiyun 	aidmp_t *ai;
1255*4882a593Smuzhiyun 	uint32 w;
1256*4882a593Smuzhiyun 
1257*4882a593Smuzhiyun 	if (BCM4707_DMP()) {
1258*4882a593Smuzhiyun 		SI_ERROR(("%s: Accessing CHIPCOMMONB DMP register (ioctrl) on 4707\n",
1259*4882a593Smuzhiyun 			__FUNCTION__));
1260*4882a593Smuzhiyun 		return;
1261*4882a593Smuzhiyun 	}
1262*4882a593Smuzhiyun 	if (PMU_DMP()) {
1263*4882a593Smuzhiyun 		SI_ERROR(("%s: Accessing PMU DMP register (ioctrl)\n",
1264*4882a593Smuzhiyun 			__FUNCTION__));
1265*4882a593Smuzhiyun 		return;
1266*4882a593Smuzhiyun 	}
1267*4882a593Smuzhiyun 
1268*4882a593Smuzhiyun 	ASSERT(GOODREGS(sii->curwrap));
1269*4882a593Smuzhiyun 	ai = sii->curwrap;
1270*4882a593Smuzhiyun 
1271*4882a593Smuzhiyun 	ASSERT((val & ~mask) == 0);
1272*4882a593Smuzhiyun 
1273*4882a593Smuzhiyun 	if (mask || val) {
1274*4882a593Smuzhiyun 		w = ((R_REG(sii->osh, &ai->ioctrl) & ~mask) | val);
1275*4882a593Smuzhiyun 		W_REG(sii->osh, &ai->ioctrl, w);
1276*4882a593Smuzhiyun 	}
1277*4882a593Smuzhiyun }
1278*4882a593Smuzhiyun 
1279*4882a593Smuzhiyun uint32
ai_core_cflags(si_t * sih,uint32 mask,uint32 val)1280*4882a593Smuzhiyun ai_core_cflags(si_t *sih, uint32 mask, uint32 val)
1281*4882a593Smuzhiyun {
1282*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
1283*4882a593Smuzhiyun 	aidmp_t *ai;
1284*4882a593Smuzhiyun 	uint32 w;
1285*4882a593Smuzhiyun 
1286*4882a593Smuzhiyun 	if (BCM4707_DMP()) {
1287*4882a593Smuzhiyun 		SI_ERROR(("%s: Accessing CHIPCOMMONB DMP register (ioctrl) on 4707\n",
1288*4882a593Smuzhiyun 			__FUNCTION__));
1289*4882a593Smuzhiyun 		return 0;
1290*4882a593Smuzhiyun 	}
1291*4882a593Smuzhiyun 
1292*4882a593Smuzhiyun 	if (PMU_DMP()) {
1293*4882a593Smuzhiyun 		SI_ERROR(("%s: Accessing PMU DMP register (ioctrl)\n",
1294*4882a593Smuzhiyun 			__FUNCTION__));
1295*4882a593Smuzhiyun 		return 0;
1296*4882a593Smuzhiyun 	}
1297*4882a593Smuzhiyun 	ASSERT(GOODREGS(sii->curwrap));
1298*4882a593Smuzhiyun 	ai = sii->curwrap;
1299*4882a593Smuzhiyun 
1300*4882a593Smuzhiyun 	ASSERT((val & ~mask) == 0);
1301*4882a593Smuzhiyun 
1302*4882a593Smuzhiyun 	if (mask || val) {
1303*4882a593Smuzhiyun 		w = ((R_REG(sii->osh, &ai->ioctrl) & ~mask) | val);
1304*4882a593Smuzhiyun 		W_REG(sii->osh, &ai->ioctrl, w);
1305*4882a593Smuzhiyun 	}
1306*4882a593Smuzhiyun 
1307*4882a593Smuzhiyun 	return R_REG(sii->osh, &ai->ioctrl);
1308*4882a593Smuzhiyun }
1309*4882a593Smuzhiyun 
1310*4882a593Smuzhiyun uint32
ai_core_sflags(si_t * sih,uint32 mask,uint32 val)1311*4882a593Smuzhiyun ai_core_sflags(si_t *sih, uint32 mask, uint32 val)
1312*4882a593Smuzhiyun {
1313*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
1314*4882a593Smuzhiyun 	aidmp_t *ai;
1315*4882a593Smuzhiyun 	uint32 w;
1316*4882a593Smuzhiyun 
1317*4882a593Smuzhiyun 	if (BCM4707_DMP()) {
1318*4882a593Smuzhiyun 		SI_ERROR(("%s: Accessing CHIPCOMMONB DMP register (ioctrl) on 4707\n",
1319*4882a593Smuzhiyun 			__FUNCTION__));
1320*4882a593Smuzhiyun 		return 0;
1321*4882a593Smuzhiyun 	}
1322*4882a593Smuzhiyun 	if (PMU_DMP()) {
1323*4882a593Smuzhiyun 		SI_ERROR(("%s: Accessing PMU DMP register (ioctrl)\n",
1324*4882a593Smuzhiyun 			__FUNCTION__));
1325*4882a593Smuzhiyun 		return 0;
1326*4882a593Smuzhiyun 	}
1327*4882a593Smuzhiyun 
1328*4882a593Smuzhiyun 	ASSERT(GOODREGS(sii->curwrap));
1329*4882a593Smuzhiyun 	ai = sii->curwrap;
1330*4882a593Smuzhiyun 
1331*4882a593Smuzhiyun 	ASSERT((val & ~mask) == 0);
1332*4882a593Smuzhiyun 	ASSERT((mask & ~SISF_CORE_BITS) == 0);
1333*4882a593Smuzhiyun 
1334*4882a593Smuzhiyun 	if (mask || val) {
1335*4882a593Smuzhiyun 		w = ((R_REG(sii->osh, &ai->iostatus) & ~mask) | val);
1336*4882a593Smuzhiyun 		W_REG(sii->osh, &ai->iostatus, w);
1337*4882a593Smuzhiyun 	}
1338*4882a593Smuzhiyun 
1339*4882a593Smuzhiyun 	return R_REG(sii->osh, &ai->iostatus);
1340*4882a593Smuzhiyun }
1341*4882a593Smuzhiyun 
1342*4882a593Smuzhiyun #if defined(BCMDBG_PHYDUMP)
1343*4882a593Smuzhiyun /* print interesting aidmp registers */
1344*4882a593Smuzhiyun void
ai_dumpregs(si_t * sih,struct bcmstrbuf * b)1345*4882a593Smuzhiyun ai_dumpregs(si_t *sih, struct bcmstrbuf *b)
1346*4882a593Smuzhiyun {
1347*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
1348*4882a593Smuzhiyun 	osl_t *osh;
1349*4882a593Smuzhiyun 	aidmp_t *ai;
1350*4882a593Smuzhiyun 	uint i;
1351*4882a593Smuzhiyun 	uint32 prev_value = 0;
1352*4882a593Smuzhiyun 	axi_wrapper_t * axi_wrapper = sii->axi_wrapper;
1353*4882a593Smuzhiyun 	uint32 cfg_reg = 0;
1354*4882a593Smuzhiyun 	uint bar0_win_offset = 0;
1355*4882a593Smuzhiyun 
1356*4882a593Smuzhiyun 	osh = sii->osh;
1357*4882a593Smuzhiyun 
1358*4882a593Smuzhiyun 	/* Save and restore wrapper access window */
1359*4882a593Smuzhiyun 	if (BUSTYPE(sii->pub.bustype) == PCI_BUS) {
1360*4882a593Smuzhiyun 		if (PCIE_GEN2(sii)) {
1361*4882a593Smuzhiyun 			cfg_reg = PCIE2_BAR0_CORE2_WIN2;
1362*4882a593Smuzhiyun 			bar0_win_offset = PCIE2_BAR0_CORE2_WIN2_OFFSET;
1363*4882a593Smuzhiyun 		} else {
1364*4882a593Smuzhiyun 			cfg_reg = PCI_BAR0_WIN2;
1365*4882a593Smuzhiyun 			bar0_win_offset = PCI_BAR0_WIN2_OFFSET;
1366*4882a593Smuzhiyun 		}
1367*4882a593Smuzhiyun 
1368*4882a593Smuzhiyun 		prev_value = OSL_PCI_READ_CONFIG(osh, cfg_reg, 4);
1369*4882a593Smuzhiyun 
1370*4882a593Smuzhiyun 		if (prev_value == ID32_INVALID) {
1371*4882a593Smuzhiyun 			SI_PRINT(("%s, PCI_BAR0_WIN2 - %x\n", __FUNCTION__, prev_value));
1372*4882a593Smuzhiyun 			return;
1373*4882a593Smuzhiyun 		}
1374*4882a593Smuzhiyun 	}
1375*4882a593Smuzhiyun 
1376*4882a593Smuzhiyun 	bcm_bprintf(b, "ChipNum:%x, ChipRev;%x, BusType:%x, BoardType:%x, BoardVendor:%x\n\n",
1377*4882a593Smuzhiyun 		sih->chip, sih->chiprev, sih->bustype, sih->boardtype, sih->boardvendor);
1378*4882a593Smuzhiyun 
1379*4882a593Smuzhiyun 	for (i = 0; i < sii->axi_num_wrappers; i++) {
1380*4882a593Smuzhiyun 
1381*4882a593Smuzhiyun 		if (BUSTYPE(sii->pub.bustype) == PCI_BUS) {
1382*4882a593Smuzhiyun 			/* Set BAR0 window to bridge wapper base address */
1383*4882a593Smuzhiyun 			OSL_PCI_WRITE_CONFIG(osh,
1384*4882a593Smuzhiyun 				cfg_reg, 4, axi_wrapper[i].wrapper_addr);
1385*4882a593Smuzhiyun 
1386*4882a593Smuzhiyun 			ai = (aidmp_t *) ((volatile uint8*)sii->curmap + bar0_win_offset);
1387*4882a593Smuzhiyun 		} else {
1388*4882a593Smuzhiyun 			ai = (aidmp_t *)(uintptr) axi_wrapper[i].wrapper_addr;
1389*4882a593Smuzhiyun 		}
1390*4882a593Smuzhiyun 
1391*4882a593Smuzhiyun 		bcm_bprintf(b, "core 0x%x: core_rev:%d, %s_WR ADDR:%x \n", axi_wrapper[i].cid,
1392*4882a593Smuzhiyun 			axi_wrapper[i].rev,
1393*4882a593Smuzhiyun 			axi_wrapper[i].wrapper_type == AI_SLAVE_WRAPPER ? "SLAVE" : "MASTER",
1394*4882a593Smuzhiyun 			axi_wrapper[i].wrapper_addr);
1395*4882a593Smuzhiyun 
1396*4882a593Smuzhiyun 		/* BCM4707_DMP() */
1397*4882a593Smuzhiyun 		if (BCM4707_CHIP(CHIPID(sih->chip)) &&
1398*4882a593Smuzhiyun 			(axi_wrapper[i].cid == NS_CCB_CORE_ID)) {
1399*4882a593Smuzhiyun 			bcm_bprintf(b, "Skipping chipcommonb in 4707\n");
1400*4882a593Smuzhiyun 			continue;
1401*4882a593Smuzhiyun 		}
1402*4882a593Smuzhiyun 
1403*4882a593Smuzhiyun 		bcm_bprintf(b, "ioctrlset 0x%x ioctrlclear 0x%x ioctrl 0x%x iostatus 0x%x "
1404*4882a593Smuzhiyun 			    "ioctrlwidth 0x%x iostatuswidth 0x%x\n"
1405*4882a593Smuzhiyun 			    "resetctrl 0x%x resetstatus 0x%x resetreadid 0x%x resetwriteid 0x%x\n"
1406*4882a593Smuzhiyun 			    "errlogctrl 0x%x errlogdone 0x%x errlogstatus 0x%x "
1407*4882a593Smuzhiyun 			    "errlogaddrlo 0x%x errlogaddrhi 0x%x\n"
1408*4882a593Smuzhiyun 			    "errlogid 0x%x errloguser 0x%x errlogflags 0x%x\n"
1409*4882a593Smuzhiyun 			    "intstatus 0x%x config 0x%x itcr 0x%x\n\n",
1410*4882a593Smuzhiyun 			    R_REG(osh, &ai->ioctrlset),
1411*4882a593Smuzhiyun 			    R_REG(osh, &ai->ioctrlclear),
1412*4882a593Smuzhiyun 			    R_REG(osh, &ai->ioctrl),
1413*4882a593Smuzhiyun 			    R_REG(osh, &ai->iostatus),
1414*4882a593Smuzhiyun 			    R_REG(osh, &ai->ioctrlwidth),
1415*4882a593Smuzhiyun 			    R_REG(osh, &ai->iostatuswidth),
1416*4882a593Smuzhiyun 			    R_REG(osh, &ai->resetctrl),
1417*4882a593Smuzhiyun 			    R_REG(osh, &ai->resetstatus),
1418*4882a593Smuzhiyun 			    R_REG(osh, &ai->resetreadid),
1419*4882a593Smuzhiyun 			    R_REG(osh, &ai->resetwriteid),
1420*4882a593Smuzhiyun 			    R_REG(osh, &ai->errlogctrl),
1421*4882a593Smuzhiyun 			    R_REG(osh, &ai->errlogdone),
1422*4882a593Smuzhiyun 			    R_REG(osh, &ai->errlogstatus),
1423*4882a593Smuzhiyun 			    R_REG(osh, &ai->errlogaddrlo),
1424*4882a593Smuzhiyun 			    R_REG(osh, &ai->errlogaddrhi),
1425*4882a593Smuzhiyun 			    R_REG(osh, &ai->errlogid),
1426*4882a593Smuzhiyun 			    R_REG(osh, &ai->errloguser),
1427*4882a593Smuzhiyun 			    R_REG(osh, &ai->errlogflags),
1428*4882a593Smuzhiyun 			    R_REG(osh, &ai->intstatus),
1429*4882a593Smuzhiyun 			    R_REG(osh, &ai->config),
1430*4882a593Smuzhiyun 			    R_REG(osh, &ai->itcr));
1431*4882a593Smuzhiyun 	}
1432*4882a593Smuzhiyun 
1433*4882a593Smuzhiyun 	/* Restore the initial wrapper space */
1434*4882a593Smuzhiyun 	if (BUSTYPE(sii->pub.bustype) == PCI_BUS) {
1435*4882a593Smuzhiyun 		if (prev_value && cfg_reg) {
1436*4882a593Smuzhiyun 			OSL_PCI_WRITE_CONFIG(osh, cfg_reg, 4, prev_value);
1437*4882a593Smuzhiyun 		}
1438*4882a593Smuzhiyun 	}
1439*4882a593Smuzhiyun }
1440*4882a593Smuzhiyun #endif // endif
1441*4882a593Smuzhiyun 
1442*4882a593Smuzhiyun void
ai_update_backplane_timeouts(si_t * sih,bool enable,uint32 timeout_exp,uint32 cid)1443*4882a593Smuzhiyun ai_update_backplane_timeouts(si_t *sih, bool enable, uint32 timeout_exp, uint32 cid)
1444*4882a593Smuzhiyun {
1445*4882a593Smuzhiyun #if defined(AXI_TIMEOUTS) || defined(BCM_BACKPLANE_TIMEOUT)
1446*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
1447*4882a593Smuzhiyun 	aidmp_t *ai;
1448*4882a593Smuzhiyun 	uint32 i;
1449*4882a593Smuzhiyun 	axi_wrapper_t * axi_wrapper = sii->axi_wrapper;
1450*4882a593Smuzhiyun 	uint32 errlogctrl = (enable << AIELC_TO_ENAB_SHIFT) |
1451*4882a593Smuzhiyun 		((timeout_exp << AIELC_TO_EXP_SHIFT) & AIELC_TO_EXP_MASK);
1452*4882a593Smuzhiyun 
1453*4882a593Smuzhiyun #ifdef BCM_BACKPLANE_TIMEOUT
1454*4882a593Smuzhiyun 	uint32 prev_value = 0;
1455*4882a593Smuzhiyun 	osl_t *osh = sii->osh;
1456*4882a593Smuzhiyun 	uint32 cfg_reg = 0;
1457*4882a593Smuzhiyun 	uint32 offset = 0;
1458*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
1459*4882a593Smuzhiyun 
1460*4882a593Smuzhiyun 	if ((sii->axi_num_wrappers == 0) ||
1461*4882a593Smuzhiyun #ifdef BCM_BACKPLANE_TIMEOUT
1462*4882a593Smuzhiyun 		(!PCIE(sii)) ||
1463*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
1464*4882a593Smuzhiyun 		FALSE) {
1465*4882a593Smuzhiyun 		SI_VMSG((" %s, axi_num_wrappers:%d, Is_PCIE:%d, BUS_TYPE:%d, ID:%x\n",
1466*4882a593Smuzhiyun 			__FUNCTION__, sii->axi_num_wrappers, PCIE(sii),
1467*4882a593Smuzhiyun 			BUSTYPE(sii->pub.bustype), sii->pub.buscoretype));
1468*4882a593Smuzhiyun 		return;
1469*4882a593Smuzhiyun 	}
1470*4882a593Smuzhiyun 
1471*4882a593Smuzhiyun #ifdef BCM_BACKPLANE_TIMEOUT
1472*4882a593Smuzhiyun 	/* Save and restore the wrapper access window */
1473*4882a593Smuzhiyun 	if (BUSTYPE(sii->pub.bustype) == PCI_BUS) {
1474*4882a593Smuzhiyun 		if (PCIE_GEN1(sii)) {
1475*4882a593Smuzhiyun 			cfg_reg = PCI_BAR0_WIN2;
1476*4882a593Smuzhiyun 			offset = PCI_BAR0_WIN2_OFFSET;
1477*4882a593Smuzhiyun 		} else if (PCIE_GEN2(sii)) {
1478*4882a593Smuzhiyun 			cfg_reg = PCIE2_BAR0_CORE2_WIN2;
1479*4882a593Smuzhiyun 			offset = PCIE2_BAR0_CORE2_WIN2_OFFSET;
1480*4882a593Smuzhiyun 		}
1481*4882a593Smuzhiyun 		else {
1482*4882a593Smuzhiyun 			ASSERT(!"!PCIE_GEN1 && !PCIE_GEN2");
1483*4882a593Smuzhiyun 		}
1484*4882a593Smuzhiyun 
1485*4882a593Smuzhiyun 		prev_value = OSL_PCI_READ_CONFIG(osh, cfg_reg, 4);
1486*4882a593Smuzhiyun 		if (prev_value == ID32_INVALID) {
1487*4882a593Smuzhiyun 			SI_PRINT(("%s, PCI_BAR0_WIN2 - %x\n", __FUNCTION__, prev_value));
1488*4882a593Smuzhiyun 			return;
1489*4882a593Smuzhiyun 		}
1490*4882a593Smuzhiyun 	}
1491*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
1492*4882a593Smuzhiyun 
1493*4882a593Smuzhiyun 	for (i = 0; i < sii->axi_num_wrappers; ++i) {
1494*4882a593Smuzhiyun 
1495*4882a593Smuzhiyun 		if (axi_wrapper[i].wrapper_type != AI_SLAVE_WRAPPER) {
1496*4882a593Smuzhiyun 			SI_VMSG(("SKIP ENABLE BPT: MFG:%x, CID:%x, ADDR:%x\n",
1497*4882a593Smuzhiyun 				axi_wrapper[i].mfg,
1498*4882a593Smuzhiyun 				axi_wrapper[i].cid,
1499*4882a593Smuzhiyun 				axi_wrapper[i].wrapper_addr));
1500*4882a593Smuzhiyun 			continue;
1501*4882a593Smuzhiyun 		}
1502*4882a593Smuzhiyun 
1503*4882a593Smuzhiyun 		/* Update only given core if requested */
1504*4882a593Smuzhiyun 		if ((cid != 0) && (axi_wrapper[i].cid != cid)) {
1505*4882a593Smuzhiyun 			continue;
1506*4882a593Smuzhiyun 		}
1507*4882a593Smuzhiyun 
1508*4882a593Smuzhiyun #ifdef BCM_BACKPLANE_TIMEOUT
1509*4882a593Smuzhiyun 		if (BUSTYPE(sii->pub.bustype) == PCI_BUS) {
1510*4882a593Smuzhiyun 			/* Set BAR0_CORE2_WIN2 to bridge wapper base address */
1511*4882a593Smuzhiyun 			OSL_PCI_WRITE_CONFIG(osh,
1512*4882a593Smuzhiyun 				cfg_reg, 4, axi_wrapper[i].wrapper_addr);
1513*4882a593Smuzhiyun 
1514*4882a593Smuzhiyun 			/* set AI to BAR0 + Offset corresponding to Gen1 or gen2 */
1515*4882a593Smuzhiyun 			ai = (aidmp_t *) (DISCARD_QUAL(sii->curmap, uint8) + offset);
1516*4882a593Smuzhiyun 		}
1517*4882a593Smuzhiyun 		else
1518*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
1519*4882a593Smuzhiyun 		{
1520*4882a593Smuzhiyun 			ai = (aidmp_t *)(uintptr) axi_wrapper[i].wrapper_addr;
1521*4882a593Smuzhiyun 		}
1522*4882a593Smuzhiyun 
1523*4882a593Smuzhiyun 		W_REG(sii->osh, &ai->errlogctrl, errlogctrl);
1524*4882a593Smuzhiyun 
1525*4882a593Smuzhiyun 		SI_VMSG(("ENABLED BPT: MFG:%x, CID:%x, ADDR:%x, ERR_CTRL:%x\n",
1526*4882a593Smuzhiyun 			axi_wrapper[i].mfg,
1527*4882a593Smuzhiyun 			axi_wrapper[i].cid,
1528*4882a593Smuzhiyun 			axi_wrapper[i].wrapper_addr,
1529*4882a593Smuzhiyun 			R_REG(sii->osh, &ai->errlogctrl)));
1530*4882a593Smuzhiyun 	}
1531*4882a593Smuzhiyun 
1532*4882a593Smuzhiyun #ifdef BCM_BACKPLANE_TIMEOUT
1533*4882a593Smuzhiyun 	/* Restore the initial wrapper space */
1534*4882a593Smuzhiyun 	if (prev_value) {
1535*4882a593Smuzhiyun 		OSL_PCI_WRITE_CONFIG(osh, cfg_reg, 4, prev_value);
1536*4882a593Smuzhiyun 	}
1537*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
1538*4882a593Smuzhiyun 
1539*4882a593Smuzhiyun #endif /* AXI_TIMEOUTS || BCM_BACKPLANE_TIMEOUT */
1540*4882a593Smuzhiyun }
1541*4882a593Smuzhiyun 
1542*4882a593Smuzhiyun #if defined(AXI_TIMEOUTS) || defined(BCM_BACKPLANE_TIMEOUT)
1543*4882a593Smuzhiyun 
1544*4882a593Smuzhiyun /* slave error is ignored, so account for those cases */
1545*4882a593Smuzhiyun static uint32 si_ignore_errlog_cnt = 0;
1546*4882a593Smuzhiyun 
1547*4882a593Smuzhiyun static bool
ai_ignore_errlog(si_info_t * sii,aidmp_t * ai,uint32 lo_addr,uint32 hi_addr,uint32 err_axi_id,uint32 errsts)1548*4882a593Smuzhiyun ai_ignore_errlog(si_info_t *sii, aidmp_t *ai,
1549*4882a593Smuzhiyun 	uint32 lo_addr, uint32 hi_addr, uint32 err_axi_id, uint32 errsts)
1550*4882a593Smuzhiyun {
1551*4882a593Smuzhiyun 	uint32 axi_id;
1552*4882a593Smuzhiyun #ifdef BCMPCIE_BTLOG
1553*4882a593Smuzhiyun 	uint32 axi_id2 = BCM4347_UNUSED_AXI_ID;
1554*4882a593Smuzhiyun #endif	/* BCMPCIE_BTLOG */
1555*4882a593Smuzhiyun 	uint32 ignore_errsts = AIELS_SLAVE_ERR;
1556*4882a593Smuzhiyun 	uint32 ignore_hi = BT_CC_SPROM_BADREG_HI;
1557*4882a593Smuzhiyun 	uint32 ignore_lo = BT_CC_SPROM_BADREG_LO;
1558*4882a593Smuzhiyun 	uint32 ignore_size = BT_CC_SPROM_BADREG_SIZE;
1559*4882a593Smuzhiyun 
1560*4882a593Smuzhiyun 	/* ignore the BT slave errors if the errlog is to chipcommon addr 0x190 */
1561*4882a593Smuzhiyun 	switch (CHIPID(sii->pub.chip)) {
1562*4882a593Smuzhiyun 		case BCM4350_CHIP_ID:
1563*4882a593Smuzhiyun 			axi_id = BCM4350_BT_AXI_ID;
1564*4882a593Smuzhiyun 			break;
1565*4882a593Smuzhiyun 		case BCM4345_CHIP_ID:
1566*4882a593Smuzhiyun 			axi_id = BCM4345_BT_AXI_ID;
1567*4882a593Smuzhiyun 			break;
1568*4882a593Smuzhiyun 		case BCM4349_CHIP_GRPID:
1569*4882a593Smuzhiyun 			axi_id = BCM4349_BT_AXI_ID;
1570*4882a593Smuzhiyun 			break;
1571*4882a593Smuzhiyun 		case BCM4364_CHIP_ID:
1572*4882a593Smuzhiyun 		case BCM4373_CHIP_ID:
1573*4882a593Smuzhiyun 			axi_id = BCM4364_BT_AXI_ID;
1574*4882a593Smuzhiyun 			break;
1575*4882a593Smuzhiyun #ifdef BCMPCIE_BTLOG
1576*4882a593Smuzhiyun 		case BCM4347_CHIP_ID:
1577*4882a593Smuzhiyun 		case BCM4357_CHIP_ID:
1578*4882a593Smuzhiyun 			axi_id = BCM4347_CC_AXI_ID;
1579*4882a593Smuzhiyun 			axi_id2 = BCM4347_PCIE_AXI_ID;
1580*4882a593Smuzhiyun 			ignore_errsts = AIELS_TIMEOUT;
1581*4882a593Smuzhiyun 			ignore_hi = BCM4347_BT_ADDR_HI;
1582*4882a593Smuzhiyun 			ignore_lo = BCM4347_BT_ADDR_LO;
1583*4882a593Smuzhiyun 			ignore_size = BCM4347_BT_SIZE;
1584*4882a593Smuzhiyun 			break;
1585*4882a593Smuzhiyun #endif	/* BCMPCIE_BTLOG */
1586*4882a593Smuzhiyun 
1587*4882a593Smuzhiyun 		default:
1588*4882a593Smuzhiyun 			return FALSE;
1589*4882a593Smuzhiyun 	}
1590*4882a593Smuzhiyun 
1591*4882a593Smuzhiyun 	/* AXI ID check */
1592*4882a593Smuzhiyun 	err_axi_id &= AI_ERRLOGID_AXI_ID_MASK;
1593*4882a593Smuzhiyun 	if (!(err_axi_id == axi_id ||
1594*4882a593Smuzhiyun #ifdef BCMPCIE_BTLOG
1595*4882a593Smuzhiyun 	      (axi_id2 != BCM4347_UNUSED_AXI_ID && err_axi_id == axi_id2)))
1596*4882a593Smuzhiyun #else
1597*4882a593Smuzhiyun 	      FALSE))
1598*4882a593Smuzhiyun #endif	/* BCMPCIE_BTLOG */
1599*4882a593Smuzhiyun 		return FALSE;
1600*4882a593Smuzhiyun 
1601*4882a593Smuzhiyun 	/* slave errors */
1602*4882a593Smuzhiyun 	if ((errsts & AIELS_TIMEOUT_MASK) != ignore_errsts)
1603*4882a593Smuzhiyun 		return FALSE;
1604*4882a593Smuzhiyun 
1605*4882a593Smuzhiyun 	/* address range check */
1606*4882a593Smuzhiyun 	if ((hi_addr != ignore_hi) ||
1607*4882a593Smuzhiyun 	    (lo_addr < ignore_lo) || (lo_addr >= (ignore_lo + ignore_size)))
1608*4882a593Smuzhiyun 		return FALSE;
1609*4882a593Smuzhiyun 
1610*4882a593Smuzhiyun #ifdef BCMPCIE_BTLOG
1611*4882a593Smuzhiyun 	if (ignore_errsts == AIELS_TIMEOUT) {
1612*4882a593Smuzhiyun 		/* reset AXI timeout */
1613*4882a593Smuzhiyun 		ai_reset_axi_to(sii, ai);
1614*4882a593Smuzhiyun 	}
1615*4882a593Smuzhiyun #endif	/* BCMPCIE_BTLOG */
1616*4882a593Smuzhiyun 
1617*4882a593Smuzhiyun 	return TRUE;
1618*4882a593Smuzhiyun }
1619*4882a593Smuzhiyun #endif /* defined (AXI_TIMEOUTS) || defined (BCM_BACKPLANE_TIMEOUT) */
1620*4882a593Smuzhiyun 
1621*4882a593Smuzhiyun #ifdef BCM_BACKPLANE_TIMEOUT
1622*4882a593Smuzhiyun 
1623*4882a593Smuzhiyun /* Function to return the APB bridge details corresponding to the core */
1624*4882a593Smuzhiyun static bool
ai_get_apb_bridge(si_t * sih,uint32 coreidx,uint32 * apb_id,uint32 * apb_coreuinit)1625*4882a593Smuzhiyun ai_get_apb_bridge(si_t * sih, uint32 coreidx, uint32 *apb_id, uint32 * apb_coreuinit)
1626*4882a593Smuzhiyun {
1627*4882a593Smuzhiyun 	uint i;
1628*4882a593Smuzhiyun 	uint32 core_base, core_end;
1629*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
1630*4882a593Smuzhiyun 	static uint32 coreidx_cached = 0, apb_id_cached = 0, apb_coreunit_cached = 0;
1631*4882a593Smuzhiyun 	uint32 tmp_coreunit = 0;
1632*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
1633*4882a593Smuzhiyun 
1634*4882a593Smuzhiyun 	if (coreidx >= MIN(sii->numcores, SI_MAXCORES))
1635*4882a593Smuzhiyun 		return FALSE;
1636*4882a593Smuzhiyun 
1637*4882a593Smuzhiyun 	/* Most of the time apb bridge query will be for d11 core.
1638*4882a593Smuzhiyun 	 * Maintain the last cache and return if found rather than iterating the table
1639*4882a593Smuzhiyun 	 */
1640*4882a593Smuzhiyun 	if (coreidx_cached == coreidx) {
1641*4882a593Smuzhiyun 		*apb_id = apb_id_cached;
1642*4882a593Smuzhiyun 		*apb_coreuinit = apb_coreunit_cached;
1643*4882a593Smuzhiyun 		return TRUE;
1644*4882a593Smuzhiyun 	}
1645*4882a593Smuzhiyun 
1646*4882a593Smuzhiyun 	core_base = cores_info->coresba[coreidx];
1647*4882a593Smuzhiyun 	core_end = core_base + cores_info->coresba_size[coreidx];
1648*4882a593Smuzhiyun 
1649*4882a593Smuzhiyun 	for (i = 0; i < sii->numcores; i++) {
1650*4882a593Smuzhiyun 		if (cores_info->coreid[i] == APB_BRIDGE_ID) {
1651*4882a593Smuzhiyun 			uint32 apb_base;
1652*4882a593Smuzhiyun 			uint32 apb_end;
1653*4882a593Smuzhiyun 
1654*4882a593Smuzhiyun 			apb_base = cores_info->coresba[i];
1655*4882a593Smuzhiyun 			apb_end = apb_base + cores_info->coresba_size[i];
1656*4882a593Smuzhiyun 
1657*4882a593Smuzhiyun 			if ((core_base >= apb_base) &&
1658*4882a593Smuzhiyun 				(core_end <= apb_end)) {
1659*4882a593Smuzhiyun 				/* Current core is attached to this APB bridge */
1660*4882a593Smuzhiyun 				*apb_id = apb_id_cached = APB_BRIDGE_ID;
1661*4882a593Smuzhiyun 				*apb_coreuinit = apb_coreunit_cached = tmp_coreunit;
1662*4882a593Smuzhiyun 				coreidx_cached = coreidx;
1663*4882a593Smuzhiyun 				return TRUE;
1664*4882a593Smuzhiyun 			}
1665*4882a593Smuzhiyun 			/* Increment the coreunit */
1666*4882a593Smuzhiyun 			tmp_coreunit++;
1667*4882a593Smuzhiyun 		}
1668*4882a593Smuzhiyun 	}
1669*4882a593Smuzhiyun 
1670*4882a593Smuzhiyun 	return FALSE;
1671*4882a593Smuzhiyun }
1672*4882a593Smuzhiyun 
1673*4882a593Smuzhiyun uint32
ai_clear_backplane_to_fast(si_t * sih,void * addr)1674*4882a593Smuzhiyun ai_clear_backplane_to_fast(si_t *sih, void *addr)
1675*4882a593Smuzhiyun {
1676*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
1677*4882a593Smuzhiyun 	volatile void *curmap = sii->curmap;
1678*4882a593Smuzhiyun 	bool core_reg = FALSE;
1679*4882a593Smuzhiyun 
1680*4882a593Smuzhiyun 	/* Use fast path only for core register access */
1681*4882a593Smuzhiyun 	if (((uintptr)addr >= (uintptr)curmap) &&
1682*4882a593Smuzhiyun 		((uintptr)addr < ((uintptr)curmap + SI_CORE_SIZE))) {
1683*4882a593Smuzhiyun 		/* address being accessed is within current core reg map */
1684*4882a593Smuzhiyun 		core_reg = TRUE;
1685*4882a593Smuzhiyun 	}
1686*4882a593Smuzhiyun 
1687*4882a593Smuzhiyun 	if (core_reg) {
1688*4882a593Smuzhiyun 		uint32 apb_id, apb_coreuinit;
1689*4882a593Smuzhiyun 
1690*4882a593Smuzhiyun 		if (ai_get_apb_bridge(sih, si_coreidx(&sii->pub),
1691*4882a593Smuzhiyun 			&apb_id, &apb_coreuinit) == TRUE) {
1692*4882a593Smuzhiyun 			/* Found the APB bridge corresponding to current core,
1693*4882a593Smuzhiyun 			 * Check for bus errors in APB wrapper
1694*4882a593Smuzhiyun 			 */
1695*4882a593Smuzhiyun 			return ai_clear_backplane_to_per_core(sih,
1696*4882a593Smuzhiyun 				apb_id, apb_coreuinit, NULL);
1697*4882a593Smuzhiyun 		}
1698*4882a593Smuzhiyun 	}
1699*4882a593Smuzhiyun 
1700*4882a593Smuzhiyun 	/* Default is to poll for errors on all slave wrappers */
1701*4882a593Smuzhiyun 	return si_clear_backplane_to(sih);
1702*4882a593Smuzhiyun }
1703*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
1704*4882a593Smuzhiyun 
1705*4882a593Smuzhiyun #if defined(AXI_TIMEOUTS) || defined(BCM_BACKPLANE_TIMEOUT)
1706*4882a593Smuzhiyun static bool g_disable_backplane_logs = FALSE;
1707*4882a593Smuzhiyun 
1708*4882a593Smuzhiyun #if defined(ETD)
1709*4882a593Smuzhiyun static uint32 last_axi_error = AXI_WRAP_STS_NONE;
1710*4882a593Smuzhiyun static uint32 last_axi_error_core = 0;
1711*4882a593Smuzhiyun static uint32 last_axi_error_wrap = 0;
1712*4882a593Smuzhiyun #endif /* ETD */
1713*4882a593Smuzhiyun 
1714*4882a593Smuzhiyun /*
1715*4882a593Smuzhiyun  * API to clear the back plane timeout per core.
1716*4882a593Smuzhiyun  * Caller may passs optional wrapper address. If present this will be used as
1717*4882a593Smuzhiyun  * the wrapper base address. If wrapper base address is provided then caller
1718*4882a593Smuzhiyun  * must provide the coreid also.
1719*4882a593Smuzhiyun  * If both coreid and wrapper is zero, then err status of current bridge
1720*4882a593Smuzhiyun  * will be verified.
1721*4882a593Smuzhiyun  */
1722*4882a593Smuzhiyun uint32
ai_clear_backplane_to_per_core(si_t * sih,uint coreid,uint coreunit,void * wrap)1723*4882a593Smuzhiyun ai_clear_backplane_to_per_core(si_t *sih, uint coreid, uint coreunit, void *wrap)
1724*4882a593Smuzhiyun {
1725*4882a593Smuzhiyun 	int ret = AXI_WRAP_STS_NONE;
1726*4882a593Smuzhiyun 	aidmp_t *ai = NULL;
1727*4882a593Smuzhiyun 	uint32 errlog_status = 0;
1728*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
1729*4882a593Smuzhiyun 	uint32 errlog_lo = 0, errlog_hi = 0, errlog_id = 0, errlog_flags = 0;
1730*4882a593Smuzhiyun 	uint32 current_coreidx = si_coreidx(sih);
1731*4882a593Smuzhiyun 	uint32 target_coreidx = si_findcoreidx(sih, coreid, coreunit);
1732*4882a593Smuzhiyun 
1733*4882a593Smuzhiyun #if defined(BCM_BACKPLANE_TIMEOUT)
1734*4882a593Smuzhiyun 	si_axi_error_t * axi_error = sih->err_info ?
1735*4882a593Smuzhiyun 		&sih->err_info->axi_error[sih->err_info->count] : NULL;
1736*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
1737*4882a593Smuzhiyun 	bool restore_core = FALSE;
1738*4882a593Smuzhiyun 
1739*4882a593Smuzhiyun 	if ((sii->axi_num_wrappers == 0) ||
1740*4882a593Smuzhiyun #ifdef BCM_BACKPLANE_TIMEOUT
1741*4882a593Smuzhiyun 		(!PCIE(sii)) ||
1742*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
1743*4882a593Smuzhiyun 		FALSE) {
1744*4882a593Smuzhiyun 		SI_VMSG((" %s, axi_num_wrappers:%d, Is_PCIE:%d, BUS_TYPE:%d, ID:%x\n",
1745*4882a593Smuzhiyun 			__FUNCTION__, sii->axi_num_wrappers, PCIE(sii),
1746*4882a593Smuzhiyun 			BUSTYPE(sii->pub.bustype), sii->pub.buscoretype));
1747*4882a593Smuzhiyun 		return AXI_WRAP_STS_NONE;
1748*4882a593Smuzhiyun 	}
1749*4882a593Smuzhiyun 
1750*4882a593Smuzhiyun 	if (wrap != NULL) {
1751*4882a593Smuzhiyun 		ai = (aidmp_t *)wrap;
1752*4882a593Smuzhiyun 	} else if (coreid && (target_coreidx != current_coreidx)) {
1753*4882a593Smuzhiyun 
1754*4882a593Smuzhiyun 		if (ai_setcoreidx(sih, target_coreidx) == NULL) {
1755*4882a593Smuzhiyun 			/* Unable to set the core */
1756*4882a593Smuzhiyun 			SI_PRINT(("Set Code Failed: coreid:%x, unit:%d, target_coreidx:%d\n",
1757*4882a593Smuzhiyun 				coreid, coreunit, target_coreidx));
1758*4882a593Smuzhiyun 			errlog_lo = target_coreidx;
1759*4882a593Smuzhiyun 			ret = AXI_WRAP_STS_SET_CORE_FAIL;
1760*4882a593Smuzhiyun 			goto end;
1761*4882a593Smuzhiyun 		}
1762*4882a593Smuzhiyun 
1763*4882a593Smuzhiyun 		restore_core = TRUE;
1764*4882a593Smuzhiyun 		ai = (aidmp_t *)si_wrapperregs(sih);
1765*4882a593Smuzhiyun 	} else {
1766*4882a593Smuzhiyun 		/* Read error status of current wrapper */
1767*4882a593Smuzhiyun 		ai = (aidmp_t *)si_wrapperregs(sih);
1768*4882a593Smuzhiyun 
1769*4882a593Smuzhiyun 		/* Update CoreID to current Code ID */
1770*4882a593Smuzhiyun 		coreid = si_coreid(sih);
1771*4882a593Smuzhiyun 	}
1772*4882a593Smuzhiyun 
1773*4882a593Smuzhiyun 	/* read error log status */
1774*4882a593Smuzhiyun 	errlog_status = R_REG(sii->osh, &ai->errlogstatus);
1775*4882a593Smuzhiyun 
1776*4882a593Smuzhiyun 	if (errlog_status == ID32_INVALID) {
1777*4882a593Smuzhiyun 		/* Do not try to peek further */
1778*4882a593Smuzhiyun 		SI_PRINT(("%s, errlogstatus:%x - Slave Wrapper:%x\n",
1779*4882a593Smuzhiyun 			__FUNCTION__, errlog_status, coreid));
1780*4882a593Smuzhiyun 		ret = AXI_WRAP_STS_WRAP_RD_ERR;
1781*4882a593Smuzhiyun 		errlog_lo = (uint32)(uintptr)&ai->errlogstatus;
1782*4882a593Smuzhiyun 		goto end;
1783*4882a593Smuzhiyun 	}
1784*4882a593Smuzhiyun 
1785*4882a593Smuzhiyun 	if ((errlog_status & AIELS_TIMEOUT_MASK) != 0) {
1786*4882a593Smuzhiyun 		uint32 tmp;
1787*4882a593Smuzhiyun 		uint32 count = 0;
1788*4882a593Smuzhiyun 		/* set ErrDone to clear the condition */
1789*4882a593Smuzhiyun 		W_REG(sii->osh, &ai->errlogdone, AIELD_ERRDONE_MASK);
1790*4882a593Smuzhiyun 
1791*4882a593Smuzhiyun 		/* SPINWAIT on errlogstatus timeout status bits */
1792*4882a593Smuzhiyun 		while ((tmp = R_REG(sii->osh, &ai->errlogstatus)) & AIELS_TIMEOUT_MASK) {
1793*4882a593Smuzhiyun 
1794*4882a593Smuzhiyun 			if (tmp == ID32_INVALID) {
1795*4882a593Smuzhiyun 				SI_PRINT(("%s: prev errlogstatus:%x, errlogstatus:%x\n",
1796*4882a593Smuzhiyun 					__FUNCTION__, errlog_status, tmp));
1797*4882a593Smuzhiyun 				ret = AXI_WRAP_STS_WRAP_RD_ERR;
1798*4882a593Smuzhiyun 				errlog_lo = (uint32)(uintptr)&ai->errlogstatus;
1799*4882a593Smuzhiyun 				goto end;
1800*4882a593Smuzhiyun 			}
1801*4882a593Smuzhiyun 			/*
1802*4882a593Smuzhiyun 			 * Clear again, to avoid getting stuck in the loop, if a new error
1803*4882a593Smuzhiyun 			 * is logged after we cleared the first timeout
1804*4882a593Smuzhiyun 			 */
1805*4882a593Smuzhiyun 			W_REG(sii->osh, &ai->errlogdone, AIELD_ERRDONE_MASK);
1806*4882a593Smuzhiyun 
1807*4882a593Smuzhiyun 			count++;
1808*4882a593Smuzhiyun 			OSL_DELAY(10);
1809*4882a593Smuzhiyun 			if ((10 * count) > AI_REG_READ_TIMEOUT) {
1810*4882a593Smuzhiyun 				errlog_status = tmp;
1811*4882a593Smuzhiyun 				break;
1812*4882a593Smuzhiyun 			}
1813*4882a593Smuzhiyun 		}
1814*4882a593Smuzhiyun 
1815*4882a593Smuzhiyun 		errlog_lo = R_REG(sii->osh, &ai->errlogaddrlo);
1816*4882a593Smuzhiyun 		errlog_hi = R_REG(sii->osh, &ai->errlogaddrhi);
1817*4882a593Smuzhiyun 		errlog_id = R_REG(sii->osh, &ai->errlogid);
1818*4882a593Smuzhiyun 		errlog_flags = R_REG(sii->osh, &ai->errlogflags);
1819*4882a593Smuzhiyun 
1820*4882a593Smuzhiyun 		/* we are already in the error path, so OK to check for the  slave error */
1821*4882a593Smuzhiyun 		if (ai_ignore_errlog(sii, ai, errlog_lo, errlog_hi, errlog_id,
1822*4882a593Smuzhiyun 			errlog_status)) {
1823*4882a593Smuzhiyun 			si_ignore_errlog_cnt++;
1824*4882a593Smuzhiyun 			goto end;
1825*4882a593Smuzhiyun 		}
1826*4882a593Smuzhiyun 
1827*4882a593Smuzhiyun 		/* only reset APB Bridge on timeout (not slave error, or dec error) */
1828*4882a593Smuzhiyun 		switch (errlog_status & AIELS_TIMEOUT_MASK) {
1829*4882a593Smuzhiyun 			case AIELS_SLAVE_ERR:
1830*4882a593Smuzhiyun 				SI_PRINT(("AXI slave error\n"));
1831*4882a593Smuzhiyun 				ret = AXI_WRAP_STS_SLAVE_ERR;
1832*4882a593Smuzhiyun 				break;
1833*4882a593Smuzhiyun 
1834*4882a593Smuzhiyun 			case AIELS_TIMEOUT:
1835*4882a593Smuzhiyun 				ai_reset_axi_to(sii, ai);
1836*4882a593Smuzhiyun 				ret = AXI_WRAP_STS_TIMEOUT;
1837*4882a593Smuzhiyun 				break;
1838*4882a593Smuzhiyun 
1839*4882a593Smuzhiyun 			case AIELS_DECODE:
1840*4882a593Smuzhiyun 				SI_PRINT(("AXI decode error\n"));
1841*4882a593Smuzhiyun 				ret = AXI_WRAP_STS_DECODE_ERR;
1842*4882a593Smuzhiyun 				break;
1843*4882a593Smuzhiyun 			default:
1844*4882a593Smuzhiyun 				ASSERT(0);	/* should be impossible */
1845*4882a593Smuzhiyun 		}
1846*4882a593Smuzhiyun 
1847*4882a593Smuzhiyun 		SI_PRINT(("\tCoreID: %x\n", coreid));
1848*4882a593Smuzhiyun 		SI_PRINT(("\t errlog: lo 0x%08x, hi 0x%08x, id 0x%08x, flags 0x%08x"
1849*4882a593Smuzhiyun 			", status 0x%08x\n",
1850*4882a593Smuzhiyun 			errlog_lo, errlog_hi, errlog_id, errlog_flags,
1851*4882a593Smuzhiyun 			errlog_status));
1852*4882a593Smuzhiyun 	}
1853*4882a593Smuzhiyun 
1854*4882a593Smuzhiyun end:
1855*4882a593Smuzhiyun #if defined(ETD)
1856*4882a593Smuzhiyun 	if (ret != AXI_WRAP_STS_NONE) {
1857*4882a593Smuzhiyun 		last_axi_error = ret;
1858*4882a593Smuzhiyun 		last_axi_error_core = coreid;
1859*4882a593Smuzhiyun 		last_axi_error_wrap = (uint32)ai;
1860*4882a593Smuzhiyun 	}
1861*4882a593Smuzhiyun #endif /* ETD */
1862*4882a593Smuzhiyun 
1863*4882a593Smuzhiyun #if defined(BCM_BACKPLANE_TIMEOUT)
1864*4882a593Smuzhiyun 	if (axi_error && (ret != AXI_WRAP_STS_NONE)) {
1865*4882a593Smuzhiyun 		axi_error->error = ret;
1866*4882a593Smuzhiyun 		axi_error->coreid = coreid;
1867*4882a593Smuzhiyun 		axi_error->errlog_lo = errlog_lo;
1868*4882a593Smuzhiyun 		axi_error->errlog_hi = errlog_hi;
1869*4882a593Smuzhiyun 		axi_error->errlog_id = errlog_id;
1870*4882a593Smuzhiyun 		axi_error->errlog_flags = errlog_flags;
1871*4882a593Smuzhiyun 		axi_error->errlog_status = errlog_status;
1872*4882a593Smuzhiyun 		sih->err_info->count++;
1873*4882a593Smuzhiyun 
1874*4882a593Smuzhiyun 		if (sih->err_info->count == SI_MAX_ERRLOG_SIZE) {
1875*4882a593Smuzhiyun 			sih->err_info->count = SI_MAX_ERRLOG_SIZE - 1;
1876*4882a593Smuzhiyun 			SI_PRINT(("AXI Error log overflow\n"));
1877*4882a593Smuzhiyun 		}
1878*4882a593Smuzhiyun 	}
1879*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
1880*4882a593Smuzhiyun 
1881*4882a593Smuzhiyun 	if (restore_core) {
1882*4882a593Smuzhiyun 		if (ai_setcoreidx(sih, current_coreidx) == NULL) {
1883*4882a593Smuzhiyun 			/* Unable to set the core */
1884*4882a593Smuzhiyun 			return ID32_INVALID;
1885*4882a593Smuzhiyun 		}
1886*4882a593Smuzhiyun 	}
1887*4882a593Smuzhiyun 
1888*4882a593Smuzhiyun 	return ret;
1889*4882a593Smuzhiyun }
1890*4882a593Smuzhiyun 
1891*4882a593Smuzhiyun /* reset AXI timeout */
1892*4882a593Smuzhiyun static void
ai_reset_axi_to(si_info_t * sii,aidmp_t * ai)1893*4882a593Smuzhiyun ai_reset_axi_to(si_info_t *sii, aidmp_t *ai)
1894*4882a593Smuzhiyun {
1895*4882a593Smuzhiyun 	/* reset APB Bridge */
1896*4882a593Smuzhiyun 	OR_REG(sii->osh, &ai->resetctrl, AIRC_RESET);
1897*4882a593Smuzhiyun 	/* sync write */
1898*4882a593Smuzhiyun 	(void)R_REG(sii->osh, &ai->resetctrl);
1899*4882a593Smuzhiyun 	/* clear Reset bit */
1900*4882a593Smuzhiyun 	AND_REG(sii->osh, &ai->resetctrl, ~(AIRC_RESET));
1901*4882a593Smuzhiyun 	/* sync write */
1902*4882a593Smuzhiyun 	(void)R_REG(sii->osh, &ai->resetctrl);
1903*4882a593Smuzhiyun 	SI_PRINT(("AXI timeout\n"));
1904*4882a593Smuzhiyun 	if (R_REG(sii->osh, &ai->resetctrl) & AIRC_RESET) {
1905*4882a593Smuzhiyun 		SI_PRINT(("reset failed on wrapper %p\n", ai));
1906*4882a593Smuzhiyun 		g_disable_backplane_logs = TRUE;
1907*4882a593Smuzhiyun 	}
1908*4882a593Smuzhiyun }
1909*4882a593Smuzhiyun #endif /* AXI_TIMEOUTS || BCM_BACKPLANE_TIMEOUT */
1910*4882a593Smuzhiyun 
1911*4882a593Smuzhiyun /*
1912*4882a593Smuzhiyun  * This API polls all slave wrappers for errors and returns bit map of
1913*4882a593Smuzhiyun  * all reported errors.
1914*4882a593Smuzhiyun  * return - bit map of
1915*4882a593Smuzhiyun  *	AXI_WRAP_STS_NONE
1916*4882a593Smuzhiyun  *	AXI_WRAP_STS_TIMEOUT
1917*4882a593Smuzhiyun  *	AXI_WRAP_STS_SLAVE_ERR
1918*4882a593Smuzhiyun  *	AXI_WRAP_STS_DECODE_ERR
1919*4882a593Smuzhiyun  *	AXI_WRAP_STS_PCI_RD_ERR
1920*4882a593Smuzhiyun  *	AXI_WRAP_STS_WRAP_RD_ERR
1921*4882a593Smuzhiyun  *	AXI_WRAP_STS_SET_CORE_FAIL
1922*4882a593Smuzhiyun  * On timeout detection, correspondign bridge will be reset to
1923*4882a593Smuzhiyun  * unblock the bus.
1924*4882a593Smuzhiyun  * Error reported in each wrapper can be retrieved using the API
1925*4882a593Smuzhiyun  * si_get_axi_errlog_info()
1926*4882a593Smuzhiyun  */
1927*4882a593Smuzhiyun uint32
ai_clear_backplane_to(si_t * sih)1928*4882a593Smuzhiyun ai_clear_backplane_to(si_t *sih)
1929*4882a593Smuzhiyun {
1930*4882a593Smuzhiyun 	uint32 ret = 0;
1931*4882a593Smuzhiyun #if defined(AXI_TIMEOUTS) || defined(BCM_BACKPLANE_TIMEOUT)
1932*4882a593Smuzhiyun 
1933*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
1934*4882a593Smuzhiyun 	aidmp_t *ai;
1935*4882a593Smuzhiyun 	uint32 i;
1936*4882a593Smuzhiyun 	axi_wrapper_t * axi_wrapper = sii->axi_wrapper;
1937*4882a593Smuzhiyun 
1938*4882a593Smuzhiyun #ifdef BCM_BACKPLANE_TIMEOUT
1939*4882a593Smuzhiyun 	uint32 prev_value = 0;
1940*4882a593Smuzhiyun 	osl_t *osh = sii->osh;
1941*4882a593Smuzhiyun 	uint32 cfg_reg = 0;
1942*4882a593Smuzhiyun 	uint32 offset = 0;
1943*4882a593Smuzhiyun 
1944*4882a593Smuzhiyun 	if ((sii->axi_num_wrappers == 0) || (!PCIE(sii)))
1945*4882a593Smuzhiyun #else
1946*4882a593Smuzhiyun 	if (sii->axi_num_wrappers == 0)
1947*4882a593Smuzhiyun #endif // endif
1948*4882a593Smuzhiyun 	{
1949*4882a593Smuzhiyun 		SI_VMSG((" %s, axi_num_wrappers:%d, Is_PCIE:%d, BUS_TYPE:%d, ID:%x\n",
1950*4882a593Smuzhiyun 			__FUNCTION__, sii->axi_num_wrappers, PCIE(sii),
1951*4882a593Smuzhiyun 			BUSTYPE(sii->pub.bustype), sii->pub.buscoretype));
1952*4882a593Smuzhiyun 		return AXI_WRAP_STS_NONE;
1953*4882a593Smuzhiyun 	}
1954*4882a593Smuzhiyun 
1955*4882a593Smuzhiyun #ifdef BCM_BACKPLANE_TIMEOUT
1956*4882a593Smuzhiyun 	/* Save and restore wrapper access window */
1957*4882a593Smuzhiyun 	if (BUSTYPE(sii->pub.bustype) == PCI_BUS) {
1958*4882a593Smuzhiyun 		if (PCIE_GEN1(sii)) {
1959*4882a593Smuzhiyun 			cfg_reg = PCI_BAR0_WIN2;
1960*4882a593Smuzhiyun 			offset = PCI_BAR0_WIN2_OFFSET;
1961*4882a593Smuzhiyun 		} else if (PCIE_GEN2(sii)) {
1962*4882a593Smuzhiyun 			cfg_reg = PCIE2_BAR0_CORE2_WIN2;
1963*4882a593Smuzhiyun 			offset = PCIE2_BAR0_CORE2_WIN2_OFFSET;
1964*4882a593Smuzhiyun 		}
1965*4882a593Smuzhiyun 		else {
1966*4882a593Smuzhiyun 			ASSERT(!"!PCIE_GEN1 && !PCIE_GEN2");
1967*4882a593Smuzhiyun 		}
1968*4882a593Smuzhiyun 
1969*4882a593Smuzhiyun 		prev_value = OSL_PCI_READ_CONFIG(osh, cfg_reg, 4);
1970*4882a593Smuzhiyun 
1971*4882a593Smuzhiyun 		if (prev_value == ID32_INVALID) {
1972*4882a593Smuzhiyun 			si_axi_error_t * axi_error =
1973*4882a593Smuzhiyun 				sih->err_info ?
1974*4882a593Smuzhiyun 					&sih->err_info->axi_error[sih->err_info->count] :
1975*4882a593Smuzhiyun 					NULL;
1976*4882a593Smuzhiyun 
1977*4882a593Smuzhiyun 			SI_PRINT(("%s, PCI_BAR0_WIN2 - %x\n", __FUNCTION__, prev_value));
1978*4882a593Smuzhiyun 			if (axi_error) {
1979*4882a593Smuzhiyun 				axi_error->error = ret = AXI_WRAP_STS_PCI_RD_ERR;
1980*4882a593Smuzhiyun 				axi_error->errlog_lo = cfg_reg;
1981*4882a593Smuzhiyun 				sih->err_info->count++;
1982*4882a593Smuzhiyun 
1983*4882a593Smuzhiyun 				if (sih->err_info->count == SI_MAX_ERRLOG_SIZE) {
1984*4882a593Smuzhiyun 					sih->err_info->count = SI_MAX_ERRLOG_SIZE - 1;
1985*4882a593Smuzhiyun 					SI_PRINT(("AXI Error log overflow\n"));
1986*4882a593Smuzhiyun 				}
1987*4882a593Smuzhiyun 			}
1988*4882a593Smuzhiyun 
1989*4882a593Smuzhiyun 			return ret;
1990*4882a593Smuzhiyun 		}
1991*4882a593Smuzhiyun 	}
1992*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
1993*4882a593Smuzhiyun 
1994*4882a593Smuzhiyun 	for (i = 0; i < sii->axi_num_wrappers; ++i) {
1995*4882a593Smuzhiyun 		uint32 tmp;
1996*4882a593Smuzhiyun 
1997*4882a593Smuzhiyun 		if (axi_wrapper[i].wrapper_type != AI_SLAVE_WRAPPER) {
1998*4882a593Smuzhiyun 			continue;
1999*4882a593Smuzhiyun 		}
2000*4882a593Smuzhiyun 
2001*4882a593Smuzhiyun #ifdef BCM_BACKPLANE_TIMEOUT
2002*4882a593Smuzhiyun 		if (BUSTYPE(sii->pub.bustype) == PCI_BUS) {
2003*4882a593Smuzhiyun 			/* Set BAR0_CORE2_WIN2 to bridge wapper base address */
2004*4882a593Smuzhiyun 			OSL_PCI_WRITE_CONFIG(osh,
2005*4882a593Smuzhiyun 				cfg_reg, 4, axi_wrapper[i].wrapper_addr);
2006*4882a593Smuzhiyun 
2007*4882a593Smuzhiyun 			/* set AI to BAR0 + Offset corresponding to Gen1 or gen2 */
2008*4882a593Smuzhiyun 			ai = (aidmp_t *) (DISCARD_QUAL(sii->curmap, uint8) + offset);
2009*4882a593Smuzhiyun 		}
2010*4882a593Smuzhiyun 		else
2011*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
2012*4882a593Smuzhiyun 		{
2013*4882a593Smuzhiyun 			ai = (aidmp_t *)(uintptr) axi_wrapper[i].wrapper_addr;
2014*4882a593Smuzhiyun 		}
2015*4882a593Smuzhiyun 
2016*4882a593Smuzhiyun 		tmp = ai_clear_backplane_to_per_core(sih, axi_wrapper[i].cid, 0,
2017*4882a593Smuzhiyun 			DISCARD_QUAL(ai, void));
2018*4882a593Smuzhiyun 
2019*4882a593Smuzhiyun 		ret |= tmp;
2020*4882a593Smuzhiyun 	}
2021*4882a593Smuzhiyun 
2022*4882a593Smuzhiyun #ifdef BCM_BACKPLANE_TIMEOUT
2023*4882a593Smuzhiyun 	/* Restore the initial wrapper space */
2024*4882a593Smuzhiyun 	if (prev_value) {
2025*4882a593Smuzhiyun 		OSL_PCI_WRITE_CONFIG(osh, cfg_reg, 4, prev_value);
2026*4882a593Smuzhiyun 	}
2027*4882a593Smuzhiyun #endif /* BCM_BACKPLANE_TIMEOUT */
2028*4882a593Smuzhiyun 
2029*4882a593Smuzhiyun #endif /* AXI_TIMEOUTS || BCM_BACKPLANE_TIMEOUT */
2030*4882a593Smuzhiyun 
2031*4882a593Smuzhiyun 	return ret;
2032*4882a593Smuzhiyun }
2033*4882a593Smuzhiyun 
2034*4882a593Smuzhiyun uint
ai_num_slaveports(si_t * sih,uint coreidx)2035*4882a593Smuzhiyun ai_num_slaveports(si_t *sih, uint coreidx)
2036*4882a593Smuzhiyun {
2037*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
2038*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
2039*4882a593Smuzhiyun 	uint32 cib;
2040*4882a593Smuzhiyun 
2041*4882a593Smuzhiyun 	cib = cores_info->cib[coreidx];
2042*4882a593Smuzhiyun 	return ((cib & CIB_NSP_MASK) >> CIB_NSP_SHIFT);
2043*4882a593Smuzhiyun }
2044*4882a593Smuzhiyun 
2045*4882a593Smuzhiyun #ifdef UART_TRAP_DBG
2046*4882a593Smuzhiyun void
ai_dump_APB_Bridge_registers(si_t * sih)2047*4882a593Smuzhiyun ai_dump_APB_Bridge_registers(si_t *sih)
2048*4882a593Smuzhiyun {
2049*4882a593Smuzhiyun aidmp_t *ai;
2050*4882a593Smuzhiyun si_info_t *sii = SI_INFO(sih);
2051*4882a593Smuzhiyun 
2052*4882a593Smuzhiyun 	ai = (aidmp_t *) sii->br_wrapba[0];
2053*4882a593Smuzhiyun 	printf("APB Bridge 0\n");
2054*4882a593Smuzhiyun 	printf("lo 0x%08x, hi 0x%08x, id 0x%08x, flags 0x%08x",
2055*4882a593Smuzhiyun 		R_REG(sii->osh, &ai->errlogaddrlo),
2056*4882a593Smuzhiyun 		R_REG(sii->osh, &ai->errlogaddrhi),
2057*4882a593Smuzhiyun 		R_REG(sii->osh, &ai->errlogid),
2058*4882a593Smuzhiyun 		R_REG(sii->osh, &ai->errlogflags));
2059*4882a593Smuzhiyun 	printf("\n status 0x%08x\n", R_REG(sii->osh, &ai->errlogstatus));
2060*4882a593Smuzhiyun }
2061*4882a593Smuzhiyun #endif /* UART_TRAP_DBG */
2062*4882a593Smuzhiyun 
2063*4882a593Smuzhiyun void
ai_force_clocks(si_t * sih,uint clock_state)2064*4882a593Smuzhiyun ai_force_clocks(si_t *sih, uint clock_state)
2065*4882a593Smuzhiyun {
2066*4882a593Smuzhiyun 
2067*4882a593Smuzhiyun 	si_info_t *sii = SI_INFO(sih);
2068*4882a593Smuzhiyun 	aidmp_t *ai, *ai_sec = NULL;
2069*4882a593Smuzhiyun 	volatile uint32 dummy;
2070*4882a593Smuzhiyun 	uint32 ioctrl;
2071*4882a593Smuzhiyun 	si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
2072*4882a593Smuzhiyun 
2073*4882a593Smuzhiyun 	ASSERT(GOODREGS(sii->curwrap));
2074*4882a593Smuzhiyun 	ai = sii->curwrap;
2075*4882a593Smuzhiyun 	if (cores_info->wrapba2[sii->curidx])
2076*4882a593Smuzhiyun 		ai_sec = REG_MAP(cores_info->wrapba2[sii->curidx], SI_CORE_SIZE);
2077*4882a593Smuzhiyun 
2078*4882a593Smuzhiyun 	/* ensure there are no pending backplane operations */
2079*4882a593Smuzhiyun 	SPINWAIT((R_REG(sii->osh, &ai->resetstatus) != 0), 300);
2080*4882a593Smuzhiyun 
2081*4882a593Smuzhiyun 	if (clock_state == FORCE_CLK_ON) {
2082*4882a593Smuzhiyun 		ioctrl = R_REG(sii->osh, &ai->ioctrl);
2083*4882a593Smuzhiyun 		W_REG(sii->osh, &ai->ioctrl, (ioctrl | SICF_FGC));
2084*4882a593Smuzhiyun 		dummy = R_REG(sii->osh, &ai->ioctrl);
2085*4882a593Smuzhiyun 		BCM_REFERENCE(dummy);
2086*4882a593Smuzhiyun 		if (ai_sec) {
2087*4882a593Smuzhiyun 			ioctrl = R_REG(sii->osh, &ai_sec->ioctrl);
2088*4882a593Smuzhiyun 			W_REG(sii->osh, &ai_sec->ioctrl, (ioctrl | SICF_FGC));
2089*4882a593Smuzhiyun 			dummy = R_REG(sii->osh, &ai_sec->ioctrl);
2090*4882a593Smuzhiyun 			BCM_REFERENCE(dummy);
2091*4882a593Smuzhiyun 		}
2092*4882a593Smuzhiyun 	} else {
2093*4882a593Smuzhiyun 		ioctrl = R_REG(sii->osh, &ai->ioctrl);
2094*4882a593Smuzhiyun 		W_REG(sii->osh, &ai->ioctrl, (ioctrl & (~SICF_FGC)));
2095*4882a593Smuzhiyun 		dummy = R_REG(sii->osh, &ai->ioctrl);
2096*4882a593Smuzhiyun 		BCM_REFERENCE(dummy);
2097*4882a593Smuzhiyun 		if (ai_sec) {
2098*4882a593Smuzhiyun 			ioctrl = R_REG(sii->osh, &ai_sec->ioctrl);
2099*4882a593Smuzhiyun 			W_REG(sii->osh, &ai_sec->ioctrl, (ioctrl & (~SICF_FGC)));
2100*4882a593Smuzhiyun 			dummy = R_REG(sii->osh, &ai_sec->ioctrl);
2101*4882a593Smuzhiyun 			BCM_REFERENCE(dummy);
2102*4882a593Smuzhiyun 		}
2103*4882a593Smuzhiyun 	}
2104*4882a593Smuzhiyun 	/* ensure there are no pending backplane operations */
2105*4882a593Smuzhiyun 	SPINWAIT((R_REG(sii->osh, &ai->resetstatus) != 0), 300);
2106*4882a593Smuzhiyun }
2107