xref: /OK3568_Linux_fs/kernel/drivers/scsi/lpfc/lpfc.h (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /*******************************************************************
2*4882a593Smuzhiyun  * This file is part of the Emulex Linux Device Driver for         *
3*4882a593Smuzhiyun  * Fibre Channel Host Bus Adapters.                                *
4*4882a593Smuzhiyun  * Copyright (C) 2017-2020 Broadcom. All Rights Reserved. The term *
5*4882a593Smuzhiyun  * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.     *
6*4882a593Smuzhiyun  * Copyright (C) 2004-2016 Emulex.  All rights reserved.           *
7*4882a593Smuzhiyun  * EMULEX and SLI are trademarks of Emulex.                        *
8*4882a593Smuzhiyun  * www.broadcom.com                                                *
9*4882a593Smuzhiyun  * Portions Copyright (C) 2004-2005 Christoph Hellwig              *
10*4882a593Smuzhiyun  *                                                                 *
11*4882a593Smuzhiyun  * This program is free software; you can redistribute it and/or   *
12*4882a593Smuzhiyun  * modify it under the terms of version 2 of the GNU General       *
13*4882a593Smuzhiyun  * Public License as published by the Free Software Foundation.    *
14*4882a593Smuzhiyun  * This program is distributed in the hope that it will be useful. *
15*4882a593Smuzhiyun  * ALL EXPRESS OR IMPLIED CONDITIONS, REPRESENTATIONS AND          *
16*4882a593Smuzhiyun  * WARRANTIES, INCLUDING ANY IMPLIED WARRANTY OF MERCHANTABILITY,  *
17*4882a593Smuzhiyun  * FITNESS FOR A PARTICULAR PURPOSE, OR NON-INFRINGEMENT, ARE      *
18*4882a593Smuzhiyun  * DISCLAIMED, EXCEPT TO THE EXTENT THAT SUCH DISCLAIMERS ARE HELD *
19*4882a593Smuzhiyun  * TO BE LEGALLY INVALID.  See the GNU General Public License for  *
20*4882a593Smuzhiyun  * more details, a copy of which can be found in the file COPYING  *
21*4882a593Smuzhiyun  * included with this package.                                     *
22*4882a593Smuzhiyun  *******************************************************************/
23*4882a593Smuzhiyun 
24*4882a593Smuzhiyun #include <scsi/scsi_host.h>
25*4882a593Smuzhiyun #include <linux/ktime.h>
26*4882a593Smuzhiyun #include <linux/workqueue.h>
27*4882a593Smuzhiyun 
28*4882a593Smuzhiyun #if defined(CONFIG_DEBUG_FS) && !defined(CONFIG_SCSI_LPFC_DEBUG_FS)
29*4882a593Smuzhiyun #define CONFIG_SCSI_LPFC_DEBUG_FS
30*4882a593Smuzhiyun #endif
31*4882a593Smuzhiyun 
32*4882a593Smuzhiyun struct lpfc_sli2_slim;
33*4882a593Smuzhiyun 
34*4882a593Smuzhiyun #define ELX_MODEL_NAME_SIZE	80
35*4882a593Smuzhiyun 
36*4882a593Smuzhiyun #define LPFC_PCI_DEV_LP		0x1
37*4882a593Smuzhiyun #define LPFC_PCI_DEV_OC		0x2
38*4882a593Smuzhiyun 
39*4882a593Smuzhiyun #define LPFC_SLI_REV2		2
40*4882a593Smuzhiyun #define LPFC_SLI_REV3		3
41*4882a593Smuzhiyun #define LPFC_SLI_REV4		4
42*4882a593Smuzhiyun 
43*4882a593Smuzhiyun #define LPFC_MAX_TARGET		4096	/* max number of targets supported */
44*4882a593Smuzhiyun #define LPFC_MAX_DISC_THREADS	64	/* max outstanding discovery els
45*4882a593Smuzhiyun 					   requests */
46*4882a593Smuzhiyun #define LPFC_MAX_NS_RETRY	3	/* Number of retry attempts to contact
47*4882a593Smuzhiyun 					   the NameServer  before giving up. */
48*4882a593Smuzhiyun #define LPFC_CMD_PER_LUN	3	/* max outstanding cmds per lun */
49*4882a593Smuzhiyun #define LPFC_DEFAULT_SG_SEG_CNT 64	/* sg element count per scsi cmnd */
50*4882a593Smuzhiyun #define LPFC_DEFAULT_MENLO_SG_SEG_CNT 128	/* sg element count per scsi
51*4882a593Smuzhiyun 		cmnd for menlo needs nearly twice as for firmware
52*4882a593Smuzhiyun 		downloads using bsg */
53*4882a593Smuzhiyun 
54*4882a593Smuzhiyun #define LPFC_DEFAULT_XPSGL_SIZE	256
55*4882a593Smuzhiyun #define LPFC_MAX_SG_TABLESIZE	0xffff
56*4882a593Smuzhiyun #define LPFC_MIN_SG_SLI4_BUF_SZ	0x800	/* based on LPFC_DEFAULT_SG_SEG_CNT */
57*4882a593Smuzhiyun #define LPFC_MAX_BG_SLI4_SEG_CNT_DIF 128 /* sg element count for BlockGuard */
58*4882a593Smuzhiyun #define LPFC_MAX_SG_SEG_CNT_DIF 512	/* sg element count per scsi cmnd  */
59*4882a593Smuzhiyun #define LPFC_MAX_SG_SEG_CNT	4096	/* sg element count per scsi cmnd */
60*4882a593Smuzhiyun #define LPFC_MIN_SG_SEG_CNT	32	/* sg element count per scsi cmnd */
61*4882a593Smuzhiyun #define LPFC_MAX_SGL_SEG_CNT	512	/* SGL element count per scsi cmnd */
62*4882a593Smuzhiyun #define LPFC_MAX_BPL_SEG_CNT	4096	/* BPL element count per scsi cmnd */
63*4882a593Smuzhiyun #define LPFC_MAX_NVME_SEG_CNT	256	/* max SGL element cnt per NVME cmnd */
64*4882a593Smuzhiyun 
65*4882a593Smuzhiyun #define LPFC_MAX_SGE_SIZE       0x80000000 /* Maximum data allowed in a SGE */
66*4882a593Smuzhiyun #define LPFC_IOCB_LIST_CNT	2250	/* list of IOCBs for fast-path usage. */
67*4882a593Smuzhiyun #define LPFC_Q_RAMP_UP_INTERVAL 120     /* lun q_depth ramp up interval */
68*4882a593Smuzhiyun #define LPFC_VNAME_LEN		100	/* vport symbolic name length */
69*4882a593Smuzhiyun #define LPFC_TGTQ_RAMPUP_PCENT	5	/* Target queue rampup in percentage */
70*4882a593Smuzhiyun #define LPFC_MIN_TGT_QDEPTH	10
71*4882a593Smuzhiyun #define LPFC_MAX_TGT_QDEPTH	0xFFFF
72*4882a593Smuzhiyun 
73*4882a593Smuzhiyun #define  LPFC_MAX_BUCKET_COUNT 20	/* Maximum no. of buckets for stat data
74*4882a593Smuzhiyun 					   collection. */
75*4882a593Smuzhiyun /*
76*4882a593Smuzhiyun  * Following time intervals are used of adjusting SCSI device
77*4882a593Smuzhiyun  * queue depths when there are driver resource error or Firmware
78*4882a593Smuzhiyun  * resource error.
79*4882a593Smuzhiyun  */
80*4882a593Smuzhiyun /* 1 Second */
81*4882a593Smuzhiyun #define QUEUE_RAMP_DOWN_INTERVAL	(msecs_to_jiffies(1000 * 1))
82*4882a593Smuzhiyun 
83*4882a593Smuzhiyun /* Number of exchanges reserved for discovery to complete */
84*4882a593Smuzhiyun #define LPFC_DISC_IOCB_BUFF_COUNT 20
85*4882a593Smuzhiyun 
86*4882a593Smuzhiyun #define LPFC_HB_MBOX_INTERVAL   5	/* Heart beat interval in seconds. */
87*4882a593Smuzhiyun #define LPFC_HB_MBOX_TIMEOUT    30	/* Heart beat timeout  in seconds. */
88*4882a593Smuzhiyun 
89*4882a593Smuzhiyun /* Error Attention event polling interval */
90*4882a593Smuzhiyun #define LPFC_ERATT_POLL_INTERVAL	5 /* EATT poll interval in seconds */
91*4882a593Smuzhiyun 
92*4882a593Smuzhiyun /* Define macros for 64 bit support */
93*4882a593Smuzhiyun #define putPaddrLow(addr)    ((uint32_t) (0xffffffff & (u64)(addr)))
94*4882a593Smuzhiyun #define putPaddrHigh(addr)   ((uint32_t) (0xffffffff & (((u64)(addr))>>32)))
95*4882a593Smuzhiyun #define getPaddr(high, low)  ((dma_addr_t)( \
96*4882a593Smuzhiyun 			     (( (u64)(high)<<16 ) << 16)|( (u64)(low))))
97*4882a593Smuzhiyun /* Provide maximum configuration definitions. */
98*4882a593Smuzhiyun #define LPFC_DRVR_TIMEOUT	16	/* driver iocb timeout value in sec */
99*4882a593Smuzhiyun #define FC_MAX_ADPTMSG		64
100*4882a593Smuzhiyun 
101*4882a593Smuzhiyun #define MAX_HBAEVT	32
102*4882a593Smuzhiyun #define MAX_HBAS_NO_RESET 16
103*4882a593Smuzhiyun 
104*4882a593Smuzhiyun /* Number of MSI-X vectors the driver uses */
105*4882a593Smuzhiyun #define LPFC_MSIX_VECTORS	2
106*4882a593Smuzhiyun 
107*4882a593Smuzhiyun /* lpfc wait event data ready flag */
108*4882a593Smuzhiyun #define LPFC_DATA_READY		0	/* bit 0 */
109*4882a593Smuzhiyun 
110*4882a593Smuzhiyun /* queue dump line buffer size */
111*4882a593Smuzhiyun #define LPFC_LBUF_SZ		128
112*4882a593Smuzhiyun 
113*4882a593Smuzhiyun /* mailbox system shutdown options */
114*4882a593Smuzhiyun #define LPFC_MBX_NO_WAIT	0
115*4882a593Smuzhiyun #define LPFC_MBX_WAIT		1
116*4882a593Smuzhiyun 
117*4882a593Smuzhiyun enum lpfc_polling_flags {
118*4882a593Smuzhiyun 	ENABLE_FCP_RING_POLLING = 0x1,
119*4882a593Smuzhiyun 	DISABLE_FCP_RING_INT    = 0x2
120*4882a593Smuzhiyun };
121*4882a593Smuzhiyun 
122*4882a593Smuzhiyun struct perf_prof {
123*4882a593Smuzhiyun 	uint16_t cmd_cpu[40];
124*4882a593Smuzhiyun 	uint16_t rsp_cpu[40];
125*4882a593Smuzhiyun 	uint16_t qh_cpu[40];
126*4882a593Smuzhiyun 	uint16_t wqidx[40];
127*4882a593Smuzhiyun };
128*4882a593Smuzhiyun 
129*4882a593Smuzhiyun /*
130*4882a593Smuzhiyun  * Provide for FC4 TYPE x28 - NVME.  The
131*4882a593Smuzhiyun  * bit mask for FCP and NVME is 0x8 identically
132*4882a593Smuzhiyun  * because they are 32 bit positions distance.
133*4882a593Smuzhiyun  */
134*4882a593Smuzhiyun #define LPFC_FC4_TYPE_BITMASK	0x00000100
135*4882a593Smuzhiyun 
136*4882a593Smuzhiyun /* Provide DMA memory definitions the driver uses per port instance. */
137*4882a593Smuzhiyun struct lpfc_dmabuf {
138*4882a593Smuzhiyun 	struct list_head list;
139*4882a593Smuzhiyun 	void *virt;		/* virtual address ptr */
140*4882a593Smuzhiyun 	dma_addr_t phys;	/* mapped address */
141*4882a593Smuzhiyun 	uint32_t   buffer_tag;	/* used for tagged queue ring */
142*4882a593Smuzhiyun };
143*4882a593Smuzhiyun 
144*4882a593Smuzhiyun struct lpfc_nvmet_ctxbuf {
145*4882a593Smuzhiyun 	struct list_head list;
146*4882a593Smuzhiyun 	struct lpfc_async_xchg_ctx *context;
147*4882a593Smuzhiyun 	struct lpfc_iocbq *iocbq;
148*4882a593Smuzhiyun 	struct lpfc_sglq *sglq;
149*4882a593Smuzhiyun 	struct work_struct defer_work;
150*4882a593Smuzhiyun };
151*4882a593Smuzhiyun 
152*4882a593Smuzhiyun struct lpfc_dma_pool {
153*4882a593Smuzhiyun 	struct lpfc_dmabuf   *elements;
154*4882a593Smuzhiyun 	uint32_t    max_count;
155*4882a593Smuzhiyun 	uint32_t    current_count;
156*4882a593Smuzhiyun };
157*4882a593Smuzhiyun 
158*4882a593Smuzhiyun struct hbq_dmabuf {
159*4882a593Smuzhiyun 	struct lpfc_dmabuf hbuf;
160*4882a593Smuzhiyun 	struct lpfc_dmabuf dbuf;
161*4882a593Smuzhiyun 	uint16_t total_size;
162*4882a593Smuzhiyun 	uint16_t bytes_recv;
163*4882a593Smuzhiyun 	uint32_t tag;
164*4882a593Smuzhiyun 	struct lpfc_cq_event cq_event;
165*4882a593Smuzhiyun 	unsigned long time_stamp;
166*4882a593Smuzhiyun 	void *context;
167*4882a593Smuzhiyun };
168*4882a593Smuzhiyun 
169*4882a593Smuzhiyun struct rqb_dmabuf {
170*4882a593Smuzhiyun 	struct lpfc_dmabuf hbuf;
171*4882a593Smuzhiyun 	struct lpfc_dmabuf dbuf;
172*4882a593Smuzhiyun 	uint16_t total_size;
173*4882a593Smuzhiyun 	uint16_t bytes_recv;
174*4882a593Smuzhiyun 	uint16_t idx;
175*4882a593Smuzhiyun 	struct lpfc_queue *hrq;	  /* ptr to associated Header RQ */
176*4882a593Smuzhiyun 	struct lpfc_queue *drq;	  /* ptr to associated Data RQ */
177*4882a593Smuzhiyun };
178*4882a593Smuzhiyun 
179*4882a593Smuzhiyun /* Priority bit.  Set value to exceed low water mark in lpfc_mem. */
180*4882a593Smuzhiyun #define MEM_PRI		0x100
181*4882a593Smuzhiyun 
182*4882a593Smuzhiyun 
183*4882a593Smuzhiyun /****************************************************************************/
184*4882a593Smuzhiyun /*      Device VPD save area                                                */
185*4882a593Smuzhiyun /****************************************************************************/
186*4882a593Smuzhiyun typedef struct lpfc_vpd {
187*4882a593Smuzhiyun 	uint32_t status;	/* vpd status value */
188*4882a593Smuzhiyun 	uint32_t length;	/* number of bytes actually returned */
189*4882a593Smuzhiyun 	struct {
190*4882a593Smuzhiyun 		uint32_t rsvd1;	/* Revision numbers */
191*4882a593Smuzhiyun 		uint32_t biuRev;
192*4882a593Smuzhiyun 		uint32_t smRev;
193*4882a593Smuzhiyun 		uint32_t smFwRev;
194*4882a593Smuzhiyun 		uint32_t endecRev;
195*4882a593Smuzhiyun 		uint16_t rBit;
196*4882a593Smuzhiyun 		uint8_t fcphHigh;
197*4882a593Smuzhiyun 		uint8_t fcphLow;
198*4882a593Smuzhiyun 		uint8_t feaLevelHigh;
199*4882a593Smuzhiyun 		uint8_t feaLevelLow;
200*4882a593Smuzhiyun 		uint32_t postKernRev;
201*4882a593Smuzhiyun 		uint32_t opFwRev;
202*4882a593Smuzhiyun 		uint8_t opFwName[16];
203*4882a593Smuzhiyun 		uint32_t sli1FwRev;
204*4882a593Smuzhiyun 		uint8_t sli1FwName[16];
205*4882a593Smuzhiyun 		uint32_t sli2FwRev;
206*4882a593Smuzhiyun 		uint8_t sli2FwName[16];
207*4882a593Smuzhiyun 	} rev;
208*4882a593Smuzhiyun 	struct {
209*4882a593Smuzhiyun #ifdef __BIG_ENDIAN_BITFIELD
210*4882a593Smuzhiyun 		uint32_t rsvd3  :20;  /* Reserved                             */
211*4882a593Smuzhiyun 		uint32_t rsvd2	: 3;  /* Reserved                             */
212*4882a593Smuzhiyun 		uint32_t cbg	: 1;  /* Configure BlockGuard                 */
213*4882a593Smuzhiyun 		uint32_t cmv	: 1;  /* Configure Max VPIs                   */
214*4882a593Smuzhiyun 		uint32_t ccrp   : 1;  /* Config Command Ring Polling          */
215*4882a593Smuzhiyun 		uint32_t csah   : 1;  /* Configure Synchronous Abort Handling */
216*4882a593Smuzhiyun 		uint32_t chbs   : 1;  /* Cofigure Host Backing store          */
217*4882a593Smuzhiyun 		uint32_t cinb   : 1;  /* Enable Interrupt Notification Block  */
218*4882a593Smuzhiyun 		uint32_t cerbm	: 1;  /* Configure Enhanced Receive Buf Mgmt  */
219*4882a593Smuzhiyun 		uint32_t cmx	: 1;  /* Configure Max XRIs                   */
220*4882a593Smuzhiyun 		uint32_t cmr	: 1;  /* Configure Max RPIs                   */
221*4882a593Smuzhiyun #else	/*  __LITTLE_ENDIAN */
222*4882a593Smuzhiyun 		uint32_t cmr	: 1;  /* Configure Max RPIs                   */
223*4882a593Smuzhiyun 		uint32_t cmx	: 1;  /* Configure Max XRIs                   */
224*4882a593Smuzhiyun 		uint32_t cerbm	: 1;  /* Configure Enhanced Receive Buf Mgmt  */
225*4882a593Smuzhiyun 		uint32_t cinb   : 1;  /* Enable Interrupt Notification Block  */
226*4882a593Smuzhiyun 		uint32_t chbs   : 1;  /* Cofigure Host Backing store          */
227*4882a593Smuzhiyun 		uint32_t csah   : 1;  /* Configure Synchronous Abort Handling */
228*4882a593Smuzhiyun 		uint32_t ccrp   : 1;  /* Config Command Ring Polling          */
229*4882a593Smuzhiyun 		uint32_t cmv	: 1;  /* Configure Max VPIs                   */
230*4882a593Smuzhiyun 		uint32_t cbg	: 1;  /* Configure BlockGuard                 */
231*4882a593Smuzhiyun 		uint32_t rsvd2	: 3;  /* Reserved                             */
232*4882a593Smuzhiyun 		uint32_t rsvd3  :20;  /* Reserved                             */
233*4882a593Smuzhiyun #endif
234*4882a593Smuzhiyun 	} sli3Feat;
235*4882a593Smuzhiyun } lpfc_vpd_t;
236*4882a593Smuzhiyun 
237*4882a593Smuzhiyun 
238*4882a593Smuzhiyun /*
239*4882a593Smuzhiyun  * lpfc stat counters
240*4882a593Smuzhiyun  */
241*4882a593Smuzhiyun struct lpfc_stats {
242*4882a593Smuzhiyun 	/* Statistics for ELS commands */
243*4882a593Smuzhiyun 	uint32_t elsLogiCol;
244*4882a593Smuzhiyun 	uint32_t elsRetryExceeded;
245*4882a593Smuzhiyun 	uint32_t elsXmitRetry;
246*4882a593Smuzhiyun 	uint32_t elsDelayRetry;
247*4882a593Smuzhiyun 	uint32_t elsRcvDrop;
248*4882a593Smuzhiyun 	uint32_t elsRcvFrame;
249*4882a593Smuzhiyun 	uint32_t elsRcvRSCN;
250*4882a593Smuzhiyun 	uint32_t elsRcvRNID;
251*4882a593Smuzhiyun 	uint32_t elsRcvFARP;
252*4882a593Smuzhiyun 	uint32_t elsRcvFARPR;
253*4882a593Smuzhiyun 	uint32_t elsRcvFLOGI;
254*4882a593Smuzhiyun 	uint32_t elsRcvPLOGI;
255*4882a593Smuzhiyun 	uint32_t elsRcvADISC;
256*4882a593Smuzhiyun 	uint32_t elsRcvPDISC;
257*4882a593Smuzhiyun 	uint32_t elsRcvFAN;
258*4882a593Smuzhiyun 	uint32_t elsRcvLOGO;
259*4882a593Smuzhiyun 	uint32_t elsRcvPRLO;
260*4882a593Smuzhiyun 	uint32_t elsRcvPRLI;
261*4882a593Smuzhiyun 	uint32_t elsRcvLIRR;
262*4882a593Smuzhiyun 	uint32_t elsRcvRLS;
263*4882a593Smuzhiyun 	uint32_t elsRcvRPL;
264*4882a593Smuzhiyun 	uint32_t elsRcvRRQ;
265*4882a593Smuzhiyun 	uint32_t elsRcvRTV;
266*4882a593Smuzhiyun 	uint32_t elsRcvECHO;
267*4882a593Smuzhiyun 	uint32_t elsRcvLCB;
268*4882a593Smuzhiyun 	uint32_t elsRcvRDP;
269*4882a593Smuzhiyun 	uint32_t elsXmitFLOGI;
270*4882a593Smuzhiyun 	uint32_t elsXmitFDISC;
271*4882a593Smuzhiyun 	uint32_t elsXmitPLOGI;
272*4882a593Smuzhiyun 	uint32_t elsXmitPRLI;
273*4882a593Smuzhiyun 	uint32_t elsXmitADISC;
274*4882a593Smuzhiyun 	uint32_t elsXmitLOGO;
275*4882a593Smuzhiyun 	uint32_t elsXmitSCR;
276*4882a593Smuzhiyun 	uint32_t elsXmitRSCN;
277*4882a593Smuzhiyun 	uint32_t elsXmitRNID;
278*4882a593Smuzhiyun 	uint32_t elsXmitFARP;
279*4882a593Smuzhiyun 	uint32_t elsXmitFARPR;
280*4882a593Smuzhiyun 	uint32_t elsXmitACC;
281*4882a593Smuzhiyun 	uint32_t elsXmitLSRJT;
282*4882a593Smuzhiyun 
283*4882a593Smuzhiyun 	uint32_t frameRcvBcast;
284*4882a593Smuzhiyun 	uint32_t frameRcvMulti;
285*4882a593Smuzhiyun 	uint32_t strayXmitCmpl;
286*4882a593Smuzhiyun 	uint32_t frameXmitDelay;
287*4882a593Smuzhiyun 	uint32_t xriCmdCmpl;
288*4882a593Smuzhiyun 	uint32_t xriStatErr;
289*4882a593Smuzhiyun 	uint32_t LinkUp;
290*4882a593Smuzhiyun 	uint32_t LinkDown;
291*4882a593Smuzhiyun 	uint32_t LinkMultiEvent;
292*4882a593Smuzhiyun 	uint32_t NoRcvBuf;
293*4882a593Smuzhiyun 	uint32_t fcpCmd;
294*4882a593Smuzhiyun 	uint32_t fcpCmpl;
295*4882a593Smuzhiyun 	uint32_t fcpRspErr;
296*4882a593Smuzhiyun 	uint32_t fcpRemoteStop;
297*4882a593Smuzhiyun 	uint32_t fcpPortRjt;
298*4882a593Smuzhiyun 	uint32_t fcpPortBusy;
299*4882a593Smuzhiyun 	uint32_t fcpError;
300*4882a593Smuzhiyun 	uint32_t fcpLocalErr;
301*4882a593Smuzhiyun };
302*4882a593Smuzhiyun 
303*4882a593Smuzhiyun struct lpfc_hba;
304*4882a593Smuzhiyun 
305*4882a593Smuzhiyun 
306*4882a593Smuzhiyun enum discovery_state {
307*4882a593Smuzhiyun 	LPFC_VPORT_UNKNOWN     =  0,    /* vport state is unknown */
308*4882a593Smuzhiyun 	LPFC_VPORT_FAILED      =  1,    /* vport has failed */
309*4882a593Smuzhiyun 	LPFC_LOCAL_CFG_LINK    =  6,    /* local NPORT Id configured */
310*4882a593Smuzhiyun 	LPFC_FLOGI             =  7,    /* FLOGI sent to Fabric */
311*4882a593Smuzhiyun 	LPFC_FDISC             =  8,    /* FDISC sent for vport */
312*4882a593Smuzhiyun 	LPFC_FABRIC_CFG_LINK   =  9,    /* Fabric assigned NPORT Id
313*4882a593Smuzhiyun 				         * configured */
314*4882a593Smuzhiyun 	LPFC_NS_REG            =  10,   /* Register with NameServer */
315*4882a593Smuzhiyun 	LPFC_NS_QRY            =  11,   /* Query NameServer for NPort ID list */
316*4882a593Smuzhiyun 	LPFC_BUILD_DISC_LIST   =  12,   /* Build ADISC and PLOGI lists for
317*4882a593Smuzhiyun 				         * device authentication / discovery */
318*4882a593Smuzhiyun 	LPFC_DISC_AUTH         =  13,   /* Processing ADISC list */
319*4882a593Smuzhiyun 	LPFC_VPORT_READY       =  32,
320*4882a593Smuzhiyun };
321*4882a593Smuzhiyun 
322*4882a593Smuzhiyun enum hba_state {
323*4882a593Smuzhiyun 	LPFC_LINK_UNKNOWN    =   0,   /* HBA state is unknown */
324*4882a593Smuzhiyun 	LPFC_WARM_START      =   1,   /* HBA state after selective reset */
325*4882a593Smuzhiyun 	LPFC_INIT_START      =   2,   /* Initial state after board reset */
326*4882a593Smuzhiyun 	LPFC_INIT_MBX_CMDS   =   3,   /* Initialize HBA with mbox commands */
327*4882a593Smuzhiyun 	LPFC_LINK_DOWN       =   4,   /* HBA initialized, link is down */
328*4882a593Smuzhiyun 	LPFC_LINK_UP         =   5,   /* Link is up  - issue READ_LA */
329*4882a593Smuzhiyun 	LPFC_CLEAR_LA        =   6,   /* authentication cmplt - issue
330*4882a593Smuzhiyun 				       * CLEAR_LA */
331*4882a593Smuzhiyun 	LPFC_HBA_READY       =  32,
332*4882a593Smuzhiyun 	LPFC_HBA_ERROR       =  -1
333*4882a593Smuzhiyun };
334*4882a593Smuzhiyun 
335*4882a593Smuzhiyun struct lpfc_trunk_link_state {
336*4882a593Smuzhiyun 	enum hba_state state;
337*4882a593Smuzhiyun 	uint8_t fault;
338*4882a593Smuzhiyun };
339*4882a593Smuzhiyun 
340*4882a593Smuzhiyun struct lpfc_trunk_link  {
341*4882a593Smuzhiyun 	struct lpfc_trunk_link_state link0,
342*4882a593Smuzhiyun 				     link1,
343*4882a593Smuzhiyun 				     link2,
344*4882a593Smuzhiyun 				     link3;
345*4882a593Smuzhiyun };
346*4882a593Smuzhiyun 
347*4882a593Smuzhiyun struct lpfc_vport {
348*4882a593Smuzhiyun 	struct lpfc_hba *phba;
349*4882a593Smuzhiyun 	struct list_head listentry;
350*4882a593Smuzhiyun 	uint8_t port_type;
351*4882a593Smuzhiyun #define LPFC_PHYSICAL_PORT 1
352*4882a593Smuzhiyun #define LPFC_NPIV_PORT  2
353*4882a593Smuzhiyun #define LPFC_FABRIC_PORT 3
354*4882a593Smuzhiyun 	enum discovery_state port_state;
355*4882a593Smuzhiyun 
356*4882a593Smuzhiyun 	uint16_t vpi;
357*4882a593Smuzhiyun 	uint16_t vfi;
358*4882a593Smuzhiyun 	uint8_t vpi_state;
359*4882a593Smuzhiyun #define LPFC_VPI_REGISTERED	0x1
360*4882a593Smuzhiyun 
361*4882a593Smuzhiyun 	uint32_t fc_flag;	/* FC flags */
362*4882a593Smuzhiyun /* Several of these flags are HBA centric and should be moved to
363*4882a593Smuzhiyun  * phba->link_flag (e.g. FC_PTP, FC_PUBLIC_LOOP)
364*4882a593Smuzhiyun  */
365*4882a593Smuzhiyun #define FC_PT2PT                0x1	 /* pt2pt with no fabric */
366*4882a593Smuzhiyun #define FC_PT2PT_PLOGI          0x2	 /* pt2pt initiate PLOGI */
367*4882a593Smuzhiyun #define FC_DISC_TMO             0x4	 /* Discovery timer running */
368*4882a593Smuzhiyun #define FC_PUBLIC_LOOP          0x8	 /* Public loop */
369*4882a593Smuzhiyun #define FC_LBIT                 0x10	 /* LOGIN bit in loopinit set */
370*4882a593Smuzhiyun #define FC_RSCN_MODE            0x20	 /* RSCN cmd rcv'ed */
371*4882a593Smuzhiyun #define FC_NLP_MORE             0x40	 /* More node to process in node tbl */
372*4882a593Smuzhiyun #define FC_OFFLINE_MODE         0x80	 /* Interface is offline for diag */
373*4882a593Smuzhiyun #define FC_FABRIC               0x100	 /* We are fabric attached */
374*4882a593Smuzhiyun #define FC_VPORT_LOGO_RCVD      0x200    /* LOGO received on vport */
375*4882a593Smuzhiyun #define FC_RSCN_DISCOVERY       0x400	 /* Auth all devices after RSCN */
376*4882a593Smuzhiyun #define FC_LOGO_RCVD_DID_CHNG   0x800    /* FDISC on phys port detect DID chng*/
377*4882a593Smuzhiyun #define FC_PT2PT_NO_NVME        0x1000   /* Don't send NVME PRLI */
378*4882a593Smuzhiyun #define FC_SCSI_SCAN_TMO        0x4000	 /* scsi scan timer running */
379*4882a593Smuzhiyun #define FC_ABORT_DISCOVERY      0x8000	 /* we want to abort discovery */
380*4882a593Smuzhiyun #define FC_NDISC_ACTIVE         0x10000	 /* NPort discovery active */
381*4882a593Smuzhiyun #define FC_BYPASSED_MODE        0x20000	 /* NPort is in bypassed mode */
382*4882a593Smuzhiyun #define FC_VPORT_NEEDS_REG_VPI	0x80000  /* Needs to have its vpi registered */
383*4882a593Smuzhiyun #define FC_RSCN_DEFERRED	0x100000 /* A deferred RSCN being processed */
384*4882a593Smuzhiyun #define FC_VPORT_NEEDS_INIT_VPI 0x200000 /* Need to INIT_VPI before FDISC */
385*4882a593Smuzhiyun #define FC_VPORT_CVL_RCVD	0x400000 /* VLink failed due to CVL	 */
386*4882a593Smuzhiyun #define FC_VFI_REGISTERED	0x800000 /* VFI is registered */
387*4882a593Smuzhiyun #define FC_FDISC_COMPLETED	0x1000000/* FDISC completed */
388*4882a593Smuzhiyun #define FC_DISC_DELAYED		0x2000000/* Delay NPort discovery */
389*4882a593Smuzhiyun 
390*4882a593Smuzhiyun 	uint32_t ct_flags;
391*4882a593Smuzhiyun #define FC_CT_RFF_ID		0x1	 /* RFF_ID accepted by switch */
392*4882a593Smuzhiyun #define FC_CT_RNN_ID		0x2	 /* RNN_ID accepted by switch */
393*4882a593Smuzhiyun #define FC_CT_RSNN_NN		0x4	 /* RSNN_NN accepted by switch */
394*4882a593Smuzhiyun #define FC_CT_RSPN_ID		0x8	 /* RSPN_ID accepted by switch */
395*4882a593Smuzhiyun #define FC_CT_RFT_ID		0x10	 /* RFT_ID accepted by switch */
396*4882a593Smuzhiyun 
397*4882a593Smuzhiyun 	struct list_head fc_nodes;
398*4882a593Smuzhiyun 
399*4882a593Smuzhiyun 	/* Keep counters for the number of entries in each list. */
400*4882a593Smuzhiyun 	uint16_t fc_plogi_cnt;
401*4882a593Smuzhiyun 	uint16_t fc_adisc_cnt;
402*4882a593Smuzhiyun 	uint16_t fc_reglogin_cnt;
403*4882a593Smuzhiyun 	uint16_t fc_prli_cnt;
404*4882a593Smuzhiyun 	uint16_t fc_unmap_cnt;
405*4882a593Smuzhiyun 	uint16_t fc_map_cnt;
406*4882a593Smuzhiyun 	uint16_t fc_npr_cnt;
407*4882a593Smuzhiyun 	uint16_t fc_unused_cnt;
408*4882a593Smuzhiyun 	struct serv_parm fc_sparam;	/* buffer for our service parameters */
409*4882a593Smuzhiyun 
410*4882a593Smuzhiyun 	uint32_t fc_myDID;	/* fibre channel S_ID */
411*4882a593Smuzhiyun 	uint32_t fc_prevDID;	/* previous fibre channel S_ID */
412*4882a593Smuzhiyun 	struct lpfc_name fabric_portname;
413*4882a593Smuzhiyun 	struct lpfc_name fabric_nodename;
414*4882a593Smuzhiyun 
415*4882a593Smuzhiyun 	int32_t stopped;   /* HBA has not been restarted since last ERATT */
416*4882a593Smuzhiyun 	uint8_t fc_linkspeed;	/* Link speed after last READ_LA */
417*4882a593Smuzhiyun 
418*4882a593Smuzhiyun 	uint32_t num_disc_nodes;	/* in addition to hba_state */
419*4882a593Smuzhiyun 	uint32_t gidft_inp;		/* cnt of outstanding GID_FTs */
420*4882a593Smuzhiyun 
421*4882a593Smuzhiyun 	uint32_t fc_nlp_cnt;	/* outstanding NODELIST requests */
422*4882a593Smuzhiyun 	uint32_t fc_rscn_id_cnt;	/* count of RSCNs payloads in list */
423*4882a593Smuzhiyun 	uint32_t fc_rscn_flush;		/* flag use of fc_rscn_id_list */
424*4882a593Smuzhiyun 	struct lpfc_dmabuf *fc_rscn_id_list[FC_MAX_HOLD_RSCN];
425*4882a593Smuzhiyun 	struct lpfc_name fc_nodename;	/* fc nodename */
426*4882a593Smuzhiyun 	struct lpfc_name fc_portname;	/* fc portname */
427*4882a593Smuzhiyun 
428*4882a593Smuzhiyun 	struct lpfc_work_evt disc_timeout_evt;
429*4882a593Smuzhiyun 
430*4882a593Smuzhiyun 	struct timer_list fc_disctmo;	/* Discovery rescue timer */
431*4882a593Smuzhiyun 	uint8_t fc_ns_retry;	/* retries for fabric nameserver */
432*4882a593Smuzhiyun 	uint32_t fc_prli_sent;	/* cntr for outstanding PRLIs */
433*4882a593Smuzhiyun 
434*4882a593Smuzhiyun 	spinlock_t work_port_lock;
435*4882a593Smuzhiyun 	uint32_t work_port_events; /* Timeout to be handled  */
436*4882a593Smuzhiyun #define WORKER_DISC_TMO                0x1	/* vport: Discovery timeout */
437*4882a593Smuzhiyun #define WORKER_ELS_TMO                 0x2	/* vport: ELS timeout */
438*4882a593Smuzhiyun #define WORKER_DELAYED_DISC_TMO        0x8	/* vport: delayed discovery */
439*4882a593Smuzhiyun 
440*4882a593Smuzhiyun #define WORKER_MBOX_TMO                0x100	/* hba: MBOX timeout */
441*4882a593Smuzhiyun #define WORKER_HB_TMO                  0x200	/* hba: Heart beat timeout */
442*4882a593Smuzhiyun #define WORKER_FABRIC_BLOCK_TMO        0x400	/* hba: fabric block timeout */
443*4882a593Smuzhiyun #define WORKER_RAMP_DOWN_QUEUE         0x800	/* hba: Decrease Q depth */
444*4882a593Smuzhiyun #define WORKER_RAMP_UP_QUEUE           0x1000	/* hba: Increase Q depth */
445*4882a593Smuzhiyun #define WORKER_SERVICE_TXQ             0x2000	/* hba: IOCBs on the txq */
446*4882a593Smuzhiyun 
447*4882a593Smuzhiyun 	struct timer_list els_tmofunc;
448*4882a593Smuzhiyun 	struct timer_list delayed_disc_tmo;
449*4882a593Smuzhiyun 
450*4882a593Smuzhiyun 	int unreg_vpi_cmpl;
451*4882a593Smuzhiyun 
452*4882a593Smuzhiyun 	uint8_t load_flag;
453*4882a593Smuzhiyun #define FC_LOADING		0x1	/* HBA in process of loading drvr */
454*4882a593Smuzhiyun #define FC_UNLOADING		0x2	/* HBA in process of unloading drvr */
455*4882a593Smuzhiyun #define FC_ALLOW_FDMI		0x4	/* port is ready for FDMI requests */
456*4882a593Smuzhiyun 	/* Vport Config Parameters */
457*4882a593Smuzhiyun 	uint32_t cfg_scan_down;
458*4882a593Smuzhiyun 	uint32_t cfg_lun_queue_depth;
459*4882a593Smuzhiyun 	uint32_t cfg_nodev_tmo;
460*4882a593Smuzhiyun 	uint32_t cfg_devloss_tmo;
461*4882a593Smuzhiyun 	uint32_t cfg_restrict_login;
462*4882a593Smuzhiyun 	uint32_t cfg_peer_port_login;
463*4882a593Smuzhiyun 	uint32_t cfg_fcp_class;
464*4882a593Smuzhiyun 	uint32_t cfg_use_adisc;
465*4882a593Smuzhiyun 	uint32_t cfg_discovery_threads;
466*4882a593Smuzhiyun 	uint32_t cfg_log_verbose;
467*4882a593Smuzhiyun 	uint32_t cfg_enable_fc4_type;
468*4882a593Smuzhiyun 	uint32_t cfg_max_luns;
469*4882a593Smuzhiyun 	uint32_t cfg_enable_da_id;
470*4882a593Smuzhiyun 	uint32_t cfg_max_scsicmpl_time;
471*4882a593Smuzhiyun 	uint32_t cfg_tgt_queue_depth;
472*4882a593Smuzhiyun 	uint32_t cfg_first_burst_size;
473*4882a593Smuzhiyun 	uint32_t dev_loss_tmo_changed;
474*4882a593Smuzhiyun 
475*4882a593Smuzhiyun 	struct fc_vport *fc_vport;
476*4882a593Smuzhiyun 
477*4882a593Smuzhiyun #ifdef CONFIG_SCSI_LPFC_DEBUG_FS
478*4882a593Smuzhiyun 	struct dentry *debug_disc_trc;
479*4882a593Smuzhiyun 	struct dentry *debug_nodelist;
480*4882a593Smuzhiyun 	struct dentry *debug_nvmestat;
481*4882a593Smuzhiyun 	struct dentry *debug_scsistat;
482*4882a593Smuzhiyun 	struct dentry *debug_ioktime;
483*4882a593Smuzhiyun 	struct dentry *debug_hdwqstat;
484*4882a593Smuzhiyun 	struct dentry *vport_debugfs_root;
485*4882a593Smuzhiyun 	struct lpfc_debugfs_trc *disc_trc;
486*4882a593Smuzhiyun 	atomic_t disc_trc_cnt;
487*4882a593Smuzhiyun #endif
488*4882a593Smuzhiyun 	uint8_t stat_data_enabled;
489*4882a593Smuzhiyun 	uint8_t stat_data_blocked;
490*4882a593Smuzhiyun 	struct list_head rcv_buffer_list;
491*4882a593Smuzhiyun 	unsigned long rcv_buffer_time_stamp;
492*4882a593Smuzhiyun 	uint32_t vport_flag;
493*4882a593Smuzhiyun #define STATIC_VPORT	1
494*4882a593Smuzhiyun #define FAWWPN_SET	2
495*4882a593Smuzhiyun #define FAWWPN_PARAM_CHG	4
496*4882a593Smuzhiyun 
497*4882a593Smuzhiyun 	uint16_t fdmi_num_disc;
498*4882a593Smuzhiyun 	uint32_t fdmi_hba_mask;
499*4882a593Smuzhiyun 	uint32_t fdmi_port_mask;
500*4882a593Smuzhiyun 
501*4882a593Smuzhiyun 	/* There is a single nvme instance per vport. */
502*4882a593Smuzhiyun 	struct nvme_fc_local_port *localport;
503*4882a593Smuzhiyun 	uint8_t  nvmei_support; /* driver supports NVME Initiator */
504*4882a593Smuzhiyun 	uint32_t last_fcp_wqidx;
505*4882a593Smuzhiyun 	uint32_t rcv_flogi_cnt; /* How many unsol FLOGIs ACK'd. */
506*4882a593Smuzhiyun };
507*4882a593Smuzhiyun 
508*4882a593Smuzhiyun struct hbq_s {
509*4882a593Smuzhiyun 	uint16_t entry_count;	  /* Current number of HBQ slots */
510*4882a593Smuzhiyun 	uint16_t buffer_count;	  /* Current number of buffers posted */
511*4882a593Smuzhiyun 	uint32_t next_hbqPutIdx;  /* Index to next HBQ slot to use */
512*4882a593Smuzhiyun 	uint32_t hbqPutIdx;	  /* HBQ slot to use */
513*4882a593Smuzhiyun 	uint32_t local_hbqGetIdx; /* Local copy of Get index from Port */
514*4882a593Smuzhiyun 	void    *hbq_virt;	  /* Virtual ptr to this hbq */
515*4882a593Smuzhiyun 	struct list_head hbq_buffer_list;  /* buffers assigned to this HBQ */
516*4882a593Smuzhiyun 				  /* Callback for HBQ buffer allocation */
517*4882a593Smuzhiyun 	struct hbq_dmabuf *(*hbq_alloc_buffer) (struct lpfc_hba *);
518*4882a593Smuzhiyun 				  /* Callback for HBQ buffer free */
519*4882a593Smuzhiyun 	void               (*hbq_free_buffer) (struct lpfc_hba *,
520*4882a593Smuzhiyun 					       struct hbq_dmabuf *);
521*4882a593Smuzhiyun };
522*4882a593Smuzhiyun 
523*4882a593Smuzhiyun /* this matches the position in the lpfc_hbq_defs array */
524*4882a593Smuzhiyun #define LPFC_ELS_HBQ	0
525*4882a593Smuzhiyun #define LPFC_MAX_HBQS	1
526*4882a593Smuzhiyun 
527*4882a593Smuzhiyun enum hba_temp_state {
528*4882a593Smuzhiyun 	HBA_NORMAL_TEMP,
529*4882a593Smuzhiyun 	HBA_OVER_TEMP
530*4882a593Smuzhiyun };
531*4882a593Smuzhiyun 
532*4882a593Smuzhiyun enum intr_type_t {
533*4882a593Smuzhiyun 	NONE = 0,
534*4882a593Smuzhiyun 	INTx,
535*4882a593Smuzhiyun 	MSI,
536*4882a593Smuzhiyun 	MSIX,
537*4882a593Smuzhiyun };
538*4882a593Smuzhiyun 
539*4882a593Smuzhiyun #define LPFC_CT_CTX_MAX		64
540*4882a593Smuzhiyun struct unsol_rcv_ct_ctx {
541*4882a593Smuzhiyun 	uint32_t ctxt_id;
542*4882a593Smuzhiyun 	uint32_t SID;
543*4882a593Smuzhiyun 	uint32_t valid;
544*4882a593Smuzhiyun #define UNSOL_INVALID		0
545*4882a593Smuzhiyun #define UNSOL_VALID		1
546*4882a593Smuzhiyun 	uint16_t oxid;
547*4882a593Smuzhiyun 	uint16_t rxid;
548*4882a593Smuzhiyun };
549*4882a593Smuzhiyun 
550*4882a593Smuzhiyun #define LPFC_USER_LINK_SPEED_AUTO	0	/* auto select (default)*/
551*4882a593Smuzhiyun #define LPFC_USER_LINK_SPEED_1G		1	/* 1 Gigabaud */
552*4882a593Smuzhiyun #define LPFC_USER_LINK_SPEED_2G		2	/* 2 Gigabaud */
553*4882a593Smuzhiyun #define LPFC_USER_LINK_SPEED_4G		4	/* 4 Gigabaud */
554*4882a593Smuzhiyun #define LPFC_USER_LINK_SPEED_8G		8	/* 8 Gigabaud */
555*4882a593Smuzhiyun #define LPFC_USER_LINK_SPEED_10G	10	/* 10 Gigabaud */
556*4882a593Smuzhiyun #define LPFC_USER_LINK_SPEED_16G	16	/* 16 Gigabaud */
557*4882a593Smuzhiyun #define LPFC_USER_LINK_SPEED_32G	32	/* 32 Gigabaud */
558*4882a593Smuzhiyun #define LPFC_USER_LINK_SPEED_64G	64	/* 64 Gigabaud */
559*4882a593Smuzhiyun #define LPFC_USER_LINK_SPEED_MAX	LPFC_USER_LINK_SPEED_64G
560*4882a593Smuzhiyun 
561*4882a593Smuzhiyun #define LPFC_LINK_SPEED_STRING "0, 1, 2, 4, 8, 10, 16, 32, 64"
562*4882a593Smuzhiyun 
563*4882a593Smuzhiyun enum nemb_type {
564*4882a593Smuzhiyun 	nemb_mse = 1,
565*4882a593Smuzhiyun 	nemb_hbd
566*4882a593Smuzhiyun };
567*4882a593Smuzhiyun 
568*4882a593Smuzhiyun enum mbox_type {
569*4882a593Smuzhiyun 	mbox_rd = 1,
570*4882a593Smuzhiyun 	mbox_wr
571*4882a593Smuzhiyun };
572*4882a593Smuzhiyun 
573*4882a593Smuzhiyun enum dma_type {
574*4882a593Smuzhiyun 	dma_mbox = 1,
575*4882a593Smuzhiyun 	dma_ebuf
576*4882a593Smuzhiyun };
577*4882a593Smuzhiyun 
578*4882a593Smuzhiyun enum sta_type {
579*4882a593Smuzhiyun 	sta_pre_addr = 1,
580*4882a593Smuzhiyun 	sta_pos_addr
581*4882a593Smuzhiyun };
582*4882a593Smuzhiyun 
583*4882a593Smuzhiyun struct lpfc_mbox_ext_buf_ctx {
584*4882a593Smuzhiyun 	uint32_t state;
585*4882a593Smuzhiyun #define LPFC_BSG_MBOX_IDLE		0
586*4882a593Smuzhiyun #define LPFC_BSG_MBOX_HOST              1
587*4882a593Smuzhiyun #define LPFC_BSG_MBOX_PORT		2
588*4882a593Smuzhiyun #define LPFC_BSG_MBOX_DONE		3
589*4882a593Smuzhiyun #define LPFC_BSG_MBOX_ABTS		4
590*4882a593Smuzhiyun 	enum nemb_type nembType;
591*4882a593Smuzhiyun 	enum mbox_type mboxType;
592*4882a593Smuzhiyun 	uint32_t numBuf;
593*4882a593Smuzhiyun 	uint32_t mbxTag;
594*4882a593Smuzhiyun 	uint32_t seqNum;
595*4882a593Smuzhiyun 	struct lpfc_dmabuf *mbx_dmabuf;
596*4882a593Smuzhiyun 	struct list_head ext_dmabuf_list;
597*4882a593Smuzhiyun };
598*4882a593Smuzhiyun 
599*4882a593Smuzhiyun struct lpfc_epd_pool {
600*4882a593Smuzhiyun 	/* Expedite pool */
601*4882a593Smuzhiyun 	struct list_head list;
602*4882a593Smuzhiyun 	u32 count;
603*4882a593Smuzhiyun 	spinlock_t lock;	/* lock for expedite pool */
604*4882a593Smuzhiyun };
605*4882a593Smuzhiyun 
606*4882a593Smuzhiyun enum ras_state {
607*4882a593Smuzhiyun 	INACTIVE,
608*4882a593Smuzhiyun 	REG_INPROGRESS,
609*4882a593Smuzhiyun 	ACTIVE
610*4882a593Smuzhiyun };
611*4882a593Smuzhiyun 
612*4882a593Smuzhiyun struct lpfc_ras_fwlog {
613*4882a593Smuzhiyun 	uint8_t *fwlog_buff;
614*4882a593Smuzhiyun 	uint32_t fw_buffcount; /* Buffer size posted to FW */
615*4882a593Smuzhiyun #define LPFC_RAS_BUFF_ENTERIES  16      /* Each entry can hold max of 64k */
616*4882a593Smuzhiyun #define LPFC_RAS_MAX_ENTRY_SIZE (64 * 1024)
617*4882a593Smuzhiyun #define LPFC_RAS_MIN_BUFF_POST_SIZE (256 * 1024)
618*4882a593Smuzhiyun #define LPFC_RAS_MAX_BUFF_POST_SIZE (1024 * 1024)
619*4882a593Smuzhiyun 	uint32_t fw_loglevel; /* Log level set */
620*4882a593Smuzhiyun 	struct lpfc_dmabuf lwpd;
621*4882a593Smuzhiyun 	struct list_head fwlog_buff_list;
622*4882a593Smuzhiyun 
623*4882a593Smuzhiyun 	/* RAS support status on adapter */
624*4882a593Smuzhiyun 	bool ras_hwsupport; /* RAS Support available on HW or not */
625*4882a593Smuzhiyun 	bool ras_enabled;   /* Ras Enabled for the function */
626*4882a593Smuzhiyun #define LPFC_RAS_DISABLE_LOGGING 0x00
627*4882a593Smuzhiyun #define LPFC_RAS_ENABLE_LOGGING 0x01
628*4882a593Smuzhiyun 	enum ras_state state;    /* RAS logging running state */
629*4882a593Smuzhiyun };
630*4882a593Smuzhiyun 
631*4882a593Smuzhiyun #define DBG_LOG_STR_SZ 256
632*4882a593Smuzhiyun #define DBG_LOG_SZ 256
633*4882a593Smuzhiyun 
634*4882a593Smuzhiyun struct dbg_log_ent {
635*4882a593Smuzhiyun 	char log[DBG_LOG_STR_SZ];
636*4882a593Smuzhiyun 	u64     t_ns;
637*4882a593Smuzhiyun };
638*4882a593Smuzhiyun 
639*4882a593Smuzhiyun enum lpfc_irq_chann_mode {
640*4882a593Smuzhiyun 	/* Assign IRQs to all possible cpus that have hardware queues */
641*4882a593Smuzhiyun 	NORMAL_MODE,
642*4882a593Smuzhiyun 
643*4882a593Smuzhiyun 	/* Assign IRQs only to cpus on the same numa node as HBA */
644*4882a593Smuzhiyun 	NUMA_MODE,
645*4882a593Smuzhiyun 
646*4882a593Smuzhiyun 	/* Assign IRQs only on non-hyperthreaded CPUs. This is the
647*4882a593Smuzhiyun 	 * same as normal_mode, but assign IRQS only on physical CPUs.
648*4882a593Smuzhiyun 	 */
649*4882a593Smuzhiyun 	NHT_MODE,
650*4882a593Smuzhiyun };
651*4882a593Smuzhiyun 
652*4882a593Smuzhiyun struct lpfc_hba {
653*4882a593Smuzhiyun 	/* SCSI interface function jump table entries */
654*4882a593Smuzhiyun 	struct lpfc_io_buf * (*lpfc_get_scsi_buf)
655*4882a593Smuzhiyun 		(struct lpfc_hba *phba, struct lpfc_nodelist *ndlp,
656*4882a593Smuzhiyun 		struct scsi_cmnd *cmnd);
657*4882a593Smuzhiyun 	int (*lpfc_scsi_prep_dma_buf)
658*4882a593Smuzhiyun 		(struct lpfc_hba *, struct lpfc_io_buf *);
659*4882a593Smuzhiyun 	void (*lpfc_scsi_unprep_dma_buf)
660*4882a593Smuzhiyun 		(struct lpfc_hba *, struct lpfc_io_buf *);
661*4882a593Smuzhiyun 	void (*lpfc_release_scsi_buf)
662*4882a593Smuzhiyun 		(struct lpfc_hba *, struct lpfc_io_buf *);
663*4882a593Smuzhiyun 	void (*lpfc_rampdown_queue_depth)
664*4882a593Smuzhiyun 		(struct lpfc_hba *);
665*4882a593Smuzhiyun 	void (*lpfc_scsi_prep_cmnd)
666*4882a593Smuzhiyun 		(struct lpfc_vport *, struct lpfc_io_buf *,
667*4882a593Smuzhiyun 		 struct lpfc_nodelist *);
668*4882a593Smuzhiyun 
669*4882a593Smuzhiyun 	/* IOCB interface function jump table entries */
670*4882a593Smuzhiyun 	int (*__lpfc_sli_issue_iocb)
671*4882a593Smuzhiyun 		(struct lpfc_hba *, uint32_t,
672*4882a593Smuzhiyun 		 struct lpfc_iocbq *, uint32_t);
673*4882a593Smuzhiyun 	void (*__lpfc_sli_release_iocbq)(struct lpfc_hba *,
674*4882a593Smuzhiyun 			 struct lpfc_iocbq *);
675*4882a593Smuzhiyun 	int (*lpfc_hba_down_post)(struct lpfc_hba *phba);
676*4882a593Smuzhiyun 	IOCB_t * (*lpfc_get_iocb_from_iocbq)
677*4882a593Smuzhiyun 		(struct lpfc_iocbq *);
678*4882a593Smuzhiyun 	void (*lpfc_scsi_cmd_iocb_cmpl)
679*4882a593Smuzhiyun 		(struct lpfc_hba *, struct lpfc_iocbq *, struct lpfc_iocbq *);
680*4882a593Smuzhiyun 
681*4882a593Smuzhiyun 	/* MBOX interface function jump table entries */
682*4882a593Smuzhiyun 	int (*lpfc_sli_issue_mbox)
683*4882a593Smuzhiyun 		(struct lpfc_hba *, LPFC_MBOXQ_t *, uint32_t);
684*4882a593Smuzhiyun 
685*4882a593Smuzhiyun 	/* Slow-path IOCB process function jump table entries */
686*4882a593Smuzhiyun 	void (*lpfc_sli_handle_slow_ring_event)
687*4882a593Smuzhiyun 		(struct lpfc_hba *phba, struct lpfc_sli_ring *pring,
688*4882a593Smuzhiyun 		 uint32_t mask);
689*4882a593Smuzhiyun 
690*4882a593Smuzhiyun 	/* INIT device interface function jump table entries */
691*4882a593Smuzhiyun 	int (*lpfc_sli_hbq_to_firmware)
692*4882a593Smuzhiyun 		(struct lpfc_hba *, uint32_t, struct hbq_dmabuf *);
693*4882a593Smuzhiyun 	int (*lpfc_sli_brdrestart)
694*4882a593Smuzhiyun 		(struct lpfc_hba *);
695*4882a593Smuzhiyun 	int (*lpfc_sli_brdready)
696*4882a593Smuzhiyun 		(struct lpfc_hba *, uint32_t);
697*4882a593Smuzhiyun 	void (*lpfc_handle_eratt)
698*4882a593Smuzhiyun 		(struct lpfc_hba *);
699*4882a593Smuzhiyun 	void (*lpfc_stop_port)
700*4882a593Smuzhiyun 		(struct lpfc_hba *);
701*4882a593Smuzhiyun 	int (*lpfc_hba_init_link)
702*4882a593Smuzhiyun 		(struct lpfc_hba *, uint32_t);
703*4882a593Smuzhiyun 	int (*lpfc_hba_down_link)
704*4882a593Smuzhiyun 		(struct lpfc_hba *, uint32_t);
705*4882a593Smuzhiyun 	int (*lpfc_selective_reset)
706*4882a593Smuzhiyun 		(struct lpfc_hba *);
707*4882a593Smuzhiyun 
708*4882a593Smuzhiyun 	int (*lpfc_bg_scsi_prep_dma_buf)
709*4882a593Smuzhiyun 		(struct lpfc_hba *, struct lpfc_io_buf *);
710*4882a593Smuzhiyun 	/* Add new entries here */
711*4882a593Smuzhiyun 
712*4882a593Smuzhiyun 	/* expedite pool */
713*4882a593Smuzhiyun 	struct lpfc_epd_pool epd_pool;
714*4882a593Smuzhiyun 
715*4882a593Smuzhiyun 	/* SLI4 specific HBA data structure */
716*4882a593Smuzhiyun 	struct lpfc_sli4_hba sli4_hba;
717*4882a593Smuzhiyun 
718*4882a593Smuzhiyun 	struct workqueue_struct *wq;
719*4882a593Smuzhiyun 	struct delayed_work     eq_delay_work;
720*4882a593Smuzhiyun 
721*4882a593Smuzhiyun #define LPFC_IDLE_STAT_DELAY 1000
722*4882a593Smuzhiyun 	struct delayed_work	idle_stat_delay_work;
723*4882a593Smuzhiyun 
724*4882a593Smuzhiyun 	struct lpfc_sli sli;
725*4882a593Smuzhiyun 	uint8_t pci_dev_grp;	/* lpfc PCI dev group: 0x0, 0x1, 0x2,... */
726*4882a593Smuzhiyun 	uint32_t sli_rev;		/* SLI2, SLI3, or SLI4 */
727*4882a593Smuzhiyun 	uint32_t sli3_options;		/* Mask of enabled SLI3 options */
728*4882a593Smuzhiyun #define LPFC_SLI3_HBQ_ENABLED		0x01
729*4882a593Smuzhiyun #define LPFC_SLI3_NPIV_ENABLED		0x02
730*4882a593Smuzhiyun #define LPFC_SLI3_VPORT_TEARDOWN	0x04
731*4882a593Smuzhiyun #define LPFC_SLI3_CRP_ENABLED		0x08
732*4882a593Smuzhiyun #define LPFC_SLI3_BG_ENABLED		0x20
733*4882a593Smuzhiyun #define LPFC_SLI3_DSS_ENABLED		0x40
734*4882a593Smuzhiyun #define LPFC_SLI4_PERFH_ENABLED		0x80
735*4882a593Smuzhiyun #define LPFC_SLI4_PHWQ_ENABLED		0x100
736*4882a593Smuzhiyun 	uint32_t iocb_cmd_size;
737*4882a593Smuzhiyun 	uint32_t iocb_rsp_size;
738*4882a593Smuzhiyun 
739*4882a593Smuzhiyun 	struct lpfc_trunk_link  trunk_link;
740*4882a593Smuzhiyun 	enum hba_state link_state;
741*4882a593Smuzhiyun 	uint32_t link_flag;	/* link state flags */
742*4882a593Smuzhiyun #define LS_LOOPBACK_MODE      0x1	/* NPort is in Loopback mode */
743*4882a593Smuzhiyun 					/* This flag is set while issuing */
744*4882a593Smuzhiyun 					/* INIT_LINK mailbox command */
745*4882a593Smuzhiyun #define LS_NPIV_FAB_SUPPORTED 0x2	/* Fabric supports NPIV */
746*4882a593Smuzhiyun #define LS_IGNORE_ERATT       0x4	/* intr handler should ignore ERATT */
747*4882a593Smuzhiyun #define LS_MDS_LINK_DOWN      0x8	/* MDS Diagnostics Link Down */
748*4882a593Smuzhiyun #define LS_MDS_LOOPBACK      0x10	/* MDS Diagnostics Link Up (Loopback) */
749*4882a593Smuzhiyun 
750*4882a593Smuzhiyun 	uint32_t hba_flag;	/* hba generic flags */
751*4882a593Smuzhiyun #define HBA_ERATT_HANDLED	0x1 /* This flag is set when eratt handled */
752*4882a593Smuzhiyun #define DEFER_ERATT		0x2 /* Deferred error attention in progress */
753*4882a593Smuzhiyun #define HBA_FCOE_MODE		0x4 /* HBA function in FCoE Mode */
754*4882a593Smuzhiyun #define HBA_SP_QUEUE_EVT	0x8 /* Slow-path qevt posted to worker thread*/
755*4882a593Smuzhiyun #define HBA_POST_RECEIVE_BUFFER 0x10 /* Rcv buffers need to be posted */
756*4882a593Smuzhiyun #define HBA_PERSISTENT_TOPO	0x20 /* Persistent topology support in hba */
757*4882a593Smuzhiyun #define ELS_XRI_ABORT_EVENT	0x40 /* ELS_XRI abort event was queued */
758*4882a593Smuzhiyun #define ASYNC_EVENT		0x80
759*4882a593Smuzhiyun #define LINK_DISABLED		0x100 /* Link disabled by user */
760*4882a593Smuzhiyun #define FCF_TS_INPROG           0x200 /* FCF table scan in progress */
761*4882a593Smuzhiyun #define FCF_RR_INPROG           0x400 /* FCF roundrobin flogi in progress */
762*4882a593Smuzhiyun #define HBA_FIP_SUPPORT		0x800 /* FIP support in HBA */
763*4882a593Smuzhiyun #define HBA_AER_ENABLED		0x1000 /* AER enabled with HBA */
764*4882a593Smuzhiyun #define HBA_DEVLOSS_TMO         0x2000 /* HBA in devloss timeout */
765*4882a593Smuzhiyun #define HBA_RRQ_ACTIVE		0x4000 /* process the rrq active list */
766*4882a593Smuzhiyun #define HBA_IOQ_FLUSH		0x8000 /* FCP/NVME I/O queues being flushed */
767*4882a593Smuzhiyun #define HBA_RECOVERABLE_UE	0x20000 /* Firmware supports recoverable UE */
768*4882a593Smuzhiyun #define HBA_FORCED_LINK_SPEED	0x40000 /*
769*4882a593Smuzhiyun 					 * Firmware supports Forced Link Speed
770*4882a593Smuzhiyun 					 * capability
771*4882a593Smuzhiyun 					 */
772*4882a593Smuzhiyun #define HBA_FLOGI_ISSUED	0x100000 /* FLOGI was issued */
773*4882a593Smuzhiyun #define HBA_DEFER_FLOGI		0x800000 /* Defer FLOGI till read_sparm cmpl */
774*4882a593Smuzhiyun 
775*4882a593Smuzhiyun 	struct completion *fw_dump_cmpl; /* cmpl event tracker for fw_dump */
776*4882a593Smuzhiyun 	uint32_t fcp_ring_in_use; /* When polling test if intr-hndlr active*/
777*4882a593Smuzhiyun 	struct lpfc_dmabuf slim2p;
778*4882a593Smuzhiyun 
779*4882a593Smuzhiyun 	MAILBOX_t *mbox;
780*4882a593Smuzhiyun 	uint32_t *mbox_ext;
781*4882a593Smuzhiyun 	struct lpfc_mbox_ext_buf_ctx mbox_ext_buf_ctx;
782*4882a593Smuzhiyun 	uint32_t ha_copy;
783*4882a593Smuzhiyun 	struct _PCB *pcb;
784*4882a593Smuzhiyun 	struct _IOCB *IOCBs;
785*4882a593Smuzhiyun 
786*4882a593Smuzhiyun 	struct lpfc_dmabuf hbqslimp;
787*4882a593Smuzhiyun 
788*4882a593Smuzhiyun 	uint16_t pci_cfg_value;
789*4882a593Smuzhiyun 
790*4882a593Smuzhiyun 	uint8_t fc_linkspeed;	/* Link speed after last READ_LA */
791*4882a593Smuzhiyun 
792*4882a593Smuzhiyun 	uint32_t fc_eventTag;	/* event tag for link attention */
793*4882a593Smuzhiyun 	uint32_t link_events;
794*4882a593Smuzhiyun 
795*4882a593Smuzhiyun 	/* These fields used to be binfo */
796*4882a593Smuzhiyun 	uint32_t fc_pref_DID;	/* preferred D_ID */
797*4882a593Smuzhiyun 	uint8_t  fc_pref_ALPA;	/* preferred AL_PA */
798*4882a593Smuzhiyun 	uint32_t fc_edtovResol; /* E_D_TOV timer resolution */
799*4882a593Smuzhiyun 	uint32_t fc_edtov;	/* E_D_TOV timer value */
800*4882a593Smuzhiyun 	uint32_t fc_arbtov;	/* ARB_TOV timer value */
801*4882a593Smuzhiyun 	uint32_t fc_ratov;	/* R_A_TOV timer value */
802*4882a593Smuzhiyun 	uint32_t fc_rttov;	/* R_T_TOV timer value */
803*4882a593Smuzhiyun 	uint32_t fc_altov;	/* AL_TOV timer value */
804*4882a593Smuzhiyun 	uint32_t fc_crtov;	/* C_R_TOV timer value */
805*4882a593Smuzhiyun 
806*4882a593Smuzhiyun 	struct serv_parm fc_fabparam;	/* fabric service parameters buffer */
807*4882a593Smuzhiyun 	uint8_t alpa_map[128];	/* AL_PA map from READ_LA */
808*4882a593Smuzhiyun 
809*4882a593Smuzhiyun 	uint32_t lmt;
810*4882a593Smuzhiyun 
811*4882a593Smuzhiyun 	uint32_t fc_topology;	/* link topology, from LINK INIT */
812*4882a593Smuzhiyun 	uint32_t fc_topology_changed;	/* link topology, from LINK INIT */
813*4882a593Smuzhiyun 
814*4882a593Smuzhiyun 	struct lpfc_stats fc_stat;
815*4882a593Smuzhiyun 
816*4882a593Smuzhiyun 	struct lpfc_nodelist fc_fcpnodev; /* nodelist entry for no device */
817*4882a593Smuzhiyun 	uint32_t nport_event_cnt;	/* timestamp for nlplist entry */
818*4882a593Smuzhiyun 
819*4882a593Smuzhiyun 	uint8_t  wwnn[8];
820*4882a593Smuzhiyun 	uint8_t  wwpn[8];
821*4882a593Smuzhiyun 	uint32_t RandomData[7];
822*4882a593Smuzhiyun 	uint8_t  fcp_embed_io;
823*4882a593Smuzhiyun 	uint8_t  nvme_support;	/* Firmware supports NVME */
824*4882a593Smuzhiyun 	uint8_t  nvmet_support;	/* driver supports NVMET */
825*4882a593Smuzhiyun #define LPFC_NVMET_MAX_PORTS	32
826*4882a593Smuzhiyun 	uint8_t  mds_diags_support;
827*4882a593Smuzhiyun 	uint8_t  bbcredit_support;
828*4882a593Smuzhiyun 	uint8_t  enab_exp_wqcq_pages;
829*4882a593Smuzhiyun 	u8	 nsler; /* Firmware supports FC-NVMe-2 SLER */
830*4882a593Smuzhiyun 
831*4882a593Smuzhiyun 	/* HBA Config Parameters */
832*4882a593Smuzhiyun 	uint32_t cfg_ack0;
833*4882a593Smuzhiyun 	uint32_t cfg_xri_rebalancing;
834*4882a593Smuzhiyun 	uint32_t cfg_xpsgl;
835*4882a593Smuzhiyun 	uint32_t cfg_enable_npiv;
836*4882a593Smuzhiyun 	uint32_t cfg_enable_rrq;
837*4882a593Smuzhiyun 	uint32_t cfg_topology;
838*4882a593Smuzhiyun 	uint32_t cfg_link_speed;
839*4882a593Smuzhiyun #define LPFC_FCF_FOV 1		/* Fast fcf failover */
840*4882a593Smuzhiyun #define LPFC_FCF_PRIORITY 2	/* Priority fcf failover */
841*4882a593Smuzhiyun 	uint32_t cfg_fcf_failover_policy;
842*4882a593Smuzhiyun 	uint32_t cfg_fcp_io_sched;
843*4882a593Smuzhiyun 	uint32_t cfg_ns_query;
844*4882a593Smuzhiyun 	uint32_t cfg_fcp2_no_tgt_reset;
845*4882a593Smuzhiyun 	uint32_t cfg_cr_delay;
846*4882a593Smuzhiyun 	uint32_t cfg_cr_count;
847*4882a593Smuzhiyun 	uint32_t cfg_multi_ring_support;
848*4882a593Smuzhiyun 	uint32_t cfg_multi_ring_rctl;
849*4882a593Smuzhiyun 	uint32_t cfg_multi_ring_type;
850*4882a593Smuzhiyun 	uint32_t cfg_poll;
851*4882a593Smuzhiyun 	uint32_t cfg_poll_tmo;
852*4882a593Smuzhiyun 	uint32_t cfg_task_mgmt_tmo;
853*4882a593Smuzhiyun 	uint32_t cfg_use_msi;
854*4882a593Smuzhiyun 	uint32_t cfg_auto_imax;
855*4882a593Smuzhiyun 	uint32_t cfg_fcp_imax;
856*4882a593Smuzhiyun 	uint32_t cfg_force_rscn;
857*4882a593Smuzhiyun 	uint32_t cfg_cq_poll_threshold;
858*4882a593Smuzhiyun 	uint32_t cfg_cq_max_proc_limit;
859*4882a593Smuzhiyun 	uint32_t cfg_fcp_cpu_map;
860*4882a593Smuzhiyun 	uint32_t cfg_fcp_mq_threshold;
861*4882a593Smuzhiyun 	uint32_t cfg_hdw_queue;
862*4882a593Smuzhiyun 	uint32_t cfg_irq_chann;
863*4882a593Smuzhiyun 	uint32_t cfg_suppress_rsp;
864*4882a593Smuzhiyun 	uint32_t cfg_nvme_oas;
865*4882a593Smuzhiyun 	uint32_t cfg_nvme_embed_cmd;
866*4882a593Smuzhiyun 	uint32_t cfg_nvmet_mrq_post;
867*4882a593Smuzhiyun 	uint32_t cfg_nvmet_mrq;
868*4882a593Smuzhiyun 	uint32_t cfg_enable_nvmet;
869*4882a593Smuzhiyun 	uint32_t cfg_nvme_enable_fb;
870*4882a593Smuzhiyun 	uint32_t cfg_nvmet_fb_size;
871*4882a593Smuzhiyun 	uint32_t cfg_total_seg_cnt;
872*4882a593Smuzhiyun 	uint32_t cfg_sg_seg_cnt;
873*4882a593Smuzhiyun 	uint32_t cfg_nvme_seg_cnt;
874*4882a593Smuzhiyun 	uint32_t cfg_scsi_seg_cnt;
875*4882a593Smuzhiyun 	uint32_t cfg_sg_dma_buf_size;
876*4882a593Smuzhiyun 	uint64_t cfg_soft_wwnn;
877*4882a593Smuzhiyun 	uint64_t cfg_soft_wwpn;
878*4882a593Smuzhiyun 	uint32_t cfg_hba_queue_depth;
879*4882a593Smuzhiyun 	uint32_t cfg_enable_hba_reset;
880*4882a593Smuzhiyun 	uint32_t cfg_enable_hba_heartbeat;
881*4882a593Smuzhiyun 	uint32_t cfg_fof;
882*4882a593Smuzhiyun 	uint32_t cfg_EnableXLane;
883*4882a593Smuzhiyun 	uint8_t cfg_oas_tgt_wwpn[8];
884*4882a593Smuzhiyun 	uint8_t cfg_oas_vpt_wwpn[8];
885*4882a593Smuzhiyun 	uint32_t cfg_oas_lun_state;
886*4882a593Smuzhiyun #define OAS_LUN_ENABLE	1
887*4882a593Smuzhiyun #define OAS_LUN_DISABLE	0
888*4882a593Smuzhiyun 	uint32_t cfg_oas_lun_status;
889*4882a593Smuzhiyun #define OAS_LUN_STATUS_EXISTS	0x01
890*4882a593Smuzhiyun 	uint32_t cfg_oas_flags;
891*4882a593Smuzhiyun #define OAS_FIND_ANY_VPORT	0x01
892*4882a593Smuzhiyun #define OAS_FIND_ANY_TARGET	0x02
893*4882a593Smuzhiyun #define OAS_LUN_VALID	0x04
894*4882a593Smuzhiyun 	uint32_t cfg_oas_priority;
895*4882a593Smuzhiyun 	uint32_t cfg_XLanePriority;
896*4882a593Smuzhiyun 	uint32_t cfg_enable_bg;
897*4882a593Smuzhiyun 	uint32_t cfg_prot_mask;
898*4882a593Smuzhiyun 	uint32_t cfg_prot_guard;
899*4882a593Smuzhiyun 	uint32_t cfg_hostmem_hgp;
900*4882a593Smuzhiyun 	uint32_t cfg_log_verbose;
901*4882a593Smuzhiyun 	uint32_t cfg_enable_fc4_type;
902*4882a593Smuzhiyun #define LPFC_ENABLE_FCP  1
903*4882a593Smuzhiyun #define LPFC_ENABLE_NVME 2
904*4882a593Smuzhiyun #define LPFC_ENABLE_BOTH 3
905*4882a593Smuzhiyun #if (IS_ENABLED(CONFIG_NVME_FC))
906*4882a593Smuzhiyun #define LPFC_MAX_ENBL_FC4_TYPE LPFC_ENABLE_BOTH
907*4882a593Smuzhiyun #define LPFC_DEF_ENBL_FC4_TYPE LPFC_ENABLE_BOTH
908*4882a593Smuzhiyun #else
909*4882a593Smuzhiyun #define LPFC_MAX_ENBL_FC4_TYPE LPFC_ENABLE_FCP
910*4882a593Smuzhiyun #define LPFC_DEF_ENBL_FC4_TYPE LPFC_ENABLE_FCP
911*4882a593Smuzhiyun #endif
912*4882a593Smuzhiyun 	uint32_t cfg_aer_support;
913*4882a593Smuzhiyun 	uint32_t cfg_sriov_nr_virtfn;
914*4882a593Smuzhiyun 	uint32_t cfg_request_firmware_upgrade;
915*4882a593Smuzhiyun 	uint32_t cfg_suppress_link_up;
916*4882a593Smuzhiyun 	uint32_t cfg_rrq_xri_bitmap_sz;
917*4882a593Smuzhiyun 	uint32_t cfg_delay_discovery;
918*4882a593Smuzhiyun 	uint32_t cfg_sli_mode;
919*4882a593Smuzhiyun #define LPFC_INITIALIZE_LINK              0	/* do normal init_link mbox */
920*4882a593Smuzhiyun #define LPFC_DELAY_INIT_LINK              1	/* layered driver hold off */
921*4882a593Smuzhiyun #define LPFC_DELAY_INIT_LINK_INDEFINITELY 2	/* wait, manual intervention */
922*4882a593Smuzhiyun 	uint32_t cfg_fdmi_on;
923*4882a593Smuzhiyun #define LPFC_FDMI_NO_SUPPORT	0	/* FDMI not supported */
924*4882a593Smuzhiyun #define LPFC_FDMI_SUPPORT	1	/* FDMI supported? */
925*4882a593Smuzhiyun 	uint32_t cfg_enable_SmartSAN;
926*4882a593Smuzhiyun 	uint32_t cfg_enable_mds_diags;
927*4882a593Smuzhiyun 	uint32_t cfg_ras_fwlog_level;
928*4882a593Smuzhiyun 	uint32_t cfg_ras_fwlog_buffsize;
929*4882a593Smuzhiyun 	uint32_t cfg_ras_fwlog_func;
930*4882a593Smuzhiyun 	uint32_t cfg_enable_bbcr;	/* Enable BB Credit Recovery */
931*4882a593Smuzhiyun 	uint32_t cfg_enable_dpp;	/* Enable Direct Packet Push */
932*4882a593Smuzhiyun 	uint32_t cfg_enable_pbde;
933*4882a593Smuzhiyun 	struct nvmet_fc_target_port *targetport;
934*4882a593Smuzhiyun 	lpfc_vpd_t vpd;		/* vital product data */
935*4882a593Smuzhiyun 
936*4882a593Smuzhiyun 	struct pci_dev *pcidev;
937*4882a593Smuzhiyun 	struct list_head      work_list;
938*4882a593Smuzhiyun 	uint32_t              work_ha;      /* Host Attention Bits for WT */
939*4882a593Smuzhiyun 	uint32_t              work_ha_mask; /* HA Bits owned by WT        */
940*4882a593Smuzhiyun 	uint32_t              work_hs;      /* HS stored in case of ERRAT */
941*4882a593Smuzhiyun 	uint32_t              work_status[2]; /* Extra status from SLIM */
942*4882a593Smuzhiyun 
943*4882a593Smuzhiyun 	wait_queue_head_t    work_waitq;
944*4882a593Smuzhiyun 	struct task_struct   *worker_thread;
945*4882a593Smuzhiyun 	unsigned long data_flags;
946*4882a593Smuzhiyun 	uint32_t border_sge_num;
947*4882a593Smuzhiyun 
948*4882a593Smuzhiyun 	uint32_t hbq_in_use;		/* HBQs in use flag */
949*4882a593Smuzhiyun 	uint32_t hbq_count;	        /* Count of configured HBQs */
950*4882a593Smuzhiyun 	struct hbq_s hbqs[LPFC_MAX_HBQS]; /* local copy of hbq indicies  */
951*4882a593Smuzhiyun 
952*4882a593Smuzhiyun 	atomic_t fcp_qidx;         /* next FCP WQ (RR Policy) */
953*4882a593Smuzhiyun 	atomic_t nvme_qidx;        /* next NVME WQ (RR Policy) */
954*4882a593Smuzhiyun 
955*4882a593Smuzhiyun 	phys_addr_t pci_bar0_map;     /* Physical address for PCI BAR0 */
956*4882a593Smuzhiyun 	phys_addr_t pci_bar1_map;     /* Physical address for PCI BAR1 */
957*4882a593Smuzhiyun 	phys_addr_t pci_bar2_map;     /* Physical address for PCI BAR2 */
958*4882a593Smuzhiyun 	void __iomem *slim_memmap_p;	/* Kernel memory mapped address for
959*4882a593Smuzhiyun 					   PCI BAR0 */
960*4882a593Smuzhiyun 	void __iomem *ctrl_regs_memmap_p;/* Kernel memory mapped address for
961*4882a593Smuzhiyun 					    PCI BAR2 */
962*4882a593Smuzhiyun 
963*4882a593Smuzhiyun 	void __iomem *pci_bar0_memmap_p; /* Kernel memory mapped address for
964*4882a593Smuzhiyun 					    PCI BAR0 with dual-ULP support */
965*4882a593Smuzhiyun 	void __iomem *pci_bar2_memmap_p; /* Kernel memory mapped address for
966*4882a593Smuzhiyun 					    PCI BAR2 with dual-ULP support */
967*4882a593Smuzhiyun 	void __iomem *pci_bar4_memmap_p; /* Kernel memory mapped address for
968*4882a593Smuzhiyun 					    PCI BAR4 with dual-ULP support */
969*4882a593Smuzhiyun #define PCI_64BIT_BAR0	0
970*4882a593Smuzhiyun #define PCI_64BIT_BAR2	2
971*4882a593Smuzhiyun #define PCI_64BIT_BAR4	4
972*4882a593Smuzhiyun 	void __iomem *MBslimaddr;	/* virtual address for mbox cmds */
973*4882a593Smuzhiyun 	void __iomem *HAregaddr;	/* virtual address for host attn reg */
974*4882a593Smuzhiyun 	void __iomem *CAregaddr;	/* virtual address for chip attn reg */
975*4882a593Smuzhiyun 	void __iomem *HSregaddr;	/* virtual address for host status
976*4882a593Smuzhiyun 					   reg */
977*4882a593Smuzhiyun 	void __iomem *HCregaddr;	/* virtual address for host ctl reg */
978*4882a593Smuzhiyun 
979*4882a593Smuzhiyun 	struct lpfc_hgp __iomem *host_gp; /* Host side get/put pointers */
980*4882a593Smuzhiyun 	struct lpfc_pgp   *port_gp;
981*4882a593Smuzhiyun 	uint32_t __iomem  *hbq_put;     /* Address in SLIM to HBQ put ptrs */
982*4882a593Smuzhiyun 	uint32_t          *hbq_get;     /* Host mem address of HBQ get ptrs */
983*4882a593Smuzhiyun 
984*4882a593Smuzhiyun 	int brd_no;			/* FC board number */
985*4882a593Smuzhiyun 	char SerialNumber[32];		/* adapter Serial Number */
986*4882a593Smuzhiyun 	char OptionROMVersion[32];	/* adapter BIOS / Fcode version */
987*4882a593Smuzhiyun 	char BIOSVersion[16];		/* Boot BIOS version */
988*4882a593Smuzhiyun 	char ModelDesc[256];		/* Model Description */
989*4882a593Smuzhiyun 	char ModelName[80];		/* Model Name */
990*4882a593Smuzhiyun 	char ProgramType[256];		/* Program Type */
991*4882a593Smuzhiyun 	char Port[20];			/* Port No */
992*4882a593Smuzhiyun 	uint8_t vpd_flag;               /* VPD data flag */
993*4882a593Smuzhiyun 
994*4882a593Smuzhiyun #define VPD_MODEL_DESC      0x1         /* valid vpd model description */
995*4882a593Smuzhiyun #define VPD_MODEL_NAME      0x2         /* valid vpd model name */
996*4882a593Smuzhiyun #define VPD_PROGRAM_TYPE    0x4         /* valid vpd program type */
997*4882a593Smuzhiyun #define VPD_PORT            0x8         /* valid vpd port data */
998*4882a593Smuzhiyun #define VPD_MASK            0xf         /* mask for any vpd data */
999*4882a593Smuzhiyun 
1000*4882a593Smuzhiyun 	uint8_t soft_wwn_enable;
1001*4882a593Smuzhiyun 
1002*4882a593Smuzhiyun 	struct timer_list fcp_poll_timer;
1003*4882a593Smuzhiyun 	struct timer_list eratt_poll;
1004*4882a593Smuzhiyun 	uint32_t eratt_poll_interval;
1005*4882a593Smuzhiyun 
1006*4882a593Smuzhiyun 	uint64_t bg_guard_err_cnt;
1007*4882a593Smuzhiyun 	uint64_t bg_apptag_err_cnt;
1008*4882a593Smuzhiyun 	uint64_t bg_reftag_err_cnt;
1009*4882a593Smuzhiyun 
1010*4882a593Smuzhiyun 	/* fastpath list. */
1011*4882a593Smuzhiyun 	spinlock_t scsi_buf_list_get_lock;  /* SCSI buf alloc list lock */
1012*4882a593Smuzhiyun 	spinlock_t scsi_buf_list_put_lock;  /* SCSI buf free list lock */
1013*4882a593Smuzhiyun 	struct list_head lpfc_scsi_buf_list_get;
1014*4882a593Smuzhiyun 	struct list_head lpfc_scsi_buf_list_put;
1015*4882a593Smuzhiyun 	uint32_t total_scsi_bufs;
1016*4882a593Smuzhiyun 	struct list_head lpfc_iocb_list;
1017*4882a593Smuzhiyun 	uint32_t total_iocbq_bufs;
1018*4882a593Smuzhiyun 	struct list_head active_rrq_list;
1019*4882a593Smuzhiyun 	spinlock_t hbalock;
1020*4882a593Smuzhiyun 
1021*4882a593Smuzhiyun 	/* dma_mem_pools */
1022*4882a593Smuzhiyun 	struct dma_pool *lpfc_sg_dma_buf_pool;
1023*4882a593Smuzhiyun 	struct dma_pool *lpfc_mbuf_pool;
1024*4882a593Smuzhiyun 	struct dma_pool *lpfc_hrb_pool;	/* header receive buffer pool */
1025*4882a593Smuzhiyun 	struct dma_pool *lpfc_drb_pool; /* data receive buffer pool */
1026*4882a593Smuzhiyun 	struct dma_pool *lpfc_nvmet_drb_pool; /* data receive buffer pool */
1027*4882a593Smuzhiyun 	struct dma_pool *lpfc_hbq_pool;	/* SLI3 hbq buffer pool */
1028*4882a593Smuzhiyun 	struct dma_pool *lpfc_cmd_rsp_buf_pool;
1029*4882a593Smuzhiyun 	struct lpfc_dma_pool lpfc_mbuf_safety_pool;
1030*4882a593Smuzhiyun 
1031*4882a593Smuzhiyun 	mempool_t *mbox_mem_pool;
1032*4882a593Smuzhiyun 	mempool_t *nlp_mem_pool;
1033*4882a593Smuzhiyun 	mempool_t *rrq_pool;
1034*4882a593Smuzhiyun 	mempool_t *active_rrq_pool;
1035*4882a593Smuzhiyun 
1036*4882a593Smuzhiyun 	struct fc_host_statistics link_stats;
1037*4882a593Smuzhiyun 	enum lpfc_irq_chann_mode irq_chann_mode;
1038*4882a593Smuzhiyun 	enum intr_type_t intr_type;
1039*4882a593Smuzhiyun 	uint32_t intr_mode;
1040*4882a593Smuzhiyun #define LPFC_INTR_ERROR	0xFFFFFFFF
1041*4882a593Smuzhiyun 	struct list_head port_list;
1042*4882a593Smuzhiyun 	spinlock_t port_list_lock;	/* lock for port_list mutations */
1043*4882a593Smuzhiyun 	struct lpfc_vport *pport;	/* physical lpfc_vport pointer */
1044*4882a593Smuzhiyun 	uint16_t max_vpi;		/* Maximum virtual nports */
1045*4882a593Smuzhiyun #define LPFC_MAX_VPI	0xFF		/* Max number VPI supported 0 - 0xff */
1046*4882a593Smuzhiyun #define LPFC_MAX_VPORTS	0x100		/* Max vports per port, with pport */
1047*4882a593Smuzhiyun 	uint16_t max_vports;            /*
1048*4882a593Smuzhiyun 					 * For IOV HBAs max_vpi can change
1049*4882a593Smuzhiyun 					 * after a reset. max_vports is max
1050*4882a593Smuzhiyun 					 * number of vports present. This can
1051*4882a593Smuzhiyun 					 * be greater than max_vpi.
1052*4882a593Smuzhiyun 					 */
1053*4882a593Smuzhiyun 	uint16_t vpi_base;
1054*4882a593Smuzhiyun 	uint16_t vfi_base;
1055*4882a593Smuzhiyun 	unsigned long *vpi_bmask;	/* vpi allocation table */
1056*4882a593Smuzhiyun 	uint16_t *vpi_ids;
1057*4882a593Smuzhiyun 	uint16_t vpi_count;
1058*4882a593Smuzhiyun 	struct list_head lpfc_vpi_blk_list;
1059*4882a593Smuzhiyun 
1060*4882a593Smuzhiyun 	/* Data structure used by fabric iocb scheduler */
1061*4882a593Smuzhiyun 	struct list_head fabric_iocb_list;
1062*4882a593Smuzhiyun 	atomic_t fabric_iocb_count;
1063*4882a593Smuzhiyun 	struct timer_list fabric_block_timer;
1064*4882a593Smuzhiyun 	unsigned long bit_flags;
1065*4882a593Smuzhiyun #define	FABRIC_COMANDS_BLOCKED	0
1066*4882a593Smuzhiyun 	atomic_t num_rsrc_err;
1067*4882a593Smuzhiyun 	atomic_t num_cmd_success;
1068*4882a593Smuzhiyun 	unsigned long last_rsrc_error_time;
1069*4882a593Smuzhiyun 	unsigned long last_ramp_down_time;
1070*4882a593Smuzhiyun #ifdef CONFIG_SCSI_LPFC_DEBUG_FS
1071*4882a593Smuzhiyun 	struct dentry *hba_debugfs_root;
1072*4882a593Smuzhiyun 	atomic_t debugfs_vport_count;
1073*4882a593Smuzhiyun 	struct dentry *debug_multixri_pools;
1074*4882a593Smuzhiyun 	struct dentry *debug_hbqinfo;
1075*4882a593Smuzhiyun 	struct dentry *debug_dumpHostSlim;
1076*4882a593Smuzhiyun 	struct dentry *debug_dumpHBASlim;
1077*4882a593Smuzhiyun 	struct dentry *debug_InjErrLBA;  /* LBA to inject errors at */
1078*4882a593Smuzhiyun 	struct dentry *debug_InjErrNPortID;  /* NPortID to inject errors at */
1079*4882a593Smuzhiyun 	struct dentry *debug_InjErrWWPN;  /* WWPN to inject errors at */
1080*4882a593Smuzhiyun 	struct dentry *debug_writeGuard; /* inject write guard_tag errors */
1081*4882a593Smuzhiyun 	struct dentry *debug_writeApp;   /* inject write app_tag errors */
1082*4882a593Smuzhiyun 	struct dentry *debug_writeRef;   /* inject write ref_tag errors */
1083*4882a593Smuzhiyun 	struct dentry *debug_readGuard;  /* inject read guard_tag errors */
1084*4882a593Smuzhiyun 	struct dentry *debug_readApp;    /* inject read app_tag errors */
1085*4882a593Smuzhiyun 	struct dentry *debug_readRef;    /* inject read ref_tag errors */
1086*4882a593Smuzhiyun 
1087*4882a593Smuzhiyun 	struct dentry *debug_nvmeio_trc;
1088*4882a593Smuzhiyun 	struct lpfc_debugfs_nvmeio_trc *nvmeio_trc;
1089*4882a593Smuzhiyun 	struct dentry *debug_hdwqinfo;
1090*4882a593Smuzhiyun #ifdef LPFC_HDWQ_LOCK_STAT
1091*4882a593Smuzhiyun 	struct dentry *debug_lockstat;
1092*4882a593Smuzhiyun #endif
1093*4882a593Smuzhiyun 	struct dentry *debug_ras_log;
1094*4882a593Smuzhiyun 	atomic_t nvmeio_trc_cnt;
1095*4882a593Smuzhiyun 	uint32_t nvmeio_trc_size;
1096*4882a593Smuzhiyun 	uint32_t nvmeio_trc_output_idx;
1097*4882a593Smuzhiyun 
1098*4882a593Smuzhiyun 	/* T10 DIF error injection */
1099*4882a593Smuzhiyun 	uint32_t lpfc_injerr_wgrd_cnt;
1100*4882a593Smuzhiyun 	uint32_t lpfc_injerr_wapp_cnt;
1101*4882a593Smuzhiyun 	uint32_t lpfc_injerr_wref_cnt;
1102*4882a593Smuzhiyun 	uint32_t lpfc_injerr_rgrd_cnt;
1103*4882a593Smuzhiyun 	uint32_t lpfc_injerr_rapp_cnt;
1104*4882a593Smuzhiyun 	uint32_t lpfc_injerr_rref_cnt;
1105*4882a593Smuzhiyun 	uint32_t lpfc_injerr_nportid;
1106*4882a593Smuzhiyun 	struct lpfc_name lpfc_injerr_wwpn;
1107*4882a593Smuzhiyun 	sector_t lpfc_injerr_lba;
1108*4882a593Smuzhiyun #define LPFC_INJERR_LBA_OFF	(sector_t)(-1)
1109*4882a593Smuzhiyun 
1110*4882a593Smuzhiyun 	struct dentry *debug_slow_ring_trc;
1111*4882a593Smuzhiyun 	struct lpfc_debugfs_trc *slow_ring_trc;
1112*4882a593Smuzhiyun 	atomic_t slow_ring_trc_cnt;
1113*4882a593Smuzhiyun 	/* iDiag debugfs sub-directory */
1114*4882a593Smuzhiyun 	struct dentry *idiag_root;
1115*4882a593Smuzhiyun 	struct dentry *idiag_pci_cfg;
1116*4882a593Smuzhiyun 	struct dentry *idiag_bar_acc;
1117*4882a593Smuzhiyun 	struct dentry *idiag_que_info;
1118*4882a593Smuzhiyun 	struct dentry *idiag_que_acc;
1119*4882a593Smuzhiyun 	struct dentry *idiag_drb_acc;
1120*4882a593Smuzhiyun 	struct dentry *idiag_ctl_acc;
1121*4882a593Smuzhiyun 	struct dentry *idiag_mbx_acc;
1122*4882a593Smuzhiyun 	struct dentry *idiag_ext_acc;
1123*4882a593Smuzhiyun 	uint8_t lpfc_idiag_last_eq;
1124*4882a593Smuzhiyun #endif
1125*4882a593Smuzhiyun 	uint16_t nvmeio_trc_on;
1126*4882a593Smuzhiyun 
1127*4882a593Smuzhiyun 	/* Used for deferred freeing of ELS data buffers */
1128*4882a593Smuzhiyun 	struct list_head elsbuf;
1129*4882a593Smuzhiyun 	int elsbuf_cnt;
1130*4882a593Smuzhiyun 	int elsbuf_prev_cnt;
1131*4882a593Smuzhiyun 
1132*4882a593Smuzhiyun 	uint8_t temp_sensor_support;
1133*4882a593Smuzhiyun 	/* Fields used for heart beat. */
1134*4882a593Smuzhiyun 	unsigned long last_completion_time;
1135*4882a593Smuzhiyun 	unsigned long skipped_hb;
1136*4882a593Smuzhiyun 	struct timer_list hb_tmofunc;
1137*4882a593Smuzhiyun 	uint8_t hb_outstanding;
1138*4882a593Smuzhiyun 	struct timer_list rrq_tmr;
1139*4882a593Smuzhiyun 	enum hba_temp_state over_temp_state;
1140*4882a593Smuzhiyun 	/* ndlp reference management */
1141*4882a593Smuzhiyun 	spinlock_t ndlp_lock;
1142*4882a593Smuzhiyun 	/*
1143*4882a593Smuzhiyun 	 * Following bit will be set for all buffer tags which are not
1144*4882a593Smuzhiyun 	 * associated with any HBQ.
1145*4882a593Smuzhiyun 	 */
1146*4882a593Smuzhiyun #define QUE_BUFTAG_BIT  (1<<31)
1147*4882a593Smuzhiyun 	uint32_t buffer_tag_count;
1148*4882a593Smuzhiyun 	int wait_4_mlo_maint_flg;
1149*4882a593Smuzhiyun 	wait_queue_head_t wait_4_mlo_m_q;
1150*4882a593Smuzhiyun 	/* data structure used for latency data collection */
1151*4882a593Smuzhiyun #define LPFC_NO_BUCKET	   0
1152*4882a593Smuzhiyun #define LPFC_LINEAR_BUCKET 1
1153*4882a593Smuzhiyun #define LPFC_POWER2_BUCKET 2
1154*4882a593Smuzhiyun 	uint8_t  bucket_type;
1155*4882a593Smuzhiyun 	uint32_t bucket_base;
1156*4882a593Smuzhiyun 	uint32_t bucket_step;
1157*4882a593Smuzhiyun 
1158*4882a593Smuzhiyun /* Maximum number of events that can be outstanding at any time*/
1159*4882a593Smuzhiyun #define LPFC_MAX_EVT_COUNT 512
1160*4882a593Smuzhiyun 	atomic_t fast_event_count;
1161*4882a593Smuzhiyun 	uint32_t fcoe_eventtag;
1162*4882a593Smuzhiyun 	uint32_t fcoe_eventtag_at_fcf_scan;
1163*4882a593Smuzhiyun 	uint32_t fcoe_cvl_eventtag;
1164*4882a593Smuzhiyun 	uint32_t fcoe_cvl_eventtag_attn;
1165*4882a593Smuzhiyun 	struct lpfc_fcf fcf;
1166*4882a593Smuzhiyun 	uint8_t fc_map[3];
1167*4882a593Smuzhiyun 	uint8_t valid_vlan;
1168*4882a593Smuzhiyun 	uint16_t vlan_id;
1169*4882a593Smuzhiyun 	struct list_head fcf_conn_rec_list;
1170*4882a593Smuzhiyun 
1171*4882a593Smuzhiyun 	bool defer_flogi_acc_flag;
1172*4882a593Smuzhiyun 	uint16_t defer_flogi_acc_rx_id;
1173*4882a593Smuzhiyun 	uint16_t defer_flogi_acc_ox_id;
1174*4882a593Smuzhiyun 
1175*4882a593Smuzhiyun 	spinlock_t ct_ev_lock; /* synchronize access to ct_ev_waiters */
1176*4882a593Smuzhiyun 	struct list_head ct_ev_waiters;
1177*4882a593Smuzhiyun 	struct unsol_rcv_ct_ctx ct_ctx[LPFC_CT_CTX_MAX];
1178*4882a593Smuzhiyun 	uint32_t ctx_idx;
1179*4882a593Smuzhiyun 
1180*4882a593Smuzhiyun 	/* RAS Support */
1181*4882a593Smuzhiyun 	struct lpfc_ras_fwlog ras_fwlog;
1182*4882a593Smuzhiyun 
1183*4882a593Smuzhiyun 	uint8_t menlo_flag;	/* menlo generic flags */
1184*4882a593Smuzhiyun #define HBA_MENLO_SUPPORT	0x1 /* HBA supports menlo commands */
1185*4882a593Smuzhiyun 	uint32_t iocb_cnt;
1186*4882a593Smuzhiyun 	uint32_t iocb_max;
1187*4882a593Smuzhiyun 	atomic_t sdev_cnt;
1188*4882a593Smuzhiyun 	spinlock_t devicelock;	/* lock for luns list */
1189*4882a593Smuzhiyun 	mempool_t *device_data_mem_pool;
1190*4882a593Smuzhiyun 	struct list_head luns;
1191*4882a593Smuzhiyun #define LPFC_TRANSGRESSION_HIGH_TEMPERATURE	0x0080
1192*4882a593Smuzhiyun #define LPFC_TRANSGRESSION_LOW_TEMPERATURE	0x0040
1193*4882a593Smuzhiyun #define LPFC_TRANSGRESSION_HIGH_VOLTAGE		0x0020
1194*4882a593Smuzhiyun #define LPFC_TRANSGRESSION_LOW_VOLTAGE		0x0010
1195*4882a593Smuzhiyun #define LPFC_TRANSGRESSION_HIGH_TXBIAS		0x0008
1196*4882a593Smuzhiyun #define LPFC_TRANSGRESSION_LOW_TXBIAS		0x0004
1197*4882a593Smuzhiyun #define LPFC_TRANSGRESSION_HIGH_TXPOWER		0x0002
1198*4882a593Smuzhiyun #define LPFC_TRANSGRESSION_LOW_TXPOWER		0x0001
1199*4882a593Smuzhiyun #define LPFC_TRANSGRESSION_HIGH_RXPOWER		0x8000
1200*4882a593Smuzhiyun #define LPFC_TRANSGRESSION_LOW_RXPOWER		0x4000
1201*4882a593Smuzhiyun 	uint16_t sfp_alarm;
1202*4882a593Smuzhiyun 	uint16_t sfp_warning;
1203*4882a593Smuzhiyun 
1204*4882a593Smuzhiyun #ifdef CONFIG_SCSI_LPFC_DEBUG_FS
1205*4882a593Smuzhiyun 	uint16_t hdwqstat_on;
1206*4882a593Smuzhiyun #define LPFC_CHECK_OFF		0
1207*4882a593Smuzhiyun #define LPFC_CHECK_NVME_IO	1
1208*4882a593Smuzhiyun #define LPFC_CHECK_NVMET_IO	2
1209*4882a593Smuzhiyun #define LPFC_CHECK_SCSI_IO	4
1210*4882a593Smuzhiyun 	uint16_t ktime_on;
1211*4882a593Smuzhiyun 	uint64_t ktime_data_samples;
1212*4882a593Smuzhiyun 	uint64_t ktime_status_samples;
1213*4882a593Smuzhiyun 	uint64_t ktime_last_cmd;
1214*4882a593Smuzhiyun 	uint64_t ktime_seg1_total;
1215*4882a593Smuzhiyun 	uint64_t ktime_seg1_min;
1216*4882a593Smuzhiyun 	uint64_t ktime_seg1_max;
1217*4882a593Smuzhiyun 	uint64_t ktime_seg2_total;
1218*4882a593Smuzhiyun 	uint64_t ktime_seg2_min;
1219*4882a593Smuzhiyun 	uint64_t ktime_seg2_max;
1220*4882a593Smuzhiyun 	uint64_t ktime_seg3_total;
1221*4882a593Smuzhiyun 	uint64_t ktime_seg3_min;
1222*4882a593Smuzhiyun 	uint64_t ktime_seg3_max;
1223*4882a593Smuzhiyun 	uint64_t ktime_seg4_total;
1224*4882a593Smuzhiyun 	uint64_t ktime_seg4_min;
1225*4882a593Smuzhiyun 	uint64_t ktime_seg4_max;
1226*4882a593Smuzhiyun 	uint64_t ktime_seg5_total;
1227*4882a593Smuzhiyun 	uint64_t ktime_seg5_min;
1228*4882a593Smuzhiyun 	uint64_t ktime_seg5_max;
1229*4882a593Smuzhiyun 	uint64_t ktime_seg6_total;
1230*4882a593Smuzhiyun 	uint64_t ktime_seg6_min;
1231*4882a593Smuzhiyun 	uint64_t ktime_seg6_max;
1232*4882a593Smuzhiyun 	uint64_t ktime_seg7_total;
1233*4882a593Smuzhiyun 	uint64_t ktime_seg7_min;
1234*4882a593Smuzhiyun 	uint64_t ktime_seg7_max;
1235*4882a593Smuzhiyun 	uint64_t ktime_seg8_total;
1236*4882a593Smuzhiyun 	uint64_t ktime_seg8_min;
1237*4882a593Smuzhiyun 	uint64_t ktime_seg8_max;
1238*4882a593Smuzhiyun 	uint64_t ktime_seg9_total;
1239*4882a593Smuzhiyun 	uint64_t ktime_seg9_min;
1240*4882a593Smuzhiyun 	uint64_t ktime_seg9_max;
1241*4882a593Smuzhiyun 	uint64_t ktime_seg10_total;
1242*4882a593Smuzhiyun 	uint64_t ktime_seg10_min;
1243*4882a593Smuzhiyun 	uint64_t ktime_seg10_max;
1244*4882a593Smuzhiyun #endif
1245*4882a593Smuzhiyun 
1246*4882a593Smuzhiyun 	struct hlist_node cpuhp;	/* used for cpuhp per hba callback */
1247*4882a593Smuzhiyun 	struct timer_list cpuhp_poll_timer;
1248*4882a593Smuzhiyun 	struct list_head poll_list;	/* slowpath eq polling list */
1249*4882a593Smuzhiyun #define LPFC_POLL_HB	1		/* slowpath heartbeat */
1250*4882a593Smuzhiyun #define LPFC_POLL_FASTPATH	0	/* called from fastpath */
1251*4882a593Smuzhiyun #define LPFC_POLL_SLOWPATH	1	/* called from slowpath */
1252*4882a593Smuzhiyun 
1253*4882a593Smuzhiyun 	char os_host_name[MAXHOSTNAMELEN];
1254*4882a593Smuzhiyun 
1255*4882a593Smuzhiyun 	/* SCSI host template information - for physical port */
1256*4882a593Smuzhiyun 	struct scsi_host_template port_template;
1257*4882a593Smuzhiyun 	/* SCSI host template information - for all vports */
1258*4882a593Smuzhiyun 	struct scsi_host_template vport_template;
1259*4882a593Smuzhiyun 	atomic_t dbg_log_idx;
1260*4882a593Smuzhiyun 	atomic_t dbg_log_cnt;
1261*4882a593Smuzhiyun 	atomic_t dbg_log_dmping;
1262*4882a593Smuzhiyun 	struct dbg_log_ent dbg_log[DBG_LOG_SZ];
1263*4882a593Smuzhiyun };
1264*4882a593Smuzhiyun 
1265*4882a593Smuzhiyun static inline struct Scsi_Host *
lpfc_shost_from_vport(struct lpfc_vport * vport)1266*4882a593Smuzhiyun lpfc_shost_from_vport(struct lpfc_vport *vport)
1267*4882a593Smuzhiyun {
1268*4882a593Smuzhiyun 	return container_of((void *) vport, struct Scsi_Host, hostdata[0]);
1269*4882a593Smuzhiyun }
1270*4882a593Smuzhiyun 
1271*4882a593Smuzhiyun static inline void
lpfc_set_loopback_flag(struct lpfc_hba * phba)1272*4882a593Smuzhiyun lpfc_set_loopback_flag(struct lpfc_hba *phba)
1273*4882a593Smuzhiyun {
1274*4882a593Smuzhiyun 	if (phba->cfg_topology == FLAGS_LOCAL_LB)
1275*4882a593Smuzhiyun 		phba->link_flag |= LS_LOOPBACK_MODE;
1276*4882a593Smuzhiyun 	else
1277*4882a593Smuzhiyun 		phba->link_flag &= ~LS_LOOPBACK_MODE;
1278*4882a593Smuzhiyun }
1279*4882a593Smuzhiyun 
1280*4882a593Smuzhiyun static inline int
lpfc_is_link_up(struct lpfc_hba * phba)1281*4882a593Smuzhiyun lpfc_is_link_up(struct lpfc_hba *phba)
1282*4882a593Smuzhiyun {
1283*4882a593Smuzhiyun 	return  phba->link_state == LPFC_LINK_UP ||
1284*4882a593Smuzhiyun 		phba->link_state == LPFC_CLEAR_LA ||
1285*4882a593Smuzhiyun 		phba->link_state == LPFC_HBA_READY;
1286*4882a593Smuzhiyun }
1287*4882a593Smuzhiyun 
1288*4882a593Smuzhiyun static inline void
lpfc_worker_wake_up(struct lpfc_hba * phba)1289*4882a593Smuzhiyun lpfc_worker_wake_up(struct lpfc_hba *phba)
1290*4882a593Smuzhiyun {
1291*4882a593Smuzhiyun 	/* Set the lpfc data pending flag */
1292*4882a593Smuzhiyun 	set_bit(LPFC_DATA_READY, &phba->data_flags);
1293*4882a593Smuzhiyun 
1294*4882a593Smuzhiyun 	/* Wake up worker thread */
1295*4882a593Smuzhiyun 	wake_up(&phba->work_waitq);
1296*4882a593Smuzhiyun 	return;
1297*4882a593Smuzhiyun }
1298*4882a593Smuzhiyun 
1299*4882a593Smuzhiyun static inline int
lpfc_readl(void __iomem * addr,uint32_t * data)1300*4882a593Smuzhiyun lpfc_readl(void __iomem *addr, uint32_t *data)
1301*4882a593Smuzhiyun {
1302*4882a593Smuzhiyun 	uint32_t temp;
1303*4882a593Smuzhiyun 	temp = readl(addr);
1304*4882a593Smuzhiyun 	if (temp == 0xffffffff)
1305*4882a593Smuzhiyun 		return -EIO;
1306*4882a593Smuzhiyun 	*data = temp;
1307*4882a593Smuzhiyun 	return 0;
1308*4882a593Smuzhiyun }
1309*4882a593Smuzhiyun 
1310*4882a593Smuzhiyun static inline int
lpfc_sli_read_hs(struct lpfc_hba * phba)1311*4882a593Smuzhiyun lpfc_sli_read_hs(struct lpfc_hba *phba)
1312*4882a593Smuzhiyun {
1313*4882a593Smuzhiyun 	/*
1314*4882a593Smuzhiyun 	 * There was a link/board error. Read the status register to retrieve
1315*4882a593Smuzhiyun 	 * the error event and process it.
1316*4882a593Smuzhiyun 	 */
1317*4882a593Smuzhiyun 	phba->sli.slistat.err_attn_event++;
1318*4882a593Smuzhiyun 
1319*4882a593Smuzhiyun 	/* Save status info and check for unplug error */
1320*4882a593Smuzhiyun 	if (lpfc_readl(phba->HSregaddr, &phba->work_hs) ||
1321*4882a593Smuzhiyun 		lpfc_readl(phba->MBslimaddr + 0xa8, &phba->work_status[0]) ||
1322*4882a593Smuzhiyun 		lpfc_readl(phba->MBslimaddr + 0xac, &phba->work_status[1])) {
1323*4882a593Smuzhiyun 		return -EIO;
1324*4882a593Smuzhiyun 	}
1325*4882a593Smuzhiyun 
1326*4882a593Smuzhiyun 	/* Clear chip Host Attention error bit */
1327*4882a593Smuzhiyun 	writel(HA_ERATT, phba->HAregaddr);
1328*4882a593Smuzhiyun 	readl(phba->HAregaddr); /* flush */
1329*4882a593Smuzhiyun 	phba->pport->stopped = 1;
1330*4882a593Smuzhiyun 
1331*4882a593Smuzhiyun 	return 0;
1332*4882a593Smuzhiyun }
1333*4882a593Smuzhiyun 
1334*4882a593Smuzhiyun static inline struct lpfc_sli_ring *
lpfc_phba_elsring(struct lpfc_hba * phba)1335*4882a593Smuzhiyun lpfc_phba_elsring(struct lpfc_hba *phba)
1336*4882a593Smuzhiyun {
1337*4882a593Smuzhiyun 	/* Return NULL if sli_rev has become invalid due to bad fw */
1338*4882a593Smuzhiyun 	if (phba->sli_rev != LPFC_SLI_REV4  &&
1339*4882a593Smuzhiyun 	    phba->sli_rev != LPFC_SLI_REV3  &&
1340*4882a593Smuzhiyun 	    phba->sli_rev != LPFC_SLI_REV2)
1341*4882a593Smuzhiyun 		return NULL;
1342*4882a593Smuzhiyun 
1343*4882a593Smuzhiyun 	if (phba->sli_rev == LPFC_SLI_REV4) {
1344*4882a593Smuzhiyun 		if (phba->sli4_hba.els_wq)
1345*4882a593Smuzhiyun 			return phba->sli4_hba.els_wq->pring;
1346*4882a593Smuzhiyun 		else
1347*4882a593Smuzhiyun 			return NULL;
1348*4882a593Smuzhiyun 	}
1349*4882a593Smuzhiyun 	return &phba->sli.sli3_ring[LPFC_ELS_RING];
1350*4882a593Smuzhiyun }
1351*4882a593Smuzhiyun 
1352*4882a593Smuzhiyun /**
1353*4882a593Smuzhiyun  * lpfc_next_online_cpu - Finds next online CPU on cpumask
1354*4882a593Smuzhiyun  * @mask: Pointer to phba's cpumask member.
1355*4882a593Smuzhiyun  * @start: starting cpu index
1356*4882a593Smuzhiyun  *
1357*4882a593Smuzhiyun  * Note: If no valid cpu found, then nr_cpu_ids is returned.
1358*4882a593Smuzhiyun  *
1359*4882a593Smuzhiyun  **/
1360*4882a593Smuzhiyun static inline unsigned int
lpfc_next_online_cpu(const struct cpumask * mask,unsigned int start)1361*4882a593Smuzhiyun lpfc_next_online_cpu(const struct cpumask *mask, unsigned int start)
1362*4882a593Smuzhiyun {
1363*4882a593Smuzhiyun 	unsigned int cpu_it;
1364*4882a593Smuzhiyun 
1365*4882a593Smuzhiyun 	for_each_cpu_wrap(cpu_it, mask, start) {
1366*4882a593Smuzhiyun 		if (cpu_online(cpu_it))
1367*4882a593Smuzhiyun 			break;
1368*4882a593Smuzhiyun 	}
1369*4882a593Smuzhiyun 
1370*4882a593Smuzhiyun 	return cpu_it;
1371*4882a593Smuzhiyun }
1372*4882a593Smuzhiyun /**
1373*4882a593Smuzhiyun  * lpfc_sli4_mod_hba_eq_delay - update EQ delay
1374*4882a593Smuzhiyun  * @phba: Pointer to HBA context object.
1375*4882a593Smuzhiyun  * @q: The Event Queue to update.
1376*4882a593Smuzhiyun  * @delay: The delay value (in us) to be written.
1377*4882a593Smuzhiyun  *
1378*4882a593Smuzhiyun  **/
1379*4882a593Smuzhiyun static inline void
lpfc_sli4_mod_hba_eq_delay(struct lpfc_hba * phba,struct lpfc_queue * eq,u32 delay)1380*4882a593Smuzhiyun lpfc_sli4_mod_hba_eq_delay(struct lpfc_hba *phba, struct lpfc_queue *eq,
1381*4882a593Smuzhiyun 			   u32 delay)
1382*4882a593Smuzhiyun {
1383*4882a593Smuzhiyun 	struct lpfc_register reg_data;
1384*4882a593Smuzhiyun 
1385*4882a593Smuzhiyun 	reg_data.word0 = 0;
1386*4882a593Smuzhiyun 	bf_set(lpfc_sliport_eqdelay_id, &reg_data, eq->queue_id);
1387*4882a593Smuzhiyun 	bf_set(lpfc_sliport_eqdelay_delay, &reg_data, delay);
1388*4882a593Smuzhiyun 	writel(reg_data.word0, phba->sli4_hba.u.if_type2.EQDregaddr);
1389*4882a593Smuzhiyun 	eq->q_mode = delay;
1390*4882a593Smuzhiyun }
1391*4882a593Smuzhiyun 
1392*4882a593Smuzhiyun 
1393*4882a593Smuzhiyun /*
1394*4882a593Smuzhiyun  * Macro that declares tables and a routine to perform enum type to
1395*4882a593Smuzhiyun  * ascii string lookup.
1396*4882a593Smuzhiyun  *
1397*4882a593Smuzhiyun  * Defines a <key,value> table for an enum. Uses xxx_INIT defines for
1398*4882a593Smuzhiyun  * the enum to populate the table.  Macro defines a routine (named
1399*4882a593Smuzhiyun  * by caller) that will search all elements of the table for the key
1400*4882a593Smuzhiyun  * and return the name string if found or "Unrecognized" if not found.
1401*4882a593Smuzhiyun  */
1402*4882a593Smuzhiyun #define DECLARE_ENUM2STR_LOOKUP(routine, enum_name, enum_init)		\
1403*4882a593Smuzhiyun static struct {								\
1404*4882a593Smuzhiyun 	enum enum_name		value;					\
1405*4882a593Smuzhiyun 	char			*name;					\
1406*4882a593Smuzhiyun } fc_##enum_name##_e2str_names[] = enum_init;				\
1407*4882a593Smuzhiyun static const char *routine(enum enum_name table_key)			\
1408*4882a593Smuzhiyun {									\
1409*4882a593Smuzhiyun 	int i;								\
1410*4882a593Smuzhiyun 	char *name = "Unrecognized";					\
1411*4882a593Smuzhiyun 									\
1412*4882a593Smuzhiyun 	for (i = 0; i < ARRAY_SIZE(fc_##enum_name##_e2str_names); i++) {\
1413*4882a593Smuzhiyun 		if (fc_##enum_name##_e2str_names[i].value == table_key) {\
1414*4882a593Smuzhiyun 			name = fc_##enum_name##_e2str_names[i].name;	\
1415*4882a593Smuzhiyun 			break;						\
1416*4882a593Smuzhiyun 		}							\
1417*4882a593Smuzhiyun 	}								\
1418*4882a593Smuzhiyun 	return name;							\
1419*4882a593Smuzhiyun }
1420