xref: /OK3568_Linux_fs/kernel/drivers/edac/i7core_edac.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun /* Intel i7 core/Nehalem Memory Controller kernel module
3*4882a593Smuzhiyun  *
4*4882a593Smuzhiyun  * This driver supports the memory controllers found on the Intel
5*4882a593Smuzhiyun  * processor families i7core, i7core 7xx/8xx, i5core, Xeon 35xx,
6*4882a593Smuzhiyun  * Xeon 55xx and Xeon 56xx also known as Nehalem, Nehalem-EP, Lynnfield
7*4882a593Smuzhiyun  * and Westmere-EP.
8*4882a593Smuzhiyun  *
9*4882a593Smuzhiyun  * Copyright (c) 2009-2010 by:
10*4882a593Smuzhiyun  *	 Mauro Carvalho Chehab
11*4882a593Smuzhiyun  *
12*4882a593Smuzhiyun  * Red Hat Inc. https://www.redhat.com
13*4882a593Smuzhiyun  *
14*4882a593Smuzhiyun  * Forked and adapted from the i5400_edac driver
15*4882a593Smuzhiyun  *
16*4882a593Smuzhiyun  * Based on the following public Intel datasheets:
17*4882a593Smuzhiyun  * Intel Core i7 Processor Extreme Edition and Intel Core i7 Processor
18*4882a593Smuzhiyun  * Datasheet, Volume 2:
19*4882a593Smuzhiyun  *	http://download.intel.com/design/processor/datashts/320835.pdf
20*4882a593Smuzhiyun  * Intel Xeon Processor 5500 Series Datasheet Volume 2
21*4882a593Smuzhiyun  *	http://www.intel.com/Assets/PDF/datasheet/321322.pdf
22*4882a593Smuzhiyun  * also available at:
23*4882a593Smuzhiyun  * 	http://www.arrownac.com/manufacturers/intel/s/nehalem/5500-datasheet-v2.pdf
24*4882a593Smuzhiyun  */
25*4882a593Smuzhiyun 
26*4882a593Smuzhiyun #include <linux/module.h>
27*4882a593Smuzhiyun #include <linux/init.h>
28*4882a593Smuzhiyun #include <linux/pci.h>
29*4882a593Smuzhiyun #include <linux/pci_ids.h>
30*4882a593Smuzhiyun #include <linux/slab.h>
31*4882a593Smuzhiyun #include <linux/delay.h>
32*4882a593Smuzhiyun #include <linux/dmi.h>
33*4882a593Smuzhiyun #include <linux/edac.h>
34*4882a593Smuzhiyun #include <linux/mmzone.h>
35*4882a593Smuzhiyun #include <linux/smp.h>
36*4882a593Smuzhiyun #include <asm/mce.h>
37*4882a593Smuzhiyun #include <asm/processor.h>
38*4882a593Smuzhiyun #include <asm/div64.h>
39*4882a593Smuzhiyun 
40*4882a593Smuzhiyun #include "edac_module.h"
41*4882a593Smuzhiyun 
42*4882a593Smuzhiyun /* Static vars */
43*4882a593Smuzhiyun static LIST_HEAD(i7core_edac_list);
44*4882a593Smuzhiyun static DEFINE_MUTEX(i7core_edac_lock);
45*4882a593Smuzhiyun static int probed;
46*4882a593Smuzhiyun 
47*4882a593Smuzhiyun static int use_pci_fixup;
48*4882a593Smuzhiyun module_param(use_pci_fixup, int, 0444);
49*4882a593Smuzhiyun MODULE_PARM_DESC(use_pci_fixup, "Enable PCI fixup to seek for hidden devices");
50*4882a593Smuzhiyun /*
51*4882a593Smuzhiyun  * This is used for Nehalem-EP and Nehalem-EX devices, where the non-core
52*4882a593Smuzhiyun  * registers start at bus 255, and are not reported by BIOS.
53*4882a593Smuzhiyun  * We currently find devices with only 2 sockets. In order to support more QPI
54*4882a593Smuzhiyun  * Quick Path Interconnect, just increment this number.
55*4882a593Smuzhiyun  */
56*4882a593Smuzhiyun #define MAX_SOCKET_BUSES	2
57*4882a593Smuzhiyun 
58*4882a593Smuzhiyun 
59*4882a593Smuzhiyun /*
60*4882a593Smuzhiyun  * Alter this version for the module when modifications are made
61*4882a593Smuzhiyun  */
62*4882a593Smuzhiyun #define I7CORE_REVISION    " Ver: 1.0.0"
63*4882a593Smuzhiyun #define EDAC_MOD_STR      "i7core_edac"
64*4882a593Smuzhiyun 
65*4882a593Smuzhiyun /*
66*4882a593Smuzhiyun  * Debug macros
67*4882a593Smuzhiyun  */
68*4882a593Smuzhiyun #define i7core_printk(level, fmt, arg...)			\
69*4882a593Smuzhiyun 	edac_printk(level, "i7core", fmt, ##arg)
70*4882a593Smuzhiyun 
71*4882a593Smuzhiyun #define i7core_mc_printk(mci, level, fmt, arg...)		\
72*4882a593Smuzhiyun 	edac_mc_chipset_printk(mci, level, "i7core", fmt, ##arg)
73*4882a593Smuzhiyun 
74*4882a593Smuzhiyun /*
75*4882a593Smuzhiyun  * i7core Memory Controller Registers
76*4882a593Smuzhiyun  */
77*4882a593Smuzhiyun 
78*4882a593Smuzhiyun 	/* OFFSETS for Device 0 Function 0 */
79*4882a593Smuzhiyun 
80*4882a593Smuzhiyun #define MC_CFG_CONTROL	0x90
81*4882a593Smuzhiyun   #define MC_CFG_UNLOCK		0x02
82*4882a593Smuzhiyun   #define MC_CFG_LOCK		0x00
83*4882a593Smuzhiyun 
84*4882a593Smuzhiyun 	/* OFFSETS for Device 3 Function 0 */
85*4882a593Smuzhiyun 
86*4882a593Smuzhiyun #define MC_CONTROL	0x48
87*4882a593Smuzhiyun #define MC_STATUS	0x4c
88*4882a593Smuzhiyun #define MC_MAX_DOD	0x64
89*4882a593Smuzhiyun 
90*4882a593Smuzhiyun /*
91*4882a593Smuzhiyun  * OFFSETS for Device 3 Function 4, as indicated on Xeon 5500 datasheet:
92*4882a593Smuzhiyun  * http://www.arrownac.com/manufacturers/intel/s/nehalem/5500-datasheet-v2.pdf
93*4882a593Smuzhiyun  */
94*4882a593Smuzhiyun 
95*4882a593Smuzhiyun #define MC_TEST_ERR_RCV1	0x60
96*4882a593Smuzhiyun   #define DIMM2_COR_ERR(r)			((r) & 0x7fff)
97*4882a593Smuzhiyun 
98*4882a593Smuzhiyun #define MC_TEST_ERR_RCV0	0x64
99*4882a593Smuzhiyun   #define DIMM1_COR_ERR(r)			(((r) >> 16) & 0x7fff)
100*4882a593Smuzhiyun   #define DIMM0_COR_ERR(r)			((r) & 0x7fff)
101*4882a593Smuzhiyun 
102*4882a593Smuzhiyun /* OFFSETS for Device 3 Function 2, as indicated on Xeon 5500 datasheet */
103*4882a593Smuzhiyun #define MC_SSRCONTROL		0x48
104*4882a593Smuzhiyun   #define SSR_MODE_DISABLE	0x00
105*4882a593Smuzhiyun   #define SSR_MODE_ENABLE	0x01
106*4882a593Smuzhiyun   #define SSR_MODE_MASK		0x03
107*4882a593Smuzhiyun 
108*4882a593Smuzhiyun #define MC_SCRUB_CONTROL	0x4c
109*4882a593Smuzhiyun   #define STARTSCRUB		(1 << 24)
110*4882a593Smuzhiyun   #define SCRUBINTERVAL_MASK    0xffffff
111*4882a593Smuzhiyun 
112*4882a593Smuzhiyun #define MC_COR_ECC_CNT_0	0x80
113*4882a593Smuzhiyun #define MC_COR_ECC_CNT_1	0x84
114*4882a593Smuzhiyun #define MC_COR_ECC_CNT_2	0x88
115*4882a593Smuzhiyun #define MC_COR_ECC_CNT_3	0x8c
116*4882a593Smuzhiyun #define MC_COR_ECC_CNT_4	0x90
117*4882a593Smuzhiyun #define MC_COR_ECC_CNT_5	0x94
118*4882a593Smuzhiyun 
119*4882a593Smuzhiyun #define DIMM_TOP_COR_ERR(r)			(((r) >> 16) & 0x7fff)
120*4882a593Smuzhiyun #define DIMM_BOT_COR_ERR(r)			((r) & 0x7fff)
121*4882a593Smuzhiyun 
122*4882a593Smuzhiyun 
123*4882a593Smuzhiyun 	/* OFFSETS for Devices 4,5 and 6 Function 0 */
124*4882a593Smuzhiyun 
125*4882a593Smuzhiyun #define MC_CHANNEL_DIMM_INIT_PARAMS 0x58
126*4882a593Smuzhiyun   #define THREE_DIMMS_PRESENT		(1 << 24)
127*4882a593Smuzhiyun   #define SINGLE_QUAD_RANK_PRESENT	(1 << 23)
128*4882a593Smuzhiyun   #define QUAD_RANK_PRESENT		(1 << 22)
129*4882a593Smuzhiyun   #define REGISTERED_DIMM		(1 << 15)
130*4882a593Smuzhiyun 
131*4882a593Smuzhiyun #define MC_CHANNEL_MAPPER	0x60
132*4882a593Smuzhiyun   #define RDLCH(r, ch)		((((r) >> (3 + (ch * 6))) & 0x07) - 1)
133*4882a593Smuzhiyun   #define WRLCH(r, ch)		((((r) >> (ch * 6)) & 0x07) - 1)
134*4882a593Smuzhiyun 
135*4882a593Smuzhiyun #define MC_CHANNEL_RANK_PRESENT 0x7c
136*4882a593Smuzhiyun   #define RANK_PRESENT_MASK		0xffff
137*4882a593Smuzhiyun 
138*4882a593Smuzhiyun #define MC_CHANNEL_ADDR_MATCH	0xf0
139*4882a593Smuzhiyun #define MC_CHANNEL_ERROR_MASK	0xf8
140*4882a593Smuzhiyun #define MC_CHANNEL_ERROR_INJECT	0xfc
141*4882a593Smuzhiyun   #define INJECT_ADDR_PARITY	0x10
142*4882a593Smuzhiyun   #define INJECT_ECC		0x08
143*4882a593Smuzhiyun   #define MASK_CACHELINE	0x06
144*4882a593Smuzhiyun   #define MASK_FULL_CACHELINE	0x06
145*4882a593Smuzhiyun   #define MASK_MSB32_CACHELINE	0x04
146*4882a593Smuzhiyun   #define MASK_LSB32_CACHELINE	0x02
147*4882a593Smuzhiyun   #define NO_MASK_CACHELINE	0x00
148*4882a593Smuzhiyun   #define REPEAT_EN		0x01
149*4882a593Smuzhiyun 
150*4882a593Smuzhiyun 	/* OFFSETS for Devices 4,5 and 6 Function 1 */
151*4882a593Smuzhiyun 
152*4882a593Smuzhiyun #define MC_DOD_CH_DIMM0		0x48
153*4882a593Smuzhiyun #define MC_DOD_CH_DIMM1		0x4c
154*4882a593Smuzhiyun #define MC_DOD_CH_DIMM2		0x50
155*4882a593Smuzhiyun   #define RANKOFFSET_MASK	((1 << 12) | (1 << 11) | (1 << 10))
156*4882a593Smuzhiyun   #define RANKOFFSET(x)		((x & RANKOFFSET_MASK) >> 10)
157*4882a593Smuzhiyun   #define DIMM_PRESENT_MASK	(1 << 9)
158*4882a593Smuzhiyun   #define DIMM_PRESENT(x)	(((x) & DIMM_PRESENT_MASK) >> 9)
159*4882a593Smuzhiyun   #define MC_DOD_NUMBANK_MASK		((1 << 8) | (1 << 7))
160*4882a593Smuzhiyun   #define MC_DOD_NUMBANK(x)		(((x) & MC_DOD_NUMBANK_MASK) >> 7)
161*4882a593Smuzhiyun   #define MC_DOD_NUMRANK_MASK		((1 << 6) | (1 << 5))
162*4882a593Smuzhiyun   #define MC_DOD_NUMRANK(x)		(((x) & MC_DOD_NUMRANK_MASK) >> 5)
163*4882a593Smuzhiyun   #define MC_DOD_NUMROW_MASK		((1 << 4) | (1 << 3) | (1 << 2))
164*4882a593Smuzhiyun   #define MC_DOD_NUMROW(x)		(((x) & MC_DOD_NUMROW_MASK) >> 2)
165*4882a593Smuzhiyun   #define MC_DOD_NUMCOL_MASK		3
166*4882a593Smuzhiyun   #define MC_DOD_NUMCOL(x)		((x) & MC_DOD_NUMCOL_MASK)
167*4882a593Smuzhiyun 
168*4882a593Smuzhiyun #define MC_RANK_PRESENT		0x7c
169*4882a593Smuzhiyun 
170*4882a593Smuzhiyun #define MC_SAG_CH_0	0x80
171*4882a593Smuzhiyun #define MC_SAG_CH_1	0x84
172*4882a593Smuzhiyun #define MC_SAG_CH_2	0x88
173*4882a593Smuzhiyun #define MC_SAG_CH_3	0x8c
174*4882a593Smuzhiyun #define MC_SAG_CH_4	0x90
175*4882a593Smuzhiyun #define MC_SAG_CH_5	0x94
176*4882a593Smuzhiyun #define MC_SAG_CH_6	0x98
177*4882a593Smuzhiyun #define MC_SAG_CH_7	0x9c
178*4882a593Smuzhiyun 
179*4882a593Smuzhiyun #define MC_RIR_LIMIT_CH_0	0x40
180*4882a593Smuzhiyun #define MC_RIR_LIMIT_CH_1	0x44
181*4882a593Smuzhiyun #define MC_RIR_LIMIT_CH_2	0x48
182*4882a593Smuzhiyun #define MC_RIR_LIMIT_CH_3	0x4C
183*4882a593Smuzhiyun #define MC_RIR_LIMIT_CH_4	0x50
184*4882a593Smuzhiyun #define MC_RIR_LIMIT_CH_5	0x54
185*4882a593Smuzhiyun #define MC_RIR_LIMIT_CH_6	0x58
186*4882a593Smuzhiyun #define MC_RIR_LIMIT_CH_7	0x5C
187*4882a593Smuzhiyun #define MC_RIR_LIMIT_MASK	((1 << 10) - 1)
188*4882a593Smuzhiyun 
189*4882a593Smuzhiyun #define MC_RIR_WAY_CH		0x80
190*4882a593Smuzhiyun   #define MC_RIR_WAY_OFFSET_MASK	(((1 << 14) - 1) & ~0x7)
191*4882a593Smuzhiyun   #define MC_RIR_WAY_RANK_MASK		0x7
192*4882a593Smuzhiyun 
193*4882a593Smuzhiyun /*
194*4882a593Smuzhiyun  * i7core structs
195*4882a593Smuzhiyun  */
196*4882a593Smuzhiyun 
197*4882a593Smuzhiyun #define NUM_CHANS 3
198*4882a593Smuzhiyun #define MAX_DIMMS 3		/* Max DIMMS per channel */
199*4882a593Smuzhiyun #define MAX_MCR_FUNC  4
200*4882a593Smuzhiyun #define MAX_CHAN_FUNC 3
201*4882a593Smuzhiyun 
202*4882a593Smuzhiyun struct i7core_info {
203*4882a593Smuzhiyun 	u32	mc_control;
204*4882a593Smuzhiyun 	u32	mc_status;
205*4882a593Smuzhiyun 	u32	max_dod;
206*4882a593Smuzhiyun 	u32	ch_map;
207*4882a593Smuzhiyun };
208*4882a593Smuzhiyun 
209*4882a593Smuzhiyun 
210*4882a593Smuzhiyun struct i7core_inject {
211*4882a593Smuzhiyun 	int	enable;
212*4882a593Smuzhiyun 
213*4882a593Smuzhiyun 	u32	section;
214*4882a593Smuzhiyun 	u32	type;
215*4882a593Smuzhiyun 	u32	eccmask;
216*4882a593Smuzhiyun 
217*4882a593Smuzhiyun 	/* Error address mask */
218*4882a593Smuzhiyun 	int channel, dimm, rank, bank, page, col;
219*4882a593Smuzhiyun };
220*4882a593Smuzhiyun 
221*4882a593Smuzhiyun struct i7core_channel {
222*4882a593Smuzhiyun 	bool		is_3dimms_present;
223*4882a593Smuzhiyun 	bool		is_single_4rank;
224*4882a593Smuzhiyun 	bool		has_4rank;
225*4882a593Smuzhiyun 	u32		dimms;
226*4882a593Smuzhiyun };
227*4882a593Smuzhiyun 
228*4882a593Smuzhiyun struct pci_id_descr {
229*4882a593Smuzhiyun 	int			dev;
230*4882a593Smuzhiyun 	int			func;
231*4882a593Smuzhiyun 	int 			dev_id;
232*4882a593Smuzhiyun 	int			optional;
233*4882a593Smuzhiyun };
234*4882a593Smuzhiyun 
235*4882a593Smuzhiyun struct pci_id_table {
236*4882a593Smuzhiyun 	const struct pci_id_descr	*descr;
237*4882a593Smuzhiyun 	int				n_devs;
238*4882a593Smuzhiyun };
239*4882a593Smuzhiyun 
240*4882a593Smuzhiyun struct i7core_dev {
241*4882a593Smuzhiyun 	struct list_head	list;
242*4882a593Smuzhiyun 	u8			socket;
243*4882a593Smuzhiyun 	struct pci_dev		**pdev;
244*4882a593Smuzhiyun 	int			n_devs;
245*4882a593Smuzhiyun 	struct mem_ctl_info	*mci;
246*4882a593Smuzhiyun };
247*4882a593Smuzhiyun 
248*4882a593Smuzhiyun struct i7core_pvt {
249*4882a593Smuzhiyun 	struct device *addrmatch_dev, *chancounts_dev;
250*4882a593Smuzhiyun 
251*4882a593Smuzhiyun 	struct pci_dev	*pci_noncore;
252*4882a593Smuzhiyun 	struct pci_dev	*pci_mcr[MAX_MCR_FUNC + 1];
253*4882a593Smuzhiyun 	struct pci_dev	*pci_ch[NUM_CHANS][MAX_CHAN_FUNC + 1];
254*4882a593Smuzhiyun 
255*4882a593Smuzhiyun 	struct i7core_dev *i7core_dev;
256*4882a593Smuzhiyun 
257*4882a593Smuzhiyun 	struct i7core_info	info;
258*4882a593Smuzhiyun 	struct i7core_inject	inject;
259*4882a593Smuzhiyun 	struct i7core_channel	channel[NUM_CHANS];
260*4882a593Smuzhiyun 
261*4882a593Smuzhiyun 	int		ce_count_available;
262*4882a593Smuzhiyun 
263*4882a593Smuzhiyun 			/* ECC corrected errors counts per udimm */
264*4882a593Smuzhiyun 	unsigned long	udimm_ce_count[MAX_DIMMS];
265*4882a593Smuzhiyun 	int		udimm_last_ce_count[MAX_DIMMS];
266*4882a593Smuzhiyun 			/* ECC corrected errors counts per rdimm */
267*4882a593Smuzhiyun 	unsigned long	rdimm_ce_count[NUM_CHANS][MAX_DIMMS];
268*4882a593Smuzhiyun 	int		rdimm_last_ce_count[NUM_CHANS][MAX_DIMMS];
269*4882a593Smuzhiyun 
270*4882a593Smuzhiyun 	bool		is_registered, enable_scrub;
271*4882a593Smuzhiyun 
272*4882a593Smuzhiyun 	/* DCLK Frequency used for computing scrub rate */
273*4882a593Smuzhiyun 	int			dclk_freq;
274*4882a593Smuzhiyun 
275*4882a593Smuzhiyun 	/* Struct to control EDAC polling */
276*4882a593Smuzhiyun 	struct edac_pci_ctl_info *i7core_pci;
277*4882a593Smuzhiyun };
278*4882a593Smuzhiyun 
279*4882a593Smuzhiyun #define PCI_DESCR(device, function, device_id)	\
280*4882a593Smuzhiyun 	.dev = (device),			\
281*4882a593Smuzhiyun 	.func = (function),			\
282*4882a593Smuzhiyun 	.dev_id = (device_id)
283*4882a593Smuzhiyun 
284*4882a593Smuzhiyun static const struct pci_id_descr pci_dev_descr_i7core_nehalem[] = {
285*4882a593Smuzhiyun 		/* Memory controller */
286*4882a593Smuzhiyun 	{ PCI_DESCR(3, 0, PCI_DEVICE_ID_INTEL_I7_MCR)     },
287*4882a593Smuzhiyun 	{ PCI_DESCR(3, 1, PCI_DEVICE_ID_INTEL_I7_MC_TAD)  },
288*4882a593Smuzhiyun 			/* Exists only for RDIMM */
289*4882a593Smuzhiyun 	{ PCI_DESCR(3, 2, PCI_DEVICE_ID_INTEL_I7_MC_RAS), .optional = 1  },
290*4882a593Smuzhiyun 	{ PCI_DESCR(3, 4, PCI_DEVICE_ID_INTEL_I7_MC_TEST) },
291*4882a593Smuzhiyun 
292*4882a593Smuzhiyun 		/* Channel 0 */
293*4882a593Smuzhiyun 	{ PCI_DESCR(4, 0, PCI_DEVICE_ID_INTEL_I7_MC_CH0_CTRL) },
294*4882a593Smuzhiyun 	{ PCI_DESCR(4, 1, PCI_DEVICE_ID_INTEL_I7_MC_CH0_ADDR) },
295*4882a593Smuzhiyun 	{ PCI_DESCR(4, 2, PCI_DEVICE_ID_INTEL_I7_MC_CH0_RANK) },
296*4882a593Smuzhiyun 	{ PCI_DESCR(4, 3, PCI_DEVICE_ID_INTEL_I7_MC_CH0_TC)   },
297*4882a593Smuzhiyun 
298*4882a593Smuzhiyun 		/* Channel 1 */
299*4882a593Smuzhiyun 	{ PCI_DESCR(5, 0, PCI_DEVICE_ID_INTEL_I7_MC_CH1_CTRL) },
300*4882a593Smuzhiyun 	{ PCI_DESCR(5, 1, PCI_DEVICE_ID_INTEL_I7_MC_CH1_ADDR) },
301*4882a593Smuzhiyun 	{ PCI_DESCR(5, 2, PCI_DEVICE_ID_INTEL_I7_MC_CH1_RANK) },
302*4882a593Smuzhiyun 	{ PCI_DESCR(5, 3, PCI_DEVICE_ID_INTEL_I7_MC_CH1_TC)   },
303*4882a593Smuzhiyun 
304*4882a593Smuzhiyun 		/* Channel 2 */
305*4882a593Smuzhiyun 	{ PCI_DESCR(6, 0, PCI_DEVICE_ID_INTEL_I7_MC_CH2_CTRL) },
306*4882a593Smuzhiyun 	{ PCI_DESCR(6, 1, PCI_DEVICE_ID_INTEL_I7_MC_CH2_ADDR) },
307*4882a593Smuzhiyun 	{ PCI_DESCR(6, 2, PCI_DEVICE_ID_INTEL_I7_MC_CH2_RANK) },
308*4882a593Smuzhiyun 	{ PCI_DESCR(6, 3, PCI_DEVICE_ID_INTEL_I7_MC_CH2_TC)   },
309*4882a593Smuzhiyun 
310*4882a593Smuzhiyun 		/* Generic Non-core registers */
311*4882a593Smuzhiyun 	/*
312*4882a593Smuzhiyun 	 * This is the PCI device on i7core and on Xeon 35xx (8086:2c41)
313*4882a593Smuzhiyun 	 * On Xeon 55xx, however, it has a different id (8086:2c40). So,
314*4882a593Smuzhiyun 	 * the probing code needs to test for the other address in case of
315*4882a593Smuzhiyun 	 * failure of this one
316*4882a593Smuzhiyun 	 */
317*4882a593Smuzhiyun 	{ PCI_DESCR(0, 0, PCI_DEVICE_ID_INTEL_I7_NONCORE)  },
318*4882a593Smuzhiyun 
319*4882a593Smuzhiyun };
320*4882a593Smuzhiyun 
321*4882a593Smuzhiyun static const struct pci_id_descr pci_dev_descr_lynnfield[] = {
322*4882a593Smuzhiyun 	{ PCI_DESCR( 3, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_MCR)         },
323*4882a593Smuzhiyun 	{ PCI_DESCR( 3, 1, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_TAD)      },
324*4882a593Smuzhiyun 	{ PCI_DESCR( 3, 4, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_TEST)     },
325*4882a593Smuzhiyun 
326*4882a593Smuzhiyun 	{ PCI_DESCR( 4, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_CTRL) },
327*4882a593Smuzhiyun 	{ PCI_DESCR( 4, 1, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_ADDR) },
328*4882a593Smuzhiyun 	{ PCI_DESCR( 4, 2, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_RANK) },
329*4882a593Smuzhiyun 	{ PCI_DESCR( 4, 3, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_TC)   },
330*4882a593Smuzhiyun 
331*4882a593Smuzhiyun 	{ PCI_DESCR( 5, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_CTRL) },
332*4882a593Smuzhiyun 	{ PCI_DESCR( 5, 1, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_ADDR) },
333*4882a593Smuzhiyun 	{ PCI_DESCR( 5, 2, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_RANK) },
334*4882a593Smuzhiyun 	{ PCI_DESCR( 5, 3, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_TC)   },
335*4882a593Smuzhiyun 
336*4882a593Smuzhiyun 	/*
337*4882a593Smuzhiyun 	 * This is the PCI device has an alternate address on some
338*4882a593Smuzhiyun 	 * processors like Core i7 860
339*4882a593Smuzhiyun 	 */
340*4882a593Smuzhiyun 	{ PCI_DESCR( 0, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_NONCORE)     },
341*4882a593Smuzhiyun };
342*4882a593Smuzhiyun 
343*4882a593Smuzhiyun static const struct pci_id_descr pci_dev_descr_i7core_westmere[] = {
344*4882a593Smuzhiyun 		/* Memory controller */
345*4882a593Smuzhiyun 	{ PCI_DESCR(3, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_MCR_REV2)     },
346*4882a593Smuzhiyun 	{ PCI_DESCR(3, 1, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_TAD_REV2)  },
347*4882a593Smuzhiyun 			/* Exists only for RDIMM */
348*4882a593Smuzhiyun 	{ PCI_DESCR(3, 2, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_RAS_REV2), .optional = 1  },
349*4882a593Smuzhiyun 	{ PCI_DESCR(3, 4, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_TEST_REV2) },
350*4882a593Smuzhiyun 
351*4882a593Smuzhiyun 		/* Channel 0 */
352*4882a593Smuzhiyun 	{ PCI_DESCR(4, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_CTRL_REV2) },
353*4882a593Smuzhiyun 	{ PCI_DESCR(4, 1, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_ADDR_REV2) },
354*4882a593Smuzhiyun 	{ PCI_DESCR(4, 2, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_RANK_REV2) },
355*4882a593Smuzhiyun 	{ PCI_DESCR(4, 3, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_TC_REV2)   },
356*4882a593Smuzhiyun 
357*4882a593Smuzhiyun 		/* Channel 1 */
358*4882a593Smuzhiyun 	{ PCI_DESCR(5, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_CTRL_REV2) },
359*4882a593Smuzhiyun 	{ PCI_DESCR(5, 1, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_ADDR_REV2) },
360*4882a593Smuzhiyun 	{ PCI_DESCR(5, 2, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_RANK_REV2) },
361*4882a593Smuzhiyun 	{ PCI_DESCR(5, 3, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_TC_REV2)   },
362*4882a593Smuzhiyun 
363*4882a593Smuzhiyun 		/* Channel 2 */
364*4882a593Smuzhiyun 	{ PCI_DESCR(6, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH2_CTRL_REV2) },
365*4882a593Smuzhiyun 	{ PCI_DESCR(6, 1, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH2_ADDR_REV2) },
366*4882a593Smuzhiyun 	{ PCI_DESCR(6, 2, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH2_RANK_REV2) },
367*4882a593Smuzhiyun 	{ PCI_DESCR(6, 3, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH2_TC_REV2)   },
368*4882a593Smuzhiyun 
369*4882a593Smuzhiyun 		/* Generic Non-core registers */
370*4882a593Smuzhiyun 	{ PCI_DESCR(0, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_NONCORE_REV2)  },
371*4882a593Smuzhiyun 
372*4882a593Smuzhiyun };
373*4882a593Smuzhiyun 
374*4882a593Smuzhiyun #define PCI_ID_TABLE_ENTRY(A) { .descr=A, .n_devs = ARRAY_SIZE(A) }
375*4882a593Smuzhiyun static const struct pci_id_table pci_dev_table[] = {
376*4882a593Smuzhiyun 	PCI_ID_TABLE_ENTRY(pci_dev_descr_i7core_nehalem),
377*4882a593Smuzhiyun 	PCI_ID_TABLE_ENTRY(pci_dev_descr_lynnfield),
378*4882a593Smuzhiyun 	PCI_ID_TABLE_ENTRY(pci_dev_descr_i7core_westmere),
379*4882a593Smuzhiyun 	{0,}			/* 0 terminated list. */
380*4882a593Smuzhiyun };
381*4882a593Smuzhiyun 
382*4882a593Smuzhiyun /*
383*4882a593Smuzhiyun  *	pci_device_id	table for which devices we are looking for
384*4882a593Smuzhiyun  */
385*4882a593Smuzhiyun static const struct pci_device_id i7core_pci_tbl[] = {
386*4882a593Smuzhiyun 	{PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_X58_HUB_MGMT)},
387*4882a593Smuzhiyun 	{PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_LYNNFIELD_QPI_LINK0)},
388*4882a593Smuzhiyun 	{0,}			/* 0 terminated list. */
389*4882a593Smuzhiyun };
390*4882a593Smuzhiyun 
391*4882a593Smuzhiyun /****************************************************************************
392*4882a593Smuzhiyun 			Ancillary status routines
393*4882a593Smuzhiyun  ****************************************************************************/
394*4882a593Smuzhiyun 
395*4882a593Smuzhiyun 	/* MC_CONTROL bits */
396*4882a593Smuzhiyun #define CH_ACTIVE(pvt, ch)	((pvt)->info.mc_control & (1 << (8 + ch)))
397*4882a593Smuzhiyun #define ECCx8(pvt)		((pvt)->info.mc_control & (1 << 1))
398*4882a593Smuzhiyun 
399*4882a593Smuzhiyun 	/* MC_STATUS bits */
400*4882a593Smuzhiyun #define ECC_ENABLED(pvt)	((pvt)->info.mc_status & (1 << 4))
401*4882a593Smuzhiyun #define CH_DISABLED(pvt, ch)	((pvt)->info.mc_status & (1 << ch))
402*4882a593Smuzhiyun 
403*4882a593Smuzhiyun 	/* MC_MAX_DOD read functions */
numdimms(u32 dimms)404*4882a593Smuzhiyun static inline int numdimms(u32 dimms)
405*4882a593Smuzhiyun {
406*4882a593Smuzhiyun 	return (dimms & 0x3) + 1;
407*4882a593Smuzhiyun }
408*4882a593Smuzhiyun 
numrank(u32 rank)409*4882a593Smuzhiyun static inline int numrank(u32 rank)
410*4882a593Smuzhiyun {
411*4882a593Smuzhiyun 	static const int ranks[] = { 1, 2, 4, -EINVAL };
412*4882a593Smuzhiyun 
413*4882a593Smuzhiyun 	return ranks[rank & 0x3];
414*4882a593Smuzhiyun }
415*4882a593Smuzhiyun 
numbank(u32 bank)416*4882a593Smuzhiyun static inline int numbank(u32 bank)
417*4882a593Smuzhiyun {
418*4882a593Smuzhiyun 	static const int banks[] = { 4, 8, 16, -EINVAL };
419*4882a593Smuzhiyun 
420*4882a593Smuzhiyun 	return banks[bank & 0x3];
421*4882a593Smuzhiyun }
422*4882a593Smuzhiyun 
numrow(u32 row)423*4882a593Smuzhiyun static inline int numrow(u32 row)
424*4882a593Smuzhiyun {
425*4882a593Smuzhiyun 	static const int rows[] = {
426*4882a593Smuzhiyun 		1 << 12, 1 << 13, 1 << 14, 1 << 15,
427*4882a593Smuzhiyun 		1 << 16, -EINVAL, -EINVAL, -EINVAL,
428*4882a593Smuzhiyun 	};
429*4882a593Smuzhiyun 
430*4882a593Smuzhiyun 	return rows[row & 0x7];
431*4882a593Smuzhiyun }
432*4882a593Smuzhiyun 
numcol(u32 col)433*4882a593Smuzhiyun static inline int numcol(u32 col)
434*4882a593Smuzhiyun {
435*4882a593Smuzhiyun 	static const int cols[] = {
436*4882a593Smuzhiyun 		1 << 10, 1 << 11, 1 << 12, -EINVAL,
437*4882a593Smuzhiyun 	};
438*4882a593Smuzhiyun 	return cols[col & 0x3];
439*4882a593Smuzhiyun }
440*4882a593Smuzhiyun 
get_i7core_dev(u8 socket)441*4882a593Smuzhiyun static struct i7core_dev *get_i7core_dev(u8 socket)
442*4882a593Smuzhiyun {
443*4882a593Smuzhiyun 	struct i7core_dev *i7core_dev;
444*4882a593Smuzhiyun 
445*4882a593Smuzhiyun 	list_for_each_entry(i7core_dev, &i7core_edac_list, list) {
446*4882a593Smuzhiyun 		if (i7core_dev->socket == socket)
447*4882a593Smuzhiyun 			return i7core_dev;
448*4882a593Smuzhiyun 	}
449*4882a593Smuzhiyun 
450*4882a593Smuzhiyun 	return NULL;
451*4882a593Smuzhiyun }
452*4882a593Smuzhiyun 
alloc_i7core_dev(u8 socket,const struct pci_id_table * table)453*4882a593Smuzhiyun static struct i7core_dev *alloc_i7core_dev(u8 socket,
454*4882a593Smuzhiyun 					   const struct pci_id_table *table)
455*4882a593Smuzhiyun {
456*4882a593Smuzhiyun 	struct i7core_dev *i7core_dev;
457*4882a593Smuzhiyun 
458*4882a593Smuzhiyun 	i7core_dev = kzalloc(sizeof(*i7core_dev), GFP_KERNEL);
459*4882a593Smuzhiyun 	if (!i7core_dev)
460*4882a593Smuzhiyun 		return NULL;
461*4882a593Smuzhiyun 
462*4882a593Smuzhiyun 	i7core_dev->pdev = kcalloc(table->n_devs, sizeof(*i7core_dev->pdev),
463*4882a593Smuzhiyun 				   GFP_KERNEL);
464*4882a593Smuzhiyun 	if (!i7core_dev->pdev) {
465*4882a593Smuzhiyun 		kfree(i7core_dev);
466*4882a593Smuzhiyun 		return NULL;
467*4882a593Smuzhiyun 	}
468*4882a593Smuzhiyun 
469*4882a593Smuzhiyun 	i7core_dev->socket = socket;
470*4882a593Smuzhiyun 	i7core_dev->n_devs = table->n_devs;
471*4882a593Smuzhiyun 	list_add_tail(&i7core_dev->list, &i7core_edac_list);
472*4882a593Smuzhiyun 
473*4882a593Smuzhiyun 	return i7core_dev;
474*4882a593Smuzhiyun }
475*4882a593Smuzhiyun 
free_i7core_dev(struct i7core_dev * i7core_dev)476*4882a593Smuzhiyun static void free_i7core_dev(struct i7core_dev *i7core_dev)
477*4882a593Smuzhiyun {
478*4882a593Smuzhiyun 	list_del(&i7core_dev->list);
479*4882a593Smuzhiyun 	kfree(i7core_dev->pdev);
480*4882a593Smuzhiyun 	kfree(i7core_dev);
481*4882a593Smuzhiyun }
482*4882a593Smuzhiyun 
483*4882a593Smuzhiyun /****************************************************************************
484*4882a593Smuzhiyun 			Memory check routines
485*4882a593Smuzhiyun  ****************************************************************************/
486*4882a593Smuzhiyun 
get_dimm_config(struct mem_ctl_info * mci)487*4882a593Smuzhiyun static int get_dimm_config(struct mem_ctl_info *mci)
488*4882a593Smuzhiyun {
489*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
490*4882a593Smuzhiyun 	struct pci_dev *pdev;
491*4882a593Smuzhiyun 	int i, j;
492*4882a593Smuzhiyun 	enum edac_type mode;
493*4882a593Smuzhiyun 	enum mem_type mtype;
494*4882a593Smuzhiyun 	struct dimm_info *dimm;
495*4882a593Smuzhiyun 
496*4882a593Smuzhiyun 	/* Get data from the MC register, function 0 */
497*4882a593Smuzhiyun 	pdev = pvt->pci_mcr[0];
498*4882a593Smuzhiyun 	if (!pdev)
499*4882a593Smuzhiyun 		return -ENODEV;
500*4882a593Smuzhiyun 
501*4882a593Smuzhiyun 	/* Device 3 function 0 reads */
502*4882a593Smuzhiyun 	pci_read_config_dword(pdev, MC_CONTROL, &pvt->info.mc_control);
503*4882a593Smuzhiyun 	pci_read_config_dword(pdev, MC_STATUS, &pvt->info.mc_status);
504*4882a593Smuzhiyun 	pci_read_config_dword(pdev, MC_MAX_DOD, &pvt->info.max_dod);
505*4882a593Smuzhiyun 	pci_read_config_dword(pdev, MC_CHANNEL_MAPPER, &pvt->info.ch_map);
506*4882a593Smuzhiyun 
507*4882a593Smuzhiyun 	edac_dbg(0, "QPI %d control=0x%08x status=0x%08x dod=0x%08x map=0x%08x\n",
508*4882a593Smuzhiyun 		 pvt->i7core_dev->socket, pvt->info.mc_control,
509*4882a593Smuzhiyun 		 pvt->info.mc_status, pvt->info.max_dod, pvt->info.ch_map);
510*4882a593Smuzhiyun 
511*4882a593Smuzhiyun 	if (ECC_ENABLED(pvt)) {
512*4882a593Smuzhiyun 		edac_dbg(0, "ECC enabled with x%d SDCC\n", ECCx8(pvt) ? 8 : 4);
513*4882a593Smuzhiyun 		if (ECCx8(pvt))
514*4882a593Smuzhiyun 			mode = EDAC_S8ECD8ED;
515*4882a593Smuzhiyun 		else
516*4882a593Smuzhiyun 			mode = EDAC_S4ECD4ED;
517*4882a593Smuzhiyun 	} else {
518*4882a593Smuzhiyun 		edac_dbg(0, "ECC disabled\n");
519*4882a593Smuzhiyun 		mode = EDAC_NONE;
520*4882a593Smuzhiyun 	}
521*4882a593Smuzhiyun 
522*4882a593Smuzhiyun 	/* FIXME: need to handle the error codes */
523*4882a593Smuzhiyun 	edac_dbg(0, "DOD Max limits: DIMMS: %d, %d-ranked, %d-banked x%x x 0x%x\n",
524*4882a593Smuzhiyun 		 numdimms(pvt->info.max_dod),
525*4882a593Smuzhiyun 		 numrank(pvt->info.max_dod >> 2),
526*4882a593Smuzhiyun 		 numbank(pvt->info.max_dod >> 4),
527*4882a593Smuzhiyun 		 numrow(pvt->info.max_dod >> 6),
528*4882a593Smuzhiyun 		 numcol(pvt->info.max_dod >> 9));
529*4882a593Smuzhiyun 
530*4882a593Smuzhiyun 	for (i = 0; i < NUM_CHANS; i++) {
531*4882a593Smuzhiyun 		u32 data, dimm_dod[3], value[8];
532*4882a593Smuzhiyun 
533*4882a593Smuzhiyun 		if (!pvt->pci_ch[i][0])
534*4882a593Smuzhiyun 			continue;
535*4882a593Smuzhiyun 
536*4882a593Smuzhiyun 		if (!CH_ACTIVE(pvt, i)) {
537*4882a593Smuzhiyun 			edac_dbg(0, "Channel %i is not active\n", i);
538*4882a593Smuzhiyun 			continue;
539*4882a593Smuzhiyun 		}
540*4882a593Smuzhiyun 		if (CH_DISABLED(pvt, i)) {
541*4882a593Smuzhiyun 			edac_dbg(0, "Channel %i is disabled\n", i);
542*4882a593Smuzhiyun 			continue;
543*4882a593Smuzhiyun 		}
544*4882a593Smuzhiyun 
545*4882a593Smuzhiyun 		/* Devices 4-6 function 0 */
546*4882a593Smuzhiyun 		pci_read_config_dword(pvt->pci_ch[i][0],
547*4882a593Smuzhiyun 				MC_CHANNEL_DIMM_INIT_PARAMS, &data);
548*4882a593Smuzhiyun 
549*4882a593Smuzhiyun 
550*4882a593Smuzhiyun 		if (data & THREE_DIMMS_PRESENT)
551*4882a593Smuzhiyun 			pvt->channel[i].is_3dimms_present = true;
552*4882a593Smuzhiyun 
553*4882a593Smuzhiyun 		if (data & SINGLE_QUAD_RANK_PRESENT)
554*4882a593Smuzhiyun 			pvt->channel[i].is_single_4rank = true;
555*4882a593Smuzhiyun 
556*4882a593Smuzhiyun 		if (data & QUAD_RANK_PRESENT)
557*4882a593Smuzhiyun 			pvt->channel[i].has_4rank = true;
558*4882a593Smuzhiyun 
559*4882a593Smuzhiyun 		if (data & REGISTERED_DIMM)
560*4882a593Smuzhiyun 			mtype = MEM_RDDR3;
561*4882a593Smuzhiyun 		else
562*4882a593Smuzhiyun 			mtype = MEM_DDR3;
563*4882a593Smuzhiyun 
564*4882a593Smuzhiyun 		/* Devices 4-6 function 1 */
565*4882a593Smuzhiyun 		pci_read_config_dword(pvt->pci_ch[i][1],
566*4882a593Smuzhiyun 				MC_DOD_CH_DIMM0, &dimm_dod[0]);
567*4882a593Smuzhiyun 		pci_read_config_dword(pvt->pci_ch[i][1],
568*4882a593Smuzhiyun 				MC_DOD_CH_DIMM1, &dimm_dod[1]);
569*4882a593Smuzhiyun 		pci_read_config_dword(pvt->pci_ch[i][1],
570*4882a593Smuzhiyun 				MC_DOD_CH_DIMM2, &dimm_dod[2]);
571*4882a593Smuzhiyun 
572*4882a593Smuzhiyun 		edac_dbg(0, "Ch%d phy rd%d, wr%d (0x%08x): %s%s%s%cDIMMs\n",
573*4882a593Smuzhiyun 			 i,
574*4882a593Smuzhiyun 			 RDLCH(pvt->info.ch_map, i), WRLCH(pvt->info.ch_map, i),
575*4882a593Smuzhiyun 			 data,
576*4882a593Smuzhiyun 			 pvt->channel[i].is_3dimms_present ? "3DIMMS " : "",
577*4882a593Smuzhiyun 			 pvt->channel[i].is_3dimms_present ? "SINGLE_4R " : "",
578*4882a593Smuzhiyun 			 pvt->channel[i].has_4rank ? "HAS_4R " : "",
579*4882a593Smuzhiyun 			 (data & REGISTERED_DIMM) ? 'R' : 'U');
580*4882a593Smuzhiyun 
581*4882a593Smuzhiyun 		for (j = 0; j < 3; j++) {
582*4882a593Smuzhiyun 			u32 banks, ranks, rows, cols;
583*4882a593Smuzhiyun 			u32 size, npages;
584*4882a593Smuzhiyun 
585*4882a593Smuzhiyun 			if (!DIMM_PRESENT(dimm_dod[j]))
586*4882a593Smuzhiyun 				continue;
587*4882a593Smuzhiyun 
588*4882a593Smuzhiyun 			dimm = edac_get_dimm(mci, i, j, 0);
589*4882a593Smuzhiyun 			banks = numbank(MC_DOD_NUMBANK(dimm_dod[j]));
590*4882a593Smuzhiyun 			ranks = numrank(MC_DOD_NUMRANK(dimm_dod[j]));
591*4882a593Smuzhiyun 			rows = numrow(MC_DOD_NUMROW(dimm_dod[j]));
592*4882a593Smuzhiyun 			cols = numcol(MC_DOD_NUMCOL(dimm_dod[j]));
593*4882a593Smuzhiyun 
594*4882a593Smuzhiyun 			/* DDR3 has 8 I/O banks */
595*4882a593Smuzhiyun 			size = (rows * cols * banks * ranks) >> (20 - 3);
596*4882a593Smuzhiyun 
597*4882a593Smuzhiyun 			edac_dbg(0, "\tdimm %d %d MiB offset: %x, bank: %d, rank: %d, row: %#x, col: %#x\n",
598*4882a593Smuzhiyun 				 j, size,
599*4882a593Smuzhiyun 				 RANKOFFSET(dimm_dod[j]),
600*4882a593Smuzhiyun 				 banks, ranks, rows, cols);
601*4882a593Smuzhiyun 
602*4882a593Smuzhiyun 			npages = MiB_TO_PAGES(size);
603*4882a593Smuzhiyun 
604*4882a593Smuzhiyun 			dimm->nr_pages = npages;
605*4882a593Smuzhiyun 
606*4882a593Smuzhiyun 			switch (banks) {
607*4882a593Smuzhiyun 			case 4:
608*4882a593Smuzhiyun 				dimm->dtype = DEV_X4;
609*4882a593Smuzhiyun 				break;
610*4882a593Smuzhiyun 			case 8:
611*4882a593Smuzhiyun 				dimm->dtype = DEV_X8;
612*4882a593Smuzhiyun 				break;
613*4882a593Smuzhiyun 			case 16:
614*4882a593Smuzhiyun 				dimm->dtype = DEV_X16;
615*4882a593Smuzhiyun 				break;
616*4882a593Smuzhiyun 			default:
617*4882a593Smuzhiyun 				dimm->dtype = DEV_UNKNOWN;
618*4882a593Smuzhiyun 			}
619*4882a593Smuzhiyun 
620*4882a593Smuzhiyun 			snprintf(dimm->label, sizeof(dimm->label),
621*4882a593Smuzhiyun 				 "CPU#%uChannel#%u_DIMM#%u",
622*4882a593Smuzhiyun 				 pvt->i7core_dev->socket, i, j);
623*4882a593Smuzhiyun 			dimm->grain = 8;
624*4882a593Smuzhiyun 			dimm->edac_mode = mode;
625*4882a593Smuzhiyun 			dimm->mtype = mtype;
626*4882a593Smuzhiyun 		}
627*4882a593Smuzhiyun 
628*4882a593Smuzhiyun 		pci_read_config_dword(pdev, MC_SAG_CH_0, &value[0]);
629*4882a593Smuzhiyun 		pci_read_config_dword(pdev, MC_SAG_CH_1, &value[1]);
630*4882a593Smuzhiyun 		pci_read_config_dword(pdev, MC_SAG_CH_2, &value[2]);
631*4882a593Smuzhiyun 		pci_read_config_dword(pdev, MC_SAG_CH_3, &value[3]);
632*4882a593Smuzhiyun 		pci_read_config_dword(pdev, MC_SAG_CH_4, &value[4]);
633*4882a593Smuzhiyun 		pci_read_config_dword(pdev, MC_SAG_CH_5, &value[5]);
634*4882a593Smuzhiyun 		pci_read_config_dword(pdev, MC_SAG_CH_6, &value[6]);
635*4882a593Smuzhiyun 		pci_read_config_dword(pdev, MC_SAG_CH_7, &value[7]);
636*4882a593Smuzhiyun 		edac_dbg(1, "\t[%i] DIVBY3\tREMOVED\tOFFSET\n", i);
637*4882a593Smuzhiyun 		for (j = 0; j < 8; j++)
638*4882a593Smuzhiyun 			edac_dbg(1, "\t\t%#x\t%#x\t%#x\n",
639*4882a593Smuzhiyun 				 (value[j] >> 27) & 0x1,
640*4882a593Smuzhiyun 				 (value[j] >> 24) & 0x7,
641*4882a593Smuzhiyun 				 (value[j] & ((1 << 24) - 1)));
642*4882a593Smuzhiyun 	}
643*4882a593Smuzhiyun 
644*4882a593Smuzhiyun 	return 0;
645*4882a593Smuzhiyun }
646*4882a593Smuzhiyun 
647*4882a593Smuzhiyun /****************************************************************************
648*4882a593Smuzhiyun 			Error insertion routines
649*4882a593Smuzhiyun  ****************************************************************************/
650*4882a593Smuzhiyun 
651*4882a593Smuzhiyun #define to_mci(k) container_of(k, struct mem_ctl_info, dev)
652*4882a593Smuzhiyun 
653*4882a593Smuzhiyun /* The i7core has independent error injection features per channel.
654*4882a593Smuzhiyun    However, to have a simpler code, we don't allow enabling error injection
655*4882a593Smuzhiyun    on more than one channel.
656*4882a593Smuzhiyun    Also, since a change at an inject parameter will be applied only at enable,
657*4882a593Smuzhiyun    we're disabling error injection on all write calls to the sysfs nodes that
658*4882a593Smuzhiyun    controls the error code injection.
659*4882a593Smuzhiyun  */
disable_inject(const struct mem_ctl_info * mci)660*4882a593Smuzhiyun static int disable_inject(const struct mem_ctl_info *mci)
661*4882a593Smuzhiyun {
662*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
663*4882a593Smuzhiyun 
664*4882a593Smuzhiyun 	pvt->inject.enable = 0;
665*4882a593Smuzhiyun 
666*4882a593Smuzhiyun 	if (!pvt->pci_ch[pvt->inject.channel][0])
667*4882a593Smuzhiyun 		return -ENODEV;
668*4882a593Smuzhiyun 
669*4882a593Smuzhiyun 	pci_write_config_dword(pvt->pci_ch[pvt->inject.channel][0],
670*4882a593Smuzhiyun 				MC_CHANNEL_ERROR_INJECT, 0);
671*4882a593Smuzhiyun 
672*4882a593Smuzhiyun 	return 0;
673*4882a593Smuzhiyun }
674*4882a593Smuzhiyun 
675*4882a593Smuzhiyun /*
676*4882a593Smuzhiyun  * i7core inject inject.section
677*4882a593Smuzhiyun  *
678*4882a593Smuzhiyun  *	accept and store error injection inject.section value
679*4882a593Smuzhiyun  *	bit 0 - refers to the lower 32-byte half cacheline
680*4882a593Smuzhiyun  *	bit 1 - refers to the upper 32-byte half cacheline
681*4882a593Smuzhiyun  */
i7core_inject_section_store(struct device * dev,struct device_attribute * mattr,const char * data,size_t count)682*4882a593Smuzhiyun static ssize_t i7core_inject_section_store(struct device *dev,
683*4882a593Smuzhiyun 					   struct device_attribute *mattr,
684*4882a593Smuzhiyun 					   const char *data, size_t count)
685*4882a593Smuzhiyun {
686*4882a593Smuzhiyun 	struct mem_ctl_info *mci = to_mci(dev);
687*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
688*4882a593Smuzhiyun 	unsigned long value;
689*4882a593Smuzhiyun 	int rc;
690*4882a593Smuzhiyun 
691*4882a593Smuzhiyun 	if (pvt->inject.enable)
692*4882a593Smuzhiyun 		disable_inject(mci);
693*4882a593Smuzhiyun 
694*4882a593Smuzhiyun 	rc = kstrtoul(data, 10, &value);
695*4882a593Smuzhiyun 	if ((rc < 0) || (value > 3))
696*4882a593Smuzhiyun 		return -EIO;
697*4882a593Smuzhiyun 
698*4882a593Smuzhiyun 	pvt->inject.section = (u32) value;
699*4882a593Smuzhiyun 	return count;
700*4882a593Smuzhiyun }
701*4882a593Smuzhiyun 
i7core_inject_section_show(struct device * dev,struct device_attribute * mattr,char * data)702*4882a593Smuzhiyun static ssize_t i7core_inject_section_show(struct device *dev,
703*4882a593Smuzhiyun 					  struct device_attribute *mattr,
704*4882a593Smuzhiyun 					  char *data)
705*4882a593Smuzhiyun {
706*4882a593Smuzhiyun 	struct mem_ctl_info *mci = to_mci(dev);
707*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
708*4882a593Smuzhiyun 	return sprintf(data, "0x%08x\n", pvt->inject.section);
709*4882a593Smuzhiyun }
710*4882a593Smuzhiyun 
711*4882a593Smuzhiyun /*
712*4882a593Smuzhiyun  * i7core inject.type
713*4882a593Smuzhiyun  *
714*4882a593Smuzhiyun  *	accept and store error injection inject.section value
715*4882a593Smuzhiyun  *	bit 0 - repeat enable - Enable error repetition
716*4882a593Smuzhiyun  *	bit 1 - inject ECC error
717*4882a593Smuzhiyun  *	bit 2 - inject parity error
718*4882a593Smuzhiyun  */
i7core_inject_type_store(struct device * dev,struct device_attribute * mattr,const char * data,size_t count)719*4882a593Smuzhiyun static ssize_t i7core_inject_type_store(struct device *dev,
720*4882a593Smuzhiyun 					struct device_attribute *mattr,
721*4882a593Smuzhiyun 					const char *data, size_t count)
722*4882a593Smuzhiyun {
723*4882a593Smuzhiyun 	struct mem_ctl_info *mci = to_mci(dev);
724*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
725*4882a593Smuzhiyun 	unsigned long value;
726*4882a593Smuzhiyun 	int rc;
727*4882a593Smuzhiyun 
728*4882a593Smuzhiyun 	if (pvt->inject.enable)
729*4882a593Smuzhiyun 		disable_inject(mci);
730*4882a593Smuzhiyun 
731*4882a593Smuzhiyun 	rc = kstrtoul(data, 10, &value);
732*4882a593Smuzhiyun 	if ((rc < 0) || (value > 7))
733*4882a593Smuzhiyun 		return -EIO;
734*4882a593Smuzhiyun 
735*4882a593Smuzhiyun 	pvt->inject.type = (u32) value;
736*4882a593Smuzhiyun 	return count;
737*4882a593Smuzhiyun }
738*4882a593Smuzhiyun 
i7core_inject_type_show(struct device * dev,struct device_attribute * mattr,char * data)739*4882a593Smuzhiyun static ssize_t i7core_inject_type_show(struct device *dev,
740*4882a593Smuzhiyun 				       struct device_attribute *mattr,
741*4882a593Smuzhiyun 				       char *data)
742*4882a593Smuzhiyun {
743*4882a593Smuzhiyun 	struct mem_ctl_info *mci = to_mci(dev);
744*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
745*4882a593Smuzhiyun 
746*4882a593Smuzhiyun 	return sprintf(data, "0x%08x\n", pvt->inject.type);
747*4882a593Smuzhiyun }
748*4882a593Smuzhiyun 
749*4882a593Smuzhiyun /*
750*4882a593Smuzhiyun  * i7core_inject_inject.eccmask_store
751*4882a593Smuzhiyun  *
752*4882a593Smuzhiyun  * The type of error (UE/CE) will depend on the inject.eccmask value:
753*4882a593Smuzhiyun  *   Any bits set to a 1 will flip the corresponding ECC bit
754*4882a593Smuzhiyun  *   Correctable errors can be injected by flipping 1 bit or the bits within
755*4882a593Smuzhiyun  *   a symbol pair (2 consecutive aligned 8-bit pairs - i.e. 7:0 and 15:8 or
756*4882a593Smuzhiyun  *   23:16 and 31:24). Flipping bits in two symbol pairs will cause an
757*4882a593Smuzhiyun  *   uncorrectable error to be injected.
758*4882a593Smuzhiyun  */
i7core_inject_eccmask_store(struct device * dev,struct device_attribute * mattr,const char * data,size_t count)759*4882a593Smuzhiyun static ssize_t i7core_inject_eccmask_store(struct device *dev,
760*4882a593Smuzhiyun 					   struct device_attribute *mattr,
761*4882a593Smuzhiyun 					   const char *data, size_t count)
762*4882a593Smuzhiyun {
763*4882a593Smuzhiyun 	struct mem_ctl_info *mci = to_mci(dev);
764*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
765*4882a593Smuzhiyun 	unsigned long value;
766*4882a593Smuzhiyun 	int rc;
767*4882a593Smuzhiyun 
768*4882a593Smuzhiyun 	if (pvt->inject.enable)
769*4882a593Smuzhiyun 		disable_inject(mci);
770*4882a593Smuzhiyun 
771*4882a593Smuzhiyun 	rc = kstrtoul(data, 10, &value);
772*4882a593Smuzhiyun 	if (rc < 0)
773*4882a593Smuzhiyun 		return -EIO;
774*4882a593Smuzhiyun 
775*4882a593Smuzhiyun 	pvt->inject.eccmask = (u32) value;
776*4882a593Smuzhiyun 	return count;
777*4882a593Smuzhiyun }
778*4882a593Smuzhiyun 
i7core_inject_eccmask_show(struct device * dev,struct device_attribute * mattr,char * data)779*4882a593Smuzhiyun static ssize_t i7core_inject_eccmask_show(struct device *dev,
780*4882a593Smuzhiyun 					  struct device_attribute *mattr,
781*4882a593Smuzhiyun 					  char *data)
782*4882a593Smuzhiyun {
783*4882a593Smuzhiyun 	struct mem_ctl_info *mci = to_mci(dev);
784*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
785*4882a593Smuzhiyun 
786*4882a593Smuzhiyun 	return sprintf(data, "0x%08x\n", pvt->inject.eccmask);
787*4882a593Smuzhiyun }
788*4882a593Smuzhiyun 
789*4882a593Smuzhiyun /*
790*4882a593Smuzhiyun  * i7core_addrmatch
791*4882a593Smuzhiyun  *
792*4882a593Smuzhiyun  * The type of error (UE/CE) will depend on the inject.eccmask value:
793*4882a593Smuzhiyun  *   Any bits set to a 1 will flip the corresponding ECC bit
794*4882a593Smuzhiyun  *   Correctable errors can be injected by flipping 1 bit or the bits within
795*4882a593Smuzhiyun  *   a symbol pair (2 consecutive aligned 8-bit pairs - i.e. 7:0 and 15:8 or
796*4882a593Smuzhiyun  *   23:16 and 31:24). Flipping bits in two symbol pairs will cause an
797*4882a593Smuzhiyun  *   uncorrectable error to be injected.
798*4882a593Smuzhiyun  */
799*4882a593Smuzhiyun 
800*4882a593Smuzhiyun #define DECLARE_ADDR_MATCH(param, limit)			\
801*4882a593Smuzhiyun static ssize_t i7core_inject_store_##param(			\
802*4882a593Smuzhiyun 	struct device *dev,					\
803*4882a593Smuzhiyun 	struct device_attribute *mattr,				\
804*4882a593Smuzhiyun 	const char *data, size_t count)				\
805*4882a593Smuzhiyun {								\
806*4882a593Smuzhiyun 	struct mem_ctl_info *mci = dev_get_drvdata(dev);	\
807*4882a593Smuzhiyun 	struct i7core_pvt *pvt;					\
808*4882a593Smuzhiyun 	long value;						\
809*4882a593Smuzhiyun 	int rc;							\
810*4882a593Smuzhiyun 								\
811*4882a593Smuzhiyun 	edac_dbg(1, "\n");					\
812*4882a593Smuzhiyun 	pvt = mci->pvt_info;					\
813*4882a593Smuzhiyun 								\
814*4882a593Smuzhiyun 	if (pvt->inject.enable)					\
815*4882a593Smuzhiyun 		disable_inject(mci);				\
816*4882a593Smuzhiyun 								\
817*4882a593Smuzhiyun 	if (!strcasecmp(data, "any") || !strcasecmp(data, "any\n"))\
818*4882a593Smuzhiyun 		value = -1;					\
819*4882a593Smuzhiyun 	else {							\
820*4882a593Smuzhiyun 		rc = kstrtoul(data, 10, &value);		\
821*4882a593Smuzhiyun 		if ((rc < 0) || (value >= limit))		\
822*4882a593Smuzhiyun 			return -EIO;				\
823*4882a593Smuzhiyun 	}							\
824*4882a593Smuzhiyun 								\
825*4882a593Smuzhiyun 	pvt->inject.param = value;				\
826*4882a593Smuzhiyun 								\
827*4882a593Smuzhiyun 	return count;						\
828*4882a593Smuzhiyun }								\
829*4882a593Smuzhiyun 								\
830*4882a593Smuzhiyun static ssize_t i7core_inject_show_##param(			\
831*4882a593Smuzhiyun 	struct device *dev,					\
832*4882a593Smuzhiyun 	struct device_attribute *mattr,				\
833*4882a593Smuzhiyun 	char *data)						\
834*4882a593Smuzhiyun {								\
835*4882a593Smuzhiyun 	struct mem_ctl_info *mci = dev_get_drvdata(dev);	\
836*4882a593Smuzhiyun 	struct i7core_pvt *pvt;					\
837*4882a593Smuzhiyun 								\
838*4882a593Smuzhiyun 	pvt = mci->pvt_info;					\
839*4882a593Smuzhiyun 	edac_dbg(1, "pvt=%p\n", pvt);				\
840*4882a593Smuzhiyun 	if (pvt->inject.param < 0)				\
841*4882a593Smuzhiyun 		return sprintf(data, "any\n");			\
842*4882a593Smuzhiyun 	else							\
843*4882a593Smuzhiyun 		return sprintf(data, "%d\n", pvt->inject.param);\
844*4882a593Smuzhiyun }
845*4882a593Smuzhiyun 
846*4882a593Smuzhiyun #define ATTR_ADDR_MATCH(param)					\
847*4882a593Smuzhiyun 	static DEVICE_ATTR(param, S_IRUGO | S_IWUSR,		\
848*4882a593Smuzhiyun 		    i7core_inject_show_##param,			\
849*4882a593Smuzhiyun 		    i7core_inject_store_##param)
850*4882a593Smuzhiyun 
851*4882a593Smuzhiyun DECLARE_ADDR_MATCH(channel, 3);
852*4882a593Smuzhiyun DECLARE_ADDR_MATCH(dimm, 3);
853*4882a593Smuzhiyun DECLARE_ADDR_MATCH(rank, 4);
854*4882a593Smuzhiyun DECLARE_ADDR_MATCH(bank, 32);
855*4882a593Smuzhiyun DECLARE_ADDR_MATCH(page, 0x10000);
856*4882a593Smuzhiyun DECLARE_ADDR_MATCH(col, 0x4000);
857*4882a593Smuzhiyun 
858*4882a593Smuzhiyun ATTR_ADDR_MATCH(channel);
859*4882a593Smuzhiyun ATTR_ADDR_MATCH(dimm);
860*4882a593Smuzhiyun ATTR_ADDR_MATCH(rank);
861*4882a593Smuzhiyun ATTR_ADDR_MATCH(bank);
862*4882a593Smuzhiyun ATTR_ADDR_MATCH(page);
863*4882a593Smuzhiyun ATTR_ADDR_MATCH(col);
864*4882a593Smuzhiyun 
write_and_test(struct pci_dev * dev,const int where,const u32 val)865*4882a593Smuzhiyun static int write_and_test(struct pci_dev *dev, const int where, const u32 val)
866*4882a593Smuzhiyun {
867*4882a593Smuzhiyun 	u32 read;
868*4882a593Smuzhiyun 	int count;
869*4882a593Smuzhiyun 
870*4882a593Smuzhiyun 	edac_dbg(0, "setting pci %02x:%02x.%x reg=%02x value=%08x\n",
871*4882a593Smuzhiyun 		 dev->bus->number, PCI_SLOT(dev->devfn), PCI_FUNC(dev->devfn),
872*4882a593Smuzhiyun 		 where, val);
873*4882a593Smuzhiyun 
874*4882a593Smuzhiyun 	for (count = 0; count < 10; count++) {
875*4882a593Smuzhiyun 		if (count)
876*4882a593Smuzhiyun 			msleep(100);
877*4882a593Smuzhiyun 		pci_write_config_dword(dev, where, val);
878*4882a593Smuzhiyun 		pci_read_config_dword(dev, where, &read);
879*4882a593Smuzhiyun 
880*4882a593Smuzhiyun 		if (read == val)
881*4882a593Smuzhiyun 			return 0;
882*4882a593Smuzhiyun 	}
883*4882a593Smuzhiyun 
884*4882a593Smuzhiyun 	i7core_printk(KERN_ERR, "Error during set pci %02x:%02x.%x reg=%02x "
885*4882a593Smuzhiyun 		"write=%08x. Read=%08x\n",
886*4882a593Smuzhiyun 		dev->bus->number, PCI_SLOT(dev->devfn), PCI_FUNC(dev->devfn),
887*4882a593Smuzhiyun 		where, val, read);
888*4882a593Smuzhiyun 
889*4882a593Smuzhiyun 	return -EINVAL;
890*4882a593Smuzhiyun }
891*4882a593Smuzhiyun 
892*4882a593Smuzhiyun /*
893*4882a593Smuzhiyun  * This routine prepares the Memory Controller for error injection.
894*4882a593Smuzhiyun  * The error will be injected when some process tries to write to the
895*4882a593Smuzhiyun  * memory that matches the given criteria.
896*4882a593Smuzhiyun  * The criteria can be set in terms of a mask where dimm, rank, bank, page
897*4882a593Smuzhiyun  * and col can be specified.
898*4882a593Smuzhiyun  * A -1 value for any of the mask items will make the MCU to ignore
899*4882a593Smuzhiyun  * that matching criteria for error injection.
900*4882a593Smuzhiyun  *
901*4882a593Smuzhiyun  * It should be noticed that the error will only happen after a write operation
902*4882a593Smuzhiyun  * on a memory that matches the condition. if REPEAT_EN is not enabled at
903*4882a593Smuzhiyun  * inject mask, then it will produce just one error. Otherwise, it will repeat
904*4882a593Smuzhiyun  * until the injectmask would be cleaned.
905*4882a593Smuzhiyun  *
906*4882a593Smuzhiyun  * FIXME: This routine assumes that MAXNUMDIMMS value of MC_MAX_DOD
907*4882a593Smuzhiyun  *    is reliable enough to check if the MC is using the
908*4882a593Smuzhiyun  *    three channels. However, this is not clear at the datasheet.
909*4882a593Smuzhiyun  */
i7core_inject_enable_store(struct device * dev,struct device_attribute * mattr,const char * data,size_t count)910*4882a593Smuzhiyun static ssize_t i7core_inject_enable_store(struct device *dev,
911*4882a593Smuzhiyun 					  struct device_attribute *mattr,
912*4882a593Smuzhiyun 					  const char *data, size_t count)
913*4882a593Smuzhiyun {
914*4882a593Smuzhiyun 	struct mem_ctl_info *mci = to_mci(dev);
915*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
916*4882a593Smuzhiyun 	u32 injectmask;
917*4882a593Smuzhiyun 	u64 mask = 0;
918*4882a593Smuzhiyun 	int  rc;
919*4882a593Smuzhiyun 	long enable;
920*4882a593Smuzhiyun 
921*4882a593Smuzhiyun 	if (!pvt->pci_ch[pvt->inject.channel][0])
922*4882a593Smuzhiyun 		return 0;
923*4882a593Smuzhiyun 
924*4882a593Smuzhiyun 	rc = kstrtoul(data, 10, &enable);
925*4882a593Smuzhiyun 	if ((rc < 0))
926*4882a593Smuzhiyun 		return 0;
927*4882a593Smuzhiyun 
928*4882a593Smuzhiyun 	if (enable) {
929*4882a593Smuzhiyun 		pvt->inject.enable = 1;
930*4882a593Smuzhiyun 	} else {
931*4882a593Smuzhiyun 		disable_inject(mci);
932*4882a593Smuzhiyun 		return count;
933*4882a593Smuzhiyun 	}
934*4882a593Smuzhiyun 
935*4882a593Smuzhiyun 	/* Sets pvt->inject.dimm mask */
936*4882a593Smuzhiyun 	if (pvt->inject.dimm < 0)
937*4882a593Smuzhiyun 		mask |= 1LL << 41;
938*4882a593Smuzhiyun 	else {
939*4882a593Smuzhiyun 		if (pvt->channel[pvt->inject.channel].dimms > 2)
940*4882a593Smuzhiyun 			mask |= (pvt->inject.dimm & 0x3LL) << 35;
941*4882a593Smuzhiyun 		else
942*4882a593Smuzhiyun 			mask |= (pvt->inject.dimm & 0x1LL) << 36;
943*4882a593Smuzhiyun 	}
944*4882a593Smuzhiyun 
945*4882a593Smuzhiyun 	/* Sets pvt->inject.rank mask */
946*4882a593Smuzhiyun 	if (pvt->inject.rank < 0)
947*4882a593Smuzhiyun 		mask |= 1LL << 40;
948*4882a593Smuzhiyun 	else {
949*4882a593Smuzhiyun 		if (pvt->channel[pvt->inject.channel].dimms > 2)
950*4882a593Smuzhiyun 			mask |= (pvt->inject.rank & 0x1LL) << 34;
951*4882a593Smuzhiyun 		else
952*4882a593Smuzhiyun 			mask |= (pvt->inject.rank & 0x3LL) << 34;
953*4882a593Smuzhiyun 	}
954*4882a593Smuzhiyun 
955*4882a593Smuzhiyun 	/* Sets pvt->inject.bank mask */
956*4882a593Smuzhiyun 	if (pvt->inject.bank < 0)
957*4882a593Smuzhiyun 		mask |= 1LL << 39;
958*4882a593Smuzhiyun 	else
959*4882a593Smuzhiyun 		mask |= (pvt->inject.bank & 0x15LL) << 30;
960*4882a593Smuzhiyun 
961*4882a593Smuzhiyun 	/* Sets pvt->inject.page mask */
962*4882a593Smuzhiyun 	if (pvt->inject.page < 0)
963*4882a593Smuzhiyun 		mask |= 1LL << 38;
964*4882a593Smuzhiyun 	else
965*4882a593Smuzhiyun 		mask |= (pvt->inject.page & 0xffff) << 14;
966*4882a593Smuzhiyun 
967*4882a593Smuzhiyun 	/* Sets pvt->inject.column mask */
968*4882a593Smuzhiyun 	if (pvt->inject.col < 0)
969*4882a593Smuzhiyun 		mask |= 1LL << 37;
970*4882a593Smuzhiyun 	else
971*4882a593Smuzhiyun 		mask |= (pvt->inject.col & 0x3fff);
972*4882a593Smuzhiyun 
973*4882a593Smuzhiyun 	/*
974*4882a593Smuzhiyun 	 * bit    0: REPEAT_EN
975*4882a593Smuzhiyun 	 * bits 1-2: MASK_HALF_CACHELINE
976*4882a593Smuzhiyun 	 * bit    3: INJECT_ECC
977*4882a593Smuzhiyun 	 * bit    4: INJECT_ADDR_PARITY
978*4882a593Smuzhiyun 	 */
979*4882a593Smuzhiyun 
980*4882a593Smuzhiyun 	injectmask = (pvt->inject.type & 1) |
981*4882a593Smuzhiyun 		     (pvt->inject.section & 0x3) << 1 |
982*4882a593Smuzhiyun 		     (pvt->inject.type & 0x6) << (3 - 1);
983*4882a593Smuzhiyun 
984*4882a593Smuzhiyun 	/* Unlock writes to registers - this register is write only */
985*4882a593Smuzhiyun 	pci_write_config_dword(pvt->pci_noncore,
986*4882a593Smuzhiyun 			       MC_CFG_CONTROL, 0x2);
987*4882a593Smuzhiyun 
988*4882a593Smuzhiyun 	write_and_test(pvt->pci_ch[pvt->inject.channel][0],
989*4882a593Smuzhiyun 			       MC_CHANNEL_ADDR_MATCH, mask);
990*4882a593Smuzhiyun 	write_and_test(pvt->pci_ch[pvt->inject.channel][0],
991*4882a593Smuzhiyun 			       MC_CHANNEL_ADDR_MATCH + 4, mask >> 32L);
992*4882a593Smuzhiyun 
993*4882a593Smuzhiyun 	write_and_test(pvt->pci_ch[pvt->inject.channel][0],
994*4882a593Smuzhiyun 			       MC_CHANNEL_ERROR_MASK, pvt->inject.eccmask);
995*4882a593Smuzhiyun 
996*4882a593Smuzhiyun 	write_and_test(pvt->pci_ch[pvt->inject.channel][0],
997*4882a593Smuzhiyun 			       MC_CHANNEL_ERROR_INJECT, injectmask);
998*4882a593Smuzhiyun 
999*4882a593Smuzhiyun 	/*
1000*4882a593Smuzhiyun 	 * This is something undocumented, based on my tests
1001*4882a593Smuzhiyun 	 * Without writing 8 to this register, errors aren't injected. Not sure
1002*4882a593Smuzhiyun 	 * why.
1003*4882a593Smuzhiyun 	 */
1004*4882a593Smuzhiyun 	pci_write_config_dword(pvt->pci_noncore,
1005*4882a593Smuzhiyun 			       MC_CFG_CONTROL, 8);
1006*4882a593Smuzhiyun 
1007*4882a593Smuzhiyun 	edac_dbg(0, "Error inject addr match 0x%016llx, ecc 0x%08x, inject 0x%08x\n",
1008*4882a593Smuzhiyun 		 mask, pvt->inject.eccmask, injectmask);
1009*4882a593Smuzhiyun 
1010*4882a593Smuzhiyun 
1011*4882a593Smuzhiyun 	return count;
1012*4882a593Smuzhiyun }
1013*4882a593Smuzhiyun 
i7core_inject_enable_show(struct device * dev,struct device_attribute * mattr,char * data)1014*4882a593Smuzhiyun static ssize_t i7core_inject_enable_show(struct device *dev,
1015*4882a593Smuzhiyun 					 struct device_attribute *mattr,
1016*4882a593Smuzhiyun 					 char *data)
1017*4882a593Smuzhiyun {
1018*4882a593Smuzhiyun 	struct mem_ctl_info *mci = to_mci(dev);
1019*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
1020*4882a593Smuzhiyun 	u32 injectmask;
1021*4882a593Smuzhiyun 
1022*4882a593Smuzhiyun 	if (!pvt->pci_ch[pvt->inject.channel][0])
1023*4882a593Smuzhiyun 		return 0;
1024*4882a593Smuzhiyun 
1025*4882a593Smuzhiyun 	pci_read_config_dword(pvt->pci_ch[pvt->inject.channel][0],
1026*4882a593Smuzhiyun 			       MC_CHANNEL_ERROR_INJECT, &injectmask);
1027*4882a593Smuzhiyun 
1028*4882a593Smuzhiyun 	edac_dbg(0, "Inject error read: 0x%018x\n", injectmask);
1029*4882a593Smuzhiyun 
1030*4882a593Smuzhiyun 	if (injectmask & 0x0c)
1031*4882a593Smuzhiyun 		pvt->inject.enable = 1;
1032*4882a593Smuzhiyun 
1033*4882a593Smuzhiyun 	return sprintf(data, "%d\n", pvt->inject.enable);
1034*4882a593Smuzhiyun }
1035*4882a593Smuzhiyun 
1036*4882a593Smuzhiyun #define DECLARE_COUNTER(param)					\
1037*4882a593Smuzhiyun static ssize_t i7core_show_counter_##param(			\
1038*4882a593Smuzhiyun 	struct device *dev,					\
1039*4882a593Smuzhiyun 	struct device_attribute *mattr,				\
1040*4882a593Smuzhiyun 	char *data)						\
1041*4882a593Smuzhiyun {								\
1042*4882a593Smuzhiyun 	struct mem_ctl_info *mci = dev_get_drvdata(dev);	\
1043*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;			\
1044*4882a593Smuzhiyun 								\
1045*4882a593Smuzhiyun 	edac_dbg(1, "\n");					\
1046*4882a593Smuzhiyun 	if (!pvt->ce_count_available || (pvt->is_registered))	\
1047*4882a593Smuzhiyun 		return sprintf(data, "data unavailable\n");	\
1048*4882a593Smuzhiyun 	return sprintf(data, "%lu\n",				\
1049*4882a593Smuzhiyun 			pvt->udimm_ce_count[param]);		\
1050*4882a593Smuzhiyun }
1051*4882a593Smuzhiyun 
1052*4882a593Smuzhiyun #define ATTR_COUNTER(param)					\
1053*4882a593Smuzhiyun 	static DEVICE_ATTR(udimm##param, S_IRUGO | S_IWUSR,	\
1054*4882a593Smuzhiyun 		    i7core_show_counter_##param,		\
1055*4882a593Smuzhiyun 		    NULL)
1056*4882a593Smuzhiyun 
1057*4882a593Smuzhiyun DECLARE_COUNTER(0);
1058*4882a593Smuzhiyun DECLARE_COUNTER(1);
1059*4882a593Smuzhiyun DECLARE_COUNTER(2);
1060*4882a593Smuzhiyun 
1061*4882a593Smuzhiyun ATTR_COUNTER(0);
1062*4882a593Smuzhiyun ATTR_COUNTER(1);
1063*4882a593Smuzhiyun ATTR_COUNTER(2);
1064*4882a593Smuzhiyun 
1065*4882a593Smuzhiyun /*
1066*4882a593Smuzhiyun  * inject_addrmatch device sysfs struct
1067*4882a593Smuzhiyun  */
1068*4882a593Smuzhiyun 
1069*4882a593Smuzhiyun static struct attribute *i7core_addrmatch_attrs[] = {
1070*4882a593Smuzhiyun 	&dev_attr_channel.attr,
1071*4882a593Smuzhiyun 	&dev_attr_dimm.attr,
1072*4882a593Smuzhiyun 	&dev_attr_rank.attr,
1073*4882a593Smuzhiyun 	&dev_attr_bank.attr,
1074*4882a593Smuzhiyun 	&dev_attr_page.attr,
1075*4882a593Smuzhiyun 	&dev_attr_col.attr,
1076*4882a593Smuzhiyun 	NULL
1077*4882a593Smuzhiyun };
1078*4882a593Smuzhiyun 
1079*4882a593Smuzhiyun static const struct attribute_group addrmatch_grp = {
1080*4882a593Smuzhiyun 	.attrs	= i7core_addrmatch_attrs,
1081*4882a593Smuzhiyun };
1082*4882a593Smuzhiyun 
1083*4882a593Smuzhiyun static const struct attribute_group *addrmatch_groups[] = {
1084*4882a593Smuzhiyun 	&addrmatch_grp,
1085*4882a593Smuzhiyun 	NULL
1086*4882a593Smuzhiyun };
1087*4882a593Smuzhiyun 
addrmatch_release(struct device * device)1088*4882a593Smuzhiyun static void addrmatch_release(struct device *device)
1089*4882a593Smuzhiyun {
1090*4882a593Smuzhiyun 	edac_dbg(1, "Releasing device %s\n", dev_name(device));
1091*4882a593Smuzhiyun 	kfree(device);
1092*4882a593Smuzhiyun }
1093*4882a593Smuzhiyun 
1094*4882a593Smuzhiyun static const struct device_type addrmatch_type = {
1095*4882a593Smuzhiyun 	.groups		= addrmatch_groups,
1096*4882a593Smuzhiyun 	.release	= addrmatch_release,
1097*4882a593Smuzhiyun };
1098*4882a593Smuzhiyun 
1099*4882a593Smuzhiyun /*
1100*4882a593Smuzhiyun  * all_channel_counts sysfs struct
1101*4882a593Smuzhiyun  */
1102*4882a593Smuzhiyun 
1103*4882a593Smuzhiyun static struct attribute *i7core_udimm_counters_attrs[] = {
1104*4882a593Smuzhiyun 	&dev_attr_udimm0.attr,
1105*4882a593Smuzhiyun 	&dev_attr_udimm1.attr,
1106*4882a593Smuzhiyun 	&dev_attr_udimm2.attr,
1107*4882a593Smuzhiyun 	NULL
1108*4882a593Smuzhiyun };
1109*4882a593Smuzhiyun 
1110*4882a593Smuzhiyun static const struct attribute_group all_channel_counts_grp = {
1111*4882a593Smuzhiyun 	.attrs	= i7core_udimm_counters_attrs,
1112*4882a593Smuzhiyun };
1113*4882a593Smuzhiyun 
1114*4882a593Smuzhiyun static const struct attribute_group *all_channel_counts_groups[] = {
1115*4882a593Smuzhiyun 	&all_channel_counts_grp,
1116*4882a593Smuzhiyun 	NULL
1117*4882a593Smuzhiyun };
1118*4882a593Smuzhiyun 
all_channel_counts_release(struct device * device)1119*4882a593Smuzhiyun static void all_channel_counts_release(struct device *device)
1120*4882a593Smuzhiyun {
1121*4882a593Smuzhiyun 	edac_dbg(1, "Releasing device %s\n", dev_name(device));
1122*4882a593Smuzhiyun 	kfree(device);
1123*4882a593Smuzhiyun }
1124*4882a593Smuzhiyun 
1125*4882a593Smuzhiyun static const struct device_type all_channel_counts_type = {
1126*4882a593Smuzhiyun 	.groups		= all_channel_counts_groups,
1127*4882a593Smuzhiyun 	.release	= all_channel_counts_release,
1128*4882a593Smuzhiyun };
1129*4882a593Smuzhiyun 
1130*4882a593Smuzhiyun /*
1131*4882a593Smuzhiyun  * inject sysfs attributes
1132*4882a593Smuzhiyun  */
1133*4882a593Smuzhiyun 
1134*4882a593Smuzhiyun static DEVICE_ATTR(inject_section, S_IRUGO | S_IWUSR,
1135*4882a593Smuzhiyun 		   i7core_inject_section_show, i7core_inject_section_store);
1136*4882a593Smuzhiyun 
1137*4882a593Smuzhiyun static DEVICE_ATTR(inject_type, S_IRUGO | S_IWUSR,
1138*4882a593Smuzhiyun 		   i7core_inject_type_show, i7core_inject_type_store);
1139*4882a593Smuzhiyun 
1140*4882a593Smuzhiyun 
1141*4882a593Smuzhiyun static DEVICE_ATTR(inject_eccmask, S_IRUGO | S_IWUSR,
1142*4882a593Smuzhiyun 		   i7core_inject_eccmask_show, i7core_inject_eccmask_store);
1143*4882a593Smuzhiyun 
1144*4882a593Smuzhiyun static DEVICE_ATTR(inject_enable, S_IRUGO | S_IWUSR,
1145*4882a593Smuzhiyun 		   i7core_inject_enable_show, i7core_inject_enable_store);
1146*4882a593Smuzhiyun 
1147*4882a593Smuzhiyun static struct attribute *i7core_dev_attrs[] = {
1148*4882a593Smuzhiyun 	&dev_attr_inject_section.attr,
1149*4882a593Smuzhiyun 	&dev_attr_inject_type.attr,
1150*4882a593Smuzhiyun 	&dev_attr_inject_eccmask.attr,
1151*4882a593Smuzhiyun 	&dev_attr_inject_enable.attr,
1152*4882a593Smuzhiyun 	NULL
1153*4882a593Smuzhiyun };
1154*4882a593Smuzhiyun 
1155*4882a593Smuzhiyun ATTRIBUTE_GROUPS(i7core_dev);
1156*4882a593Smuzhiyun 
i7core_create_sysfs_devices(struct mem_ctl_info * mci)1157*4882a593Smuzhiyun static int i7core_create_sysfs_devices(struct mem_ctl_info *mci)
1158*4882a593Smuzhiyun {
1159*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
1160*4882a593Smuzhiyun 	int rc;
1161*4882a593Smuzhiyun 
1162*4882a593Smuzhiyun 	pvt->addrmatch_dev = kzalloc(sizeof(*pvt->addrmatch_dev), GFP_KERNEL);
1163*4882a593Smuzhiyun 	if (!pvt->addrmatch_dev)
1164*4882a593Smuzhiyun 		return -ENOMEM;
1165*4882a593Smuzhiyun 
1166*4882a593Smuzhiyun 	pvt->addrmatch_dev->type = &addrmatch_type;
1167*4882a593Smuzhiyun 	pvt->addrmatch_dev->bus = mci->dev.bus;
1168*4882a593Smuzhiyun 	device_initialize(pvt->addrmatch_dev);
1169*4882a593Smuzhiyun 	pvt->addrmatch_dev->parent = &mci->dev;
1170*4882a593Smuzhiyun 	dev_set_name(pvt->addrmatch_dev, "inject_addrmatch");
1171*4882a593Smuzhiyun 	dev_set_drvdata(pvt->addrmatch_dev, mci);
1172*4882a593Smuzhiyun 
1173*4882a593Smuzhiyun 	edac_dbg(1, "creating %s\n", dev_name(pvt->addrmatch_dev));
1174*4882a593Smuzhiyun 
1175*4882a593Smuzhiyun 	rc = device_add(pvt->addrmatch_dev);
1176*4882a593Smuzhiyun 	if (rc < 0)
1177*4882a593Smuzhiyun 		goto err_put_addrmatch;
1178*4882a593Smuzhiyun 
1179*4882a593Smuzhiyun 	if (!pvt->is_registered) {
1180*4882a593Smuzhiyun 		pvt->chancounts_dev = kzalloc(sizeof(*pvt->chancounts_dev),
1181*4882a593Smuzhiyun 					      GFP_KERNEL);
1182*4882a593Smuzhiyun 		if (!pvt->chancounts_dev) {
1183*4882a593Smuzhiyun 			rc = -ENOMEM;
1184*4882a593Smuzhiyun 			goto err_del_addrmatch;
1185*4882a593Smuzhiyun 		}
1186*4882a593Smuzhiyun 
1187*4882a593Smuzhiyun 		pvt->chancounts_dev->type = &all_channel_counts_type;
1188*4882a593Smuzhiyun 		pvt->chancounts_dev->bus = mci->dev.bus;
1189*4882a593Smuzhiyun 		device_initialize(pvt->chancounts_dev);
1190*4882a593Smuzhiyun 		pvt->chancounts_dev->parent = &mci->dev;
1191*4882a593Smuzhiyun 		dev_set_name(pvt->chancounts_dev, "all_channel_counts");
1192*4882a593Smuzhiyun 		dev_set_drvdata(pvt->chancounts_dev, mci);
1193*4882a593Smuzhiyun 
1194*4882a593Smuzhiyun 		edac_dbg(1, "creating %s\n", dev_name(pvt->chancounts_dev));
1195*4882a593Smuzhiyun 
1196*4882a593Smuzhiyun 		rc = device_add(pvt->chancounts_dev);
1197*4882a593Smuzhiyun 		if (rc < 0)
1198*4882a593Smuzhiyun 			goto err_put_chancounts;
1199*4882a593Smuzhiyun 	}
1200*4882a593Smuzhiyun 	return 0;
1201*4882a593Smuzhiyun 
1202*4882a593Smuzhiyun err_put_chancounts:
1203*4882a593Smuzhiyun 	put_device(pvt->chancounts_dev);
1204*4882a593Smuzhiyun err_del_addrmatch:
1205*4882a593Smuzhiyun 	device_del(pvt->addrmatch_dev);
1206*4882a593Smuzhiyun err_put_addrmatch:
1207*4882a593Smuzhiyun 	put_device(pvt->addrmatch_dev);
1208*4882a593Smuzhiyun 
1209*4882a593Smuzhiyun 	return rc;
1210*4882a593Smuzhiyun }
1211*4882a593Smuzhiyun 
i7core_delete_sysfs_devices(struct mem_ctl_info * mci)1212*4882a593Smuzhiyun static void i7core_delete_sysfs_devices(struct mem_ctl_info *mci)
1213*4882a593Smuzhiyun {
1214*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
1215*4882a593Smuzhiyun 
1216*4882a593Smuzhiyun 	edac_dbg(1, "\n");
1217*4882a593Smuzhiyun 
1218*4882a593Smuzhiyun 	if (!pvt->is_registered) {
1219*4882a593Smuzhiyun 		device_del(pvt->chancounts_dev);
1220*4882a593Smuzhiyun 		put_device(pvt->chancounts_dev);
1221*4882a593Smuzhiyun 	}
1222*4882a593Smuzhiyun 	device_del(pvt->addrmatch_dev);
1223*4882a593Smuzhiyun 	put_device(pvt->addrmatch_dev);
1224*4882a593Smuzhiyun }
1225*4882a593Smuzhiyun 
1226*4882a593Smuzhiyun /****************************************************************************
1227*4882a593Smuzhiyun 	Device initialization routines: put/get, init/exit
1228*4882a593Smuzhiyun  ****************************************************************************/
1229*4882a593Smuzhiyun 
1230*4882a593Smuzhiyun /*
1231*4882a593Smuzhiyun  *	i7core_put_all_devices	'put' all the devices that we have
1232*4882a593Smuzhiyun  *				reserved via 'get'
1233*4882a593Smuzhiyun  */
i7core_put_devices(struct i7core_dev * i7core_dev)1234*4882a593Smuzhiyun static void i7core_put_devices(struct i7core_dev *i7core_dev)
1235*4882a593Smuzhiyun {
1236*4882a593Smuzhiyun 	int i;
1237*4882a593Smuzhiyun 
1238*4882a593Smuzhiyun 	edac_dbg(0, "\n");
1239*4882a593Smuzhiyun 	for (i = 0; i < i7core_dev->n_devs; i++) {
1240*4882a593Smuzhiyun 		struct pci_dev *pdev = i7core_dev->pdev[i];
1241*4882a593Smuzhiyun 		if (!pdev)
1242*4882a593Smuzhiyun 			continue;
1243*4882a593Smuzhiyun 		edac_dbg(0, "Removing dev %02x:%02x.%d\n",
1244*4882a593Smuzhiyun 			 pdev->bus->number,
1245*4882a593Smuzhiyun 			 PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn));
1246*4882a593Smuzhiyun 		pci_dev_put(pdev);
1247*4882a593Smuzhiyun 	}
1248*4882a593Smuzhiyun }
1249*4882a593Smuzhiyun 
i7core_put_all_devices(void)1250*4882a593Smuzhiyun static void i7core_put_all_devices(void)
1251*4882a593Smuzhiyun {
1252*4882a593Smuzhiyun 	struct i7core_dev *i7core_dev, *tmp;
1253*4882a593Smuzhiyun 
1254*4882a593Smuzhiyun 	list_for_each_entry_safe(i7core_dev, tmp, &i7core_edac_list, list) {
1255*4882a593Smuzhiyun 		i7core_put_devices(i7core_dev);
1256*4882a593Smuzhiyun 		free_i7core_dev(i7core_dev);
1257*4882a593Smuzhiyun 	}
1258*4882a593Smuzhiyun }
1259*4882a593Smuzhiyun 
i7core_xeon_pci_fixup(const struct pci_id_table * table)1260*4882a593Smuzhiyun static void __init i7core_xeon_pci_fixup(const struct pci_id_table *table)
1261*4882a593Smuzhiyun {
1262*4882a593Smuzhiyun 	struct pci_dev *pdev = NULL;
1263*4882a593Smuzhiyun 	int i;
1264*4882a593Smuzhiyun 
1265*4882a593Smuzhiyun 	/*
1266*4882a593Smuzhiyun 	 * On Xeon 55xx, the Intel Quick Path Arch Generic Non-core pci buses
1267*4882a593Smuzhiyun 	 * aren't announced by acpi. So, we need to use a legacy scan probing
1268*4882a593Smuzhiyun 	 * to detect them
1269*4882a593Smuzhiyun 	 */
1270*4882a593Smuzhiyun 	while (table && table->descr) {
1271*4882a593Smuzhiyun 		pdev = pci_get_device(PCI_VENDOR_ID_INTEL, table->descr[0].dev_id, NULL);
1272*4882a593Smuzhiyun 		if (unlikely(!pdev)) {
1273*4882a593Smuzhiyun 			for (i = 0; i < MAX_SOCKET_BUSES; i++)
1274*4882a593Smuzhiyun 				pcibios_scan_specific_bus(255-i);
1275*4882a593Smuzhiyun 		}
1276*4882a593Smuzhiyun 		pci_dev_put(pdev);
1277*4882a593Smuzhiyun 		table++;
1278*4882a593Smuzhiyun 	}
1279*4882a593Smuzhiyun }
1280*4882a593Smuzhiyun 
i7core_pci_lastbus(void)1281*4882a593Smuzhiyun static unsigned i7core_pci_lastbus(void)
1282*4882a593Smuzhiyun {
1283*4882a593Smuzhiyun 	int last_bus = 0, bus;
1284*4882a593Smuzhiyun 	struct pci_bus *b = NULL;
1285*4882a593Smuzhiyun 
1286*4882a593Smuzhiyun 	while ((b = pci_find_next_bus(b)) != NULL) {
1287*4882a593Smuzhiyun 		bus = b->number;
1288*4882a593Smuzhiyun 		edac_dbg(0, "Found bus %d\n", bus);
1289*4882a593Smuzhiyun 		if (bus > last_bus)
1290*4882a593Smuzhiyun 			last_bus = bus;
1291*4882a593Smuzhiyun 	}
1292*4882a593Smuzhiyun 
1293*4882a593Smuzhiyun 	edac_dbg(0, "Last bus %d\n", last_bus);
1294*4882a593Smuzhiyun 
1295*4882a593Smuzhiyun 	return last_bus;
1296*4882a593Smuzhiyun }
1297*4882a593Smuzhiyun 
1298*4882a593Smuzhiyun /*
1299*4882a593Smuzhiyun  *	i7core_get_all_devices	Find and perform 'get' operation on the MCH's
1300*4882a593Smuzhiyun  *			device/functions we want to reference for this driver
1301*4882a593Smuzhiyun  *
1302*4882a593Smuzhiyun  *			Need to 'get' device 16 func 1 and func 2
1303*4882a593Smuzhiyun  */
i7core_get_onedevice(struct pci_dev ** prev,const struct pci_id_table * table,const unsigned devno,const unsigned last_bus)1304*4882a593Smuzhiyun static int i7core_get_onedevice(struct pci_dev **prev,
1305*4882a593Smuzhiyun 				const struct pci_id_table *table,
1306*4882a593Smuzhiyun 				const unsigned devno,
1307*4882a593Smuzhiyun 				const unsigned last_bus)
1308*4882a593Smuzhiyun {
1309*4882a593Smuzhiyun 	struct i7core_dev *i7core_dev;
1310*4882a593Smuzhiyun 	const struct pci_id_descr *dev_descr = &table->descr[devno];
1311*4882a593Smuzhiyun 
1312*4882a593Smuzhiyun 	struct pci_dev *pdev = NULL;
1313*4882a593Smuzhiyun 	u8 bus = 0;
1314*4882a593Smuzhiyun 	u8 socket = 0;
1315*4882a593Smuzhiyun 
1316*4882a593Smuzhiyun 	pdev = pci_get_device(PCI_VENDOR_ID_INTEL,
1317*4882a593Smuzhiyun 			      dev_descr->dev_id, *prev);
1318*4882a593Smuzhiyun 
1319*4882a593Smuzhiyun 	/*
1320*4882a593Smuzhiyun 	 * On Xeon 55xx, the Intel QuickPath Arch Generic Non-core regs
1321*4882a593Smuzhiyun 	 * is at addr 8086:2c40, instead of 8086:2c41. So, we need
1322*4882a593Smuzhiyun 	 * to probe for the alternate address in case of failure
1323*4882a593Smuzhiyun 	 */
1324*4882a593Smuzhiyun 	if (dev_descr->dev_id == PCI_DEVICE_ID_INTEL_I7_NONCORE && !pdev) {
1325*4882a593Smuzhiyun 		pci_dev_get(*prev);	/* pci_get_device will put it */
1326*4882a593Smuzhiyun 		pdev = pci_get_device(PCI_VENDOR_ID_INTEL,
1327*4882a593Smuzhiyun 				      PCI_DEVICE_ID_INTEL_I7_NONCORE_ALT, *prev);
1328*4882a593Smuzhiyun 	}
1329*4882a593Smuzhiyun 
1330*4882a593Smuzhiyun 	if (dev_descr->dev_id == PCI_DEVICE_ID_INTEL_LYNNFIELD_NONCORE &&
1331*4882a593Smuzhiyun 	    !pdev) {
1332*4882a593Smuzhiyun 		pci_dev_get(*prev);	/* pci_get_device will put it */
1333*4882a593Smuzhiyun 		pdev = pci_get_device(PCI_VENDOR_ID_INTEL,
1334*4882a593Smuzhiyun 				      PCI_DEVICE_ID_INTEL_LYNNFIELD_NONCORE_ALT,
1335*4882a593Smuzhiyun 				      *prev);
1336*4882a593Smuzhiyun 	}
1337*4882a593Smuzhiyun 
1338*4882a593Smuzhiyun 	if (!pdev) {
1339*4882a593Smuzhiyun 		if (*prev) {
1340*4882a593Smuzhiyun 			*prev = pdev;
1341*4882a593Smuzhiyun 			return 0;
1342*4882a593Smuzhiyun 		}
1343*4882a593Smuzhiyun 
1344*4882a593Smuzhiyun 		if (dev_descr->optional)
1345*4882a593Smuzhiyun 			return 0;
1346*4882a593Smuzhiyun 
1347*4882a593Smuzhiyun 		if (devno == 0)
1348*4882a593Smuzhiyun 			return -ENODEV;
1349*4882a593Smuzhiyun 
1350*4882a593Smuzhiyun 		i7core_printk(KERN_INFO,
1351*4882a593Smuzhiyun 			"Device not found: dev %02x.%d PCI ID %04x:%04x\n",
1352*4882a593Smuzhiyun 			dev_descr->dev, dev_descr->func,
1353*4882a593Smuzhiyun 			PCI_VENDOR_ID_INTEL, dev_descr->dev_id);
1354*4882a593Smuzhiyun 
1355*4882a593Smuzhiyun 		/* End of list, leave */
1356*4882a593Smuzhiyun 		return -ENODEV;
1357*4882a593Smuzhiyun 	}
1358*4882a593Smuzhiyun 	bus = pdev->bus->number;
1359*4882a593Smuzhiyun 
1360*4882a593Smuzhiyun 	socket = last_bus - bus;
1361*4882a593Smuzhiyun 
1362*4882a593Smuzhiyun 	i7core_dev = get_i7core_dev(socket);
1363*4882a593Smuzhiyun 	if (!i7core_dev) {
1364*4882a593Smuzhiyun 		i7core_dev = alloc_i7core_dev(socket, table);
1365*4882a593Smuzhiyun 		if (!i7core_dev) {
1366*4882a593Smuzhiyun 			pci_dev_put(pdev);
1367*4882a593Smuzhiyun 			return -ENOMEM;
1368*4882a593Smuzhiyun 		}
1369*4882a593Smuzhiyun 	}
1370*4882a593Smuzhiyun 
1371*4882a593Smuzhiyun 	if (i7core_dev->pdev[devno]) {
1372*4882a593Smuzhiyun 		i7core_printk(KERN_ERR,
1373*4882a593Smuzhiyun 			"Duplicated device for "
1374*4882a593Smuzhiyun 			"dev %02x:%02x.%d PCI ID %04x:%04x\n",
1375*4882a593Smuzhiyun 			bus, dev_descr->dev, dev_descr->func,
1376*4882a593Smuzhiyun 			PCI_VENDOR_ID_INTEL, dev_descr->dev_id);
1377*4882a593Smuzhiyun 		pci_dev_put(pdev);
1378*4882a593Smuzhiyun 		return -ENODEV;
1379*4882a593Smuzhiyun 	}
1380*4882a593Smuzhiyun 
1381*4882a593Smuzhiyun 	i7core_dev->pdev[devno] = pdev;
1382*4882a593Smuzhiyun 
1383*4882a593Smuzhiyun 	/* Sanity check */
1384*4882a593Smuzhiyun 	if (unlikely(PCI_SLOT(pdev->devfn) != dev_descr->dev ||
1385*4882a593Smuzhiyun 			PCI_FUNC(pdev->devfn) != dev_descr->func)) {
1386*4882a593Smuzhiyun 		i7core_printk(KERN_ERR,
1387*4882a593Smuzhiyun 			"Device PCI ID %04x:%04x "
1388*4882a593Smuzhiyun 			"has dev %02x:%02x.%d instead of dev %02x:%02x.%d\n",
1389*4882a593Smuzhiyun 			PCI_VENDOR_ID_INTEL, dev_descr->dev_id,
1390*4882a593Smuzhiyun 			bus, PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn),
1391*4882a593Smuzhiyun 			bus, dev_descr->dev, dev_descr->func);
1392*4882a593Smuzhiyun 		return -ENODEV;
1393*4882a593Smuzhiyun 	}
1394*4882a593Smuzhiyun 
1395*4882a593Smuzhiyun 	/* Be sure that the device is enabled */
1396*4882a593Smuzhiyun 	if (unlikely(pci_enable_device(pdev) < 0)) {
1397*4882a593Smuzhiyun 		i7core_printk(KERN_ERR,
1398*4882a593Smuzhiyun 			"Couldn't enable "
1399*4882a593Smuzhiyun 			"dev %02x:%02x.%d PCI ID %04x:%04x\n",
1400*4882a593Smuzhiyun 			bus, dev_descr->dev, dev_descr->func,
1401*4882a593Smuzhiyun 			PCI_VENDOR_ID_INTEL, dev_descr->dev_id);
1402*4882a593Smuzhiyun 		return -ENODEV;
1403*4882a593Smuzhiyun 	}
1404*4882a593Smuzhiyun 
1405*4882a593Smuzhiyun 	edac_dbg(0, "Detected socket %d dev %02x:%02x.%d PCI ID %04x:%04x\n",
1406*4882a593Smuzhiyun 		 socket, bus, dev_descr->dev,
1407*4882a593Smuzhiyun 		 dev_descr->func,
1408*4882a593Smuzhiyun 		 PCI_VENDOR_ID_INTEL, dev_descr->dev_id);
1409*4882a593Smuzhiyun 
1410*4882a593Smuzhiyun 	/*
1411*4882a593Smuzhiyun 	 * As stated on drivers/pci/search.c, the reference count for
1412*4882a593Smuzhiyun 	 * @from is always decremented if it is not %NULL. So, as we need
1413*4882a593Smuzhiyun 	 * to get all devices up to null, we need to do a get for the device
1414*4882a593Smuzhiyun 	 */
1415*4882a593Smuzhiyun 	pci_dev_get(pdev);
1416*4882a593Smuzhiyun 
1417*4882a593Smuzhiyun 	*prev = pdev;
1418*4882a593Smuzhiyun 
1419*4882a593Smuzhiyun 	return 0;
1420*4882a593Smuzhiyun }
1421*4882a593Smuzhiyun 
i7core_get_all_devices(void)1422*4882a593Smuzhiyun static int i7core_get_all_devices(void)
1423*4882a593Smuzhiyun {
1424*4882a593Smuzhiyun 	int i, rc, last_bus;
1425*4882a593Smuzhiyun 	struct pci_dev *pdev = NULL;
1426*4882a593Smuzhiyun 	const struct pci_id_table *table = pci_dev_table;
1427*4882a593Smuzhiyun 
1428*4882a593Smuzhiyun 	last_bus = i7core_pci_lastbus();
1429*4882a593Smuzhiyun 
1430*4882a593Smuzhiyun 	while (table && table->descr) {
1431*4882a593Smuzhiyun 		for (i = 0; i < table->n_devs; i++) {
1432*4882a593Smuzhiyun 			pdev = NULL;
1433*4882a593Smuzhiyun 			do {
1434*4882a593Smuzhiyun 				rc = i7core_get_onedevice(&pdev, table, i,
1435*4882a593Smuzhiyun 							  last_bus);
1436*4882a593Smuzhiyun 				if (rc < 0) {
1437*4882a593Smuzhiyun 					if (i == 0) {
1438*4882a593Smuzhiyun 						i = table->n_devs;
1439*4882a593Smuzhiyun 						break;
1440*4882a593Smuzhiyun 					}
1441*4882a593Smuzhiyun 					i7core_put_all_devices();
1442*4882a593Smuzhiyun 					return -ENODEV;
1443*4882a593Smuzhiyun 				}
1444*4882a593Smuzhiyun 			} while (pdev);
1445*4882a593Smuzhiyun 		}
1446*4882a593Smuzhiyun 		table++;
1447*4882a593Smuzhiyun 	}
1448*4882a593Smuzhiyun 
1449*4882a593Smuzhiyun 	return 0;
1450*4882a593Smuzhiyun }
1451*4882a593Smuzhiyun 
mci_bind_devs(struct mem_ctl_info * mci,struct i7core_dev * i7core_dev)1452*4882a593Smuzhiyun static int mci_bind_devs(struct mem_ctl_info *mci,
1453*4882a593Smuzhiyun 			 struct i7core_dev *i7core_dev)
1454*4882a593Smuzhiyun {
1455*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
1456*4882a593Smuzhiyun 	struct pci_dev *pdev;
1457*4882a593Smuzhiyun 	int i, func, slot;
1458*4882a593Smuzhiyun 	char *family;
1459*4882a593Smuzhiyun 
1460*4882a593Smuzhiyun 	pvt->is_registered = false;
1461*4882a593Smuzhiyun 	pvt->enable_scrub  = false;
1462*4882a593Smuzhiyun 	for (i = 0; i < i7core_dev->n_devs; i++) {
1463*4882a593Smuzhiyun 		pdev = i7core_dev->pdev[i];
1464*4882a593Smuzhiyun 		if (!pdev)
1465*4882a593Smuzhiyun 			continue;
1466*4882a593Smuzhiyun 
1467*4882a593Smuzhiyun 		func = PCI_FUNC(pdev->devfn);
1468*4882a593Smuzhiyun 		slot = PCI_SLOT(pdev->devfn);
1469*4882a593Smuzhiyun 		if (slot == 3) {
1470*4882a593Smuzhiyun 			if (unlikely(func > MAX_MCR_FUNC))
1471*4882a593Smuzhiyun 				goto error;
1472*4882a593Smuzhiyun 			pvt->pci_mcr[func] = pdev;
1473*4882a593Smuzhiyun 		} else if (likely(slot >= 4 && slot < 4 + NUM_CHANS)) {
1474*4882a593Smuzhiyun 			if (unlikely(func > MAX_CHAN_FUNC))
1475*4882a593Smuzhiyun 				goto error;
1476*4882a593Smuzhiyun 			pvt->pci_ch[slot - 4][func] = pdev;
1477*4882a593Smuzhiyun 		} else if (!slot && !func) {
1478*4882a593Smuzhiyun 			pvt->pci_noncore = pdev;
1479*4882a593Smuzhiyun 
1480*4882a593Smuzhiyun 			/* Detect the processor family */
1481*4882a593Smuzhiyun 			switch (pdev->device) {
1482*4882a593Smuzhiyun 			case PCI_DEVICE_ID_INTEL_I7_NONCORE:
1483*4882a593Smuzhiyun 				family = "Xeon 35xx/ i7core";
1484*4882a593Smuzhiyun 				pvt->enable_scrub = false;
1485*4882a593Smuzhiyun 				break;
1486*4882a593Smuzhiyun 			case PCI_DEVICE_ID_INTEL_LYNNFIELD_NONCORE_ALT:
1487*4882a593Smuzhiyun 				family = "i7-800/i5-700";
1488*4882a593Smuzhiyun 				pvt->enable_scrub = false;
1489*4882a593Smuzhiyun 				break;
1490*4882a593Smuzhiyun 			case PCI_DEVICE_ID_INTEL_LYNNFIELD_NONCORE:
1491*4882a593Smuzhiyun 				family = "Xeon 34xx";
1492*4882a593Smuzhiyun 				pvt->enable_scrub = false;
1493*4882a593Smuzhiyun 				break;
1494*4882a593Smuzhiyun 			case PCI_DEVICE_ID_INTEL_I7_NONCORE_ALT:
1495*4882a593Smuzhiyun 				family = "Xeon 55xx";
1496*4882a593Smuzhiyun 				pvt->enable_scrub = true;
1497*4882a593Smuzhiyun 				break;
1498*4882a593Smuzhiyun 			case PCI_DEVICE_ID_INTEL_LYNNFIELD_NONCORE_REV2:
1499*4882a593Smuzhiyun 				family = "Xeon 56xx / i7-900";
1500*4882a593Smuzhiyun 				pvt->enable_scrub = true;
1501*4882a593Smuzhiyun 				break;
1502*4882a593Smuzhiyun 			default:
1503*4882a593Smuzhiyun 				family = "unknown";
1504*4882a593Smuzhiyun 				pvt->enable_scrub = false;
1505*4882a593Smuzhiyun 			}
1506*4882a593Smuzhiyun 			edac_dbg(0, "Detected a processor type %s\n", family);
1507*4882a593Smuzhiyun 		} else
1508*4882a593Smuzhiyun 			goto error;
1509*4882a593Smuzhiyun 
1510*4882a593Smuzhiyun 		edac_dbg(0, "Associated fn %d.%d, dev = %p, socket %d\n",
1511*4882a593Smuzhiyun 			 PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn),
1512*4882a593Smuzhiyun 			 pdev, i7core_dev->socket);
1513*4882a593Smuzhiyun 
1514*4882a593Smuzhiyun 		if (PCI_SLOT(pdev->devfn) == 3 &&
1515*4882a593Smuzhiyun 			PCI_FUNC(pdev->devfn) == 2)
1516*4882a593Smuzhiyun 			pvt->is_registered = true;
1517*4882a593Smuzhiyun 	}
1518*4882a593Smuzhiyun 
1519*4882a593Smuzhiyun 	return 0;
1520*4882a593Smuzhiyun 
1521*4882a593Smuzhiyun error:
1522*4882a593Smuzhiyun 	i7core_printk(KERN_ERR, "Device %d, function %d "
1523*4882a593Smuzhiyun 		      "is out of the expected range\n",
1524*4882a593Smuzhiyun 		      slot, func);
1525*4882a593Smuzhiyun 	return -EINVAL;
1526*4882a593Smuzhiyun }
1527*4882a593Smuzhiyun 
1528*4882a593Smuzhiyun /****************************************************************************
1529*4882a593Smuzhiyun 			Error check routines
1530*4882a593Smuzhiyun  ****************************************************************************/
1531*4882a593Smuzhiyun 
i7core_rdimm_update_ce_count(struct mem_ctl_info * mci,const int chan,const int new0,const int new1,const int new2)1532*4882a593Smuzhiyun static void i7core_rdimm_update_ce_count(struct mem_ctl_info *mci,
1533*4882a593Smuzhiyun 					 const int chan,
1534*4882a593Smuzhiyun 					 const int new0,
1535*4882a593Smuzhiyun 					 const int new1,
1536*4882a593Smuzhiyun 					 const int new2)
1537*4882a593Smuzhiyun {
1538*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
1539*4882a593Smuzhiyun 	int add0 = 0, add1 = 0, add2 = 0;
1540*4882a593Smuzhiyun 	/* Updates CE counters if it is not the first time here */
1541*4882a593Smuzhiyun 	if (pvt->ce_count_available) {
1542*4882a593Smuzhiyun 		/* Updates CE counters */
1543*4882a593Smuzhiyun 
1544*4882a593Smuzhiyun 		add2 = new2 - pvt->rdimm_last_ce_count[chan][2];
1545*4882a593Smuzhiyun 		add1 = new1 - pvt->rdimm_last_ce_count[chan][1];
1546*4882a593Smuzhiyun 		add0 = new0 - pvt->rdimm_last_ce_count[chan][0];
1547*4882a593Smuzhiyun 
1548*4882a593Smuzhiyun 		if (add2 < 0)
1549*4882a593Smuzhiyun 			add2 += 0x7fff;
1550*4882a593Smuzhiyun 		pvt->rdimm_ce_count[chan][2] += add2;
1551*4882a593Smuzhiyun 
1552*4882a593Smuzhiyun 		if (add1 < 0)
1553*4882a593Smuzhiyun 			add1 += 0x7fff;
1554*4882a593Smuzhiyun 		pvt->rdimm_ce_count[chan][1] += add1;
1555*4882a593Smuzhiyun 
1556*4882a593Smuzhiyun 		if (add0 < 0)
1557*4882a593Smuzhiyun 			add0 += 0x7fff;
1558*4882a593Smuzhiyun 		pvt->rdimm_ce_count[chan][0] += add0;
1559*4882a593Smuzhiyun 	} else
1560*4882a593Smuzhiyun 		pvt->ce_count_available = 1;
1561*4882a593Smuzhiyun 
1562*4882a593Smuzhiyun 	/* Store the new values */
1563*4882a593Smuzhiyun 	pvt->rdimm_last_ce_count[chan][2] = new2;
1564*4882a593Smuzhiyun 	pvt->rdimm_last_ce_count[chan][1] = new1;
1565*4882a593Smuzhiyun 	pvt->rdimm_last_ce_count[chan][0] = new0;
1566*4882a593Smuzhiyun 
1567*4882a593Smuzhiyun 	/*updated the edac core */
1568*4882a593Smuzhiyun 	if (add0 != 0)
1569*4882a593Smuzhiyun 		edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, add0,
1570*4882a593Smuzhiyun 				     0, 0, 0,
1571*4882a593Smuzhiyun 				     chan, 0, -1, "error", "");
1572*4882a593Smuzhiyun 	if (add1 != 0)
1573*4882a593Smuzhiyun 		edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, add1,
1574*4882a593Smuzhiyun 				     0, 0, 0,
1575*4882a593Smuzhiyun 				     chan, 1, -1, "error", "");
1576*4882a593Smuzhiyun 	if (add2 != 0)
1577*4882a593Smuzhiyun 		edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, add2,
1578*4882a593Smuzhiyun 				     0, 0, 0,
1579*4882a593Smuzhiyun 				     chan, 2, -1, "error", "");
1580*4882a593Smuzhiyun }
1581*4882a593Smuzhiyun 
i7core_rdimm_check_mc_ecc_err(struct mem_ctl_info * mci)1582*4882a593Smuzhiyun static void i7core_rdimm_check_mc_ecc_err(struct mem_ctl_info *mci)
1583*4882a593Smuzhiyun {
1584*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
1585*4882a593Smuzhiyun 	u32 rcv[3][2];
1586*4882a593Smuzhiyun 	int i, new0, new1, new2;
1587*4882a593Smuzhiyun 
1588*4882a593Smuzhiyun 	/*Read DEV 3: FUN 2:  MC_COR_ECC_CNT regs directly*/
1589*4882a593Smuzhiyun 	pci_read_config_dword(pvt->pci_mcr[2], MC_COR_ECC_CNT_0,
1590*4882a593Smuzhiyun 								&rcv[0][0]);
1591*4882a593Smuzhiyun 	pci_read_config_dword(pvt->pci_mcr[2], MC_COR_ECC_CNT_1,
1592*4882a593Smuzhiyun 								&rcv[0][1]);
1593*4882a593Smuzhiyun 	pci_read_config_dword(pvt->pci_mcr[2], MC_COR_ECC_CNT_2,
1594*4882a593Smuzhiyun 								&rcv[1][0]);
1595*4882a593Smuzhiyun 	pci_read_config_dword(pvt->pci_mcr[2], MC_COR_ECC_CNT_3,
1596*4882a593Smuzhiyun 								&rcv[1][1]);
1597*4882a593Smuzhiyun 	pci_read_config_dword(pvt->pci_mcr[2], MC_COR_ECC_CNT_4,
1598*4882a593Smuzhiyun 								&rcv[2][0]);
1599*4882a593Smuzhiyun 	pci_read_config_dword(pvt->pci_mcr[2], MC_COR_ECC_CNT_5,
1600*4882a593Smuzhiyun 								&rcv[2][1]);
1601*4882a593Smuzhiyun 	for (i = 0 ; i < 3; i++) {
1602*4882a593Smuzhiyun 		edac_dbg(3, "MC_COR_ECC_CNT%d = 0x%x; MC_COR_ECC_CNT%d = 0x%x\n",
1603*4882a593Smuzhiyun 			 (i * 2), rcv[i][0], (i * 2) + 1, rcv[i][1]);
1604*4882a593Smuzhiyun 		/*if the channel has 3 dimms*/
1605*4882a593Smuzhiyun 		if (pvt->channel[i].dimms > 2) {
1606*4882a593Smuzhiyun 			new0 = DIMM_BOT_COR_ERR(rcv[i][0]);
1607*4882a593Smuzhiyun 			new1 = DIMM_TOP_COR_ERR(rcv[i][0]);
1608*4882a593Smuzhiyun 			new2 = DIMM_BOT_COR_ERR(rcv[i][1]);
1609*4882a593Smuzhiyun 		} else {
1610*4882a593Smuzhiyun 			new0 = DIMM_TOP_COR_ERR(rcv[i][0]) +
1611*4882a593Smuzhiyun 					DIMM_BOT_COR_ERR(rcv[i][0]);
1612*4882a593Smuzhiyun 			new1 = DIMM_TOP_COR_ERR(rcv[i][1]) +
1613*4882a593Smuzhiyun 					DIMM_BOT_COR_ERR(rcv[i][1]);
1614*4882a593Smuzhiyun 			new2 = 0;
1615*4882a593Smuzhiyun 		}
1616*4882a593Smuzhiyun 
1617*4882a593Smuzhiyun 		i7core_rdimm_update_ce_count(mci, i, new0, new1, new2);
1618*4882a593Smuzhiyun 	}
1619*4882a593Smuzhiyun }
1620*4882a593Smuzhiyun 
1621*4882a593Smuzhiyun /* This function is based on the device 3 function 4 registers as described on:
1622*4882a593Smuzhiyun  * Intel Xeon Processor 5500 Series Datasheet Volume 2
1623*4882a593Smuzhiyun  *	http://www.intel.com/Assets/PDF/datasheet/321322.pdf
1624*4882a593Smuzhiyun  * also available at:
1625*4882a593Smuzhiyun  * 	http://www.arrownac.com/manufacturers/intel/s/nehalem/5500-datasheet-v2.pdf
1626*4882a593Smuzhiyun  */
i7core_udimm_check_mc_ecc_err(struct mem_ctl_info * mci)1627*4882a593Smuzhiyun static void i7core_udimm_check_mc_ecc_err(struct mem_ctl_info *mci)
1628*4882a593Smuzhiyun {
1629*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
1630*4882a593Smuzhiyun 	u32 rcv1, rcv0;
1631*4882a593Smuzhiyun 	int new0, new1, new2;
1632*4882a593Smuzhiyun 
1633*4882a593Smuzhiyun 	if (!pvt->pci_mcr[4]) {
1634*4882a593Smuzhiyun 		edac_dbg(0, "MCR registers not found\n");
1635*4882a593Smuzhiyun 		return;
1636*4882a593Smuzhiyun 	}
1637*4882a593Smuzhiyun 
1638*4882a593Smuzhiyun 	/* Corrected test errors */
1639*4882a593Smuzhiyun 	pci_read_config_dword(pvt->pci_mcr[4], MC_TEST_ERR_RCV1, &rcv1);
1640*4882a593Smuzhiyun 	pci_read_config_dword(pvt->pci_mcr[4], MC_TEST_ERR_RCV0, &rcv0);
1641*4882a593Smuzhiyun 
1642*4882a593Smuzhiyun 	/* Store the new values */
1643*4882a593Smuzhiyun 	new2 = DIMM2_COR_ERR(rcv1);
1644*4882a593Smuzhiyun 	new1 = DIMM1_COR_ERR(rcv0);
1645*4882a593Smuzhiyun 	new0 = DIMM0_COR_ERR(rcv0);
1646*4882a593Smuzhiyun 
1647*4882a593Smuzhiyun 	/* Updates CE counters if it is not the first time here */
1648*4882a593Smuzhiyun 	if (pvt->ce_count_available) {
1649*4882a593Smuzhiyun 		/* Updates CE counters */
1650*4882a593Smuzhiyun 		int add0, add1, add2;
1651*4882a593Smuzhiyun 
1652*4882a593Smuzhiyun 		add2 = new2 - pvt->udimm_last_ce_count[2];
1653*4882a593Smuzhiyun 		add1 = new1 - pvt->udimm_last_ce_count[1];
1654*4882a593Smuzhiyun 		add0 = new0 - pvt->udimm_last_ce_count[0];
1655*4882a593Smuzhiyun 
1656*4882a593Smuzhiyun 		if (add2 < 0)
1657*4882a593Smuzhiyun 			add2 += 0x7fff;
1658*4882a593Smuzhiyun 		pvt->udimm_ce_count[2] += add2;
1659*4882a593Smuzhiyun 
1660*4882a593Smuzhiyun 		if (add1 < 0)
1661*4882a593Smuzhiyun 			add1 += 0x7fff;
1662*4882a593Smuzhiyun 		pvt->udimm_ce_count[1] += add1;
1663*4882a593Smuzhiyun 
1664*4882a593Smuzhiyun 		if (add0 < 0)
1665*4882a593Smuzhiyun 			add0 += 0x7fff;
1666*4882a593Smuzhiyun 		pvt->udimm_ce_count[0] += add0;
1667*4882a593Smuzhiyun 
1668*4882a593Smuzhiyun 		if (add0 | add1 | add2)
1669*4882a593Smuzhiyun 			i7core_printk(KERN_ERR, "New Corrected error(s): "
1670*4882a593Smuzhiyun 				      "dimm0: +%d, dimm1: +%d, dimm2 +%d\n",
1671*4882a593Smuzhiyun 				      add0, add1, add2);
1672*4882a593Smuzhiyun 	} else
1673*4882a593Smuzhiyun 		pvt->ce_count_available = 1;
1674*4882a593Smuzhiyun 
1675*4882a593Smuzhiyun 	/* Store the new values */
1676*4882a593Smuzhiyun 	pvt->udimm_last_ce_count[2] = new2;
1677*4882a593Smuzhiyun 	pvt->udimm_last_ce_count[1] = new1;
1678*4882a593Smuzhiyun 	pvt->udimm_last_ce_count[0] = new0;
1679*4882a593Smuzhiyun }
1680*4882a593Smuzhiyun 
1681*4882a593Smuzhiyun /*
1682*4882a593Smuzhiyun  * According with tables E-11 and E-12 of chapter E.3.3 of Intel 64 and IA-32
1683*4882a593Smuzhiyun  * Architectures Software Developer’s Manual Volume 3B.
1684*4882a593Smuzhiyun  * Nehalem are defined as family 0x06, model 0x1a
1685*4882a593Smuzhiyun  *
1686*4882a593Smuzhiyun  * The MCA registers used here are the following ones:
1687*4882a593Smuzhiyun  *     struct mce field	MCA Register
1688*4882a593Smuzhiyun  *     m->status	MSR_IA32_MC8_STATUS
1689*4882a593Smuzhiyun  *     m->addr		MSR_IA32_MC8_ADDR
1690*4882a593Smuzhiyun  *     m->misc		MSR_IA32_MC8_MISC
1691*4882a593Smuzhiyun  * In the case of Nehalem, the error information is masked at .status and .misc
1692*4882a593Smuzhiyun  * fields
1693*4882a593Smuzhiyun  */
i7core_mce_output_error(struct mem_ctl_info * mci,const struct mce * m)1694*4882a593Smuzhiyun static void i7core_mce_output_error(struct mem_ctl_info *mci,
1695*4882a593Smuzhiyun 				    const struct mce *m)
1696*4882a593Smuzhiyun {
1697*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
1698*4882a593Smuzhiyun 	char *optype, *err;
1699*4882a593Smuzhiyun 	enum hw_event_mc_err_type tp_event;
1700*4882a593Smuzhiyun 	unsigned long error = m->status & 0x1ff0000l;
1701*4882a593Smuzhiyun 	bool uncorrected_error = m->mcgstatus & 1ll << 61;
1702*4882a593Smuzhiyun 	bool ripv = m->mcgstatus & 1;
1703*4882a593Smuzhiyun 	u32 optypenum = (m->status >> 4) & 0x07;
1704*4882a593Smuzhiyun 	u32 core_err_cnt = (m->status >> 38) & 0x7fff;
1705*4882a593Smuzhiyun 	u32 dimm = (m->misc >> 16) & 0x3;
1706*4882a593Smuzhiyun 	u32 channel = (m->misc >> 18) & 0x3;
1707*4882a593Smuzhiyun 	u32 syndrome = m->misc >> 32;
1708*4882a593Smuzhiyun 	u32 errnum = find_first_bit(&error, 32);
1709*4882a593Smuzhiyun 
1710*4882a593Smuzhiyun 	if (uncorrected_error) {
1711*4882a593Smuzhiyun 		core_err_cnt = 1;
1712*4882a593Smuzhiyun 		if (ripv)
1713*4882a593Smuzhiyun 			tp_event = HW_EVENT_ERR_UNCORRECTED;
1714*4882a593Smuzhiyun 		else
1715*4882a593Smuzhiyun 			tp_event = HW_EVENT_ERR_FATAL;
1716*4882a593Smuzhiyun 	} else {
1717*4882a593Smuzhiyun 		tp_event = HW_EVENT_ERR_CORRECTED;
1718*4882a593Smuzhiyun 	}
1719*4882a593Smuzhiyun 
1720*4882a593Smuzhiyun 	switch (optypenum) {
1721*4882a593Smuzhiyun 	case 0:
1722*4882a593Smuzhiyun 		optype = "generic undef request";
1723*4882a593Smuzhiyun 		break;
1724*4882a593Smuzhiyun 	case 1:
1725*4882a593Smuzhiyun 		optype = "read error";
1726*4882a593Smuzhiyun 		break;
1727*4882a593Smuzhiyun 	case 2:
1728*4882a593Smuzhiyun 		optype = "write error";
1729*4882a593Smuzhiyun 		break;
1730*4882a593Smuzhiyun 	case 3:
1731*4882a593Smuzhiyun 		optype = "addr/cmd error";
1732*4882a593Smuzhiyun 		break;
1733*4882a593Smuzhiyun 	case 4:
1734*4882a593Smuzhiyun 		optype = "scrubbing error";
1735*4882a593Smuzhiyun 		break;
1736*4882a593Smuzhiyun 	default:
1737*4882a593Smuzhiyun 		optype = "reserved";
1738*4882a593Smuzhiyun 		break;
1739*4882a593Smuzhiyun 	}
1740*4882a593Smuzhiyun 
1741*4882a593Smuzhiyun 	switch (errnum) {
1742*4882a593Smuzhiyun 	case 16:
1743*4882a593Smuzhiyun 		err = "read ECC error";
1744*4882a593Smuzhiyun 		break;
1745*4882a593Smuzhiyun 	case 17:
1746*4882a593Smuzhiyun 		err = "RAS ECC error";
1747*4882a593Smuzhiyun 		break;
1748*4882a593Smuzhiyun 	case 18:
1749*4882a593Smuzhiyun 		err = "write parity error";
1750*4882a593Smuzhiyun 		break;
1751*4882a593Smuzhiyun 	case 19:
1752*4882a593Smuzhiyun 		err = "redundancy loss";
1753*4882a593Smuzhiyun 		break;
1754*4882a593Smuzhiyun 	case 20:
1755*4882a593Smuzhiyun 		err = "reserved";
1756*4882a593Smuzhiyun 		break;
1757*4882a593Smuzhiyun 	case 21:
1758*4882a593Smuzhiyun 		err = "memory range error";
1759*4882a593Smuzhiyun 		break;
1760*4882a593Smuzhiyun 	case 22:
1761*4882a593Smuzhiyun 		err = "RTID out of range";
1762*4882a593Smuzhiyun 		break;
1763*4882a593Smuzhiyun 	case 23:
1764*4882a593Smuzhiyun 		err = "address parity error";
1765*4882a593Smuzhiyun 		break;
1766*4882a593Smuzhiyun 	case 24:
1767*4882a593Smuzhiyun 		err = "byte enable parity error";
1768*4882a593Smuzhiyun 		break;
1769*4882a593Smuzhiyun 	default:
1770*4882a593Smuzhiyun 		err = "unknown";
1771*4882a593Smuzhiyun 	}
1772*4882a593Smuzhiyun 
1773*4882a593Smuzhiyun 	/*
1774*4882a593Smuzhiyun 	 * Call the helper to output message
1775*4882a593Smuzhiyun 	 * FIXME: what to do if core_err_cnt > 1? Currently, it generates
1776*4882a593Smuzhiyun 	 * only one event
1777*4882a593Smuzhiyun 	 */
1778*4882a593Smuzhiyun 	if (uncorrected_error || !pvt->is_registered)
1779*4882a593Smuzhiyun 		edac_mc_handle_error(tp_event, mci, core_err_cnt,
1780*4882a593Smuzhiyun 				     m->addr >> PAGE_SHIFT,
1781*4882a593Smuzhiyun 				     m->addr & ~PAGE_MASK,
1782*4882a593Smuzhiyun 				     syndrome,
1783*4882a593Smuzhiyun 				     channel, dimm, -1,
1784*4882a593Smuzhiyun 				     err, optype);
1785*4882a593Smuzhiyun }
1786*4882a593Smuzhiyun 
1787*4882a593Smuzhiyun /*
1788*4882a593Smuzhiyun  *	i7core_check_error	Retrieve and process errors reported by the
1789*4882a593Smuzhiyun  *				hardware. Called by the Core module.
1790*4882a593Smuzhiyun  */
i7core_check_error(struct mem_ctl_info * mci,struct mce * m)1791*4882a593Smuzhiyun static void i7core_check_error(struct mem_ctl_info *mci, struct mce *m)
1792*4882a593Smuzhiyun {
1793*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
1794*4882a593Smuzhiyun 
1795*4882a593Smuzhiyun 	i7core_mce_output_error(mci, m);
1796*4882a593Smuzhiyun 
1797*4882a593Smuzhiyun 	/*
1798*4882a593Smuzhiyun 	 * Now, let's increment CE error counts
1799*4882a593Smuzhiyun 	 */
1800*4882a593Smuzhiyun 	if (!pvt->is_registered)
1801*4882a593Smuzhiyun 		i7core_udimm_check_mc_ecc_err(mci);
1802*4882a593Smuzhiyun 	else
1803*4882a593Smuzhiyun 		i7core_rdimm_check_mc_ecc_err(mci);
1804*4882a593Smuzhiyun }
1805*4882a593Smuzhiyun 
1806*4882a593Smuzhiyun /*
1807*4882a593Smuzhiyun  * Check that logging is enabled and that this is the right type
1808*4882a593Smuzhiyun  * of error for us to handle.
1809*4882a593Smuzhiyun  */
i7core_mce_check_error(struct notifier_block * nb,unsigned long val,void * data)1810*4882a593Smuzhiyun static int i7core_mce_check_error(struct notifier_block *nb, unsigned long val,
1811*4882a593Smuzhiyun 				  void *data)
1812*4882a593Smuzhiyun {
1813*4882a593Smuzhiyun 	struct mce *mce = (struct mce *)data;
1814*4882a593Smuzhiyun 	struct i7core_dev *i7_dev;
1815*4882a593Smuzhiyun 	struct mem_ctl_info *mci;
1816*4882a593Smuzhiyun 
1817*4882a593Smuzhiyun 	i7_dev = get_i7core_dev(mce->socketid);
1818*4882a593Smuzhiyun 	if (!i7_dev || (mce->kflags & MCE_HANDLED_CEC))
1819*4882a593Smuzhiyun 		return NOTIFY_DONE;
1820*4882a593Smuzhiyun 
1821*4882a593Smuzhiyun 	mci = i7_dev->mci;
1822*4882a593Smuzhiyun 
1823*4882a593Smuzhiyun 	/*
1824*4882a593Smuzhiyun 	 * Just let mcelog handle it if the error is
1825*4882a593Smuzhiyun 	 * outside the memory controller
1826*4882a593Smuzhiyun 	 */
1827*4882a593Smuzhiyun 	if (((mce->status & 0xffff) >> 7) != 1)
1828*4882a593Smuzhiyun 		return NOTIFY_DONE;
1829*4882a593Smuzhiyun 
1830*4882a593Smuzhiyun 	/* Bank 8 registers are the only ones that we know how to handle */
1831*4882a593Smuzhiyun 	if (mce->bank != 8)
1832*4882a593Smuzhiyun 		return NOTIFY_DONE;
1833*4882a593Smuzhiyun 
1834*4882a593Smuzhiyun 	i7core_check_error(mci, mce);
1835*4882a593Smuzhiyun 
1836*4882a593Smuzhiyun 	/* Advise mcelog that the errors were handled */
1837*4882a593Smuzhiyun 	mce->kflags |= MCE_HANDLED_EDAC;
1838*4882a593Smuzhiyun 	return NOTIFY_OK;
1839*4882a593Smuzhiyun }
1840*4882a593Smuzhiyun 
1841*4882a593Smuzhiyun static struct notifier_block i7_mce_dec = {
1842*4882a593Smuzhiyun 	.notifier_call	= i7core_mce_check_error,
1843*4882a593Smuzhiyun 	.priority	= MCE_PRIO_EDAC,
1844*4882a593Smuzhiyun };
1845*4882a593Smuzhiyun 
1846*4882a593Smuzhiyun struct memdev_dmi_entry {
1847*4882a593Smuzhiyun 	u8 type;
1848*4882a593Smuzhiyun 	u8 length;
1849*4882a593Smuzhiyun 	u16 handle;
1850*4882a593Smuzhiyun 	u16 phys_mem_array_handle;
1851*4882a593Smuzhiyun 	u16 mem_err_info_handle;
1852*4882a593Smuzhiyun 	u16 total_width;
1853*4882a593Smuzhiyun 	u16 data_width;
1854*4882a593Smuzhiyun 	u16 size;
1855*4882a593Smuzhiyun 	u8 form;
1856*4882a593Smuzhiyun 	u8 device_set;
1857*4882a593Smuzhiyun 	u8 device_locator;
1858*4882a593Smuzhiyun 	u8 bank_locator;
1859*4882a593Smuzhiyun 	u8 memory_type;
1860*4882a593Smuzhiyun 	u16 type_detail;
1861*4882a593Smuzhiyun 	u16 speed;
1862*4882a593Smuzhiyun 	u8 manufacturer;
1863*4882a593Smuzhiyun 	u8 serial_number;
1864*4882a593Smuzhiyun 	u8 asset_tag;
1865*4882a593Smuzhiyun 	u8 part_number;
1866*4882a593Smuzhiyun 	u8 attributes;
1867*4882a593Smuzhiyun 	u32 extended_size;
1868*4882a593Smuzhiyun 	u16 conf_mem_clk_speed;
1869*4882a593Smuzhiyun } __attribute__((__packed__));
1870*4882a593Smuzhiyun 
1871*4882a593Smuzhiyun 
1872*4882a593Smuzhiyun /*
1873*4882a593Smuzhiyun  * Decode the DRAM Clock Frequency, be paranoid, make sure that all
1874*4882a593Smuzhiyun  * memory devices show the same speed, and if they don't then consider
1875*4882a593Smuzhiyun  * all speeds to be invalid.
1876*4882a593Smuzhiyun  */
decode_dclk(const struct dmi_header * dh,void * _dclk_freq)1877*4882a593Smuzhiyun static void decode_dclk(const struct dmi_header *dh, void *_dclk_freq)
1878*4882a593Smuzhiyun {
1879*4882a593Smuzhiyun 	int *dclk_freq = _dclk_freq;
1880*4882a593Smuzhiyun 	u16 dmi_mem_clk_speed;
1881*4882a593Smuzhiyun 
1882*4882a593Smuzhiyun 	if (*dclk_freq == -1)
1883*4882a593Smuzhiyun 		return;
1884*4882a593Smuzhiyun 
1885*4882a593Smuzhiyun 	if (dh->type == DMI_ENTRY_MEM_DEVICE) {
1886*4882a593Smuzhiyun 		struct memdev_dmi_entry *memdev_dmi_entry =
1887*4882a593Smuzhiyun 			(struct memdev_dmi_entry *)dh;
1888*4882a593Smuzhiyun 		unsigned long conf_mem_clk_speed_offset =
1889*4882a593Smuzhiyun 			(unsigned long)&memdev_dmi_entry->conf_mem_clk_speed -
1890*4882a593Smuzhiyun 			(unsigned long)&memdev_dmi_entry->type;
1891*4882a593Smuzhiyun 		unsigned long speed_offset =
1892*4882a593Smuzhiyun 			(unsigned long)&memdev_dmi_entry->speed -
1893*4882a593Smuzhiyun 			(unsigned long)&memdev_dmi_entry->type;
1894*4882a593Smuzhiyun 
1895*4882a593Smuzhiyun 		/* Check that a DIMM is present */
1896*4882a593Smuzhiyun 		if (memdev_dmi_entry->size == 0)
1897*4882a593Smuzhiyun 			return;
1898*4882a593Smuzhiyun 
1899*4882a593Smuzhiyun 		/*
1900*4882a593Smuzhiyun 		 * Pick the configured speed if it's available, otherwise
1901*4882a593Smuzhiyun 		 * pick the DIMM speed, or we don't have a speed.
1902*4882a593Smuzhiyun 		 */
1903*4882a593Smuzhiyun 		if (memdev_dmi_entry->length > conf_mem_clk_speed_offset) {
1904*4882a593Smuzhiyun 			dmi_mem_clk_speed =
1905*4882a593Smuzhiyun 				memdev_dmi_entry->conf_mem_clk_speed;
1906*4882a593Smuzhiyun 		} else if (memdev_dmi_entry->length > speed_offset) {
1907*4882a593Smuzhiyun 			dmi_mem_clk_speed = memdev_dmi_entry->speed;
1908*4882a593Smuzhiyun 		} else {
1909*4882a593Smuzhiyun 			*dclk_freq = -1;
1910*4882a593Smuzhiyun 			return;
1911*4882a593Smuzhiyun 		}
1912*4882a593Smuzhiyun 
1913*4882a593Smuzhiyun 		if (*dclk_freq == 0) {
1914*4882a593Smuzhiyun 			/* First pass, speed was 0 */
1915*4882a593Smuzhiyun 			if (dmi_mem_clk_speed > 0) {
1916*4882a593Smuzhiyun 				/* Set speed if a valid speed is read */
1917*4882a593Smuzhiyun 				*dclk_freq = dmi_mem_clk_speed;
1918*4882a593Smuzhiyun 			} else {
1919*4882a593Smuzhiyun 				/* Otherwise we don't have a valid speed */
1920*4882a593Smuzhiyun 				*dclk_freq = -1;
1921*4882a593Smuzhiyun 			}
1922*4882a593Smuzhiyun 		} else if (*dclk_freq > 0 &&
1923*4882a593Smuzhiyun 			   *dclk_freq != dmi_mem_clk_speed) {
1924*4882a593Smuzhiyun 			/*
1925*4882a593Smuzhiyun 			 * If we have a speed, check that all DIMMS are the same
1926*4882a593Smuzhiyun 			 * speed, otherwise set the speed as invalid.
1927*4882a593Smuzhiyun 			 */
1928*4882a593Smuzhiyun 			*dclk_freq = -1;
1929*4882a593Smuzhiyun 		}
1930*4882a593Smuzhiyun 	}
1931*4882a593Smuzhiyun }
1932*4882a593Smuzhiyun 
1933*4882a593Smuzhiyun /*
1934*4882a593Smuzhiyun  * The default DCLK frequency is used as a fallback if we
1935*4882a593Smuzhiyun  * fail to find anything reliable in the DMI. The value
1936*4882a593Smuzhiyun  * is taken straight from the datasheet.
1937*4882a593Smuzhiyun  */
1938*4882a593Smuzhiyun #define DEFAULT_DCLK_FREQ 800
1939*4882a593Smuzhiyun 
get_dclk_freq(void)1940*4882a593Smuzhiyun static int get_dclk_freq(void)
1941*4882a593Smuzhiyun {
1942*4882a593Smuzhiyun 	int dclk_freq = 0;
1943*4882a593Smuzhiyun 
1944*4882a593Smuzhiyun 	dmi_walk(decode_dclk, (void *)&dclk_freq);
1945*4882a593Smuzhiyun 
1946*4882a593Smuzhiyun 	if (dclk_freq < 1)
1947*4882a593Smuzhiyun 		return DEFAULT_DCLK_FREQ;
1948*4882a593Smuzhiyun 
1949*4882a593Smuzhiyun 	return dclk_freq;
1950*4882a593Smuzhiyun }
1951*4882a593Smuzhiyun 
1952*4882a593Smuzhiyun /*
1953*4882a593Smuzhiyun  * set_sdram_scrub_rate		This routine sets byte/sec bandwidth scrub rate
1954*4882a593Smuzhiyun  *				to hardware according to SCRUBINTERVAL formula
1955*4882a593Smuzhiyun  *				found in datasheet.
1956*4882a593Smuzhiyun  */
set_sdram_scrub_rate(struct mem_ctl_info * mci,u32 new_bw)1957*4882a593Smuzhiyun static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw)
1958*4882a593Smuzhiyun {
1959*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
1960*4882a593Smuzhiyun 	struct pci_dev *pdev;
1961*4882a593Smuzhiyun 	u32 dw_scrub;
1962*4882a593Smuzhiyun 	u32 dw_ssr;
1963*4882a593Smuzhiyun 
1964*4882a593Smuzhiyun 	/* Get data from the MC register, function 2 */
1965*4882a593Smuzhiyun 	pdev = pvt->pci_mcr[2];
1966*4882a593Smuzhiyun 	if (!pdev)
1967*4882a593Smuzhiyun 		return -ENODEV;
1968*4882a593Smuzhiyun 
1969*4882a593Smuzhiyun 	pci_read_config_dword(pdev, MC_SCRUB_CONTROL, &dw_scrub);
1970*4882a593Smuzhiyun 
1971*4882a593Smuzhiyun 	if (new_bw == 0) {
1972*4882a593Smuzhiyun 		/* Prepare to disable petrol scrub */
1973*4882a593Smuzhiyun 		dw_scrub &= ~STARTSCRUB;
1974*4882a593Smuzhiyun 		/* Stop the patrol scrub engine */
1975*4882a593Smuzhiyun 		write_and_test(pdev, MC_SCRUB_CONTROL,
1976*4882a593Smuzhiyun 			       dw_scrub & ~SCRUBINTERVAL_MASK);
1977*4882a593Smuzhiyun 
1978*4882a593Smuzhiyun 		/* Get current status of scrub rate and set bit to disable */
1979*4882a593Smuzhiyun 		pci_read_config_dword(pdev, MC_SSRCONTROL, &dw_ssr);
1980*4882a593Smuzhiyun 		dw_ssr &= ~SSR_MODE_MASK;
1981*4882a593Smuzhiyun 		dw_ssr |= SSR_MODE_DISABLE;
1982*4882a593Smuzhiyun 	} else {
1983*4882a593Smuzhiyun 		const int cache_line_size = 64;
1984*4882a593Smuzhiyun 		const u32 freq_dclk_mhz = pvt->dclk_freq;
1985*4882a593Smuzhiyun 		unsigned long long scrub_interval;
1986*4882a593Smuzhiyun 		/*
1987*4882a593Smuzhiyun 		 * Translate the desired scrub rate to a register value and
1988*4882a593Smuzhiyun 		 * program the corresponding register value.
1989*4882a593Smuzhiyun 		 */
1990*4882a593Smuzhiyun 		scrub_interval = (unsigned long long)freq_dclk_mhz *
1991*4882a593Smuzhiyun 			cache_line_size * 1000000;
1992*4882a593Smuzhiyun 		do_div(scrub_interval, new_bw);
1993*4882a593Smuzhiyun 
1994*4882a593Smuzhiyun 		if (!scrub_interval || scrub_interval > SCRUBINTERVAL_MASK)
1995*4882a593Smuzhiyun 			return -EINVAL;
1996*4882a593Smuzhiyun 
1997*4882a593Smuzhiyun 		dw_scrub = SCRUBINTERVAL_MASK & scrub_interval;
1998*4882a593Smuzhiyun 
1999*4882a593Smuzhiyun 		/* Start the patrol scrub engine */
2000*4882a593Smuzhiyun 		pci_write_config_dword(pdev, MC_SCRUB_CONTROL,
2001*4882a593Smuzhiyun 				       STARTSCRUB | dw_scrub);
2002*4882a593Smuzhiyun 
2003*4882a593Smuzhiyun 		/* Get current status of scrub rate and set bit to enable */
2004*4882a593Smuzhiyun 		pci_read_config_dword(pdev, MC_SSRCONTROL, &dw_ssr);
2005*4882a593Smuzhiyun 		dw_ssr &= ~SSR_MODE_MASK;
2006*4882a593Smuzhiyun 		dw_ssr |= SSR_MODE_ENABLE;
2007*4882a593Smuzhiyun 	}
2008*4882a593Smuzhiyun 	/* Disable or enable scrubbing */
2009*4882a593Smuzhiyun 	pci_write_config_dword(pdev, MC_SSRCONTROL, dw_ssr);
2010*4882a593Smuzhiyun 
2011*4882a593Smuzhiyun 	return new_bw;
2012*4882a593Smuzhiyun }
2013*4882a593Smuzhiyun 
2014*4882a593Smuzhiyun /*
2015*4882a593Smuzhiyun  * get_sdram_scrub_rate		This routine convert current scrub rate value
2016*4882a593Smuzhiyun  *				into byte/sec bandwidth according to
2017*4882a593Smuzhiyun  *				SCRUBINTERVAL formula found in datasheet.
2018*4882a593Smuzhiyun  */
get_sdram_scrub_rate(struct mem_ctl_info * mci)2019*4882a593Smuzhiyun static int get_sdram_scrub_rate(struct mem_ctl_info *mci)
2020*4882a593Smuzhiyun {
2021*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
2022*4882a593Smuzhiyun 	struct pci_dev *pdev;
2023*4882a593Smuzhiyun 	const u32 cache_line_size = 64;
2024*4882a593Smuzhiyun 	const u32 freq_dclk_mhz = pvt->dclk_freq;
2025*4882a593Smuzhiyun 	unsigned long long scrub_rate;
2026*4882a593Smuzhiyun 	u32 scrubval;
2027*4882a593Smuzhiyun 
2028*4882a593Smuzhiyun 	/* Get data from the MC register, function 2 */
2029*4882a593Smuzhiyun 	pdev = pvt->pci_mcr[2];
2030*4882a593Smuzhiyun 	if (!pdev)
2031*4882a593Smuzhiyun 		return -ENODEV;
2032*4882a593Smuzhiyun 
2033*4882a593Smuzhiyun 	/* Get current scrub control data */
2034*4882a593Smuzhiyun 	pci_read_config_dword(pdev, MC_SCRUB_CONTROL, &scrubval);
2035*4882a593Smuzhiyun 
2036*4882a593Smuzhiyun 	/* Mask highest 8-bits to 0 */
2037*4882a593Smuzhiyun 	scrubval &=  SCRUBINTERVAL_MASK;
2038*4882a593Smuzhiyun 	if (!scrubval)
2039*4882a593Smuzhiyun 		return 0;
2040*4882a593Smuzhiyun 
2041*4882a593Smuzhiyun 	/* Calculate scrub rate value into byte/sec bandwidth */
2042*4882a593Smuzhiyun 	scrub_rate =  (unsigned long long)freq_dclk_mhz *
2043*4882a593Smuzhiyun 		1000000 * cache_line_size;
2044*4882a593Smuzhiyun 	do_div(scrub_rate, scrubval);
2045*4882a593Smuzhiyun 	return (int)scrub_rate;
2046*4882a593Smuzhiyun }
2047*4882a593Smuzhiyun 
enable_sdram_scrub_setting(struct mem_ctl_info * mci)2048*4882a593Smuzhiyun static void enable_sdram_scrub_setting(struct mem_ctl_info *mci)
2049*4882a593Smuzhiyun {
2050*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
2051*4882a593Smuzhiyun 	u32 pci_lock;
2052*4882a593Smuzhiyun 
2053*4882a593Smuzhiyun 	/* Unlock writes to pci registers */
2054*4882a593Smuzhiyun 	pci_read_config_dword(pvt->pci_noncore, MC_CFG_CONTROL, &pci_lock);
2055*4882a593Smuzhiyun 	pci_lock &= ~0x3;
2056*4882a593Smuzhiyun 	pci_write_config_dword(pvt->pci_noncore, MC_CFG_CONTROL,
2057*4882a593Smuzhiyun 			       pci_lock | MC_CFG_UNLOCK);
2058*4882a593Smuzhiyun 
2059*4882a593Smuzhiyun 	mci->set_sdram_scrub_rate = set_sdram_scrub_rate;
2060*4882a593Smuzhiyun 	mci->get_sdram_scrub_rate = get_sdram_scrub_rate;
2061*4882a593Smuzhiyun }
2062*4882a593Smuzhiyun 
disable_sdram_scrub_setting(struct mem_ctl_info * mci)2063*4882a593Smuzhiyun static void disable_sdram_scrub_setting(struct mem_ctl_info *mci)
2064*4882a593Smuzhiyun {
2065*4882a593Smuzhiyun 	struct i7core_pvt *pvt = mci->pvt_info;
2066*4882a593Smuzhiyun 	u32 pci_lock;
2067*4882a593Smuzhiyun 
2068*4882a593Smuzhiyun 	/* Lock writes to pci registers */
2069*4882a593Smuzhiyun 	pci_read_config_dword(pvt->pci_noncore, MC_CFG_CONTROL, &pci_lock);
2070*4882a593Smuzhiyun 	pci_lock &= ~0x3;
2071*4882a593Smuzhiyun 	pci_write_config_dword(pvt->pci_noncore, MC_CFG_CONTROL,
2072*4882a593Smuzhiyun 			       pci_lock | MC_CFG_LOCK);
2073*4882a593Smuzhiyun }
2074*4882a593Smuzhiyun 
i7core_pci_ctl_create(struct i7core_pvt * pvt)2075*4882a593Smuzhiyun static void i7core_pci_ctl_create(struct i7core_pvt *pvt)
2076*4882a593Smuzhiyun {
2077*4882a593Smuzhiyun 	pvt->i7core_pci = edac_pci_create_generic_ctl(
2078*4882a593Smuzhiyun 						&pvt->i7core_dev->pdev[0]->dev,
2079*4882a593Smuzhiyun 						EDAC_MOD_STR);
2080*4882a593Smuzhiyun 	if (unlikely(!pvt->i7core_pci))
2081*4882a593Smuzhiyun 		i7core_printk(KERN_WARNING,
2082*4882a593Smuzhiyun 			      "Unable to setup PCI error report via EDAC\n");
2083*4882a593Smuzhiyun }
2084*4882a593Smuzhiyun 
i7core_pci_ctl_release(struct i7core_pvt * pvt)2085*4882a593Smuzhiyun static void i7core_pci_ctl_release(struct i7core_pvt *pvt)
2086*4882a593Smuzhiyun {
2087*4882a593Smuzhiyun 	if (likely(pvt->i7core_pci))
2088*4882a593Smuzhiyun 		edac_pci_release_generic_ctl(pvt->i7core_pci);
2089*4882a593Smuzhiyun 	else
2090*4882a593Smuzhiyun 		i7core_printk(KERN_ERR,
2091*4882a593Smuzhiyun 				"Couldn't find mem_ctl_info for socket %d\n",
2092*4882a593Smuzhiyun 				pvt->i7core_dev->socket);
2093*4882a593Smuzhiyun 	pvt->i7core_pci = NULL;
2094*4882a593Smuzhiyun }
2095*4882a593Smuzhiyun 
i7core_unregister_mci(struct i7core_dev * i7core_dev)2096*4882a593Smuzhiyun static void i7core_unregister_mci(struct i7core_dev *i7core_dev)
2097*4882a593Smuzhiyun {
2098*4882a593Smuzhiyun 	struct mem_ctl_info *mci = i7core_dev->mci;
2099*4882a593Smuzhiyun 	struct i7core_pvt *pvt;
2100*4882a593Smuzhiyun 
2101*4882a593Smuzhiyun 	if (unlikely(!mci || !mci->pvt_info)) {
2102*4882a593Smuzhiyun 		edac_dbg(0, "MC: dev = %p\n", &i7core_dev->pdev[0]->dev);
2103*4882a593Smuzhiyun 
2104*4882a593Smuzhiyun 		i7core_printk(KERN_ERR, "Couldn't find mci handler\n");
2105*4882a593Smuzhiyun 		return;
2106*4882a593Smuzhiyun 	}
2107*4882a593Smuzhiyun 
2108*4882a593Smuzhiyun 	pvt = mci->pvt_info;
2109*4882a593Smuzhiyun 
2110*4882a593Smuzhiyun 	edac_dbg(0, "MC: mci = %p, dev = %p\n", mci, &i7core_dev->pdev[0]->dev);
2111*4882a593Smuzhiyun 
2112*4882a593Smuzhiyun 	/* Disable scrubrate setting */
2113*4882a593Smuzhiyun 	if (pvt->enable_scrub)
2114*4882a593Smuzhiyun 		disable_sdram_scrub_setting(mci);
2115*4882a593Smuzhiyun 
2116*4882a593Smuzhiyun 	/* Disable EDAC polling */
2117*4882a593Smuzhiyun 	i7core_pci_ctl_release(pvt);
2118*4882a593Smuzhiyun 
2119*4882a593Smuzhiyun 	/* Remove MC sysfs nodes */
2120*4882a593Smuzhiyun 	i7core_delete_sysfs_devices(mci);
2121*4882a593Smuzhiyun 	edac_mc_del_mc(mci->pdev);
2122*4882a593Smuzhiyun 
2123*4882a593Smuzhiyun 	edac_dbg(1, "%s: free mci struct\n", mci->ctl_name);
2124*4882a593Smuzhiyun 	kfree(mci->ctl_name);
2125*4882a593Smuzhiyun 	edac_mc_free(mci);
2126*4882a593Smuzhiyun 	i7core_dev->mci = NULL;
2127*4882a593Smuzhiyun }
2128*4882a593Smuzhiyun 
i7core_register_mci(struct i7core_dev * i7core_dev)2129*4882a593Smuzhiyun static int i7core_register_mci(struct i7core_dev *i7core_dev)
2130*4882a593Smuzhiyun {
2131*4882a593Smuzhiyun 	struct mem_ctl_info *mci;
2132*4882a593Smuzhiyun 	struct i7core_pvt *pvt;
2133*4882a593Smuzhiyun 	int rc;
2134*4882a593Smuzhiyun 	struct edac_mc_layer layers[2];
2135*4882a593Smuzhiyun 
2136*4882a593Smuzhiyun 	/* allocate a new MC control structure */
2137*4882a593Smuzhiyun 
2138*4882a593Smuzhiyun 	layers[0].type = EDAC_MC_LAYER_CHANNEL;
2139*4882a593Smuzhiyun 	layers[0].size = NUM_CHANS;
2140*4882a593Smuzhiyun 	layers[0].is_virt_csrow = false;
2141*4882a593Smuzhiyun 	layers[1].type = EDAC_MC_LAYER_SLOT;
2142*4882a593Smuzhiyun 	layers[1].size = MAX_DIMMS;
2143*4882a593Smuzhiyun 	layers[1].is_virt_csrow = true;
2144*4882a593Smuzhiyun 	mci = edac_mc_alloc(i7core_dev->socket, ARRAY_SIZE(layers), layers,
2145*4882a593Smuzhiyun 			    sizeof(*pvt));
2146*4882a593Smuzhiyun 	if (unlikely(!mci))
2147*4882a593Smuzhiyun 		return -ENOMEM;
2148*4882a593Smuzhiyun 
2149*4882a593Smuzhiyun 	edac_dbg(0, "MC: mci = %p, dev = %p\n", mci, &i7core_dev->pdev[0]->dev);
2150*4882a593Smuzhiyun 
2151*4882a593Smuzhiyun 	pvt = mci->pvt_info;
2152*4882a593Smuzhiyun 	memset(pvt, 0, sizeof(*pvt));
2153*4882a593Smuzhiyun 
2154*4882a593Smuzhiyun 	/* Associates i7core_dev and mci for future usage */
2155*4882a593Smuzhiyun 	pvt->i7core_dev = i7core_dev;
2156*4882a593Smuzhiyun 	i7core_dev->mci = mci;
2157*4882a593Smuzhiyun 
2158*4882a593Smuzhiyun 	/*
2159*4882a593Smuzhiyun 	 * FIXME: how to handle RDDR3 at MCI level? It is possible to have
2160*4882a593Smuzhiyun 	 * Mixed RDDR3/UDDR3 with Nehalem, provided that they are on different
2161*4882a593Smuzhiyun 	 * memory channels
2162*4882a593Smuzhiyun 	 */
2163*4882a593Smuzhiyun 	mci->mtype_cap = MEM_FLAG_DDR3;
2164*4882a593Smuzhiyun 	mci->edac_ctl_cap = EDAC_FLAG_NONE;
2165*4882a593Smuzhiyun 	mci->edac_cap = EDAC_FLAG_NONE;
2166*4882a593Smuzhiyun 	mci->mod_name = "i7core_edac.c";
2167*4882a593Smuzhiyun 
2168*4882a593Smuzhiyun 	mci->ctl_name = kasprintf(GFP_KERNEL, "i7 core #%d", i7core_dev->socket);
2169*4882a593Smuzhiyun 	if (!mci->ctl_name) {
2170*4882a593Smuzhiyun 		rc = -ENOMEM;
2171*4882a593Smuzhiyun 		goto fail1;
2172*4882a593Smuzhiyun 	}
2173*4882a593Smuzhiyun 
2174*4882a593Smuzhiyun 	mci->dev_name = pci_name(i7core_dev->pdev[0]);
2175*4882a593Smuzhiyun 	mci->ctl_page_to_phys = NULL;
2176*4882a593Smuzhiyun 
2177*4882a593Smuzhiyun 	/* Store pci devices at mci for faster access */
2178*4882a593Smuzhiyun 	rc = mci_bind_devs(mci, i7core_dev);
2179*4882a593Smuzhiyun 	if (unlikely(rc < 0))
2180*4882a593Smuzhiyun 		goto fail0;
2181*4882a593Smuzhiyun 
2182*4882a593Smuzhiyun 
2183*4882a593Smuzhiyun 	/* Get dimm basic config */
2184*4882a593Smuzhiyun 	get_dimm_config(mci);
2185*4882a593Smuzhiyun 	/* record ptr to the generic device */
2186*4882a593Smuzhiyun 	mci->pdev = &i7core_dev->pdev[0]->dev;
2187*4882a593Smuzhiyun 
2188*4882a593Smuzhiyun 	/* Enable scrubrate setting */
2189*4882a593Smuzhiyun 	if (pvt->enable_scrub)
2190*4882a593Smuzhiyun 		enable_sdram_scrub_setting(mci);
2191*4882a593Smuzhiyun 
2192*4882a593Smuzhiyun 	/* add this new MC control structure to EDAC's list of MCs */
2193*4882a593Smuzhiyun 	if (unlikely(edac_mc_add_mc_with_groups(mci, i7core_dev_groups))) {
2194*4882a593Smuzhiyun 		edac_dbg(0, "MC: failed edac_mc_add_mc()\n");
2195*4882a593Smuzhiyun 		/* FIXME: perhaps some code should go here that disables error
2196*4882a593Smuzhiyun 		 * reporting if we just enabled it
2197*4882a593Smuzhiyun 		 */
2198*4882a593Smuzhiyun 
2199*4882a593Smuzhiyun 		rc = -EINVAL;
2200*4882a593Smuzhiyun 		goto fail0;
2201*4882a593Smuzhiyun 	}
2202*4882a593Smuzhiyun 	if (i7core_create_sysfs_devices(mci)) {
2203*4882a593Smuzhiyun 		edac_dbg(0, "MC: failed to create sysfs nodes\n");
2204*4882a593Smuzhiyun 		edac_mc_del_mc(mci->pdev);
2205*4882a593Smuzhiyun 		rc = -EINVAL;
2206*4882a593Smuzhiyun 		goto fail0;
2207*4882a593Smuzhiyun 	}
2208*4882a593Smuzhiyun 
2209*4882a593Smuzhiyun 	/* Default error mask is any memory */
2210*4882a593Smuzhiyun 	pvt->inject.channel = 0;
2211*4882a593Smuzhiyun 	pvt->inject.dimm = -1;
2212*4882a593Smuzhiyun 	pvt->inject.rank = -1;
2213*4882a593Smuzhiyun 	pvt->inject.bank = -1;
2214*4882a593Smuzhiyun 	pvt->inject.page = -1;
2215*4882a593Smuzhiyun 	pvt->inject.col = -1;
2216*4882a593Smuzhiyun 
2217*4882a593Smuzhiyun 	/* allocating generic PCI control info */
2218*4882a593Smuzhiyun 	i7core_pci_ctl_create(pvt);
2219*4882a593Smuzhiyun 
2220*4882a593Smuzhiyun 	/* DCLK for scrub rate setting */
2221*4882a593Smuzhiyun 	pvt->dclk_freq = get_dclk_freq();
2222*4882a593Smuzhiyun 
2223*4882a593Smuzhiyun 	return 0;
2224*4882a593Smuzhiyun 
2225*4882a593Smuzhiyun fail0:
2226*4882a593Smuzhiyun 	kfree(mci->ctl_name);
2227*4882a593Smuzhiyun 
2228*4882a593Smuzhiyun fail1:
2229*4882a593Smuzhiyun 	edac_mc_free(mci);
2230*4882a593Smuzhiyun 	i7core_dev->mci = NULL;
2231*4882a593Smuzhiyun 	return rc;
2232*4882a593Smuzhiyun }
2233*4882a593Smuzhiyun 
2234*4882a593Smuzhiyun /*
2235*4882a593Smuzhiyun  *	i7core_probe	Probe for ONE instance of device to see if it is
2236*4882a593Smuzhiyun  *			present.
2237*4882a593Smuzhiyun  *	return:
2238*4882a593Smuzhiyun  *		0 for FOUND a device
2239*4882a593Smuzhiyun  *		< 0 for error code
2240*4882a593Smuzhiyun  */
2241*4882a593Smuzhiyun 
i7core_probe(struct pci_dev * pdev,const struct pci_device_id * id)2242*4882a593Smuzhiyun static int i7core_probe(struct pci_dev *pdev, const struct pci_device_id *id)
2243*4882a593Smuzhiyun {
2244*4882a593Smuzhiyun 	int rc, count = 0;
2245*4882a593Smuzhiyun 	struct i7core_dev *i7core_dev;
2246*4882a593Smuzhiyun 
2247*4882a593Smuzhiyun 	/* get the pci devices we want to reserve for our use */
2248*4882a593Smuzhiyun 	mutex_lock(&i7core_edac_lock);
2249*4882a593Smuzhiyun 
2250*4882a593Smuzhiyun 	/*
2251*4882a593Smuzhiyun 	 * All memory controllers are allocated at the first pass.
2252*4882a593Smuzhiyun 	 */
2253*4882a593Smuzhiyun 	if (unlikely(probed >= 1)) {
2254*4882a593Smuzhiyun 		mutex_unlock(&i7core_edac_lock);
2255*4882a593Smuzhiyun 		return -ENODEV;
2256*4882a593Smuzhiyun 	}
2257*4882a593Smuzhiyun 	probed++;
2258*4882a593Smuzhiyun 
2259*4882a593Smuzhiyun 	rc = i7core_get_all_devices();
2260*4882a593Smuzhiyun 	if (unlikely(rc < 0))
2261*4882a593Smuzhiyun 		goto fail0;
2262*4882a593Smuzhiyun 
2263*4882a593Smuzhiyun 	list_for_each_entry(i7core_dev, &i7core_edac_list, list) {
2264*4882a593Smuzhiyun 		count++;
2265*4882a593Smuzhiyun 		rc = i7core_register_mci(i7core_dev);
2266*4882a593Smuzhiyun 		if (unlikely(rc < 0))
2267*4882a593Smuzhiyun 			goto fail1;
2268*4882a593Smuzhiyun 	}
2269*4882a593Smuzhiyun 
2270*4882a593Smuzhiyun 	/*
2271*4882a593Smuzhiyun 	 * Nehalem-EX uses a different memory controller. However, as the
2272*4882a593Smuzhiyun 	 * memory controller is not visible on some Nehalem/Nehalem-EP, we
2273*4882a593Smuzhiyun 	 * need to indirectly probe via a X58 PCI device. The same devices
2274*4882a593Smuzhiyun 	 * are found on (some) Nehalem-EX. So, on those machines, the
2275*4882a593Smuzhiyun 	 * probe routine needs to return -ENODEV, as the actual Memory
2276*4882a593Smuzhiyun 	 * Controller registers won't be detected.
2277*4882a593Smuzhiyun 	 */
2278*4882a593Smuzhiyun 	if (!count) {
2279*4882a593Smuzhiyun 		rc = -ENODEV;
2280*4882a593Smuzhiyun 		goto fail1;
2281*4882a593Smuzhiyun 	}
2282*4882a593Smuzhiyun 
2283*4882a593Smuzhiyun 	i7core_printk(KERN_INFO,
2284*4882a593Smuzhiyun 		      "Driver loaded, %d memory controller(s) found.\n",
2285*4882a593Smuzhiyun 		      count);
2286*4882a593Smuzhiyun 
2287*4882a593Smuzhiyun 	mutex_unlock(&i7core_edac_lock);
2288*4882a593Smuzhiyun 	return 0;
2289*4882a593Smuzhiyun 
2290*4882a593Smuzhiyun fail1:
2291*4882a593Smuzhiyun 	list_for_each_entry(i7core_dev, &i7core_edac_list, list)
2292*4882a593Smuzhiyun 		i7core_unregister_mci(i7core_dev);
2293*4882a593Smuzhiyun 
2294*4882a593Smuzhiyun 	i7core_put_all_devices();
2295*4882a593Smuzhiyun fail0:
2296*4882a593Smuzhiyun 	mutex_unlock(&i7core_edac_lock);
2297*4882a593Smuzhiyun 	return rc;
2298*4882a593Smuzhiyun }
2299*4882a593Smuzhiyun 
2300*4882a593Smuzhiyun /*
2301*4882a593Smuzhiyun  *	i7core_remove	destructor for one instance of device
2302*4882a593Smuzhiyun  *
2303*4882a593Smuzhiyun  */
i7core_remove(struct pci_dev * pdev)2304*4882a593Smuzhiyun static void i7core_remove(struct pci_dev *pdev)
2305*4882a593Smuzhiyun {
2306*4882a593Smuzhiyun 	struct i7core_dev *i7core_dev;
2307*4882a593Smuzhiyun 
2308*4882a593Smuzhiyun 	edac_dbg(0, "\n");
2309*4882a593Smuzhiyun 
2310*4882a593Smuzhiyun 	/*
2311*4882a593Smuzhiyun 	 * we have a trouble here: pdev value for removal will be wrong, since
2312*4882a593Smuzhiyun 	 * it will point to the X58 register used to detect that the machine
2313*4882a593Smuzhiyun 	 * is a Nehalem or upper design. However, due to the way several PCI
2314*4882a593Smuzhiyun 	 * devices are grouped together to provide MC functionality, we need
2315*4882a593Smuzhiyun 	 * to use a different method for releasing the devices
2316*4882a593Smuzhiyun 	 */
2317*4882a593Smuzhiyun 
2318*4882a593Smuzhiyun 	mutex_lock(&i7core_edac_lock);
2319*4882a593Smuzhiyun 
2320*4882a593Smuzhiyun 	if (unlikely(!probed)) {
2321*4882a593Smuzhiyun 		mutex_unlock(&i7core_edac_lock);
2322*4882a593Smuzhiyun 		return;
2323*4882a593Smuzhiyun 	}
2324*4882a593Smuzhiyun 
2325*4882a593Smuzhiyun 	list_for_each_entry(i7core_dev, &i7core_edac_list, list)
2326*4882a593Smuzhiyun 		i7core_unregister_mci(i7core_dev);
2327*4882a593Smuzhiyun 
2328*4882a593Smuzhiyun 	/* Release PCI resources */
2329*4882a593Smuzhiyun 	i7core_put_all_devices();
2330*4882a593Smuzhiyun 
2331*4882a593Smuzhiyun 	probed--;
2332*4882a593Smuzhiyun 
2333*4882a593Smuzhiyun 	mutex_unlock(&i7core_edac_lock);
2334*4882a593Smuzhiyun }
2335*4882a593Smuzhiyun 
2336*4882a593Smuzhiyun MODULE_DEVICE_TABLE(pci, i7core_pci_tbl);
2337*4882a593Smuzhiyun 
2338*4882a593Smuzhiyun /*
2339*4882a593Smuzhiyun  *	i7core_driver	pci_driver structure for this module
2340*4882a593Smuzhiyun  *
2341*4882a593Smuzhiyun  */
2342*4882a593Smuzhiyun static struct pci_driver i7core_driver = {
2343*4882a593Smuzhiyun 	.name     = "i7core_edac",
2344*4882a593Smuzhiyun 	.probe    = i7core_probe,
2345*4882a593Smuzhiyun 	.remove   = i7core_remove,
2346*4882a593Smuzhiyun 	.id_table = i7core_pci_tbl,
2347*4882a593Smuzhiyun };
2348*4882a593Smuzhiyun 
2349*4882a593Smuzhiyun /*
2350*4882a593Smuzhiyun  *	i7core_init		Module entry function
2351*4882a593Smuzhiyun  *			Try to initialize this module for its devices
2352*4882a593Smuzhiyun  */
i7core_init(void)2353*4882a593Smuzhiyun static int __init i7core_init(void)
2354*4882a593Smuzhiyun {
2355*4882a593Smuzhiyun 	int pci_rc;
2356*4882a593Smuzhiyun 
2357*4882a593Smuzhiyun 	edac_dbg(2, "\n");
2358*4882a593Smuzhiyun 
2359*4882a593Smuzhiyun 	/* Ensure that the OPSTATE is set correctly for POLL or NMI */
2360*4882a593Smuzhiyun 	opstate_init();
2361*4882a593Smuzhiyun 
2362*4882a593Smuzhiyun 	if (use_pci_fixup)
2363*4882a593Smuzhiyun 		i7core_xeon_pci_fixup(pci_dev_table);
2364*4882a593Smuzhiyun 
2365*4882a593Smuzhiyun 	pci_rc = pci_register_driver(&i7core_driver);
2366*4882a593Smuzhiyun 
2367*4882a593Smuzhiyun 	if (pci_rc >= 0) {
2368*4882a593Smuzhiyun 		mce_register_decode_chain(&i7_mce_dec);
2369*4882a593Smuzhiyun 		return 0;
2370*4882a593Smuzhiyun 	}
2371*4882a593Smuzhiyun 
2372*4882a593Smuzhiyun 	i7core_printk(KERN_ERR, "Failed to register device with error %d.\n",
2373*4882a593Smuzhiyun 		      pci_rc);
2374*4882a593Smuzhiyun 
2375*4882a593Smuzhiyun 	return pci_rc;
2376*4882a593Smuzhiyun }
2377*4882a593Smuzhiyun 
2378*4882a593Smuzhiyun /*
2379*4882a593Smuzhiyun  *	i7core_exit()	Module exit function
2380*4882a593Smuzhiyun  *			Unregister the driver
2381*4882a593Smuzhiyun  */
i7core_exit(void)2382*4882a593Smuzhiyun static void __exit i7core_exit(void)
2383*4882a593Smuzhiyun {
2384*4882a593Smuzhiyun 	edac_dbg(2, "\n");
2385*4882a593Smuzhiyun 	pci_unregister_driver(&i7core_driver);
2386*4882a593Smuzhiyun 	mce_unregister_decode_chain(&i7_mce_dec);
2387*4882a593Smuzhiyun }
2388*4882a593Smuzhiyun 
2389*4882a593Smuzhiyun module_init(i7core_init);
2390*4882a593Smuzhiyun module_exit(i7core_exit);
2391*4882a593Smuzhiyun 
2392*4882a593Smuzhiyun MODULE_LICENSE("GPL");
2393*4882a593Smuzhiyun MODULE_AUTHOR("Mauro Carvalho Chehab");
2394*4882a593Smuzhiyun MODULE_AUTHOR("Red Hat Inc. (https://www.redhat.com)");
2395*4882a593Smuzhiyun MODULE_DESCRIPTION("MC Driver for Intel i7 Core memory controllers - "
2396*4882a593Smuzhiyun 		   I7CORE_REVISION);
2397*4882a593Smuzhiyun 
2398*4882a593Smuzhiyun module_param(edac_op_state, int, 0444);
2399*4882a593Smuzhiyun MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI");
2400