1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * PCI Express Downstream Port Containment services driver
4*4882a593Smuzhiyun * Author: Keith Busch <keith.busch@intel.com>
5*4882a593Smuzhiyun *
6*4882a593Smuzhiyun * Copyright (C) 2016 Intel Corp.
7*4882a593Smuzhiyun */
8*4882a593Smuzhiyun
9*4882a593Smuzhiyun #define dev_fmt(fmt) "DPC: " fmt
10*4882a593Smuzhiyun
11*4882a593Smuzhiyun #include <linux/aer.h>
12*4882a593Smuzhiyun #include <linux/delay.h>
13*4882a593Smuzhiyun #include <linux/interrupt.h>
14*4882a593Smuzhiyun #include <linux/init.h>
15*4882a593Smuzhiyun #include <linux/pci.h>
16*4882a593Smuzhiyun
17*4882a593Smuzhiyun #include "portdrv.h"
18*4882a593Smuzhiyun #include "../pci.h"
19*4882a593Smuzhiyun
20*4882a593Smuzhiyun static const char * const rp_pio_error_string[] = {
21*4882a593Smuzhiyun "Configuration Request received UR Completion", /* Bit Position 0 */
22*4882a593Smuzhiyun "Configuration Request received CA Completion", /* Bit Position 1 */
23*4882a593Smuzhiyun "Configuration Request Completion Timeout", /* Bit Position 2 */
24*4882a593Smuzhiyun NULL,
25*4882a593Smuzhiyun NULL,
26*4882a593Smuzhiyun NULL,
27*4882a593Smuzhiyun NULL,
28*4882a593Smuzhiyun NULL,
29*4882a593Smuzhiyun "I/O Request received UR Completion", /* Bit Position 8 */
30*4882a593Smuzhiyun "I/O Request received CA Completion", /* Bit Position 9 */
31*4882a593Smuzhiyun "I/O Request Completion Timeout", /* Bit Position 10 */
32*4882a593Smuzhiyun NULL,
33*4882a593Smuzhiyun NULL,
34*4882a593Smuzhiyun NULL,
35*4882a593Smuzhiyun NULL,
36*4882a593Smuzhiyun NULL,
37*4882a593Smuzhiyun "Memory Request received UR Completion", /* Bit Position 16 */
38*4882a593Smuzhiyun "Memory Request received CA Completion", /* Bit Position 17 */
39*4882a593Smuzhiyun "Memory Request Completion Timeout", /* Bit Position 18 */
40*4882a593Smuzhiyun };
41*4882a593Smuzhiyun
pci_save_dpc_state(struct pci_dev * dev)42*4882a593Smuzhiyun void pci_save_dpc_state(struct pci_dev *dev)
43*4882a593Smuzhiyun {
44*4882a593Smuzhiyun struct pci_cap_saved_state *save_state;
45*4882a593Smuzhiyun u16 *cap;
46*4882a593Smuzhiyun
47*4882a593Smuzhiyun if (!pci_is_pcie(dev))
48*4882a593Smuzhiyun return;
49*4882a593Smuzhiyun
50*4882a593Smuzhiyun save_state = pci_find_saved_ext_cap(dev, PCI_EXT_CAP_ID_DPC);
51*4882a593Smuzhiyun if (!save_state)
52*4882a593Smuzhiyun return;
53*4882a593Smuzhiyun
54*4882a593Smuzhiyun cap = (u16 *)&save_state->cap.data[0];
55*4882a593Smuzhiyun pci_read_config_word(dev, dev->dpc_cap + PCI_EXP_DPC_CTL, cap);
56*4882a593Smuzhiyun }
57*4882a593Smuzhiyun
pci_restore_dpc_state(struct pci_dev * dev)58*4882a593Smuzhiyun void pci_restore_dpc_state(struct pci_dev *dev)
59*4882a593Smuzhiyun {
60*4882a593Smuzhiyun struct pci_cap_saved_state *save_state;
61*4882a593Smuzhiyun u16 *cap;
62*4882a593Smuzhiyun
63*4882a593Smuzhiyun if (!pci_is_pcie(dev))
64*4882a593Smuzhiyun return;
65*4882a593Smuzhiyun
66*4882a593Smuzhiyun save_state = pci_find_saved_ext_cap(dev, PCI_EXT_CAP_ID_DPC);
67*4882a593Smuzhiyun if (!save_state)
68*4882a593Smuzhiyun return;
69*4882a593Smuzhiyun
70*4882a593Smuzhiyun cap = (u16 *)&save_state->cap.data[0];
71*4882a593Smuzhiyun pci_write_config_word(dev, dev->dpc_cap + PCI_EXP_DPC_CTL, *cap);
72*4882a593Smuzhiyun }
73*4882a593Smuzhiyun
74*4882a593Smuzhiyun static DECLARE_WAIT_QUEUE_HEAD(dpc_completed_waitqueue);
75*4882a593Smuzhiyun
76*4882a593Smuzhiyun #ifdef CONFIG_HOTPLUG_PCI_PCIE
dpc_completed(struct pci_dev * pdev)77*4882a593Smuzhiyun static bool dpc_completed(struct pci_dev *pdev)
78*4882a593Smuzhiyun {
79*4882a593Smuzhiyun u16 status;
80*4882a593Smuzhiyun
81*4882a593Smuzhiyun pci_read_config_word(pdev, pdev->dpc_cap + PCI_EXP_DPC_STATUS, &status);
82*4882a593Smuzhiyun if ((status != 0xffff) && (status & PCI_EXP_DPC_STATUS_TRIGGER))
83*4882a593Smuzhiyun return false;
84*4882a593Smuzhiyun
85*4882a593Smuzhiyun if (test_bit(PCI_DPC_RECOVERING, &pdev->priv_flags))
86*4882a593Smuzhiyun return false;
87*4882a593Smuzhiyun
88*4882a593Smuzhiyun return true;
89*4882a593Smuzhiyun }
90*4882a593Smuzhiyun
91*4882a593Smuzhiyun /**
92*4882a593Smuzhiyun * pci_dpc_recovered - whether DPC triggered and has recovered successfully
93*4882a593Smuzhiyun * @pdev: PCI device
94*4882a593Smuzhiyun *
95*4882a593Smuzhiyun * Return true if DPC was triggered for @pdev and has recovered successfully.
96*4882a593Smuzhiyun * Wait for recovery if it hasn't completed yet. Called from the PCIe hotplug
97*4882a593Smuzhiyun * driver to recognize and ignore Link Down/Up events caused by DPC.
98*4882a593Smuzhiyun */
pci_dpc_recovered(struct pci_dev * pdev)99*4882a593Smuzhiyun bool pci_dpc_recovered(struct pci_dev *pdev)
100*4882a593Smuzhiyun {
101*4882a593Smuzhiyun struct pci_host_bridge *host;
102*4882a593Smuzhiyun
103*4882a593Smuzhiyun if (!pdev->dpc_cap)
104*4882a593Smuzhiyun return false;
105*4882a593Smuzhiyun
106*4882a593Smuzhiyun /*
107*4882a593Smuzhiyun * Synchronization between hotplug and DPC is not supported
108*4882a593Smuzhiyun * if DPC is owned by firmware and EDR is not enabled.
109*4882a593Smuzhiyun */
110*4882a593Smuzhiyun host = pci_find_host_bridge(pdev->bus);
111*4882a593Smuzhiyun if (!host->native_dpc && !IS_ENABLED(CONFIG_PCIE_EDR))
112*4882a593Smuzhiyun return false;
113*4882a593Smuzhiyun
114*4882a593Smuzhiyun /*
115*4882a593Smuzhiyun * Need a timeout in case DPC never completes due to failure of
116*4882a593Smuzhiyun * dpc_wait_rp_inactive(). The spec doesn't mandate a time limit,
117*4882a593Smuzhiyun * but reports indicate that DPC completes within 4 seconds.
118*4882a593Smuzhiyun */
119*4882a593Smuzhiyun wait_event_timeout(dpc_completed_waitqueue, dpc_completed(pdev),
120*4882a593Smuzhiyun msecs_to_jiffies(4000));
121*4882a593Smuzhiyun
122*4882a593Smuzhiyun return test_and_clear_bit(PCI_DPC_RECOVERED, &pdev->priv_flags);
123*4882a593Smuzhiyun }
124*4882a593Smuzhiyun #endif /* CONFIG_HOTPLUG_PCI_PCIE */
125*4882a593Smuzhiyun
dpc_wait_rp_inactive(struct pci_dev * pdev)126*4882a593Smuzhiyun static int dpc_wait_rp_inactive(struct pci_dev *pdev)
127*4882a593Smuzhiyun {
128*4882a593Smuzhiyun unsigned long timeout = jiffies + HZ;
129*4882a593Smuzhiyun u16 cap = pdev->dpc_cap, status;
130*4882a593Smuzhiyun
131*4882a593Smuzhiyun pci_read_config_word(pdev, cap + PCI_EXP_DPC_STATUS, &status);
132*4882a593Smuzhiyun while (status & PCI_EXP_DPC_RP_BUSY &&
133*4882a593Smuzhiyun !time_after(jiffies, timeout)) {
134*4882a593Smuzhiyun msleep(10);
135*4882a593Smuzhiyun pci_read_config_word(pdev, cap + PCI_EXP_DPC_STATUS, &status);
136*4882a593Smuzhiyun }
137*4882a593Smuzhiyun if (status & PCI_EXP_DPC_RP_BUSY) {
138*4882a593Smuzhiyun pci_warn(pdev, "root port still busy\n");
139*4882a593Smuzhiyun return -EBUSY;
140*4882a593Smuzhiyun }
141*4882a593Smuzhiyun return 0;
142*4882a593Smuzhiyun }
143*4882a593Smuzhiyun
dpc_reset_link(struct pci_dev * pdev)144*4882a593Smuzhiyun pci_ers_result_t dpc_reset_link(struct pci_dev *pdev)
145*4882a593Smuzhiyun {
146*4882a593Smuzhiyun pci_ers_result_t ret;
147*4882a593Smuzhiyun u16 cap;
148*4882a593Smuzhiyun
149*4882a593Smuzhiyun set_bit(PCI_DPC_RECOVERING, &pdev->priv_flags);
150*4882a593Smuzhiyun
151*4882a593Smuzhiyun /*
152*4882a593Smuzhiyun * DPC disables the Link automatically in hardware, so it has
153*4882a593Smuzhiyun * already been reset by the time we get here.
154*4882a593Smuzhiyun */
155*4882a593Smuzhiyun cap = pdev->dpc_cap;
156*4882a593Smuzhiyun
157*4882a593Smuzhiyun /*
158*4882a593Smuzhiyun * Wait until the Link is inactive, then clear DPC Trigger Status
159*4882a593Smuzhiyun * to allow the Port to leave DPC.
160*4882a593Smuzhiyun */
161*4882a593Smuzhiyun if (!pcie_wait_for_link(pdev, false))
162*4882a593Smuzhiyun pci_info(pdev, "Data Link Layer Link Active not cleared in 1000 msec\n");
163*4882a593Smuzhiyun
164*4882a593Smuzhiyun if (pdev->dpc_rp_extensions && dpc_wait_rp_inactive(pdev)) {
165*4882a593Smuzhiyun clear_bit(PCI_DPC_RECOVERED, &pdev->priv_flags);
166*4882a593Smuzhiyun ret = PCI_ERS_RESULT_DISCONNECT;
167*4882a593Smuzhiyun goto out;
168*4882a593Smuzhiyun }
169*4882a593Smuzhiyun
170*4882a593Smuzhiyun pci_write_config_word(pdev, cap + PCI_EXP_DPC_STATUS,
171*4882a593Smuzhiyun PCI_EXP_DPC_STATUS_TRIGGER);
172*4882a593Smuzhiyun
173*4882a593Smuzhiyun if (!pcie_wait_for_link(pdev, true)) {
174*4882a593Smuzhiyun pci_info(pdev, "Data Link Layer Link Active not set in 1000 msec\n");
175*4882a593Smuzhiyun clear_bit(PCI_DPC_RECOVERED, &pdev->priv_flags);
176*4882a593Smuzhiyun ret = PCI_ERS_RESULT_DISCONNECT;
177*4882a593Smuzhiyun } else {
178*4882a593Smuzhiyun set_bit(PCI_DPC_RECOVERED, &pdev->priv_flags);
179*4882a593Smuzhiyun ret = PCI_ERS_RESULT_RECOVERED;
180*4882a593Smuzhiyun }
181*4882a593Smuzhiyun out:
182*4882a593Smuzhiyun clear_bit(PCI_DPC_RECOVERING, &pdev->priv_flags);
183*4882a593Smuzhiyun wake_up_all(&dpc_completed_waitqueue);
184*4882a593Smuzhiyun return ret;
185*4882a593Smuzhiyun }
186*4882a593Smuzhiyun
dpc_process_rp_pio_error(struct pci_dev * pdev)187*4882a593Smuzhiyun static void dpc_process_rp_pio_error(struct pci_dev *pdev)
188*4882a593Smuzhiyun {
189*4882a593Smuzhiyun u16 cap = pdev->dpc_cap, dpc_status, first_error;
190*4882a593Smuzhiyun u32 status, mask, sev, syserr, exc, dw0, dw1, dw2, dw3, log, prefix;
191*4882a593Smuzhiyun int i;
192*4882a593Smuzhiyun
193*4882a593Smuzhiyun pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_STATUS, &status);
194*4882a593Smuzhiyun pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_MASK, &mask);
195*4882a593Smuzhiyun pci_err(pdev, "rp_pio_status: %#010x, rp_pio_mask: %#010x\n",
196*4882a593Smuzhiyun status, mask);
197*4882a593Smuzhiyun
198*4882a593Smuzhiyun pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_SEVERITY, &sev);
199*4882a593Smuzhiyun pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_SYSERROR, &syserr);
200*4882a593Smuzhiyun pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_EXCEPTION, &exc);
201*4882a593Smuzhiyun pci_err(pdev, "RP PIO severity=%#010x, syserror=%#010x, exception=%#010x\n",
202*4882a593Smuzhiyun sev, syserr, exc);
203*4882a593Smuzhiyun
204*4882a593Smuzhiyun /* Get First Error Pointer */
205*4882a593Smuzhiyun pci_read_config_word(pdev, cap + PCI_EXP_DPC_STATUS, &dpc_status);
206*4882a593Smuzhiyun first_error = (dpc_status & 0x1f00) >> 8;
207*4882a593Smuzhiyun
208*4882a593Smuzhiyun for (i = 0; i < ARRAY_SIZE(rp_pio_error_string); i++) {
209*4882a593Smuzhiyun if ((status & ~mask) & (1 << i))
210*4882a593Smuzhiyun pci_err(pdev, "[%2d] %s%s\n", i, rp_pio_error_string[i],
211*4882a593Smuzhiyun first_error == i ? " (First)" : "");
212*4882a593Smuzhiyun }
213*4882a593Smuzhiyun
214*4882a593Smuzhiyun if (pdev->dpc_rp_log_size < 4)
215*4882a593Smuzhiyun goto clear_status;
216*4882a593Smuzhiyun pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_HEADER_LOG,
217*4882a593Smuzhiyun &dw0);
218*4882a593Smuzhiyun pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_HEADER_LOG + 4,
219*4882a593Smuzhiyun &dw1);
220*4882a593Smuzhiyun pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_HEADER_LOG + 8,
221*4882a593Smuzhiyun &dw2);
222*4882a593Smuzhiyun pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_HEADER_LOG + 12,
223*4882a593Smuzhiyun &dw3);
224*4882a593Smuzhiyun pci_err(pdev, "TLP Header: %#010x %#010x %#010x %#010x\n",
225*4882a593Smuzhiyun dw0, dw1, dw2, dw3);
226*4882a593Smuzhiyun
227*4882a593Smuzhiyun if (pdev->dpc_rp_log_size < 5)
228*4882a593Smuzhiyun goto clear_status;
229*4882a593Smuzhiyun pci_read_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_IMPSPEC_LOG, &log);
230*4882a593Smuzhiyun pci_err(pdev, "RP PIO ImpSpec Log %#010x\n", log);
231*4882a593Smuzhiyun
232*4882a593Smuzhiyun for (i = 0; i < pdev->dpc_rp_log_size - 5; i++) {
233*4882a593Smuzhiyun pci_read_config_dword(pdev,
234*4882a593Smuzhiyun cap + PCI_EXP_DPC_RP_PIO_TLPPREFIX_LOG, &prefix);
235*4882a593Smuzhiyun pci_err(pdev, "TLP Prefix Header: dw%d, %#010x\n", i, prefix);
236*4882a593Smuzhiyun }
237*4882a593Smuzhiyun clear_status:
238*4882a593Smuzhiyun pci_write_config_dword(pdev, cap + PCI_EXP_DPC_RP_PIO_STATUS, status);
239*4882a593Smuzhiyun }
240*4882a593Smuzhiyun
dpc_get_aer_uncorrect_severity(struct pci_dev * dev,struct aer_err_info * info)241*4882a593Smuzhiyun static int dpc_get_aer_uncorrect_severity(struct pci_dev *dev,
242*4882a593Smuzhiyun struct aer_err_info *info)
243*4882a593Smuzhiyun {
244*4882a593Smuzhiyun int pos = dev->aer_cap;
245*4882a593Smuzhiyun u32 status, mask, sev;
246*4882a593Smuzhiyun
247*4882a593Smuzhiyun pci_read_config_dword(dev, pos + PCI_ERR_UNCOR_STATUS, &status);
248*4882a593Smuzhiyun pci_read_config_dword(dev, pos + PCI_ERR_UNCOR_MASK, &mask);
249*4882a593Smuzhiyun status &= ~mask;
250*4882a593Smuzhiyun if (!status)
251*4882a593Smuzhiyun return 0;
252*4882a593Smuzhiyun
253*4882a593Smuzhiyun pci_read_config_dword(dev, pos + PCI_ERR_UNCOR_SEVER, &sev);
254*4882a593Smuzhiyun status &= sev;
255*4882a593Smuzhiyun if (status)
256*4882a593Smuzhiyun info->severity = AER_FATAL;
257*4882a593Smuzhiyun else
258*4882a593Smuzhiyun info->severity = AER_NONFATAL;
259*4882a593Smuzhiyun
260*4882a593Smuzhiyun return 1;
261*4882a593Smuzhiyun }
262*4882a593Smuzhiyun
dpc_process_error(struct pci_dev * pdev)263*4882a593Smuzhiyun void dpc_process_error(struct pci_dev *pdev)
264*4882a593Smuzhiyun {
265*4882a593Smuzhiyun u16 cap = pdev->dpc_cap, status, source, reason, ext_reason;
266*4882a593Smuzhiyun struct aer_err_info info;
267*4882a593Smuzhiyun
268*4882a593Smuzhiyun pci_read_config_word(pdev, cap + PCI_EXP_DPC_STATUS, &status);
269*4882a593Smuzhiyun pci_read_config_word(pdev, cap + PCI_EXP_DPC_SOURCE_ID, &source);
270*4882a593Smuzhiyun
271*4882a593Smuzhiyun pci_info(pdev, "containment event, status:%#06x source:%#06x\n",
272*4882a593Smuzhiyun status, source);
273*4882a593Smuzhiyun
274*4882a593Smuzhiyun reason = (status & PCI_EXP_DPC_STATUS_TRIGGER_RSN) >> 1;
275*4882a593Smuzhiyun ext_reason = (status & PCI_EXP_DPC_STATUS_TRIGGER_RSN_EXT) >> 5;
276*4882a593Smuzhiyun pci_warn(pdev, "%s detected\n",
277*4882a593Smuzhiyun (reason == 0) ? "unmasked uncorrectable error" :
278*4882a593Smuzhiyun (reason == 1) ? "ERR_NONFATAL" :
279*4882a593Smuzhiyun (reason == 2) ? "ERR_FATAL" :
280*4882a593Smuzhiyun (ext_reason == 0) ? "RP PIO error" :
281*4882a593Smuzhiyun (ext_reason == 1) ? "software trigger" :
282*4882a593Smuzhiyun "reserved error");
283*4882a593Smuzhiyun
284*4882a593Smuzhiyun /* show RP PIO error detail information */
285*4882a593Smuzhiyun if (pdev->dpc_rp_extensions && reason == 3 && ext_reason == 0)
286*4882a593Smuzhiyun dpc_process_rp_pio_error(pdev);
287*4882a593Smuzhiyun else if (reason == 0 &&
288*4882a593Smuzhiyun dpc_get_aer_uncorrect_severity(pdev, &info) &&
289*4882a593Smuzhiyun aer_get_device_error_info(pdev, &info)) {
290*4882a593Smuzhiyun aer_print_error(pdev, &info);
291*4882a593Smuzhiyun pci_aer_clear_nonfatal_status(pdev);
292*4882a593Smuzhiyun pci_aer_clear_fatal_status(pdev);
293*4882a593Smuzhiyun }
294*4882a593Smuzhiyun }
295*4882a593Smuzhiyun
dpc_handler(int irq,void * context)296*4882a593Smuzhiyun static irqreturn_t dpc_handler(int irq, void *context)
297*4882a593Smuzhiyun {
298*4882a593Smuzhiyun struct pci_dev *pdev = context;
299*4882a593Smuzhiyun
300*4882a593Smuzhiyun dpc_process_error(pdev);
301*4882a593Smuzhiyun
302*4882a593Smuzhiyun /* We configure DPC so it only triggers on ERR_FATAL */
303*4882a593Smuzhiyun pcie_do_recovery(pdev, pci_channel_io_frozen, dpc_reset_link);
304*4882a593Smuzhiyun
305*4882a593Smuzhiyun return IRQ_HANDLED;
306*4882a593Smuzhiyun }
307*4882a593Smuzhiyun
dpc_irq(int irq,void * context)308*4882a593Smuzhiyun static irqreturn_t dpc_irq(int irq, void *context)
309*4882a593Smuzhiyun {
310*4882a593Smuzhiyun struct pci_dev *pdev = context;
311*4882a593Smuzhiyun u16 cap = pdev->dpc_cap, status;
312*4882a593Smuzhiyun
313*4882a593Smuzhiyun pci_read_config_word(pdev, cap + PCI_EXP_DPC_STATUS, &status);
314*4882a593Smuzhiyun
315*4882a593Smuzhiyun if (!(status & PCI_EXP_DPC_STATUS_INTERRUPT) || status == (u16)(~0))
316*4882a593Smuzhiyun return IRQ_NONE;
317*4882a593Smuzhiyun
318*4882a593Smuzhiyun pci_write_config_word(pdev, cap + PCI_EXP_DPC_STATUS,
319*4882a593Smuzhiyun PCI_EXP_DPC_STATUS_INTERRUPT);
320*4882a593Smuzhiyun if (status & PCI_EXP_DPC_STATUS_TRIGGER)
321*4882a593Smuzhiyun return IRQ_WAKE_THREAD;
322*4882a593Smuzhiyun return IRQ_HANDLED;
323*4882a593Smuzhiyun }
324*4882a593Smuzhiyun
pci_dpc_init(struct pci_dev * pdev)325*4882a593Smuzhiyun void pci_dpc_init(struct pci_dev *pdev)
326*4882a593Smuzhiyun {
327*4882a593Smuzhiyun u16 cap;
328*4882a593Smuzhiyun
329*4882a593Smuzhiyun pdev->dpc_cap = pci_find_ext_capability(pdev, PCI_EXT_CAP_ID_DPC);
330*4882a593Smuzhiyun if (!pdev->dpc_cap)
331*4882a593Smuzhiyun return;
332*4882a593Smuzhiyun
333*4882a593Smuzhiyun pci_read_config_word(pdev, pdev->dpc_cap + PCI_EXP_DPC_CAP, &cap);
334*4882a593Smuzhiyun if (!(cap & PCI_EXP_DPC_CAP_RP_EXT))
335*4882a593Smuzhiyun return;
336*4882a593Smuzhiyun
337*4882a593Smuzhiyun pdev->dpc_rp_extensions = true;
338*4882a593Smuzhiyun pdev->dpc_rp_log_size = (cap & PCI_EXP_DPC_RP_PIO_LOG_SIZE) >> 8;
339*4882a593Smuzhiyun if (pdev->dpc_rp_log_size < 4 || pdev->dpc_rp_log_size > 9) {
340*4882a593Smuzhiyun pci_err(pdev, "RP PIO log size %u is invalid\n",
341*4882a593Smuzhiyun pdev->dpc_rp_log_size);
342*4882a593Smuzhiyun pdev->dpc_rp_log_size = 0;
343*4882a593Smuzhiyun }
344*4882a593Smuzhiyun }
345*4882a593Smuzhiyun
346*4882a593Smuzhiyun #define FLAG(x, y) (((x) & (y)) ? '+' : '-')
dpc_probe(struct pcie_device * dev)347*4882a593Smuzhiyun static int dpc_probe(struct pcie_device *dev)
348*4882a593Smuzhiyun {
349*4882a593Smuzhiyun struct pci_dev *pdev = dev->port;
350*4882a593Smuzhiyun struct device *device = &dev->device;
351*4882a593Smuzhiyun int status;
352*4882a593Smuzhiyun u16 ctl, cap;
353*4882a593Smuzhiyun
354*4882a593Smuzhiyun if (!pcie_aer_is_native(pdev) && !pcie_ports_dpc_native)
355*4882a593Smuzhiyun return -ENOTSUPP;
356*4882a593Smuzhiyun
357*4882a593Smuzhiyun status = devm_request_threaded_irq(device, dev->irq, dpc_irq,
358*4882a593Smuzhiyun dpc_handler, IRQF_SHARED,
359*4882a593Smuzhiyun "pcie-dpc", pdev);
360*4882a593Smuzhiyun if (status) {
361*4882a593Smuzhiyun pci_warn(pdev, "request IRQ%d failed: %d\n", dev->irq,
362*4882a593Smuzhiyun status);
363*4882a593Smuzhiyun return status;
364*4882a593Smuzhiyun }
365*4882a593Smuzhiyun
366*4882a593Smuzhiyun pci_read_config_word(pdev, pdev->dpc_cap + PCI_EXP_DPC_CAP, &cap);
367*4882a593Smuzhiyun pci_read_config_word(pdev, pdev->dpc_cap + PCI_EXP_DPC_CTL, &ctl);
368*4882a593Smuzhiyun
369*4882a593Smuzhiyun ctl = (ctl & 0xfff4) | PCI_EXP_DPC_CTL_EN_FATAL | PCI_EXP_DPC_CTL_INT_EN;
370*4882a593Smuzhiyun pci_write_config_word(pdev, pdev->dpc_cap + PCI_EXP_DPC_CTL, ctl);
371*4882a593Smuzhiyun pci_info(pdev, "enabled with IRQ %d\n", dev->irq);
372*4882a593Smuzhiyun
373*4882a593Smuzhiyun pci_info(pdev, "error containment capabilities: Int Msg #%d, RPExt%c PoisonedTLP%c SwTrigger%c RP PIO Log %d, DL_ActiveErr%c\n",
374*4882a593Smuzhiyun cap & PCI_EXP_DPC_IRQ, FLAG(cap, PCI_EXP_DPC_CAP_RP_EXT),
375*4882a593Smuzhiyun FLAG(cap, PCI_EXP_DPC_CAP_POISONED_TLP),
376*4882a593Smuzhiyun FLAG(cap, PCI_EXP_DPC_CAP_SW_TRIGGER), pdev->dpc_rp_log_size,
377*4882a593Smuzhiyun FLAG(cap, PCI_EXP_DPC_CAP_DL_ACTIVE));
378*4882a593Smuzhiyun
379*4882a593Smuzhiyun pci_add_ext_cap_save_buffer(pdev, PCI_EXT_CAP_ID_DPC, sizeof(u16));
380*4882a593Smuzhiyun return status;
381*4882a593Smuzhiyun }
382*4882a593Smuzhiyun
dpc_remove(struct pcie_device * dev)383*4882a593Smuzhiyun static void dpc_remove(struct pcie_device *dev)
384*4882a593Smuzhiyun {
385*4882a593Smuzhiyun struct pci_dev *pdev = dev->port;
386*4882a593Smuzhiyun u16 ctl;
387*4882a593Smuzhiyun
388*4882a593Smuzhiyun pci_read_config_word(pdev, pdev->dpc_cap + PCI_EXP_DPC_CTL, &ctl);
389*4882a593Smuzhiyun ctl &= ~(PCI_EXP_DPC_CTL_EN_FATAL | PCI_EXP_DPC_CTL_INT_EN);
390*4882a593Smuzhiyun pci_write_config_word(pdev, pdev->dpc_cap + PCI_EXP_DPC_CTL, ctl);
391*4882a593Smuzhiyun }
392*4882a593Smuzhiyun
393*4882a593Smuzhiyun static struct pcie_port_service_driver dpcdriver = {
394*4882a593Smuzhiyun .name = "dpc",
395*4882a593Smuzhiyun .port_type = PCIE_ANY_PORT,
396*4882a593Smuzhiyun .service = PCIE_PORT_SERVICE_DPC,
397*4882a593Smuzhiyun .probe = dpc_probe,
398*4882a593Smuzhiyun .remove = dpc_remove,
399*4882a593Smuzhiyun };
400*4882a593Smuzhiyun
pcie_dpc_init(void)401*4882a593Smuzhiyun int __init pcie_dpc_init(void)
402*4882a593Smuzhiyun {
403*4882a593Smuzhiyun return pcie_port_service_register(&dpcdriver);
404*4882a593Smuzhiyun }
405