1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * Copyright (C) 2005-2008, PA Semi, Inc
4*4882a593Smuzhiyun *
5*4882a593Smuzhiyun * Maintained by: Olof Johansson <olof@lixom.net>
6*4882a593Smuzhiyun */
7*4882a593Smuzhiyun
8*4882a593Smuzhiyun #undef DEBUG
9*4882a593Smuzhiyun
10*4882a593Smuzhiyun #include <linux/memblock.h>
11*4882a593Smuzhiyun #include <linux/types.h>
12*4882a593Smuzhiyun #include <linux/spinlock.h>
13*4882a593Smuzhiyun #include <linux/pci.h>
14*4882a593Smuzhiyun #include <asm/iommu.h>
15*4882a593Smuzhiyun #include <asm/machdep.h>
16*4882a593Smuzhiyun #include <asm/firmware.h>
17*4882a593Smuzhiyun
18*4882a593Smuzhiyun #include "pasemi.h"
19*4882a593Smuzhiyun
20*4882a593Smuzhiyun #define IOBMAP_PAGE_SHIFT 12
21*4882a593Smuzhiyun #define IOBMAP_PAGE_SIZE (1 << IOBMAP_PAGE_SHIFT)
22*4882a593Smuzhiyun #define IOBMAP_PAGE_MASK (IOBMAP_PAGE_SIZE - 1)
23*4882a593Smuzhiyun
24*4882a593Smuzhiyun #define IOB_BASE 0xe0000000
25*4882a593Smuzhiyun #define IOB_SIZE 0x3000
26*4882a593Smuzhiyun /* Configuration registers */
27*4882a593Smuzhiyun #define IOBCAP_REG 0x40
28*4882a593Smuzhiyun #define IOBCOM_REG 0x100
29*4882a593Smuzhiyun /* Enable IOB address translation */
30*4882a593Smuzhiyun #define IOBCOM_ATEN 0x00000100
31*4882a593Smuzhiyun
32*4882a593Smuzhiyun /* Address decode configuration register */
33*4882a593Smuzhiyun #define IOB_AD_REG 0x14c
34*4882a593Smuzhiyun /* IOBCOM_AD_REG fields */
35*4882a593Smuzhiyun #define IOB_AD_VGPRT 0x00000e00
36*4882a593Smuzhiyun #define IOB_AD_VGAEN 0x00000100
37*4882a593Smuzhiyun /* Direct mapping settings */
38*4882a593Smuzhiyun #define IOB_AD_MPSEL_MASK 0x00000030
39*4882a593Smuzhiyun #define IOB_AD_MPSEL_B38 0x00000000
40*4882a593Smuzhiyun #define IOB_AD_MPSEL_B40 0x00000010
41*4882a593Smuzhiyun #define IOB_AD_MPSEL_B42 0x00000020
42*4882a593Smuzhiyun /* Translation window size / enable */
43*4882a593Smuzhiyun #define IOB_AD_TRNG_MASK 0x00000003
44*4882a593Smuzhiyun #define IOB_AD_TRNG_256M 0x00000000
45*4882a593Smuzhiyun #define IOB_AD_TRNG_2G 0x00000001
46*4882a593Smuzhiyun #define IOB_AD_TRNG_128G 0x00000003
47*4882a593Smuzhiyun
48*4882a593Smuzhiyun #define IOB_TABLEBASE_REG 0x154
49*4882a593Smuzhiyun
50*4882a593Smuzhiyun /* Base of the 64 4-byte L1 registers */
51*4882a593Smuzhiyun #define IOB_XLT_L1_REGBASE 0x2b00
52*4882a593Smuzhiyun
53*4882a593Smuzhiyun /* Register to invalidate TLB entries */
54*4882a593Smuzhiyun #define IOB_AT_INVAL_TLB_REG 0x2d00
55*4882a593Smuzhiyun
56*4882a593Smuzhiyun /* The top two bits of the level 1 entry contains valid and type flags */
57*4882a593Smuzhiyun #define IOBMAP_L1E_V 0x40000000
58*4882a593Smuzhiyun #define IOBMAP_L1E_V_B 0x80000000
59*4882a593Smuzhiyun
60*4882a593Smuzhiyun /* For big page entries, the bottom two bits contains flags */
61*4882a593Smuzhiyun #define IOBMAP_L1E_BIG_CACHED 0x00000002
62*4882a593Smuzhiyun #define IOBMAP_L1E_BIG_PRIORITY 0x00000001
63*4882a593Smuzhiyun
64*4882a593Smuzhiyun /* For regular level 2 entries, top 2 bits contain valid and cache flags */
65*4882a593Smuzhiyun #define IOBMAP_L2E_V 0x80000000
66*4882a593Smuzhiyun #define IOBMAP_L2E_V_CACHED 0xc0000000
67*4882a593Smuzhiyun
68*4882a593Smuzhiyun static void __iomem *iob;
69*4882a593Smuzhiyun static u32 iob_l1_emptyval;
70*4882a593Smuzhiyun static u32 iob_l2_emptyval;
71*4882a593Smuzhiyun static u32 *iob_l2_base;
72*4882a593Smuzhiyun
73*4882a593Smuzhiyun static struct iommu_table iommu_table_iobmap;
74*4882a593Smuzhiyun static int iommu_table_iobmap_inited;
75*4882a593Smuzhiyun
iobmap_build(struct iommu_table * tbl,long index,long npages,unsigned long uaddr,enum dma_data_direction direction,unsigned long attrs)76*4882a593Smuzhiyun static int iobmap_build(struct iommu_table *tbl, long index,
77*4882a593Smuzhiyun long npages, unsigned long uaddr,
78*4882a593Smuzhiyun enum dma_data_direction direction,
79*4882a593Smuzhiyun unsigned long attrs)
80*4882a593Smuzhiyun {
81*4882a593Smuzhiyun u32 *ip;
82*4882a593Smuzhiyun u32 rpn;
83*4882a593Smuzhiyun unsigned long bus_addr;
84*4882a593Smuzhiyun
85*4882a593Smuzhiyun pr_debug("iobmap: build at: %lx, %lx, addr: %lx\n", index, npages, uaddr);
86*4882a593Smuzhiyun
87*4882a593Smuzhiyun bus_addr = (tbl->it_offset + index) << IOBMAP_PAGE_SHIFT;
88*4882a593Smuzhiyun
89*4882a593Smuzhiyun ip = ((u32 *)tbl->it_base) + index;
90*4882a593Smuzhiyun
91*4882a593Smuzhiyun while (npages--) {
92*4882a593Smuzhiyun rpn = __pa(uaddr) >> IOBMAP_PAGE_SHIFT;
93*4882a593Smuzhiyun
94*4882a593Smuzhiyun *(ip++) = IOBMAP_L2E_V | rpn;
95*4882a593Smuzhiyun /* invalidate tlb, can be optimized more */
96*4882a593Smuzhiyun out_le32(iob+IOB_AT_INVAL_TLB_REG, bus_addr >> 14);
97*4882a593Smuzhiyun
98*4882a593Smuzhiyun uaddr += IOBMAP_PAGE_SIZE;
99*4882a593Smuzhiyun bus_addr += IOBMAP_PAGE_SIZE;
100*4882a593Smuzhiyun }
101*4882a593Smuzhiyun return 0;
102*4882a593Smuzhiyun }
103*4882a593Smuzhiyun
104*4882a593Smuzhiyun
iobmap_free(struct iommu_table * tbl,long index,long npages)105*4882a593Smuzhiyun static void iobmap_free(struct iommu_table *tbl, long index,
106*4882a593Smuzhiyun long npages)
107*4882a593Smuzhiyun {
108*4882a593Smuzhiyun u32 *ip;
109*4882a593Smuzhiyun unsigned long bus_addr;
110*4882a593Smuzhiyun
111*4882a593Smuzhiyun pr_debug("iobmap: free at: %lx, %lx\n", index, npages);
112*4882a593Smuzhiyun
113*4882a593Smuzhiyun bus_addr = (tbl->it_offset + index) << IOBMAP_PAGE_SHIFT;
114*4882a593Smuzhiyun
115*4882a593Smuzhiyun ip = ((u32 *)tbl->it_base) + index;
116*4882a593Smuzhiyun
117*4882a593Smuzhiyun while (npages--) {
118*4882a593Smuzhiyun *(ip++) = iob_l2_emptyval;
119*4882a593Smuzhiyun /* invalidate tlb, can be optimized more */
120*4882a593Smuzhiyun out_le32(iob+IOB_AT_INVAL_TLB_REG, bus_addr >> 14);
121*4882a593Smuzhiyun bus_addr += IOBMAP_PAGE_SIZE;
122*4882a593Smuzhiyun }
123*4882a593Smuzhiyun }
124*4882a593Smuzhiyun
125*4882a593Smuzhiyun static struct iommu_table_ops iommu_table_iobmap_ops = {
126*4882a593Smuzhiyun .set = iobmap_build,
127*4882a593Smuzhiyun .clear = iobmap_free
128*4882a593Smuzhiyun };
129*4882a593Smuzhiyun
iommu_table_iobmap_setup(void)130*4882a593Smuzhiyun static void iommu_table_iobmap_setup(void)
131*4882a593Smuzhiyun {
132*4882a593Smuzhiyun pr_debug(" -> %s\n", __func__);
133*4882a593Smuzhiyun iommu_table_iobmap.it_busno = 0;
134*4882a593Smuzhiyun iommu_table_iobmap.it_offset = 0;
135*4882a593Smuzhiyun iommu_table_iobmap.it_page_shift = IOBMAP_PAGE_SHIFT;
136*4882a593Smuzhiyun
137*4882a593Smuzhiyun /* it_size is in number of entries */
138*4882a593Smuzhiyun iommu_table_iobmap.it_size =
139*4882a593Smuzhiyun 0x80000000 >> iommu_table_iobmap.it_page_shift;
140*4882a593Smuzhiyun
141*4882a593Smuzhiyun /* Initialize the common IOMMU code */
142*4882a593Smuzhiyun iommu_table_iobmap.it_base = (unsigned long)iob_l2_base;
143*4882a593Smuzhiyun iommu_table_iobmap.it_index = 0;
144*4882a593Smuzhiyun /* XXXOJN tune this to avoid IOB cache invals.
145*4882a593Smuzhiyun * Should probably be 8 (64 bytes)
146*4882a593Smuzhiyun */
147*4882a593Smuzhiyun iommu_table_iobmap.it_blocksize = 4;
148*4882a593Smuzhiyun iommu_table_iobmap.it_ops = &iommu_table_iobmap_ops;
149*4882a593Smuzhiyun iommu_init_table(&iommu_table_iobmap, 0, 0, 0);
150*4882a593Smuzhiyun pr_debug(" <- %s\n", __func__);
151*4882a593Smuzhiyun }
152*4882a593Smuzhiyun
153*4882a593Smuzhiyun
154*4882a593Smuzhiyun
pci_dma_bus_setup_pasemi(struct pci_bus * bus)155*4882a593Smuzhiyun static void pci_dma_bus_setup_pasemi(struct pci_bus *bus)
156*4882a593Smuzhiyun {
157*4882a593Smuzhiyun pr_debug("pci_dma_bus_setup, bus %p, bus->self %p\n", bus, bus->self);
158*4882a593Smuzhiyun
159*4882a593Smuzhiyun if (!iommu_table_iobmap_inited) {
160*4882a593Smuzhiyun iommu_table_iobmap_inited = 1;
161*4882a593Smuzhiyun iommu_table_iobmap_setup();
162*4882a593Smuzhiyun }
163*4882a593Smuzhiyun }
164*4882a593Smuzhiyun
165*4882a593Smuzhiyun
pci_dma_dev_setup_pasemi(struct pci_dev * dev)166*4882a593Smuzhiyun static void pci_dma_dev_setup_pasemi(struct pci_dev *dev)
167*4882a593Smuzhiyun {
168*4882a593Smuzhiyun pr_debug("pci_dma_dev_setup, dev %p (%s)\n", dev, pci_name(dev));
169*4882a593Smuzhiyun
170*4882a593Smuzhiyun #if !defined(CONFIG_PPC_PASEMI_IOMMU_DMA_FORCE)
171*4882a593Smuzhiyun /* For non-LPAR environment, don't translate anything for the DMA
172*4882a593Smuzhiyun * engine. The exception to this is if the user has enabled
173*4882a593Smuzhiyun * CONFIG_PPC_PASEMI_IOMMU_DMA_FORCE at build time.
174*4882a593Smuzhiyun */
175*4882a593Smuzhiyun if (dev->vendor == 0x1959 && dev->device == 0xa007 &&
176*4882a593Smuzhiyun !firmware_has_feature(FW_FEATURE_LPAR)) {
177*4882a593Smuzhiyun dev->dev.dma_ops = NULL;
178*4882a593Smuzhiyun /*
179*4882a593Smuzhiyun * Set the coherent DMA mask to prevent the iommu
180*4882a593Smuzhiyun * being used unnecessarily
181*4882a593Smuzhiyun */
182*4882a593Smuzhiyun dev->dev.coherent_dma_mask = DMA_BIT_MASK(44);
183*4882a593Smuzhiyun return;
184*4882a593Smuzhiyun }
185*4882a593Smuzhiyun #endif
186*4882a593Smuzhiyun
187*4882a593Smuzhiyun set_iommu_table_base(&dev->dev, &iommu_table_iobmap);
188*4882a593Smuzhiyun }
189*4882a593Smuzhiyun
iob_init(struct device_node * dn)190*4882a593Smuzhiyun static int __init iob_init(struct device_node *dn)
191*4882a593Smuzhiyun {
192*4882a593Smuzhiyun unsigned long tmp;
193*4882a593Smuzhiyun u32 regword;
194*4882a593Smuzhiyun int i;
195*4882a593Smuzhiyun
196*4882a593Smuzhiyun pr_debug(" -> %s\n", __func__);
197*4882a593Smuzhiyun
198*4882a593Smuzhiyun /* For 2G space, 8x64 pages (2^21 bytes) is max total l2 size */
199*4882a593Smuzhiyun iob_l2_base = memblock_alloc_try_nid_raw(1UL << 21, 1UL << 21,
200*4882a593Smuzhiyun MEMBLOCK_LOW_LIMIT, 0x80000000,
201*4882a593Smuzhiyun NUMA_NO_NODE);
202*4882a593Smuzhiyun if (!iob_l2_base)
203*4882a593Smuzhiyun panic("%s: Failed to allocate %lu bytes align=0x%lx max_addr=%x\n",
204*4882a593Smuzhiyun __func__, 1UL << 21, 1UL << 21, 0x80000000);
205*4882a593Smuzhiyun
206*4882a593Smuzhiyun pr_info("IOBMAP L2 allocated at: %p\n", iob_l2_base);
207*4882a593Smuzhiyun
208*4882a593Smuzhiyun /* Allocate a spare page to map all invalid IOTLB pages. */
209*4882a593Smuzhiyun tmp = memblock_phys_alloc(IOBMAP_PAGE_SIZE, IOBMAP_PAGE_SIZE);
210*4882a593Smuzhiyun if (!tmp)
211*4882a593Smuzhiyun panic("IOBMAP: Cannot allocate spare page!");
212*4882a593Smuzhiyun /* Empty l1 is marked invalid */
213*4882a593Smuzhiyun iob_l1_emptyval = 0;
214*4882a593Smuzhiyun /* Empty l2 is mapped to dummy page */
215*4882a593Smuzhiyun iob_l2_emptyval = IOBMAP_L2E_V | (tmp >> IOBMAP_PAGE_SHIFT);
216*4882a593Smuzhiyun
217*4882a593Smuzhiyun iob = ioremap(IOB_BASE, IOB_SIZE);
218*4882a593Smuzhiyun if (!iob)
219*4882a593Smuzhiyun panic("IOBMAP: Cannot map registers!");
220*4882a593Smuzhiyun
221*4882a593Smuzhiyun /* setup direct mapping of the L1 entries */
222*4882a593Smuzhiyun for (i = 0; i < 64; i++) {
223*4882a593Smuzhiyun /* Each L1 covers 32MB, i.e. 8K entries = 32K of ram */
224*4882a593Smuzhiyun regword = IOBMAP_L1E_V | (__pa(iob_l2_base + i*0x2000) >> 12);
225*4882a593Smuzhiyun out_le32(iob+IOB_XLT_L1_REGBASE+i*4, regword);
226*4882a593Smuzhiyun }
227*4882a593Smuzhiyun
228*4882a593Smuzhiyun /* set 2GB translation window, based at 0 */
229*4882a593Smuzhiyun regword = in_le32(iob+IOB_AD_REG);
230*4882a593Smuzhiyun regword &= ~IOB_AD_TRNG_MASK;
231*4882a593Smuzhiyun regword |= IOB_AD_TRNG_2G;
232*4882a593Smuzhiyun out_le32(iob+IOB_AD_REG, regword);
233*4882a593Smuzhiyun
234*4882a593Smuzhiyun /* Enable translation */
235*4882a593Smuzhiyun regword = in_le32(iob+IOBCOM_REG);
236*4882a593Smuzhiyun regword |= IOBCOM_ATEN;
237*4882a593Smuzhiyun out_le32(iob+IOBCOM_REG, regword);
238*4882a593Smuzhiyun
239*4882a593Smuzhiyun pr_debug(" <- %s\n", __func__);
240*4882a593Smuzhiyun
241*4882a593Smuzhiyun return 0;
242*4882a593Smuzhiyun }
243*4882a593Smuzhiyun
244*4882a593Smuzhiyun
245*4882a593Smuzhiyun /* These are called very early. */
iommu_init_early_pasemi(void)246*4882a593Smuzhiyun void __init iommu_init_early_pasemi(void)
247*4882a593Smuzhiyun {
248*4882a593Smuzhiyun int iommu_off;
249*4882a593Smuzhiyun
250*4882a593Smuzhiyun #ifndef CONFIG_PPC_PASEMI_IOMMU
251*4882a593Smuzhiyun iommu_off = 1;
252*4882a593Smuzhiyun #else
253*4882a593Smuzhiyun iommu_off = of_chosen &&
254*4882a593Smuzhiyun of_get_property(of_chosen, "linux,iommu-off", NULL);
255*4882a593Smuzhiyun #endif
256*4882a593Smuzhiyun if (iommu_off)
257*4882a593Smuzhiyun return;
258*4882a593Smuzhiyun
259*4882a593Smuzhiyun iob_init(NULL);
260*4882a593Smuzhiyun
261*4882a593Smuzhiyun pasemi_pci_controller_ops.dma_dev_setup = pci_dma_dev_setup_pasemi;
262*4882a593Smuzhiyun pasemi_pci_controller_ops.dma_bus_setup = pci_dma_bus_setup_pasemi;
263*4882a593Smuzhiyun set_pci_dma_ops(&dma_iommu_ops);
264*4882a593Smuzhiyun }
265