1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0+
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * TCE helpers for IODA PCI/PCIe on PowerNV platforms
4*4882a593Smuzhiyun *
5*4882a593Smuzhiyun * Copyright 2018 IBM Corp.
6*4882a593Smuzhiyun *
7*4882a593Smuzhiyun * This program is free software; you can redistribute it and/or
8*4882a593Smuzhiyun * modify it under the terms of the GNU General Public License
9*4882a593Smuzhiyun * as published by the Free Software Foundation; either version
10*4882a593Smuzhiyun * 2 of the License, or (at your option) any later version.
11*4882a593Smuzhiyun */
12*4882a593Smuzhiyun
13*4882a593Smuzhiyun #include <linux/kernel.h>
14*4882a593Smuzhiyun #include <linux/iommu.h>
15*4882a593Smuzhiyun
16*4882a593Smuzhiyun #include <asm/iommu.h>
17*4882a593Smuzhiyun #include <asm/tce.h>
18*4882a593Smuzhiyun #include "pci.h"
19*4882a593Smuzhiyun
pnv_ioda_parse_tce_sizes(struct pnv_phb * phb)20*4882a593Smuzhiyun unsigned long pnv_ioda_parse_tce_sizes(struct pnv_phb *phb)
21*4882a593Smuzhiyun {
22*4882a593Smuzhiyun struct pci_controller *hose = phb->hose;
23*4882a593Smuzhiyun struct device_node *dn = hose->dn;
24*4882a593Smuzhiyun unsigned long mask = 0;
25*4882a593Smuzhiyun int i, rc, count;
26*4882a593Smuzhiyun u32 val;
27*4882a593Smuzhiyun
28*4882a593Smuzhiyun count = of_property_count_u32_elems(dn, "ibm,supported-tce-sizes");
29*4882a593Smuzhiyun if (count <= 0) {
30*4882a593Smuzhiyun mask = SZ_4K | SZ_64K;
31*4882a593Smuzhiyun /* Add 16M for POWER8 by default */
32*4882a593Smuzhiyun if (cpu_has_feature(CPU_FTR_ARCH_207S) &&
33*4882a593Smuzhiyun !cpu_has_feature(CPU_FTR_ARCH_300))
34*4882a593Smuzhiyun mask |= SZ_16M | SZ_256M;
35*4882a593Smuzhiyun return mask;
36*4882a593Smuzhiyun }
37*4882a593Smuzhiyun
38*4882a593Smuzhiyun for (i = 0; i < count; i++) {
39*4882a593Smuzhiyun rc = of_property_read_u32_index(dn, "ibm,supported-tce-sizes",
40*4882a593Smuzhiyun i, &val);
41*4882a593Smuzhiyun if (rc == 0)
42*4882a593Smuzhiyun mask |= 1ULL << val;
43*4882a593Smuzhiyun }
44*4882a593Smuzhiyun
45*4882a593Smuzhiyun return mask;
46*4882a593Smuzhiyun }
47*4882a593Smuzhiyun
pnv_pci_setup_iommu_table(struct iommu_table * tbl,void * tce_mem,u64 tce_size,u64 dma_offset,unsigned int page_shift)48*4882a593Smuzhiyun void pnv_pci_setup_iommu_table(struct iommu_table *tbl,
49*4882a593Smuzhiyun void *tce_mem, u64 tce_size,
50*4882a593Smuzhiyun u64 dma_offset, unsigned int page_shift)
51*4882a593Smuzhiyun {
52*4882a593Smuzhiyun tbl->it_blocksize = 16;
53*4882a593Smuzhiyun tbl->it_base = (unsigned long)tce_mem;
54*4882a593Smuzhiyun tbl->it_page_shift = page_shift;
55*4882a593Smuzhiyun tbl->it_offset = dma_offset >> tbl->it_page_shift;
56*4882a593Smuzhiyun tbl->it_index = 0;
57*4882a593Smuzhiyun tbl->it_size = tce_size >> 3;
58*4882a593Smuzhiyun tbl->it_busno = 0;
59*4882a593Smuzhiyun tbl->it_type = TCE_PCI;
60*4882a593Smuzhiyun }
61*4882a593Smuzhiyun
pnv_alloc_tce_level(int nid,unsigned int shift)62*4882a593Smuzhiyun static __be64 *pnv_alloc_tce_level(int nid, unsigned int shift)
63*4882a593Smuzhiyun {
64*4882a593Smuzhiyun struct page *tce_mem = NULL;
65*4882a593Smuzhiyun __be64 *addr;
66*4882a593Smuzhiyun
67*4882a593Smuzhiyun tce_mem = alloc_pages_node(nid, GFP_ATOMIC | __GFP_NOWARN,
68*4882a593Smuzhiyun shift - PAGE_SHIFT);
69*4882a593Smuzhiyun if (!tce_mem) {
70*4882a593Smuzhiyun pr_err("Failed to allocate a TCE memory, level shift=%d\n",
71*4882a593Smuzhiyun shift);
72*4882a593Smuzhiyun return NULL;
73*4882a593Smuzhiyun }
74*4882a593Smuzhiyun addr = page_address(tce_mem);
75*4882a593Smuzhiyun memset(addr, 0, 1UL << shift);
76*4882a593Smuzhiyun
77*4882a593Smuzhiyun return addr;
78*4882a593Smuzhiyun }
79*4882a593Smuzhiyun
80*4882a593Smuzhiyun static void pnv_pci_ioda2_table_do_free_pages(__be64 *addr,
81*4882a593Smuzhiyun unsigned long size, unsigned int levels);
82*4882a593Smuzhiyun
pnv_tce(struct iommu_table * tbl,bool user,long idx,bool alloc)83*4882a593Smuzhiyun static __be64 *pnv_tce(struct iommu_table *tbl, bool user, long idx, bool alloc)
84*4882a593Smuzhiyun {
85*4882a593Smuzhiyun __be64 *tmp = user ? tbl->it_userspace : (__be64 *) tbl->it_base;
86*4882a593Smuzhiyun int level = tbl->it_indirect_levels;
87*4882a593Smuzhiyun const long shift = ilog2(tbl->it_level_size);
88*4882a593Smuzhiyun unsigned long mask = (tbl->it_level_size - 1) << (level * shift);
89*4882a593Smuzhiyun
90*4882a593Smuzhiyun while (level) {
91*4882a593Smuzhiyun int n = (idx & mask) >> (level * shift);
92*4882a593Smuzhiyun unsigned long oldtce, tce = be64_to_cpu(READ_ONCE(tmp[n]));
93*4882a593Smuzhiyun
94*4882a593Smuzhiyun if (!tce) {
95*4882a593Smuzhiyun __be64 *tmp2;
96*4882a593Smuzhiyun
97*4882a593Smuzhiyun if (!alloc)
98*4882a593Smuzhiyun return NULL;
99*4882a593Smuzhiyun
100*4882a593Smuzhiyun tmp2 = pnv_alloc_tce_level(tbl->it_nid,
101*4882a593Smuzhiyun ilog2(tbl->it_level_size) + 3);
102*4882a593Smuzhiyun if (!tmp2)
103*4882a593Smuzhiyun return NULL;
104*4882a593Smuzhiyun
105*4882a593Smuzhiyun tce = __pa(tmp2) | TCE_PCI_READ | TCE_PCI_WRITE;
106*4882a593Smuzhiyun oldtce = be64_to_cpu(cmpxchg(&tmp[n], 0,
107*4882a593Smuzhiyun cpu_to_be64(tce)));
108*4882a593Smuzhiyun if (oldtce) {
109*4882a593Smuzhiyun pnv_pci_ioda2_table_do_free_pages(tmp2,
110*4882a593Smuzhiyun ilog2(tbl->it_level_size) + 3, 1);
111*4882a593Smuzhiyun tce = oldtce;
112*4882a593Smuzhiyun }
113*4882a593Smuzhiyun }
114*4882a593Smuzhiyun
115*4882a593Smuzhiyun tmp = __va(tce & ~(TCE_PCI_READ | TCE_PCI_WRITE));
116*4882a593Smuzhiyun idx &= ~mask;
117*4882a593Smuzhiyun mask >>= shift;
118*4882a593Smuzhiyun --level;
119*4882a593Smuzhiyun }
120*4882a593Smuzhiyun
121*4882a593Smuzhiyun return tmp + idx;
122*4882a593Smuzhiyun }
123*4882a593Smuzhiyun
pnv_tce_build(struct iommu_table * tbl,long index,long npages,unsigned long uaddr,enum dma_data_direction direction,unsigned long attrs)124*4882a593Smuzhiyun int pnv_tce_build(struct iommu_table *tbl, long index, long npages,
125*4882a593Smuzhiyun unsigned long uaddr, enum dma_data_direction direction,
126*4882a593Smuzhiyun unsigned long attrs)
127*4882a593Smuzhiyun {
128*4882a593Smuzhiyun u64 proto_tce = iommu_direction_to_tce_perm(direction);
129*4882a593Smuzhiyun u64 rpn = __pa(uaddr) >> tbl->it_page_shift;
130*4882a593Smuzhiyun long i;
131*4882a593Smuzhiyun
132*4882a593Smuzhiyun if (proto_tce & TCE_PCI_WRITE)
133*4882a593Smuzhiyun proto_tce |= TCE_PCI_READ;
134*4882a593Smuzhiyun
135*4882a593Smuzhiyun for (i = 0; i < npages; i++) {
136*4882a593Smuzhiyun unsigned long newtce = proto_tce |
137*4882a593Smuzhiyun ((rpn + i) << tbl->it_page_shift);
138*4882a593Smuzhiyun unsigned long idx = index - tbl->it_offset + i;
139*4882a593Smuzhiyun
140*4882a593Smuzhiyun *(pnv_tce(tbl, false, idx, true)) = cpu_to_be64(newtce);
141*4882a593Smuzhiyun }
142*4882a593Smuzhiyun
143*4882a593Smuzhiyun return 0;
144*4882a593Smuzhiyun }
145*4882a593Smuzhiyun
146*4882a593Smuzhiyun #ifdef CONFIG_IOMMU_API
pnv_tce_xchg(struct iommu_table * tbl,long index,unsigned long * hpa,enum dma_data_direction * direction,bool alloc)147*4882a593Smuzhiyun int pnv_tce_xchg(struct iommu_table *tbl, long index,
148*4882a593Smuzhiyun unsigned long *hpa, enum dma_data_direction *direction,
149*4882a593Smuzhiyun bool alloc)
150*4882a593Smuzhiyun {
151*4882a593Smuzhiyun u64 proto_tce = iommu_direction_to_tce_perm(*direction);
152*4882a593Smuzhiyun unsigned long newtce = *hpa | proto_tce, oldtce;
153*4882a593Smuzhiyun unsigned long idx = index - tbl->it_offset;
154*4882a593Smuzhiyun __be64 *ptce = NULL;
155*4882a593Smuzhiyun
156*4882a593Smuzhiyun BUG_ON(*hpa & ~IOMMU_PAGE_MASK(tbl));
157*4882a593Smuzhiyun
158*4882a593Smuzhiyun if (*direction == DMA_NONE) {
159*4882a593Smuzhiyun ptce = pnv_tce(tbl, false, idx, false);
160*4882a593Smuzhiyun if (!ptce) {
161*4882a593Smuzhiyun *hpa = 0;
162*4882a593Smuzhiyun return 0;
163*4882a593Smuzhiyun }
164*4882a593Smuzhiyun }
165*4882a593Smuzhiyun
166*4882a593Smuzhiyun if (!ptce) {
167*4882a593Smuzhiyun ptce = pnv_tce(tbl, false, idx, alloc);
168*4882a593Smuzhiyun if (!ptce)
169*4882a593Smuzhiyun return -ENOMEM;
170*4882a593Smuzhiyun }
171*4882a593Smuzhiyun
172*4882a593Smuzhiyun if (newtce & TCE_PCI_WRITE)
173*4882a593Smuzhiyun newtce |= TCE_PCI_READ;
174*4882a593Smuzhiyun
175*4882a593Smuzhiyun oldtce = be64_to_cpu(xchg(ptce, cpu_to_be64(newtce)));
176*4882a593Smuzhiyun *hpa = oldtce & ~(TCE_PCI_READ | TCE_PCI_WRITE);
177*4882a593Smuzhiyun *direction = iommu_tce_direction(oldtce);
178*4882a593Smuzhiyun
179*4882a593Smuzhiyun return 0;
180*4882a593Smuzhiyun }
181*4882a593Smuzhiyun
pnv_tce_useraddrptr(struct iommu_table * tbl,long index,bool alloc)182*4882a593Smuzhiyun __be64 *pnv_tce_useraddrptr(struct iommu_table *tbl, long index, bool alloc)
183*4882a593Smuzhiyun {
184*4882a593Smuzhiyun if (WARN_ON_ONCE(!tbl->it_userspace))
185*4882a593Smuzhiyun return NULL;
186*4882a593Smuzhiyun
187*4882a593Smuzhiyun return pnv_tce(tbl, true, index - tbl->it_offset, alloc);
188*4882a593Smuzhiyun }
189*4882a593Smuzhiyun #endif
190*4882a593Smuzhiyun
pnv_tce_free(struct iommu_table * tbl,long index,long npages)191*4882a593Smuzhiyun void pnv_tce_free(struct iommu_table *tbl, long index, long npages)
192*4882a593Smuzhiyun {
193*4882a593Smuzhiyun long i;
194*4882a593Smuzhiyun
195*4882a593Smuzhiyun for (i = 0; i < npages; i++) {
196*4882a593Smuzhiyun unsigned long idx = index - tbl->it_offset + i;
197*4882a593Smuzhiyun __be64 *ptce = pnv_tce(tbl, false, idx, false);
198*4882a593Smuzhiyun
199*4882a593Smuzhiyun if (ptce)
200*4882a593Smuzhiyun *ptce = cpu_to_be64(0);
201*4882a593Smuzhiyun else
202*4882a593Smuzhiyun /* Skip the rest of the level */
203*4882a593Smuzhiyun i |= tbl->it_level_size - 1;
204*4882a593Smuzhiyun }
205*4882a593Smuzhiyun }
206*4882a593Smuzhiyun
pnv_tce_get(struct iommu_table * tbl,long index)207*4882a593Smuzhiyun unsigned long pnv_tce_get(struct iommu_table *tbl, long index)
208*4882a593Smuzhiyun {
209*4882a593Smuzhiyun __be64 *ptce = pnv_tce(tbl, false, index - tbl->it_offset, false);
210*4882a593Smuzhiyun
211*4882a593Smuzhiyun if (!ptce)
212*4882a593Smuzhiyun return 0;
213*4882a593Smuzhiyun
214*4882a593Smuzhiyun return be64_to_cpu(*ptce);
215*4882a593Smuzhiyun }
216*4882a593Smuzhiyun
pnv_pci_ioda2_table_do_free_pages(__be64 * addr,unsigned long size,unsigned int levels)217*4882a593Smuzhiyun static void pnv_pci_ioda2_table_do_free_pages(__be64 *addr,
218*4882a593Smuzhiyun unsigned long size, unsigned int levels)
219*4882a593Smuzhiyun {
220*4882a593Smuzhiyun const unsigned long addr_ul = (unsigned long) addr &
221*4882a593Smuzhiyun ~(TCE_PCI_READ | TCE_PCI_WRITE);
222*4882a593Smuzhiyun
223*4882a593Smuzhiyun if (levels) {
224*4882a593Smuzhiyun long i;
225*4882a593Smuzhiyun u64 *tmp = (u64 *) addr_ul;
226*4882a593Smuzhiyun
227*4882a593Smuzhiyun for (i = 0; i < size; ++i) {
228*4882a593Smuzhiyun unsigned long hpa = be64_to_cpu(tmp[i]);
229*4882a593Smuzhiyun
230*4882a593Smuzhiyun if (!(hpa & (TCE_PCI_READ | TCE_PCI_WRITE)))
231*4882a593Smuzhiyun continue;
232*4882a593Smuzhiyun
233*4882a593Smuzhiyun pnv_pci_ioda2_table_do_free_pages(__va(hpa), size,
234*4882a593Smuzhiyun levels - 1);
235*4882a593Smuzhiyun }
236*4882a593Smuzhiyun }
237*4882a593Smuzhiyun
238*4882a593Smuzhiyun free_pages(addr_ul, get_order(size << 3));
239*4882a593Smuzhiyun }
240*4882a593Smuzhiyun
pnv_pci_ioda2_table_free_pages(struct iommu_table * tbl)241*4882a593Smuzhiyun void pnv_pci_ioda2_table_free_pages(struct iommu_table *tbl)
242*4882a593Smuzhiyun {
243*4882a593Smuzhiyun const unsigned long size = tbl->it_indirect_levels ?
244*4882a593Smuzhiyun tbl->it_level_size : tbl->it_size;
245*4882a593Smuzhiyun
246*4882a593Smuzhiyun if (!tbl->it_size)
247*4882a593Smuzhiyun return;
248*4882a593Smuzhiyun
249*4882a593Smuzhiyun pnv_pci_ioda2_table_do_free_pages((__be64 *)tbl->it_base, size,
250*4882a593Smuzhiyun tbl->it_indirect_levels);
251*4882a593Smuzhiyun if (tbl->it_userspace) {
252*4882a593Smuzhiyun pnv_pci_ioda2_table_do_free_pages(tbl->it_userspace, size,
253*4882a593Smuzhiyun tbl->it_indirect_levels);
254*4882a593Smuzhiyun }
255*4882a593Smuzhiyun }
256*4882a593Smuzhiyun
pnv_pci_ioda2_table_do_alloc_pages(int nid,unsigned int shift,unsigned int levels,unsigned long limit,unsigned long * current_offset,unsigned long * total_allocated)257*4882a593Smuzhiyun static __be64 *pnv_pci_ioda2_table_do_alloc_pages(int nid, unsigned int shift,
258*4882a593Smuzhiyun unsigned int levels, unsigned long limit,
259*4882a593Smuzhiyun unsigned long *current_offset, unsigned long *total_allocated)
260*4882a593Smuzhiyun {
261*4882a593Smuzhiyun __be64 *addr, *tmp;
262*4882a593Smuzhiyun unsigned long allocated = 1UL << shift;
263*4882a593Smuzhiyun unsigned int entries = 1UL << (shift - 3);
264*4882a593Smuzhiyun long i;
265*4882a593Smuzhiyun
266*4882a593Smuzhiyun addr = pnv_alloc_tce_level(nid, shift);
267*4882a593Smuzhiyun *total_allocated += allocated;
268*4882a593Smuzhiyun
269*4882a593Smuzhiyun --levels;
270*4882a593Smuzhiyun if (!levels) {
271*4882a593Smuzhiyun *current_offset += allocated;
272*4882a593Smuzhiyun return addr;
273*4882a593Smuzhiyun }
274*4882a593Smuzhiyun
275*4882a593Smuzhiyun for (i = 0; i < entries; ++i) {
276*4882a593Smuzhiyun tmp = pnv_pci_ioda2_table_do_alloc_pages(nid, shift,
277*4882a593Smuzhiyun levels, limit, current_offset, total_allocated);
278*4882a593Smuzhiyun if (!tmp)
279*4882a593Smuzhiyun break;
280*4882a593Smuzhiyun
281*4882a593Smuzhiyun addr[i] = cpu_to_be64(__pa(tmp) |
282*4882a593Smuzhiyun TCE_PCI_READ | TCE_PCI_WRITE);
283*4882a593Smuzhiyun
284*4882a593Smuzhiyun if (*current_offset >= limit)
285*4882a593Smuzhiyun break;
286*4882a593Smuzhiyun }
287*4882a593Smuzhiyun
288*4882a593Smuzhiyun return addr;
289*4882a593Smuzhiyun }
290*4882a593Smuzhiyun
pnv_pci_ioda2_table_alloc_pages(int nid,__u64 bus_offset,__u32 page_shift,__u64 window_size,__u32 levels,bool alloc_userspace_copy,struct iommu_table * tbl)291*4882a593Smuzhiyun long pnv_pci_ioda2_table_alloc_pages(int nid, __u64 bus_offset,
292*4882a593Smuzhiyun __u32 page_shift, __u64 window_size, __u32 levels,
293*4882a593Smuzhiyun bool alloc_userspace_copy, struct iommu_table *tbl)
294*4882a593Smuzhiyun {
295*4882a593Smuzhiyun void *addr, *uas = NULL;
296*4882a593Smuzhiyun unsigned long offset = 0, level_shift, total_allocated = 0;
297*4882a593Smuzhiyun unsigned long total_allocated_uas = 0;
298*4882a593Smuzhiyun const unsigned int window_shift = ilog2(window_size);
299*4882a593Smuzhiyun unsigned int entries_shift = window_shift - page_shift;
300*4882a593Smuzhiyun unsigned int table_shift = max_t(unsigned int, entries_shift + 3,
301*4882a593Smuzhiyun PAGE_SHIFT);
302*4882a593Smuzhiyun const unsigned long tce_table_size = 1UL << table_shift;
303*4882a593Smuzhiyun
304*4882a593Smuzhiyun if (!levels || (levels > POWERNV_IOMMU_MAX_LEVELS))
305*4882a593Smuzhiyun return -EINVAL;
306*4882a593Smuzhiyun
307*4882a593Smuzhiyun if (!is_power_of_2(window_size))
308*4882a593Smuzhiyun return -EINVAL;
309*4882a593Smuzhiyun
310*4882a593Smuzhiyun /* Adjust direct table size from window_size and levels */
311*4882a593Smuzhiyun entries_shift = (entries_shift + levels - 1) / levels;
312*4882a593Smuzhiyun level_shift = entries_shift + 3;
313*4882a593Smuzhiyun level_shift = max_t(unsigned int, level_shift, PAGE_SHIFT);
314*4882a593Smuzhiyun
315*4882a593Smuzhiyun if ((level_shift - 3) * levels + page_shift >= 55)
316*4882a593Smuzhiyun return -EINVAL;
317*4882a593Smuzhiyun
318*4882a593Smuzhiyun /* Allocate TCE table */
319*4882a593Smuzhiyun addr = pnv_pci_ioda2_table_do_alloc_pages(nid, level_shift,
320*4882a593Smuzhiyun 1, tce_table_size, &offset, &total_allocated);
321*4882a593Smuzhiyun
322*4882a593Smuzhiyun /* addr==NULL means that the first level allocation failed */
323*4882a593Smuzhiyun if (!addr)
324*4882a593Smuzhiyun return -ENOMEM;
325*4882a593Smuzhiyun
326*4882a593Smuzhiyun /*
327*4882a593Smuzhiyun * First level was allocated but some lower level failed as
328*4882a593Smuzhiyun * we did not allocate as much as we wanted,
329*4882a593Smuzhiyun * release partially allocated table.
330*4882a593Smuzhiyun */
331*4882a593Smuzhiyun if (levels == 1 && offset < tce_table_size)
332*4882a593Smuzhiyun goto free_tces_exit;
333*4882a593Smuzhiyun
334*4882a593Smuzhiyun /* Allocate userspace view of the TCE table */
335*4882a593Smuzhiyun if (alloc_userspace_copy) {
336*4882a593Smuzhiyun offset = 0;
337*4882a593Smuzhiyun uas = pnv_pci_ioda2_table_do_alloc_pages(nid, level_shift,
338*4882a593Smuzhiyun 1, tce_table_size, &offset,
339*4882a593Smuzhiyun &total_allocated_uas);
340*4882a593Smuzhiyun if (!uas)
341*4882a593Smuzhiyun goto free_tces_exit;
342*4882a593Smuzhiyun if (levels == 1 && (offset < tce_table_size ||
343*4882a593Smuzhiyun total_allocated_uas != total_allocated))
344*4882a593Smuzhiyun goto free_uas_exit;
345*4882a593Smuzhiyun }
346*4882a593Smuzhiyun
347*4882a593Smuzhiyun /* Setup linux iommu table */
348*4882a593Smuzhiyun pnv_pci_setup_iommu_table(tbl, addr, tce_table_size, bus_offset,
349*4882a593Smuzhiyun page_shift);
350*4882a593Smuzhiyun tbl->it_level_size = 1ULL << (level_shift - 3);
351*4882a593Smuzhiyun tbl->it_indirect_levels = levels - 1;
352*4882a593Smuzhiyun tbl->it_userspace = uas;
353*4882a593Smuzhiyun tbl->it_nid = nid;
354*4882a593Smuzhiyun
355*4882a593Smuzhiyun pr_debug("Created TCE table: ws=%08llx ts=%lx @%08llx base=%lx uas=%p levels=%d/%d\n",
356*4882a593Smuzhiyun window_size, tce_table_size, bus_offset, tbl->it_base,
357*4882a593Smuzhiyun tbl->it_userspace, 1, levels);
358*4882a593Smuzhiyun
359*4882a593Smuzhiyun return 0;
360*4882a593Smuzhiyun
361*4882a593Smuzhiyun free_uas_exit:
362*4882a593Smuzhiyun pnv_pci_ioda2_table_do_free_pages(uas,
363*4882a593Smuzhiyun 1ULL << (level_shift - 3), levels - 1);
364*4882a593Smuzhiyun free_tces_exit:
365*4882a593Smuzhiyun pnv_pci_ioda2_table_do_free_pages(addr,
366*4882a593Smuzhiyun 1ULL << (level_shift - 3), levels - 1);
367*4882a593Smuzhiyun
368*4882a593Smuzhiyun return -ENOMEM;
369*4882a593Smuzhiyun }
370*4882a593Smuzhiyun
pnv_pci_unlink_table_and_group(struct iommu_table * tbl,struct iommu_table_group * table_group)371*4882a593Smuzhiyun void pnv_pci_unlink_table_and_group(struct iommu_table *tbl,
372*4882a593Smuzhiyun struct iommu_table_group *table_group)
373*4882a593Smuzhiyun {
374*4882a593Smuzhiyun long i;
375*4882a593Smuzhiyun bool found;
376*4882a593Smuzhiyun struct iommu_table_group_link *tgl;
377*4882a593Smuzhiyun
378*4882a593Smuzhiyun if (!tbl || !table_group)
379*4882a593Smuzhiyun return;
380*4882a593Smuzhiyun
381*4882a593Smuzhiyun /* Remove link to a group from table's list of attached groups */
382*4882a593Smuzhiyun found = false;
383*4882a593Smuzhiyun list_for_each_entry_rcu(tgl, &tbl->it_group_list, next) {
384*4882a593Smuzhiyun if (tgl->table_group == table_group) {
385*4882a593Smuzhiyun list_del_rcu(&tgl->next);
386*4882a593Smuzhiyun kfree_rcu(tgl, rcu);
387*4882a593Smuzhiyun found = true;
388*4882a593Smuzhiyun break;
389*4882a593Smuzhiyun }
390*4882a593Smuzhiyun }
391*4882a593Smuzhiyun if (WARN_ON(!found))
392*4882a593Smuzhiyun return;
393*4882a593Smuzhiyun
394*4882a593Smuzhiyun /* Clean a pointer to iommu_table in iommu_table_group::tables[] */
395*4882a593Smuzhiyun found = false;
396*4882a593Smuzhiyun for (i = 0; i < IOMMU_TABLE_GROUP_MAX_TABLES; ++i) {
397*4882a593Smuzhiyun if (table_group->tables[i] == tbl) {
398*4882a593Smuzhiyun iommu_tce_table_put(tbl);
399*4882a593Smuzhiyun table_group->tables[i] = NULL;
400*4882a593Smuzhiyun found = true;
401*4882a593Smuzhiyun break;
402*4882a593Smuzhiyun }
403*4882a593Smuzhiyun }
404*4882a593Smuzhiyun WARN_ON(!found);
405*4882a593Smuzhiyun }
406*4882a593Smuzhiyun
pnv_pci_link_table_and_group(int node,int num,struct iommu_table * tbl,struct iommu_table_group * table_group)407*4882a593Smuzhiyun long pnv_pci_link_table_and_group(int node, int num,
408*4882a593Smuzhiyun struct iommu_table *tbl,
409*4882a593Smuzhiyun struct iommu_table_group *table_group)
410*4882a593Smuzhiyun {
411*4882a593Smuzhiyun struct iommu_table_group_link *tgl = NULL;
412*4882a593Smuzhiyun
413*4882a593Smuzhiyun if (WARN_ON(!tbl || !table_group))
414*4882a593Smuzhiyun return -EINVAL;
415*4882a593Smuzhiyun
416*4882a593Smuzhiyun tgl = kzalloc_node(sizeof(struct iommu_table_group_link), GFP_KERNEL,
417*4882a593Smuzhiyun node);
418*4882a593Smuzhiyun if (!tgl)
419*4882a593Smuzhiyun return -ENOMEM;
420*4882a593Smuzhiyun
421*4882a593Smuzhiyun tgl->table_group = table_group;
422*4882a593Smuzhiyun list_add_rcu(&tgl->next, &tbl->it_group_list);
423*4882a593Smuzhiyun
424*4882a593Smuzhiyun table_group->tables[num] = iommu_tce_table_get(tbl);
425*4882a593Smuzhiyun
426*4882a593Smuzhiyun return 0;
427*4882a593Smuzhiyun }
428