Lines Matching refs:pent
51 #define lv2ent_fault(pent) ((*(pent) & 3) == 0) argument
52 #define lv2ent_small(pent) ((*(pent) & 2) == 2) argument
53 #define lv2ent_large(pent) ((*(pent) & 3) == 1) argument
100 #define lpage_phys(pent) (sect_to_phys(*(pent)) & LPAGE_MASK) argument
102 #define spage_phys(pent) (sect_to_phys(*(pent)) & SPAGE_MASK) argument
932 sysmmu_pte_t *pent; in alloc_lv2entry() local
935 pent = kmem_cache_zalloc(lv2table_kmem_cache, GFP_ATOMIC); in alloc_lv2entry()
936 BUG_ON((uintptr_t)pent & (LV2TABLE_SIZE - 1)); in alloc_lv2entry()
937 if (!pent) in alloc_lv2entry()
940 exynos_iommu_set_pte(sent, mk_lv1ent_page(virt_to_phys(pent))); in alloc_lv2entry()
941 kmemleak_ignore(pent); in alloc_lv2entry()
943 handle = dma_map_single(dma_dev, pent, LV2TABLE_SIZE, in alloc_lv2entry()
946 kmem_cache_free(lv2table_kmem_cache, pent); in alloc_lv2entry()
1018 static int lv2set_page(sysmmu_pte_t *pent, phys_addr_t paddr, size_t size, in lv2set_page() argument
1022 if (WARN_ON(!lv2ent_fault(pent))) in lv2set_page()
1025 exynos_iommu_set_pte(pent, mk_lv2ent_spage(paddr, prot)); in lv2set_page()
1029 dma_addr_t pent_base = virt_to_phys(pent); in lv2set_page()
1032 sizeof(*pent) * SPAGES_PER_LPAGE, in lv2set_page()
1034 for (i = 0; i < SPAGES_PER_LPAGE; i++, pent++) { in lv2set_page()
1035 if (WARN_ON(!lv2ent_fault(pent))) { in lv2set_page()
1037 memset(pent - i, 0, sizeof(*pent) * i); in lv2set_page()
1041 *pent = mk_lv2ent_lpage(paddr, prot); in lv2set_page()
1044 sizeof(*pent) * SPAGES_PER_LPAGE, in lv2set_page()
1099 sysmmu_pte_t *pent; in exynos_iommu_map() local
1101 pent = alloc_lv2entry(domain, entry, iova, in exynos_iommu_map()
1104 if (IS_ERR(pent)) in exynos_iommu_map()
1105 ret = PTR_ERR(pent); in exynos_iommu_map()
1107 ret = lv2set_page(pent, paddr, size, prot, in exynos_iommu_map()