Lines Matching refs:iova

30 	dma_addr_t		iova;  member
192 msi_page->iova = start; in cookie_init_hw_msi_region()
478 unsigned long shift, iova_len, iova = 0; in iommu_dma_alloc_iova() local
503 iova = alloc_iova_fast(iovad, iova_len, in iommu_dma_alloc_iova()
506 if (!iova) in iommu_dma_alloc_iova()
507 iova = alloc_iova_fast(iovad, iova_len, dma_limit >> shift, in iommu_dma_alloc_iova()
510 trace_android_vh_iommu_alloc_iova(dev, (dma_addr_t)iova << shift, size); in iommu_dma_alloc_iova()
511 trace_android_vh_iommu_iovad_alloc_iova(dev, iovad, (dma_addr_t)iova << shift, size); in iommu_dma_alloc_iova()
513 return (dma_addr_t)iova << shift; in iommu_dma_alloc_iova()
517 dma_addr_t iova, size_t size) in iommu_dma_free_iova() argument
525 queue_iova(iovad, iova_pfn(iovad, iova), in iommu_dma_free_iova()
528 free_iova_fast(iovad, iova_pfn(iovad, iova), in iommu_dma_free_iova()
531 trace_android_vh_iommu_free_iova(iova, size); in iommu_dma_free_iova()
532 trace_android_vh_iommu_iovad_free_iova(iovad, iova, size); in iommu_dma_free_iova()
564 dma_addr_t iova; in __iommu_dma_map() local
571 iova = iommu_dma_alloc_iova(domain, size, dma_mask, dev); in __iommu_dma_map()
572 if (!iova) in __iommu_dma_map()
575 if (iommu_map_atomic(domain, iova, phys - iova_off, size, prot)) { in __iommu_dma_map()
576 iommu_dma_free_iova(cookie, iova, size); in __iommu_dma_map()
579 return iova + iova_off; in __iommu_dma_map()
671 dma_addr_t iova; in iommu_dma_alloc_remap() local
696 iova = iommu_dma_alloc_iova(domain, size, dev->coherent_dma_mask, dev); in iommu_dma_alloc_remap()
697 if (!iova) in iommu_dma_alloc_remap()
711 if (iommu_map_sg_atomic(domain, iova, sgt.sgl, sgt.orig_nents, ioprot) in iommu_dma_alloc_remap()
720 *dma_handle = iova; in iommu_dma_alloc_remap()
725 __iommu_dma_unmap(dev, iova, size); in iommu_dma_alloc_remap()
729 iommu_dma_free_iova(cookie, iova, size); in iommu_dma_alloc_remap()
916 dma_addr_t iova; in iommu_dma_map_sg() local
966 iova = iommu_dma_alloc_iova(domain, iova_len, dma_get_mask(dev), dev); in iommu_dma_map_sg()
967 if (!iova) in iommu_dma_map_sg()
974 if (iommu_map_sg_atomic(domain, iova, sg, nents, prot) < iova_len) in iommu_dma_map_sg()
977 return __finalise_sg(dev, sg, nents, iova); in iommu_dma_map_sg()
980 iommu_dma_free_iova(cookie, iova, iova_len); in iommu_dma_map_sg()
1281 dma_addr_t iova; in iommu_dma_get_msi_page() local
1294 iova = iommu_dma_alloc_iova(domain, size, dma_get_mask(dev), dev); in iommu_dma_get_msi_page()
1295 if (!iova) in iommu_dma_get_msi_page()
1298 if (iommu_map(domain, iova, msi_addr, size, prot)) in iommu_dma_get_msi_page()
1303 msi_page->iova = iova; in iommu_dma_get_msi_page()
1308 iommu_dma_free_iova(cookie, iova, size); in iommu_dma_get_msi_page()
1354 msg->address_hi = upper_32_bits(msi_page->iova); in iommu_dma_compose_msi_msg()
1356 msg->address_lo += lower_32_bits(msi_page->iova); in iommu_dma_compose_msi_msg()