Lines Matching refs:phys_addr
178 __ioremap_caller(resource_size_t phys_addr, unsigned long size, in __ioremap_caller() argument
183 const resource_size_t unaligned_phys_addr = phys_addr; in __ioremap_caller()
193 last_addr = phys_addr + size - 1; in __ioremap_caller()
194 if (!size || last_addr < phys_addr) in __ioremap_caller()
197 if (!phys_addr_valid(phys_addr)) { in __ioremap_caller()
199 (unsigned long long)phys_addr); in __ioremap_caller()
204 __ioremap_check_mem(phys_addr, size, &io_desc); in __ioremap_caller()
211 &phys_addr, &last_addr); in __ioremap_caller()
218 offset = phys_addr & ~PAGE_MASK; in __ioremap_caller()
219 phys_addr &= PAGE_MASK; in __ioremap_caller()
220 size = PAGE_ALIGN(last_addr+1) - phys_addr; in __ioremap_caller()
226 phys_addr &= PHYSICAL_PAGE_MASK; in __ioremap_caller()
228 retval = memtype_reserve(phys_addr, (u64)phys_addr + size, in __ioremap_caller()
236 if (!is_new_memtype_allowed(phys_addr, size, pcm, new_pcm)) { in __ioremap_caller()
239 (unsigned long long)phys_addr, in __ioremap_caller()
240 (unsigned long long)(phys_addr + size), in __ioremap_caller()
284 area->phys_addr = phys_addr; in __ioremap_caller()
287 if (memtype_kernel_map_sync(phys_addr, size, pcm)) in __ioremap_caller()
290 if (ioremap_page_range(vaddr, vaddr + size, phys_addr, prot)) in __ioremap_caller()
307 memtype_free(phys_addr, phys_addr + size); in __ioremap_caller()
332 void __iomem *ioremap(resource_size_t phys_addr, unsigned long size) in ioremap() argument
344 return __ioremap_caller(phys_addr, size, pcm, in ioremap()
373 void __iomem *ioremap_uc(resource_size_t phys_addr, unsigned long size) in ioremap_uc() argument
377 return __ioremap_caller(phys_addr, size, pcm, in ioremap_uc()
392 void __iomem *ioremap_wc(resource_size_t phys_addr, unsigned long size) in ioremap_wc() argument
394 return __ioremap_caller(phys_addr, size, _PAGE_CACHE_MODE_WC, in ioremap_wc()
409 void __iomem *ioremap_wt(resource_size_t phys_addr, unsigned long size) in ioremap_wt() argument
411 return __ioremap_caller(phys_addr, size, _PAGE_CACHE_MODE_WT, in ioremap_wt()
416 void __iomem *ioremap_encrypted(resource_size_t phys_addr, unsigned long size) in ioremap_encrypted() argument
418 return __ioremap_caller(phys_addr, size, _PAGE_CACHE_MODE_WB, in ioremap_encrypted()
423 void __iomem *ioremap_cache(resource_size_t phys_addr, unsigned long size) in ioremap_cache() argument
425 return __ioremap_caller(phys_addr, size, _PAGE_CACHE_MODE_WB, in ioremap_cache()
430 void __iomem *ioremap_prot(resource_size_t phys_addr, unsigned long size, in ioremap_prot() argument
433 return __ioremap_caller(phys_addr, size, in ioremap_prot()
483 memtype_free(p->phys_addr, p->phys_addr + get_vm_area_size(p)); in iounmap()
547 static bool memremap_should_map_decrypted(resource_size_t phys_addr, in memremap_should_map_decrypted() argument
556 is_pmem = region_intersects(phys_addr, size, IORESOURCE_MEM, in memremap_should_map_decrypted()
566 switch (efi_mem_type(phys_addr)) { in memremap_should_map_decrypted()
568 if (efi_mem_attributes(phys_addr) & EFI_MEMORY_NV) in memremap_should_map_decrypted()
577 switch (e820__get_entry_type(phys_addr, phys_addr + size - 1)) { in memremap_should_map_decrypted()
600 static bool memremap_is_efi_data(resource_size_t phys_addr, in memremap_is_efi_data() argument
612 if (phys_addr == paddr) in memremap_is_efi_data()
618 if (phys_addr == paddr) in memremap_is_efi_data()
621 if (efi_is_table_address(phys_addr)) in memremap_is_efi_data()
624 switch (efi_mem_type(phys_addr)) { in memremap_is_efi_data()
639 static bool memremap_is_setup_data(resource_size_t phys_addr, in memremap_is_setup_data() argument
650 if (phys_addr == paddr) in memremap_is_setup_data()
663 if ((phys_addr > paddr) && (phys_addr < (paddr + len))) { in memremap_is_setup_data()
687 if ((phys_addr > paddr) && (phys_addr < (paddr + len))) in memremap_is_setup_data()
700 static bool __init early_memremap_is_setup_data(resource_size_t phys_addr, in early_memremap_is_setup_data() argument
711 if (phys_addr == paddr) in early_memremap_is_setup_data()
725 if ((phys_addr > paddr) && (phys_addr < (paddr + len))) { in early_memremap_is_setup_data()
749 if ((phys_addr > paddr) && (phys_addr < (paddr + len))) in early_memremap_is_setup_data()
763 bool arch_memremap_can_ram_remap(resource_size_t phys_addr, unsigned long size, in arch_memremap_can_ram_remap() argument
776 if (memremap_is_setup_data(phys_addr, size) || in arch_memremap_can_ram_remap()
777 memremap_is_efi_data(phys_addr, size)) in arch_memremap_can_ram_remap()
781 return !memremap_should_map_decrypted(phys_addr, size); in arch_memremap_can_ram_remap()
790 pgprot_t __init early_memremap_pgprot_adjust(resource_size_t phys_addr, in early_memremap_pgprot_adjust() argument
802 if (early_memremap_is_setup_data(phys_addr, size) || in early_memremap_pgprot_adjust()
803 memremap_is_efi_data(phys_addr, size)) in early_memremap_pgprot_adjust()
807 if (encrypted_prot && memremap_should_map_decrypted(phys_addr, size)) in early_memremap_pgprot_adjust()
814 bool phys_mem_access_encrypted(unsigned long phys_addr, unsigned long size) in phys_mem_access_encrypted() argument
816 return arch_memremap_can_ram_remap(phys_addr, size, 0); in phys_mem_access_encrypted()
821 void __init *early_memremap_encrypted(resource_size_t phys_addr, in early_memremap_encrypted() argument
824 return early_memremap_prot(phys_addr, size, __PAGE_KERNEL_ENC); in early_memremap_encrypted()
831 void __init *early_memremap_encrypted_wp(resource_size_t phys_addr, in early_memremap_encrypted_wp() argument
836 return early_memremap_prot(phys_addr, size, __PAGE_KERNEL_ENC_WP); in early_memremap_encrypted_wp()
840 void __init *early_memremap_decrypted(resource_size_t phys_addr, in early_memremap_decrypted() argument
843 return early_memremap_prot(phys_addr, size, __PAGE_KERNEL_NOENC); in early_memremap_decrypted()
850 void __init *early_memremap_decrypted_wp(resource_size_t phys_addr, in early_memremap_decrypted_wp() argument
855 return early_memremap_prot(phys_addr, size, __PAGE_KERNEL_NOENC_WP); in early_memremap_decrypted_wp()