Lines Matching refs:paddr

73 	paddr_t paddr;  member
77 #define MEMACCESS_AREA(a, s) { .paddr = a, .size = s }
169 *base = secure_only[0].paddr; in core_mmu_get_secure_memory()
183 secure_only[0].paddr = base; in core_mmu_set_secure_memory()
205 if (core_is_buffer_intersect(pa, size, a[n].paddr, a[n].size)) in _pbuf_intersects()
219 if (core_is_buffer_inside(pa, size, a[n].paddr, a[n].size)) in _pbuf_is_inside()
365 sec_sdp.paddr = tmp_addr; in dtb_get_sdp_region()
503 carve_out_phys_mem(&m, &num_elems, sec_sdp.paddr, sec_sdp.size); in core_mmu_set_discovered_nsec_ddr()
510 carve_out_phys_mem(&m, &num_elems, secure_only[n].paddr, in core_mmu_set_discovered_nsec_ddr()
598 is_sdp_mem = core_is_buffer_inside(pbuf, len, sec_sdp.paddr, in pbuf_is_sdp_mem()
651 *mobj = core_sdp_mem_alloc_mobj(sec_sdp.paddr, sec_sdp.size); in core_sdp_mem_create_mobjs()
1056 if (pbuf_intersects(nsec_shared, secure_only[n].paddr, in check_sec_nsec_mem_config()
1108 vaddr_t ram_start = secure_only[0].paddr; in collect_mem_ranges()
1180 secure_only[0].paddr + in collect_mem_ranges()
1185 ADD_PHYS_MEM(MEM_AREA_SEC_RAM_OVERALL, secure_only[n].paddr, in collect_mem_ranges()
1190 ADD_PHYS_MEM(MEM_AREA_SEC_RAM_OVERALL, secure_only[n].paddr, in collect_mem_ranges()
1247 static bool place_tee_ram_at_top(paddr_t paddr) in place_tee_ram_at_top() argument
1249 return paddr > BIT64(core_mmu_get_va_width()) / 2; in place_tee_ram_at_top()
1256 bool __weak core_mmu_prefer_tee_ram_at_top(paddr_t paddr) in core_mmu_prefer_tee_ram_at_top() argument
1258 return place_tee_ram_at_top(paddr); in core_mmu_prefer_tee_ram_at_top()
1498 vaddr_t start_addr = secure_only[0].paddr; in init_mem_map()
1938 static bool can_map_at_level(paddr_t paddr, vaddr_t vaddr, in can_map_at_level() argument
1943 if ((vaddr | paddr) & (block_size - 1)) in can_map_at_level()
1975 paddr_t paddr = mm->pa; in core_mmu_map_region() local
1982 assert(!((vaddr | paddr) & SMALL_PAGE_MASK)); in core_mmu_map_region()
1983 if (!paddr) in core_mmu_map_region()
2002 if (!can_map_at_level(paddr, vaddr, size_left, in core_mmu_map_region()
2023 core_mmu_set_entry(&tbl_info, idx, paddr, attr); in core_mmu_map_region()
2030 if (paddr) in core_mmu_map_region()
2031 paddr += block_size; in core_mmu_map_region()
2127 paddr_t paddr = pstart; in core_mmu_map_contiguous_pages() local
2135 if ((vaddr | paddr) & SMALL_PAGE_MASK) in core_mmu_map_contiguous_pages()
2167 core_mmu_set_entry(&tbl_info, idx, paddr, in core_mmu_map_contiguous_pages()
2169 paddr += SMALL_PAGE_SIZE; in core_mmu_map_contiguous_pages()
2790 b1 = secure_only[1].paddr; in core_mmu_init_phys_mem()
2793 virt_init_memory(&static_memory_map, secure_only[0].paddr, in core_mmu_init_phys_mem()
2802 phys_mem_init(0, 0, secure_only[1].paddr, secure_only[1].size); in core_mmu_init_phys_mem()
2812 ps = secure_only[1].paddr; in core_mmu_init_phys_mem()
2815 phys_mem_init(secure_only[0].paddr, secure_only[0].size, in core_mmu_init_phys_mem()