Lines Matching refs:va
63 vaddr_t va = 0; in mem_alloc_tmp() local
70 if (SUB_OVERFLOW(desc->mem_end, len, &va)) in mem_alloc_tmp()
72 va = ROUNDDOWN2(va, align); in mem_alloc_tmp()
73 if (va < desc->mem_start) in mem_alloc_tmp()
75 desc->mem_end = va; in mem_alloc_tmp()
77 asan_tag_access((void *)va, (void *)(va + len)); in mem_alloc_tmp()
79 return (void *)va; in mem_alloc_tmp()
82 static void add_padding(struct boot_mem_desc *desc, vaddr_t va) in add_padding() argument
92 if (rounded < va && va - rounded > sizeof(*pad)) { in add_padding()
95 pad->len = va - start; in add_padding()
104 vaddr_t va = 0; in mem_alloc() local
113 va = ROUNDUP2(desc->mem_start, align); in mem_alloc()
114 if (ADD_OVERFLOW(va, len, &ve)) in mem_alloc()
118 add_padding(desc, va); in mem_alloc()
121 asan_tag_access((void *)va, (void *)(va + len)); in mem_alloc()
123 return (void *)va; in mem_alloc()
143 static bool tag_padding_no_access(vaddr_t va __unused, size_t len __unused, in tag_padding_no_access()
248 void boot_mem_foreach_padding(bool (*func)(vaddr_t va, size_t len, void *ptr), in boot_mem_foreach_padding() argument
285 vaddr_t va = 0; in boot_mem_release_unused() local
311 va = ROUNDUP(boot_mem_desc->mem_start, SMALL_PAGE_SIZE); in boot_mem_release_unused()
329 DMSG("Releasing %zu bytes from va %#"PRIxVA, n, va); in boot_mem_release_unused()
332 core_mmu_unmap_pages(va, n / SMALL_PAGE_SIZE); in boot_mem_release_unused()
337 return va; in boot_mem_release_unused()
343 vaddr_t va = 0; in boot_mem_release_tmp_alloc() local
352 va = boot_mem_desc->mem_end; in boot_mem_release_tmp_alloc()
354 DMSG("Releasing %zu bytes from va %#"PRIxVA, n, va); in boot_mem_release_tmp_alloc()
358 va = ROUNDDOWN(boot_mem_desc->mem_end, SMALL_PAGE_SIZE); in boot_mem_release_tmp_alloc()
359 pa = vaddr_to_phys(va); in boot_mem_release_tmp_alloc()
370 DMSG("Releasing %zu bytes from va %#"PRIxVA, n, va); in boot_mem_release_tmp_alloc()
373 core_mmu_unmap_pages(va, n / SMALL_PAGE_SIZE); in boot_mem_release_tmp_alloc()