Lines Matching refs:va
254 struct pgt *pp, vaddr_t va) in prune_before_va() argument
256 while (p && p->vabase < va) { in prune_before_va()
280 vaddr_t va = 0; in pgt_check_avail() local
291 for (va = ROUNDDOWN(r->va, CORE_MMU_PGDIR_SIZE); in pgt_check_avail()
292 va < r->va + r->size; va += CORE_MMU_PGDIR_SIZE) { in pgt_check_avail()
294 p = prune_before_va(pgt_cache, p, pp, va); in pgt_check_avail()
298 if (p->vabase < va) { in pgt_check_avail()
306 if (p->vabase == va) in pgt_check_avail()
315 for (va = ROUNDDOWN(r->va, CORE_MMU_PGDIR_SIZE); in pgt_check_avail()
316 va < r->va + r->size; va += CORE_MMU_PGDIR_SIZE) { in pgt_check_avail()
317 if (p && p->vabase < va) { in pgt_check_avail()
323 if (p->vabase == va) in pgt_check_avail()
325 assert(p->vabase > va); in pgt_check_avail()
328 p = alloc_pgt(va); in pgt_check_avail()
480 vaddr_t va = (vaddr_t)p->tbl & ~SMALL_PAGE_MASK; in push_to_free_list() local
482 tee_pager_release_phys((void *)va, SMALL_PAGE_SIZE); in push_to_free_list()
779 vaddr_t va = 0; in pgt_alloc_unlocked() local
782 for (va = ROUNDDOWN(r->va, CORE_MMU_PGDIR_SIZE); in pgt_alloc_unlocked()
783 va < r->va + r->size; va += CORE_MMU_PGDIR_SIZE) { in pgt_alloc_unlocked()
784 if (p && p->vabase == va) in pgt_alloc_unlocked()
786 p = pop_from_some_list(va, ctx); in pgt_alloc_unlocked()
808 vaddr_t va = 0; in pgt_check_avail() local
811 for (va = ROUNDDOWN(r->va, CORE_MMU_PGDIR_SIZE); in pgt_check_avail()
812 va < r->va + r->size; va += CORE_MMU_PGDIR_SIZE) { in pgt_check_avail()
813 if (va == last_va) in pgt_check_avail()
816 last_va = va; in pgt_check_avail()