Searched refs:vabase (Results 1 – 5 of 5) sorted by relevance
91 pgt->vabase = 0; in free_pgt()131 static struct pgt *alloc_pgt(vaddr_t vabase) in alloc_pgt() argument155 pgt->vabase = vabase; in alloc_pgt()167 return core_is_buffer_inside(p->vabase, CORE_MMU_PGDIR_SIZE, begin, in pgt_entry_matches()240 vaddr_t b = MAX(p->vabase, begin); in pgt_clear_range()241 vaddr_t e = MIN(p->vabase + CORE_MMU_PGDIR_SIZE, end); in pgt_clear_range()247 idx = (b - p->vabase) / SMALL_PAGE_SIZE; in pgt_clear_range()256 while (p && p->vabase < va) { in prune_before_va()298 if (p->vabase < va) { in pgt_check_avail()306 if (p->vabase == va) in pgt_check_avail()[all …]
1905 while ((*pgt)->vabase < pg_info->va_base) { in set_pg_region()1910 assert((*pgt)->vabase == pg_info->va_base); in set_pg_region()
213 ti.va_base = p->vabase; in set_um_region()
26 vaddr_t vabase; member74 static inline struct pgt *pgt_pop_from_cache_list(vaddr_t vabase __unused, in pgt_pop_from_cache_list()79 struct pgt *pgt_pop_from_cache_list(vaddr_t vabase, struct ts_ctx *ctx);
462 return tblidx.pgt->vabase + (tblidx.idx << SMALL_PAGE_SHIFT); in tblidx2va()557 pager_tables[n].pgt.vabase = pager_tables[n].tbl_info.va_base; in tee_pager_early_init()789 idx = core_mmu_va2idx(&dir_info, pgt->vabase); in map_pgts()1905 while (p && (va & ~CORE_MMU_PGDIR_MASK) != p->vabase) in find_pgt()