Home
last modified time | relevance | path

Searched refs:vm_rb (Results 1 – 17 of 17) sorted by relevance

/OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/thirdparty/
H A Dmali_kbase_mmap.c169 vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown()
176 if (gap_start <= high_limit && vma->vm_rb.rb_right) { in kbase_unmapped_area_topdown()
178 rb_entry(vma->vm_rb.rb_right, in kbase_unmapped_area_topdown()
179 struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown()
204 if (vma->vm_rb.rb_left) { in kbase_unmapped_area_topdown()
206 rb_entry(vma->vm_rb.rb_left, in kbase_unmapped_area_topdown()
207 struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown()
216 struct rb_node *prev = &vma->vm_rb; in kbase_unmapped_area_topdown()
221 struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown()
222 if (prev == vma->vm_rb.rb_right) { in kbase_unmapped_area_topdown()
/OK3568_Linux_fs/kernel/drivers/gpu/arm/mali400/mali/linux/
H A Dmali_memory_virtual.c40 iter_node = rb_entry(*iter, struct mali_vma_node, vm_rb); in _mali_vma_offset_add_rb()
50 rb_link_node(&node->vm_rb, parent, iter); in _mali_vma_offset_add_rb()
51 rb_insert_color(&node->vm_rb, &mgr->allocation_mgr_rb); in _mali_vma_offset_add_rb()
85 rb_erase(&node->vm_rb, &mgr->allocation_mgr_rb); in mali_vma_offset_remove()
106 node = rb_entry(iter, struct mali_vma_node, vm_rb); in mali_vma_offset_search()
H A Dmali_memory_types.h125 struct rb_node vm_rb; member
/OK3568_Linux_fs/kernel/drivers/gpu/drm/
H A Ddrm_vma_manager.c282 entry = rb_entry(*iter, struct drm_vma_offset_file, vm_rb); in drm_vma_node_allow()
301 rb_link_node(&new->vm_rb, parent, iter); in drm_vma_node_allow()
302 rb_insert_color(&new->vm_rb, &node->vm_files); in drm_vma_node_allow()
335 entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb); in drm_vma_node_revoke()
338 rb_erase(&entry->vm_rb, &node->vm_files); in drm_vma_node_revoke()
376 entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb); in drm_vma_node_is_allowed()
/OK3568_Linux_fs/kernel/mm/
H A Dmmap.c333 if (vma->vm_rb.rb_left) { in vma_compute_subtree_gap()
334 subtree_gap = rb_entry(vma->vm_rb.rb_left, in vma_compute_subtree_gap()
335 struct vm_area_struct, vm_rb)->rb_subtree_gap; in vma_compute_subtree_gap()
339 if (vma->vm_rb.rb_right) { in vma_compute_subtree_gap()
340 subtree_gap = rb_entry(vma->vm_rb.rb_right, in vma_compute_subtree_gap()
341 struct vm_area_struct, vm_rb)->rb_subtree_gap; in vma_compute_subtree_gap()
357 vma = rb_entry(nd, struct vm_area_struct, vm_rb); in browse_rb()
402 vma = rb_entry(nd, struct vm_area_struct, vm_rb); in validate_mm_rb()
454 struct vm_area_struct, vm_rb, in RB_DECLARE_CALLBACKS_MAX() argument
475 vma_gap_callbacks_propagate(&vma->vm_rb, NULL);
[all …]
H A Dnommu.c442 last = rb_entry(lastp, struct vm_region, vm_rb); in validate_nommu_regions()
447 region = rb_entry(p, struct vm_region, vm_rb); in validate_nommu_regions()
448 last = rb_entry(lastp, struct vm_region, vm_rb); in validate_nommu_regions()
477 pregion = rb_entry(parent, struct vm_region, vm_rb); in add_nommu_region()
488 rb_link_node(&region->vm_rb, parent, p); in add_nommu_region()
489 rb_insert_color(&region->vm_rb, &nommu_region_tree); in add_nommu_region()
502 rb_erase(&region->vm_rb, &nommu_region_tree); in delete_nommu_region()
590 pvma = rb_entry(parent, struct vm_area_struct, vm_rb); in add_vma_to_mm()
613 rb_link_node(&vma->vm_rb, parent, p); in add_vma_to_mm()
614 rb_insert_color(&vma->vm_rb, &mm->mm_rb); in add_vma_to_mm()
[all …]
H A Dinternal.h46 int ret = RB_EMPTY_NODE(&vmf->vma->vm_rb); in vma_has_changed()
H A Dmemory.c5239 ret = !RB_EMPTY_NODE(&vma->vm_rb) && in can_reuse_spf_vma()
/OK3568_Linux_fs/kernel/mm/damon/
H A Dvaddr-test.h28 vmas[i].vm_rb.rb_left = NULL; in __link_vmas()
29 vmas[i].vm_rb.rb_right = &vmas[i + 1].vm_rb; in __link_vmas()
42 vmas[i].vm_rb.rb_right = NULL; in __link_vmas()
H A Dvaddr.c132 rbroot.rb_node = &vma->vm_rb; in __damon_va_three_regions()
134 struct vm_area_struct, vm_rb); in __damon_va_three_regions()
/OK3568_Linux_fs/kernel/fs/proc/
H A Dtask_nommu.c30 vma = rb_entry(p, struct vm_area_struct, vm_rb); in task_mem()
91 vma = rb_entry(p, struct vm_area_struct, vm_rb); in task_vsize()
109 vma = rb_entry(p, struct vm_area_struct, vm_rb); in task_statm()
195 return nommu_vma_show(m, rb_entry(p, struct vm_area_struct, vm_rb)); in show_map()
H A Dnommu.c77 return nommu_region_show(m, rb_entry(p, struct vm_region, vm_rb)); in nommu_region_list_show()
/OK3568_Linux_fs/kernel/include/linux/
H A Dmm_types.h279 struct rb_node vm_rb; /* link in global region tree */ member
318 struct rb_node vm_rb; member
H A Dmm.h1767 if (!RB_EMPTY_NODE(&vma->vm_rb)) in vm_write_begin()
/OK3568_Linux_fs/kernel/include/drm/
H A Ddrm_vma_manager.h47 struct rb_node vm_rb; member
/OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_core_linux.c2094 vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown()
2101 if (gap_start <= high_limit && vma->vm_rb.rb_right) { in kbase_unmapped_area_topdown()
2103 rb_entry(vma->vm_rb.rb_right, in kbase_unmapped_area_topdown()
2104 struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown()
2128 if (vma->vm_rb.rb_left) { in kbase_unmapped_area_topdown()
2130 rb_entry(vma->vm_rb.rb_left, in kbase_unmapped_area_topdown()
2131 struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown()
2140 struct rb_node *prev = &vma->vm_rb; in kbase_unmapped_area_topdown()
2144 struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown()
2145 if (prev == vma->vm_rb.rb_right) { in kbase_unmapped_area_topdown()
/OK3568_Linux_fs/kernel/kernel/
H A Dfork.c597 rb_link = &tmp->vm_rb.rb_right; in dup_mmap()
598 rb_parent = &tmp->vm_rb; in dup_mmap()