| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/thirdparty/ |
| H A D | mali_kbase_mmap.c | 169 vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown() 176 if (gap_start <= high_limit && vma->vm_rb.rb_right) { in kbase_unmapped_area_topdown() 178 rb_entry(vma->vm_rb.rb_right, in kbase_unmapped_area_topdown() 179 struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown() 204 if (vma->vm_rb.rb_left) { in kbase_unmapped_area_topdown() 206 rb_entry(vma->vm_rb.rb_left, in kbase_unmapped_area_topdown() 207 struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown() 216 struct rb_node *prev = &vma->vm_rb; in kbase_unmapped_area_topdown() 221 struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown() 222 if (prev == vma->vm_rb.rb_right) { in kbase_unmapped_area_topdown()
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/mali400/mali/linux/ |
| H A D | mali_memory_virtual.c | 40 iter_node = rb_entry(*iter, struct mali_vma_node, vm_rb); in _mali_vma_offset_add_rb() 50 rb_link_node(&node->vm_rb, parent, iter); in _mali_vma_offset_add_rb() 51 rb_insert_color(&node->vm_rb, &mgr->allocation_mgr_rb); in _mali_vma_offset_add_rb() 85 rb_erase(&node->vm_rb, &mgr->allocation_mgr_rb); in mali_vma_offset_remove() 106 node = rb_entry(iter, struct mali_vma_node, vm_rb); in mali_vma_offset_search()
|
| H A D | mali_memory_types.h | 125 struct rb_node vm_rb; member
|
| /OK3568_Linux_fs/kernel/drivers/gpu/drm/ |
| H A D | drm_vma_manager.c | 282 entry = rb_entry(*iter, struct drm_vma_offset_file, vm_rb); in drm_vma_node_allow() 301 rb_link_node(&new->vm_rb, parent, iter); in drm_vma_node_allow() 302 rb_insert_color(&new->vm_rb, &node->vm_files); in drm_vma_node_allow() 335 entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb); in drm_vma_node_revoke() 338 rb_erase(&entry->vm_rb, &node->vm_files); in drm_vma_node_revoke() 376 entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb); in drm_vma_node_is_allowed()
|
| /OK3568_Linux_fs/kernel/mm/ |
| H A D | mmap.c | 333 if (vma->vm_rb.rb_left) { in vma_compute_subtree_gap() 334 subtree_gap = rb_entry(vma->vm_rb.rb_left, in vma_compute_subtree_gap() 335 struct vm_area_struct, vm_rb)->rb_subtree_gap; in vma_compute_subtree_gap() 339 if (vma->vm_rb.rb_right) { in vma_compute_subtree_gap() 340 subtree_gap = rb_entry(vma->vm_rb.rb_right, in vma_compute_subtree_gap() 341 struct vm_area_struct, vm_rb)->rb_subtree_gap; in vma_compute_subtree_gap() 357 vma = rb_entry(nd, struct vm_area_struct, vm_rb); in browse_rb() 402 vma = rb_entry(nd, struct vm_area_struct, vm_rb); in validate_mm_rb() 454 struct vm_area_struct, vm_rb, in RB_DECLARE_CALLBACKS_MAX() argument 475 vma_gap_callbacks_propagate(&vma->vm_rb, NULL); [all …]
|
| H A D | nommu.c | 442 last = rb_entry(lastp, struct vm_region, vm_rb); in validate_nommu_regions() 447 region = rb_entry(p, struct vm_region, vm_rb); in validate_nommu_regions() 448 last = rb_entry(lastp, struct vm_region, vm_rb); in validate_nommu_regions() 477 pregion = rb_entry(parent, struct vm_region, vm_rb); in add_nommu_region() 488 rb_link_node(®ion->vm_rb, parent, p); in add_nommu_region() 489 rb_insert_color(®ion->vm_rb, &nommu_region_tree); in add_nommu_region() 502 rb_erase(®ion->vm_rb, &nommu_region_tree); in delete_nommu_region() 590 pvma = rb_entry(parent, struct vm_area_struct, vm_rb); in add_vma_to_mm() 613 rb_link_node(&vma->vm_rb, parent, p); in add_vma_to_mm() 614 rb_insert_color(&vma->vm_rb, &mm->mm_rb); in add_vma_to_mm() [all …]
|
| H A D | internal.h | 46 int ret = RB_EMPTY_NODE(&vmf->vma->vm_rb); in vma_has_changed()
|
| H A D | memory.c | 5239 ret = !RB_EMPTY_NODE(&vma->vm_rb) && in can_reuse_spf_vma()
|
| /OK3568_Linux_fs/kernel/mm/damon/ |
| H A D | vaddr-test.h | 28 vmas[i].vm_rb.rb_left = NULL; in __link_vmas() 29 vmas[i].vm_rb.rb_right = &vmas[i + 1].vm_rb; in __link_vmas() 42 vmas[i].vm_rb.rb_right = NULL; in __link_vmas()
|
| H A D | vaddr.c | 132 rbroot.rb_node = &vma->vm_rb; in __damon_va_three_regions() 134 struct vm_area_struct, vm_rb); in __damon_va_three_regions()
|
| /OK3568_Linux_fs/kernel/fs/proc/ |
| H A D | task_nommu.c | 30 vma = rb_entry(p, struct vm_area_struct, vm_rb); in task_mem() 91 vma = rb_entry(p, struct vm_area_struct, vm_rb); in task_vsize() 109 vma = rb_entry(p, struct vm_area_struct, vm_rb); in task_statm() 195 return nommu_vma_show(m, rb_entry(p, struct vm_area_struct, vm_rb)); in show_map()
|
| H A D | nommu.c | 77 return nommu_region_show(m, rb_entry(p, struct vm_region, vm_rb)); in nommu_region_list_show()
|
| /OK3568_Linux_fs/kernel/include/linux/ |
| H A D | mm_types.h | 279 struct rb_node vm_rb; /* link in global region tree */ member 318 struct rb_node vm_rb; member
|
| H A D | mm.h | 1767 if (!RB_EMPTY_NODE(&vma->vm_rb)) in vm_write_begin()
|
| /OK3568_Linux_fs/kernel/include/drm/ |
| H A D | drm_vma_manager.h | 47 struct rb_node vm_rb; member
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/ |
| H A D | mali_kbase_core_linux.c | 2094 vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown() 2101 if (gap_start <= high_limit && vma->vm_rb.rb_right) { in kbase_unmapped_area_topdown() 2103 rb_entry(vma->vm_rb.rb_right, in kbase_unmapped_area_topdown() 2104 struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown() 2128 if (vma->vm_rb.rb_left) { in kbase_unmapped_area_topdown() 2130 rb_entry(vma->vm_rb.rb_left, in kbase_unmapped_area_topdown() 2131 struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown() 2140 struct rb_node *prev = &vma->vm_rb; in kbase_unmapped_area_topdown() 2144 struct vm_area_struct, vm_rb); in kbase_unmapped_area_topdown() 2145 if (prev == vma->vm_rb.rb_right) { in kbase_unmapped_area_topdown()
|
| /OK3568_Linux_fs/kernel/kernel/ |
| H A D | fork.c | 597 rb_link = &tmp->vm_rb.rb_right; in dup_mmap() 598 rb_parent = &tmp->vm_rb; in dup_mmap()
|