Lines Matching refs:vm_rb

333 	if (vma->vm_rb.rb_left) {  in vma_compute_subtree_gap()
334 subtree_gap = rb_entry(vma->vm_rb.rb_left, in vma_compute_subtree_gap()
335 struct vm_area_struct, vm_rb)->rb_subtree_gap; in vma_compute_subtree_gap()
339 if (vma->vm_rb.rb_right) { in vma_compute_subtree_gap()
340 subtree_gap = rb_entry(vma->vm_rb.rb_right, in vma_compute_subtree_gap()
341 struct vm_area_struct, vm_rb)->rb_subtree_gap; in vma_compute_subtree_gap()
357 vma = rb_entry(nd, struct vm_area_struct, vm_rb); in browse_rb()
402 vma = rb_entry(nd, struct vm_area_struct, vm_rb); in validate_mm_rb()
454 struct vm_area_struct, vm_rb, in RB_DECLARE_CALLBACKS_MAX() argument
475 vma_gap_callbacks_propagate(&vma->vm_rb, NULL);
486 rb_insert_augmented(&vma->vm_rb, root, &vma_gap_callbacks); in vma_rb_insert()
498 rb_erase_augmented(&vma->vm_rb, root, &vma_gap_callbacks); in __vma_rb_erase()
505 RB_CLEAR_NODE(&vma->vm_rb); in __vma_rb_erase()
577 vma_tmp = rb_entry(__rb_parent, struct vm_area_struct, vm_rb); in find_vma_links()
592 *pprev = rb_entry(rb_prev, struct vm_area_struct, vm_rb); in find_vma_links()
689 rb_link_node(&vma->vm_rb, rb_parent, rb_link); in __vma_link_rb()
2028 vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); in unmapped_area()
2035 if (gap_end >= low_limit && vma->vm_rb.rb_left) { in unmapped_area()
2037 rb_entry(vma->vm_rb.rb_left, in unmapped_area()
2038 struct vm_area_struct, vm_rb); in unmapped_area()
2055 if (vma->vm_rb.rb_right) { in unmapped_area()
2057 rb_entry(vma->vm_rb.rb_right, in unmapped_area()
2058 struct vm_area_struct, vm_rb); in unmapped_area()
2067 struct rb_node *prev = &vma->vm_rb; in unmapped_area()
2071 struct vm_area_struct, vm_rb); in unmapped_area()
2072 if (prev == vma->vm_rb.rb_left) { in unmapped_area()
2137 vma = rb_entry(mm->mm_rb.rb_node, struct vm_area_struct, vm_rb); in unmapped_area_topdown()
2144 if (gap_start <= high_limit && vma->vm_rb.rb_right) { in unmapped_area_topdown()
2146 rb_entry(vma->vm_rb.rb_right, in unmapped_area_topdown()
2147 struct vm_area_struct, vm_rb); in unmapped_area_topdown()
2164 if (vma->vm_rb.rb_left) { in unmapped_area_topdown()
2166 rb_entry(vma->vm_rb.rb_left, in unmapped_area_topdown()
2167 struct vm_area_struct, vm_rb); in unmapped_area_topdown()
2176 struct rb_node *prev = &vma->vm_rb; in unmapped_area_topdown()
2180 struct vm_area_struct, vm_rb); in unmapped_area_topdown()
2181 if (prev == vma->vm_rb.rb_right) { in unmapped_area_topdown()
2390 tmp = rb_entry(rb_node, struct vm_area_struct, vm_rb); in __find_vma()
2464 *pprev = rb_node ? rb_entry(rb_node, struct vm_area_struct, vm_rb) : NULL; in find_vma_prev()