Lines Matching refs:src_mm

737 copy_nonpresent_pte(struct mm_struct *dst_mm, struct mm_struct *src_mm,  in copy_nonpresent_pte()  argument
755 &src_mm->mmlist); in copy_nonpresent_pte()
776 set_pte_at(src_mm, addr, src_pte, pte); in copy_nonpresent_pte()
807 set_pte_at(src_mm, addr, src_pte, pte); in copy_nonpresent_pte()
841 struct mm_struct *src_mm = src_vma->vm_mm; in copy_present_page() local
860 if (likely(!atomic_read(&src_mm->has_pinned))) in copy_present_page()
910 struct mm_struct *src_mm = src_vma->vm_mm; in copy_present_pte() local
934 ptep_set_wrprotect(src_mm, addr, src_pte); in copy_present_pte()
954 page_copy_prealloc(struct mm_struct *src_mm, struct vm_area_struct *vma, in page_copy_prealloc() argument
963 if (mem_cgroup_charge(new_page, src_mm, GFP_KERNEL)) { in page_copy_prealloc()
978 struct mm_struct *src_mm = src_vma->vm_mm; in copy_pte_range() local
997 src_ptl = pte_lockptr(src_mm, src_pmd); in copy_pte_range()
1019 entry.val = copy_nonpresent_pte(dst_mm, src_mm, in copy_pte_range()
1065 prealloc = page_copy_prealloc(src_mm, src_vma, addr); in copy_pte_range()
1085 struct mm_struct *src_mm = src_vma->vm_mm; in copy_pmd_range() local
1099 err = copy_huge_pmd(dst_mm, src_mm, dst_pmd, src_pmd, in copy_pmd_range()
1122 struct mm_struct *src_mm = src_vma->vm_mm; in copy_pud_range() local
1136 err = copy_huge_pud(dst_mm, src_mm, in copy_pud_range()
1185 struct mm_struct *src_mm = src_vma->vm_mm; in copy_page_range() local
1201 return copy_hugetlb_page_range(dst_mm, src_mm, src_vma); in copy_page_range()
1223 0, src_vma, src_mm, addr, end); in copy_page_range()
1232 mmap_assert_write_locked(src_mm); in copy_page_range()
1233 raw_write_seqcount_begin(&src_mm->write_protect_seq); in copy_page_range()
1238 src_pgd = pgd_offset(src_mm, addr); in copy_page_range()
1251 raw_write_seqcount_end(&src_mm->write_protect_seq); in copy_page_range()