Lines Matching refs:vm_area_struct

109 void reset_vma_resv_huge_pages(struct vm_area_struct *vma);
118 int copy_hugetlb_page_range(struct mm_struct *, struct mm_struct *, struct vm_area_struct *);
119 long follow_hugetlb_page(struct mm_struct *, struct vm_area_struct *,
120 struct page **, struct vm_area_struct **,
123 void unmap_hugepage_range(struct vm_area_struct *,
126 struct vm_area_struct *vma,
129 void __unmap_hugepage_range(struct mmu_gather *tlb, struct vm_area_struct *vma,
136 vm_fault_t hugetlb_fault(struct mm_struct *mm, struct vm_area_struct *vma,
140 struct vm_area_struct *dst_vma,
147 struct vm_area_struct *vma,
159 pte_t *huge_pmd_share(struct mm_struct *mm, struct vm_area_struct *vma,
169 pte_t *huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma,
173 int huge_pmd_unshare(struct mm_struct *mm, struct vm_area_struct *vma,
175 void adjust_range_if_pmd_sharing_possible(struct vm_area_struct *vma,
179 struct page *follow_huge_pd(struct vm_area_struct *vma,
182 struct page *follow_huge_pmd_pte(struct vm_area_struct *vma, unsigned long address,
191 unsigned long hugetlb_change_protection(struct vm_area_struct *vma,
195 void hugetlb_unshare_all_pmds(struct vm_area_struct *vma);
199 static inline void reset_vma_resv_huge_pages(struct vm_area_struct *vma) in reset_vma_resv_huge_pages()
215 struct vm_area_struct *vma, in huge_pmd_unshare()
222 struct vm_area_struct *vma, in adjust_range_if_pmd_sharing_possible()
228 struct vm_area_struct *vma, struct page **pages, in follow_hugetlb_page()
229 struct vm_area_struct **vmas, unsigned long *position, in follow_hugetlb_page()
244 struct mm_struct *src, struct vm_area_struct *vma) in copy_hugetlb_page_range()
263 static inline struct page *follow_huge_pd(struct vm_area_struct *vma, in follow_huge_pd()
270 static inline struct page *follow_huge_pmd_pte(struct vm_area_struct *vma, in follow_huge_pmd_pte()
320 struct vm_area_struct *dst_vma, in hugetlb_mcopy_atomic_pte()
352 struct vm_area_struct *vma, unsigned long address, in hugetlb_change_protection()
359 struct vm_area_struct *vma, unsigned long start, in __unmap_hugepage_range_final()
366 struct vm_area_struct *vma, unsigned long start, in __unmap_hugepage_range()
373 struct vm_area_struct *vma, unsigned long address, in hugetlb_fault()
380 static inline void hugetlb_unshare_all_pmds(struct vm_area_struct *vma) { } in hugetlb_unshare_all_pmds()
519 struct page *alloc_huge_page(struct vm_area_struct *vma,
523 struct page *alloc_huge_page_vma(struct hstate *h, struct vm_area_struct *vma,
558 static inline struct hstate *hstate_vma(struct vm_area_struct *vma) in hstate_vma()
568 extern unsigned long vma_kernel_pagesize(struct vm_area_struct *vma);
570 extern unsigned long vma_mmu_pagesize(struct vm_area_struct *vma);
619 static inline pte_t arch_make_huge_pte(pte_t entry, struct vm_area_struct *vma, in arch_make_huge_pte()
761 static inline pte_t huge_ptep_modify_prot_start(struct vm_area_struct *vma, in huge_ptep_modify_prot_start()
770 static inline void huge_ptep_modify_prot_commit(struct vm_area_struct *vma, in huge_ptep_modify_prot_commit()
783 static inline struct page *alloc_huge_page(struct vm_area_struct *vma, in alloc_huge_page()
798 struct vm_area_struct *vma, in alloc_huge_page_vma()
819 static inline struct hstate *hstate_vma(struct vm_area_struct *vma) in hstate_vma()
839 static inline unsigned long vma_kernel_pagesize(struct vm_area_struct *vma) in vma_kernel_pagesize()
844 static inline unsigned long vma_mmu_pagesize(struct vm_area_struct *vma) in vma_mmu_pagesize()
956 bool want_pmd_share(struct vm_area_struct *vma, unsigned long addr);