| /OK3568_Linux_fs/external/xserver/miext/sync/ |
| H A D | misync.c | 116 SyncTriggerList *ptl, *pNext; in miSyncDestroyFence() local 119 for (ptl = pFence->sync.pTriglist; ptl; ptl = pNext) { in miSyncDestroyFence() 120 (*ptl->pTrigger->CounterDestroyed) (ptl->pTrigger); in miSyncDestroyFence() 121 pNext = ptl->next; in miSyncDestroyFence() 122 free(ptl); /* destroy the trigger list as we go */ in miSyncDestroyFence() 134 SyncTriggerList *ptl, *pNext; in miSyncTriggerFence() local 139 for (ptl = pFence->sync.pTriglist; ptl; ptl = pNext) { in miSyncTriggerFence() 140 pNext = ptl->next; in miSyncTriggerFence() 141 if ((*ptl->pTrigger->CheckTrigger) (ptl->pTrigger, 0)) in miSyncTriggerFence() 142 (*ptl->pTrigger->TriggerFired) (ptl->pTrigger); in miSyncTriggerFence()
|
| /OK3568_Linux_fs/kernel/mm/ |
| H A D | page_vma_mapped.c | 50 pvmw->ptl = pte_lockptr(pvmw->vma->vm_mm, pvmw->pmd); in map_pte() 51 spin_lock(pvmw->ptl); in map_pte() 173 pvmw->ptl = huge_pte_lockptr(page_hstate(page), mm, pvmw->pte); in page_vma_mapped_walk() 174 spin_lock(pvmw->ptl); in page_vma_mapped_walk() 218 pvmw->ptl = pmd_lock(mm, pvmw->pmd); in page_vma_mapped_walk() 240 spin_unlock(pvmw->ptl); in page_vma_mapped_walk() 241 pvmw->ptl = NULL; in page_vma_mapped_walk() 250 spinlock_t *ptl = pmd_lock(mm, pvmw->pmd); in page_vma_mapped_walk() local 252 spin_unlock(ptl); in page_vma_mapped_walk() 269 if (pvmw->ptl) { in page_vma_mapped_walk() [all …]
|
| H A D | huge_memory.c | 613 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in __do_huge_pmd_anonymous_page() 627 spin_unlock(vmf->ptl); in __do_huge_pmd_anonymous_page() 643 spin_unlock(vmf->ptl); in __do_huge_pmd_anonymous_page() 650 spin_unlock(vmf->ptl); in __do_huge_pmd_anonymous_page() 739 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_anonymous_page() 744 spin_unlock(vmf->ptl); in do_huge_pmd_anonymous_page() 747 spin_unlock(vmf->ptl); in do_huge_pmd_anonymous_page() 754 spin_unlock(vmf->ptl); in do_huge_pmd_anonymous_page() 757 spin_unlock(vmf->ptl); in do_huge_pmd_anonymous_page() 778 spinlock_t *ptl; in insert_pfn_pmd() local [all …]
|
| H A D | memory.c | 461 spinlock_t *ptl; in __pte_alloc() local 481 ptl = pmd_lock(mm, pmd); in __pte_alloc() 487 spin_unlock(ptl); in __pte_alloc() 1276 spinlock_t *ptl; in zap_pte_range() local 1284 start_pte = pte_offset_map_lock(mm, pmd, addr, &ptl); in zap_pte_range() 1385 pte_unmap_unlock(start_pte, ptl); in zap_pte_range() 1426 spinlock_t *ptl = pmd_lock(tlb->mm, pmd); in zap_pmd_range() local 1432 spin_unlock(ptl); in zap_pmd_range() 1693 spinlock_t **ptl) in __get_locked_pte() argument 1699 return pte_alloc_map_lock(mm, pmd, addr, ptl); in __get_locked_pte() [all …]
|
| H A D | madvise.c | 203 spinlock_t *ptl; in swapin_walk_pmd_entry() local 205 orig_pte = pte_offset_map_lock(vma->vm_mm, pmd, start, &ptl); in swapin_walk_pmd_entry() 207 pte_unmap_unlock(orig_pte, ptl); in swapin_walk_pmd_entry() 321 spinlock_t *ptl; in madvise_cold_or_pageout_pte_range() local 336 ptl = pmd_trans_huge_lock(pmd, vma); in madvise_cold_or_pageout_pte_range() 337 if (!ptl) in madvise_cold_or_pageout_pte_range() 363 spin_unlock(ptl); in madvise_cold_or_pageout_pte_range() 393 spin_unlock(ptl); in madvise_cold_or_pageout_pte_range() 404 orig_pte = pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in madvise_cold_or_pageout_pte_range() 434 pte_unmap_unlock(orig_pte, ptl); in madvise_cold_or_pageout_pte_range() [all …]
|
| H A D | mincore.c | 102 spinlock_t *ptl; in mincore_pte_range() local 108 ptl = pmd_trans_huge_lock(pmd, vma); in mincore_pte_range() 109 if (ptl) { in mincore_pte_range() 111 spin_unlock(ptl); in mincore_pte_range() 120 ptep = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in mincore_pte_range() 150 pte_unmap_unlock(ptep - 1, ptl); in mincore_pte_range()
|
| H A D | hmm.c | 417 spinlock_t *ptl = pud_trans_huge_lock(pudp, walk->vma); in hmm_vma_walk_pud() local 419 if (!ptl) in hmm_vma_walk_pud() 427 spin_unlock(ptl); in hmm_vma_walk_pud() 438 spin_unlock(ptl); in hmm_vma_walk_pud() 450 spin_unlock(ptl); in hmm_vma_walk_pud() 464 spin_unlock(ptl); in hmm_vma_walk_pud() 483 spinlock_t *ptl; in hmm_vma_walk_hugetlb_entry() local 486 ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte); in hmm_vma_walk_hugetlb_entry() 496 spin_unlock(ptl); in hmm_vma_walk_hugetlb_entry() 504 spin_unlock(ptl); in hmm_vma_walk_hugetlb_entry()
|
| H A D | migrate.c | 291 spinlock_t *ptl) in __migration_entry_wait() argument 297 spin_lock(ptl); in __migration_entry_wait() 316 pte_unmap_unlock(ptep, ptl); in __migration_entry_wait() 321 pte_unmap_unlock(ptep, ptl); in __migration_entry_wait() 327 spinlock_t *ptl = pte_lockptr(mm, pmd); in migration_entry_wait() local 329 __migration_entry_wait(mm, ptep, ptl); in migration_entry_wait() 335 spinlock_t *ptl = huge_pte_lockptr(hstate_vma(vma), mm, pte); in migration_entry_wait_huge() local 336 __migration_entry_wait(mm, pte, ptl); in migration_entry_wait_huge() 342 spinlock_t *ptl; in pmd_migration_entry_wait() local 345 ptl = pmd_lock(mm, pmd); in pmd_migration_entry_wait() [all …]
|
| H A D | hugetlb.c | 3912 spinlock_t *ptl; in __unmap_hugepage_range() local 3943 ptl = huge_pte_lock(h, mm, ptep); in __unmap_hugepage_range() 3945 spin_unlock(ptl); in __unmap_hugepage_range() 3953 spin_unlock(ptl); in __unmap_hugepage_range() 3963 spin_unlock(ptl); in __unmap_hugepage_range() 3975 spin_unlock(ptl); in __unmap_hugepage_range() 3994 spin_unlock(ptl); in __unmap_hugepage_range() 4129 struct page *pagecache_page, spinlock_t *ptl) in hugetlb_cow() argument 4170 spin_unlock(ptl); in hugetlb_cow() 4206 spin_lock(ptl); in hugetlb_cow() [all …]
|
| H A D | gup.c | 431 spinlock_t *ptl; in follow_page_pte() local 455 ptep = pte_offset_map_lock(mm, pmd, address, &ptl); in follow_page_pte() 471 pte_unmap_unlock(ptep, ptl); in follow_page_pte() 478 pte_unmap_unlock(ptep, ptl); in follow_page_pte() 512 pte_unmap_unlock(ptep, ptl); in follow_page_pte() 578 pte_unmap_unlock(ptep, ptl); in follow_page_pte() 581 pte_unmap_unlock(ptep, ptl); in follow_page_pte() 593 spinlock_t *ptl; in follow_pmd_mask() local 637 ptl = pmd_lock(mm, pmd); in follow_pmd_mask() 639 spin_unlock(ptl); in follow_pmd_mask() [all …]
|
| H A D | userfaultfd.c | 67 spinlock_t *ptl; in mfill_atomic_install_pte() local 83 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); in mfill_atomic_install_pte() 119 pte_unmap_unlock(dst_pte, ptl); in mfill_atomic_install_pte() 189 spinlock_t *ptl; in mfill_zeropage_pte() local 196 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); in mfill_zeropage_pte() 214 pte_unmap_unlock(dst_pte, ptl); in mfill_zeropage_pte()
|
| /OK3568_Linux_fs/kernel/arch/arm/lib/ |
| H A D | uaccess_with_memcpy.c | 31 spinlock_t *ptl; in pin_page_for_write() local 60 ptl = ¤t->mm->page_table_lock; in pin_page_for_write() 61 spin_lock(ptl); in pin_page_for_write() 64 spin_unlock(ptl); in pin_page_for_write() 69 *ptlp = ptl; in pin_page_for_write() 76 pte = pte_offset_map_lock(current->mm, pmd, addr, &ptl); in pin_page_for_write() 79 pte_unmap_unlock(pte, ptl); in pin_page_for_write() 84 *ptlp = ptl; in pin_page_for_write() 107 spinlock_t *ptl; in __copy_to_user_memcpy() local 110 while (!pin_page_for_write(to, &pte, &ptl)) { in __copy_to_user_memcpy() [all …]
|
| /OK3568_Linux_fs/kernel/arch/arm/mm/ |
| H A D | fault-armv.c | 70 static inline void do_pte_lock(spinlock_t *ptl) in do_pte_lock() argument 76 spin_lock_nested(ptl, SINGLE_DEPTH_NESTING); in do_pte_lock() 79 static inline void do_pte_unlock(spinlock_t *ptl) in do_pte_unlock() argument 81 spin_unlock(ptl); in do_pte_unlock() 84 static inline void do_pte_lock(spinlock_t *ptl) {} in do_pte_lock() argument 85 static inline void do_pte_unlock(spinlock_t *ptl) {} in do_pte_unlock() argument 91 spinlock_t *ptl; in adjust_pte() local 120 ptl = pte_lockptr(vma->vm_mm, pmd); in adjust_pte() 122 do_pte_lock(ptl); in adjust_pte() 126 do_pte_unlock(ptl); in adjust_pte()
|
| /OK3568_Linux_fs/kernel/mm/damon/ |
| H A D | vaddr.c | 374 spinlock_t *ptl; in damon_mkold_pmd_entry() local 377 ptl = pmd_lock(walk->mm, pmd); in damon_mkold_pmd_entry() 379 spin_unlock(ptl); in damon_mkold_pmd_entry() 385 spin_unlock(ptl); in damon_mkold_pmd_entry() 388 spin_unlock(ptl); in damon_mkold_pmd_entry() 393 pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in damon_mkold_pmd_entry() 398 pte_unmap_unlock(pte, ptl); in damon_mkold_pmd_entry() 439 spinlock_t *ptl; in damon_mkold_hugetlb_entry() local 442 ptl = huge_pte_lock(h, walk->mm, pte); in damon_mkold_hugetlb_entry() 450 spin_unlock(ptl); in damon_mkold_hugetlb_entry() [all …]
|
| /OK3568_Linux_fs/kernel/arch/powerpc/mm/ |
| H A D | hugetlbpage.c | 46 unsigned int pshift, spinlock_t *ptl) in __hugepte_alloc() argument 81 spin_lock(ptl); in __hugepte_alloc() 101 spin_unlock(ptl); in __hugepte_alloc() 119 spinlock_t *ptl; in huge_pte_alloc() local 133 ptl = &mm->page_table_lock; in huge_pte_alloc() 143 ptl = pud_lockptr(mm, pu); in huge_pte_alloc() 154 ptl = pmd_lockptr(mm, pm); in huge_pte_alloc() 161 ptl = &mm->page_table_lock; in huge_pte_alloc() 169 ptl = pud_lockptr(mm, pu); in huge_pte_alloc() 176 ptl = pmd_lockptr(mm, pm); in huge_pte_alloc() [all …]
|
| /OK3568_Linux_fs/kernel/arch/s390/mm/ |
| H A D | pgtable.c | 776 spinlock_t *ptl; in set_guest_storage_key() local 785 ptl = pmd_lock(mm, pmdp); in set_guest_storage_key() 787 spin_unlock(ptl); in set_guest_storage_key() 799 spin_unlock(ptl); in set_guest_storage_key() 802 spin_unlock(ptl); in set_guest_storage_key() 804 ptep = pte_alloc_map_lock(mm, pmdp, addr, &ptl); in set_guest_storage_key() 832 pte_unmap_unlock(ptep, ptl); in set_guest_storage_key() 877 spinlock_t *ptl; in reset_guest_reference_bit() local 888 ptl = pmd_lock(mm, pmdp); in reset_guest_reference_bit() 890 spin_unlock(ptl); in reset_guest_reference_bit() [all …]
|
| H A D | gmap.c | 544 spinlock_t *ptl; in __gmap_link() local 600 ptl = pmd_lock(mm, pmd); in __gmap_link() 622 spin_unlock(ptl); in __gmap_link() 677 spinlock_t *ptl; in __gmap_zap() local 686 ptep = get_locked_pte(gmap->mm, vmaddr, &ptl); in __gmap_zap() 689 pte_unmap_unlock(ptep, ptl); in __gmap_zap() 851 spinlock_t **ptl) in gmap_pte_op_walk() argument 860 return pte_alloc_map_lock(gmap->mm, (pmd_t *) table, gaddr, ptl); in gmap_pte_op_walk() 896 static void gmap_pte_op_end(spinlock_t *ptl) in gmap_pte_op_end() argument 898 if (ptl) in gmap_pte_op_end() [all …]
|
| /OK3568_Linux_fs/external/mpp/mpp/codec/enc/h265/ |
| H A D | h265e_header_gen.c | 232 void codeProfileTier(H265eStream *s, ProfileTierLevel* ptl) in codeProfileTier() argument 235 h265e_stream_write_with_log(s, ptl->m_profileSpace, 2, "profile_space[]"); in codeProfileTier() 236 h265e_stream_write1_with_log(s, ptl->m_tierFlag, "tier_flag[]"); in codeProfileTier() 237 h265e_stream_write_with_log(s, ptl->m_profileIdc, 5, "profile_idc[]"); in codeProfileTier() 239 …h265e_stream_write1_with_log(s, ptl->m_profileCompatibilityFlag[j], "profile_compatibility_flag[][… in codeProfileTier() 242 …h265e_stream_write1_with_log(s, ptl->m_progressiveSourceFlag, "general_progressive_source_flag"); in codeProfileTier() 243 … h265e_stream_write1_with_log(s, ptl->m_interlacedSourceFlag, "general_interlaced_source_flag"); in codeProfileTier() 244 …h265e_stream_write1_with_log(s, ptl->m_nonPackedConstraintFlag, "general_non_packed_constraint_fla… in codeProfileTier() 245 …h265e_stream_write1_with_log(s, ptl->m_frameOnlyConstraintFlag, "general_frame_only_constraint_fla… in codeProfileTier() 252 void codePTL(H265eStream *s, H265ePTL* ptl, RK_U32 profilePresentFlag, int maxNumSubLayersMinus1) in codePTL() argument [all …]
|
| /OK3568_Linux_fs/kernel/fs/proc/ |
| H A D | task_mmu.c | 652 spinlock_t *ptl; in smaps_pte_range() local 654 ptl = pmd_trans_huge_lock(pmd, vma); in smaps_pte_range() 655 if (ptl) { in smaps_pte_range() 657 spin_unlock(ptl); in smaps_pte_range() 668 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in smaps_pte_range() 671 pte_unmap_unlock(pte - 1, ptl); in smaps_pte_range() 1207 spinlock_t *ptl; in clear_refs_pte_range() local 1210 ptl = pmd_trans_huge_lock(pmd, vma); in clear_refs_pte_range() 1211 if (ptl) { in clear_refs_pte_range() 1227 spin_unlock(ptl); in clear_refs_pte_range() [all …]
|
| /OK3568_Linux_fs/kernel/Documentation/vm/ |
| H A D | split_page_table_lock.rst | 63 This field shares storage with page->ptl. 80 page->ptl 83 page->ptl is used to access split page table lock, where 'page' is struct 92 - if size of spinlock_t is bigger then size of long, we use page->ptl as 100 Please, never access page->ptl directly -- use appropriate helper.
|
| /OK3568_Linux_fs/external/mpp/mpp/codec/dec/h265/ |
| H A D | h265d_ps.c | 266 static RK_S32 decode_profile_tier_level(HEVCContext *s, PTLCommon *ptl) in decode_profile_tier_level() argument 272 READ_BITS(gb, 2, &ptl->profile_space); in decode_profile_tier_level() 273 READ_ONEBIT(gb, &ptl->tier_flag); in decode_profile_tier_level() 274 READ_BITS(gb, 5, &ptl->profile_idc); in decode_profile_tier_level() 276 if (ptl->profile_idc == MPP_PROFILE_HEVC_MAIN) in decode_profile_tier_level() 278 else if (ptl->profile_idc == MPP_PROFILE_HEVC_MAIN_10) in decode_profile_tier_level() 280 else if (ptl->profile_idc == MPP_PROFILE_HEVC_MAIN_STILL_PICTURE) in decode_profile_tier_level() 283 mpp_log("Unknown HEVC profile: %d\n", ptl->profile_idc); in decode_profile_tier_level() 286 READ_ONEBIT(gb, & ptl->profile_compatibility_flag[i]); in decode_profile_tier_level() 287 READ_ONEBIT(gb, &ptl->progressive_source_flag); in decode_profile_tier_level() [all …]
|
| /OK3568_Linux_fs/kernel/arch/x86/kernel/ |
| H A D | ldt.c | 292 spinlock_t *ptl; in map_ldt_struct() local 326 ptep = get_locked_pte(mm, va, &ptl); in map_ldt_struct() 339 pte_unmap_unlock(ptep, ptl); in map_ldt_struct() 365 spinlock_t *ptl; in unmap_ldt_struct() local 369 ptep = get_locked_pte(mm, va, &ptl); in unmap_ldt_struct() 371 pte_unmap_unlock(ptep, ptl); in unmap_ldt_struct()
|
| /OK3568_Linux_fs/kernel/arch/m68k/kernel/ |
| H A D | sys_m68k.c | 474 spinlock_t *ptl; in sys_atomic_cmpxchg_32() local 490 pte = pte_offset_map_lock(mm, pmd, (unsigned long)mem, &ptl); in sys_atomic_cmpxchg_32() 493 pte_unmap_unlock(pte, ptl); in sys_atomic_cmpxchg_32() 505 pte_unmap_unlock(pte, ptl); in sys_atomic_cmpxchg_32()
|
| /OK3568_Linux_fs/kernel/include/linux/ |
| H A D | rmap.h | 233 spinlock_t *ptl; member 242 if (pvmw->ptl) in page_vma_mapped_walk_done() 243 spin_unlock(pvmw->ptl); in page_vma_mapped_walk_done()
|
| /OK3568_Linux_fs/kernel/arch/x86/xen/ |
| H A D | mmu_pv.c | 647 spinlock_t *ptl = NULL; in xen_pte_lock() local 650 ptl = ptlock_ptr(page); in xen_pte_lock() 651 spin_lock_nest_lock(ptl, &mm->page_table_lock); in xen_pte_lock() 654 return ptl; in xen_pte_lock() 659 spinlock_t *ptl = v; in xen_pte_unlock() local 660 spin_unlock(ptl); in xen_pte_unlock() 682 spinlock_t *ptl; in xen_pin_page() local 704 ptl = NULL; in xen_pin_page() 706 ptl = xen_pte_lock(page, mm); in xen_pin_page() 712 if (ptl) { in xen_pin_page() [all …]
|