Home
last modified time | relevance | path

Searched refs:i_pages (Results 1 – 25 of 36) sorted by relevance

12

/OK3568_Linux_fs/kernel/fs/nilfs2/
H A Dbtnode.c198 xa_lock_irq(&btnc->i_pages); in nilfs_btnode_prepare_change_key()
199 err = __xa_insert(&btnc->i_pages, newkey, opage, GFP_NOFS); in nilfs_btnode_prepare_change_key()
200 xa_unlock_irq(&btnc->i_pages); in nilfs_btnode_prepare_change_key()
255 xa_lock_irq(&btnc->i_pages); in nilfs_btnode_commit_change_key()
256 __xa_erase(&btnc->i_pages, oldkey); in nilfs_btnode_commit_change_key()
257 __xa_set_mark(&btnc->i_pages, newkey, PAGECACHE_TAG_DIRTY); in nilfs_btnode_commit_change_key()
258 xa_unlock_irq(&btnc->i_pages); in nilfs_btnode_commit_change_key()
286 xa_erase_irq(&btnc->i_pages, newkey); in nilfs_btnode_abort_change_key()
H A Dpage.c324 xa_lock_irq(&smap->i_pages); in nilfs_copy_back_pages()
325 p = __xa_erase(&smap->i_pages, offset); in nilfs_copy_back_pages()
328 xa_unlock_irq(&smap->i_pages); in nilfs_copy_back_pages()
330 xa_lock_irq(&dmap->i_pages); in nilfs_copy_back_pages()
331 p = __xa_store(&dmap->i_pages, offset, page, GFP_NOFS); in nilfs_copy_back_pages()
340 __xa_set_mark(&dmap->i_pages, offset, in nilfs_copy_back_pages()
343 xa_unlock_irq(&dmap->i_pages); in nilfs_copy_back_pages()
463 xa_lock_irq(&mapping->i_pages); in __nilfs_clear_page_dirty()
465 __xa_clear_mark(&mapping->i_pages, page_index(page), in __nilfs_clear_page_dirty()
467 xa_unlock_irq(&mapping->i_pages); in __nilfs_clear_page_dirty()
[all …]
/OK3568_Linux_fs/kernel/mm/
H A Dtruncate.c37 XA_STATE(xas, &mapping->i_pages, index); in __clear_shadow_entry()
49 xa_lock_irq(&mapping->i_pages); in clear_shadow_entry()
51 xa_unlock_irq(&mapping->i_pages); in clear_shadow_entry()
80 xa_lock_irq(&mapping->i_pages); in truncate_exceptional_pvec_entries()
103 xa_unlock_irq(&mapping->i_pages); in truncate_exceptional_pvec_entries()
516 xa_lock_irq(&mapping->i_pages); in truncate_inode_pages_final()
517 xa_unlock_irq(&mapping->i_pages); in truncate_inode_pages_final()
665 xa_lock_irqsave(&mapping->i_pages, flags); in invalidate_complete_page2()
671 xa_unlock_irqrestore(&mapping->i_pages, flags); in invalidate_complete_page2()
679 xa_unlock_irqrestore(&mapping->i_pages, flags); in invalidate_complete_page2()
H A Dswap_state.c134 XA_STATE_ORDER(xas, &address_space->i_pages, idx, compound_order(page)); in add_to_swap_cache()
190 XA_STATE(xas, &address_space->i_pages, idx); in __delete_from_swap_cache()
278 xa_lock_irq(&address_space->i_pages); in delete_from_swap_cache()
280 xa_unlock_irq(&address_space->i_pages); in delete_from_swap_cache()
296 XA_STATE(xas, &address_space->i_pages, curr); in clear_shadow_from_swap_cache()
298 xa_lock_irq(&address_space->i_pages); in clear_shadow_from_swap_cache()
306 xa_unlock_irq(&address_space->i_pages); in clear_shadow_from_swap_cache()
723 xa_init_flags(&space->i_pages, XA_FLAGS_LOCK_IRQ); in init_swap_address_space()
H A Dworkingset.c537 mapping = container_of(node->array, struct address_space, i_pages); in shadow_lru_isolate()
540 if (!xa_trylock(&mapping->i_pages)) { in shadow_lru_isolate()
565 xa_unlock_irq(&mapping->i_pages); in shadow_lru_isolate()
H A Dfilemap.c128 XA_STATE(xas, &mapping->i_pages, page->index); in page_cache_delete()
278 xa_lock_irqsave(&mapping->i_pages, flags); in delete_from_page_cache()
280 xa_unlock_irqrestore(&mapping->i_pages, flags); in delete_from_page_cache()
303 XA_STATE(xas, &mapping->i_pages, pvec->pages[0]->index); in page_cache_delete_batch()
357 xa_lock_irqsave(&mapping->i_pages, flags); in delete_from_page_cache_batch()
364 xa_unlock_irqrestore(&mapping->i_pages, flags); in delete_from_page_cache_batch()
482 XA_STATE(xas, &mapping->i_pages, start_byte >> PAGE_SHIFT); in filemap_range_has_page()
800 XA_STATE(xas, &mapping->i_pages, offset); in replace_page_cache_page()
840 XA_STATE(xas, &mapping->i_pages, offset); in __add_to_page_cache_locked()
1632 XA_STATE(xas, &mapping->i_pages, index); in page_cache_next_miss()
[all …]
H A Dpage-writeback.c2136 XA_STATE(xas, &mapping->i_pages, start); in tag_pages_for_writeback()
2503 xa_lock_irqsave(&mapping->i_pages, flags); in __set_page_dirty_nobuffers()
2507 __xa_set_mark(&mapping->i_pages, page_index(page), in __set_page_dirty_nobuffers()
2509 xa_unlock_irqrestore(&mapping->i_pages, flags); in __set_page_dirty_nobuffers()
2754 xa_lock_irqsave(&mapping->i_pages, flags); in test_clear_page_writeback()
2757 __xa_clear_mark(&mapping->i_pages, page_index(page), in test_clear_page_writeback()
2771 xa_unlock_irqrestore(&mapping->i_pages, flags); in test_clear_page_writeback()
2791 XA_STATE(xas, &mapping->i_pages, page_index(page)); in __test_set_page_writeback()
H A Dshmem.c424 XA_STATE(xas, &mapping->i_pages, index); in shmem_replace_entry()
446 return xa_load(&mapping->i_pages, index) == swp_to_radix_entry(swap); in shmem_confirm_swap()
684 XA_STATE_ORDER(xas, &mapping->i_pages, index, compound_order(page)); in shmem_add_to_page_cache()
759 xa_lock_irq(&mapping->i_pages); in shmem_delete_from_page_cache()
765 xa_unlock_irq(&mapping->i_pages); in shmem_delete_from_page_cache()
778 old = xa_cmpxchg_irq(&mapping->i_pages, index, radswap, NULL, 0); in shmem_free_swap()
795 XA_STATE(xas, &mapping->i_pages, start); in shmem_partial_swap_usage()
1204 XA_STATE(xas, &mapping->i_pages, start); in shmem_find_swap_entries()
1556 if (xa_find(&mapping->i_pages, &hindex, hindex + HPAGE_PMD_NR - 1, in shmem_alloc_hugepage()
1674 xa_lock_irq(&swap_mapping->i_pages); in shmem_replace_page()
[all …]
H A Dhuge_memory.c2477 xa_lock(&swap_cache->i_pages); in __split_huge_page()
2490 __xa_store(&head->mapping->i_pages, head[i].index, in __split_huge_page()
2493 __xa_store(&swap_cache->i_pages, offset + i, in __split_huge_page()
2507 xa_unlock(&swap_cache->i_pages); in __split_huge_page()
2514 xa_unlock(&head->mapping->i_pages); in __split_huge_page()
2732 XA_STATE(xas, &mapping->i_pages, page_index(head)); in split_huge_page_to_list()
2738 xa_lock(&mapping->i_pages); in split_huge_page_to_list()
2764 xa_unlock(&mapping->i_pages); in split_huge_page_to_list()
H A Dmemfd.c76 XA_STATE(xas, &mapping->i_pages, 0); in memfd_wait_for_pins()
H A Dreadahead.c213 struct page *page = xa_load(&mapping->i_pages, index + i); in page_cache_ra_unbounded()
H A Dvmscan.c896 xa_lock_irqsave(&mapping->i_pages, flags); in __remove_mapping()
937 xa_unlock_irqrestore(&mapping->i_pages, flags); in __remove_mapping()
963 xa_unlock_irqrestore(&mapping->i_pages, flags); in __remove_mapping()
972 xa_unlock_irqrestore(&mapping->i_pages, flags); in __remove_mapping()
/OK3568_Linux_fs/kernel/arch/nios2/include/asm/
H A Dcacheflush.h49 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages)
50 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages)
/OK3568_Linux_fs/kernel/arch/nds32/include/asm/
H A Dcacheflush.h43 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&(mapping)->i_pages)
44 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&(mapping)->i_pages)
/OK3568_Linux_fs/kernel/arch/csky/abiv1/inc/abi/
H A Dcacheflush.h21 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages)
22 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages)
/OK3568_Linux_fs/kernel/arch/parisc/include/asm/
H A Dcacheflush.h58 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages)
59 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages)
/OK3568_Linux_fs/kernel/include/linux/
H A Dbacking-dev.h285 !lockdep_is_held(&inode->i_mapping->i_pages.xa_lock) && in inode_to_wb()
318 xa_lock_irqsave(&inode->i_mapping->i_pages, cookie->flags); in unlocked_inode_to_wb_begin()
336 xa_unlock_irqrestore(&inode->i_mapping->i_pages, cookie->flags); in unlocked_inode_to_wb_end()
H A Dpagemap.h885 page = xa_load(&rac->mapping->i_pages, rac->_index); in readahead_page()
896 XA_STATE(xas, &rac->mapping->i_pages, 0); in __readahead_batch()
/OK3568_Linux_fs/kernel/arch/arm/include/asm/
H A Dcacheflush.h318 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages)
319 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages)
/OK3568_Linux_fs/kernel/fs/
H A Ddax.c425 xas.xa = &mapping->i_pages; in dax_lock_page()
450 XA_STATE(xas, &mapping->i_pages, page->index); in dax_unlock_page()
599 XA_STATE(xas, &mapping->i_pages, start_idx); in dax_layout_busy_page_range()
662 XA_STATE(xas, &mapping->i_pages, index); in __dax_invalidate_entry()
976 XA_STATE(xas, &mapping->i_pages, wbc->range_start >> PAGE_SHIFT); in dax_writeback_mapping_range()
1270 XA_STATE(xas, &mapping->i_pages, vmf->pgoff); in dax_iomap_pte_fault()
1492 XA_STATE_ORDER(xas, &mapping->i_pages, vmf->pgoff, PMD_ORDER); in dax_iomap_pmd_fault()
1696 XA_STATE_ORDER(xas, &mapping->i_pages, vmf->pgoff, order); in dax_insert_pfn_mkwrite()
H A Dfs-writeback.c361 XA_STATE(xas, &mapping->i_pages, 0); in inode_switch_wbs_work_fn()
389 xa_lock_irq(&mapping->i_pages); in inode_switch_wbs_work_fn()
453 xa_unlock_irq(&mapping->i_pages); in inode_switch_wbs_work_fn()
H A Dinode.c371 xa_init_flags(&mapping->i_pages, XA_FLAGS_LOCK_IRQ | XA_FLAGS_ACCOUNT); in __address_space_init_once()
530 xa_lock_irq(&inode->i_data.i_pages); in clear_inode()
533 xa_unlock_irq(&inode->i_data.i_pages); in clear_inode()
H A Dbuffer.c605 xa_lock_irqsave(&mapping->i_pages, flags); in __set_page_dirty()
609 __xa_set_mark(&mapping->i_pages, page_index(page), in __set_page_dirty()
612 xa_unlock_irqrestore(&mapping->i_pages, flags); in __set_page_dirty()
/OK3568_Linux_fs/kernel/fs/gfs2/
H A Dglops.c548 xa_lock_irq(&inode->i_data.i_pages); in inode_go_dump()
550 xa_unlock_irq(&inode->i_data.i_pages); in inode_go_dump()
/OK3568_Linux_fs/kernel/Documentation/vm/
H A Dpage_migration.rst110 5. The i_pages lock is taken. This will cause all processes trying
128 11. The i_pages lock is dropped. With that lookups in the mapping

12