| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/ |
| H A D | mali_kbase_mem_migrate.c | 167 static int kbasep_migrate_page_pt_mapped(struct page *old_page, struct page *new_page) in kbasep_migrate_page_pt_mapped() argument 177 new_dma_addr = dma_map_page(kbdev->dev, new_page, 0, PAGE_SIZE, DMA_BIDIRECTIONAL); in kbasep_migrate_page_pt_mapped() 188 as_tagged(page_to_phys(old_page)), as_tagged(page_to_phys(new_page)), old_dma_addr, in kbasep_migrate_page_pt_mapped() 198 page_md = kbase_page_private(new_page); in kbasep_migrate_page_pt_mapped() 200 __SetPageMovable(new_page, &movable_ops); in kbasep_migrate_page_pt_mapped() 204 __SetPageMovable(new_page, kbdev->mem_migrate.inode->i_mapping); in kbasep_migrate_page_pt_mapped() 208 SetPagePrivate(new_page); in kbasep_migrate_page_pt_mapped() 209 get_page(new_page); in kbasep_migrate_page_pt_mapped() 233 static int kbasep_migrate_page_allocated_mapped(struct page *old_page, struct page *new_page) in kbasep_migrate_page_allocated_mapped() argument 241 new_dma_addr = dma_map_page(kctx->kbdev->dev, new_page, 0, PAGE_SIZE, DMA_BIDIRECTIONAL); in kbasep_migrate_page_allocated_mapped() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/mali400/mali/linux/ |
| H A D | mali_memory_cow.c | 40 struct page *new_page; in mali_mem_cow_alloc_page() local 53 new_page = node->page; in mali_mem_cow_alloc_page() 58 return new_page; in mali_mem_cow_alloc_page() 114 struct page *new_page; in mali_memory_cow_os_memory() local 148 new_page = mali_mem_cow_alloc_page(); in mali_memory_cow_os_memory() 150 if (NULL == new_page) { in mali_memory_cow_os_memory() 156 _mali_page_node_add_page(page_node, new_page); in mali_memory_cow_os_memory() 297 struct page *new_page; in mali_memory_cow_modify_range() local 323 new_page = mali_mem_cow_alloc_page(); in mali_memory_cow_modify_range() 325 if (NULL == new_page) { in mali_memory_cow_modify_range() [all …]
|
| H A D | mali_memory_os_alloc.c | 155 struct page *new_page; in mali_mem_os_alloc_pages() local 219 new_page = alloc_page(flags); in mali_mem_os_alloc_pages() 221 if (unlikely(NULL == new_page)) { in mali_mem_os_alloc_pages() 231 dma_addr = dma_map_page(&mali_platform_device->dev, new_page, in mali_mem_os_alloc_pages() 235 dma_addr = dma_map_page(&mali_platform_device->dev, new_page, in mali_mem_os_alloc_pages() 241 new_page, err)); in mali_mem_os_alloc_pages() 242 __free_page(new_page); in mali_mem_os_alloc_pages() 250 SetPagePrivate(new_page); in mali_mem_os_alloc_pages() 251 set_page_private(new_page, dma_addr); in mali_mem_os_alloc_pages() 256 dma_unmap_page(&mali_platform_device->dev, page_private(new_page), in mali_mem_os_alloc_pages() [all …]
|
| /OK3568_Linux_fs/kernel/mm/ |
| H A D | khugepaged.c | 1068 struct page *new_page; in collapse_huge_page() local 1087 new_page = khugepaged_alloc_page(hpage, gfp, node); in collapse_huge_page() 1088 if (!new_page) { in collapse_huge_page() 1093 if (unlikely(mem_cgroup_charge(new_page, mm, gfp))) { in collapse_huge_page() 1097 count_memcg_page_event(new_page, THP_COLLAPSE_ALLOC); in collapse_huge_page() 1190 __collapse_huge_page_copy(pte, new_page, vma, address, pte_ptl, in collapse_huge_page() 1193 __SetPageUptodate(new_page); in collapse_huge_page() 1196 _pmd = mk_huge_pmd(new_page, vma->vm_page_prot); in collapse_huge_page() 1208 page_add_new_anon_rmap(new_page, vma, address, true); in collapse_huge_page() 1209 lru_cache_add_inactive_or_unevictable(new_page, vma); in collapse_huge_page() [all …]
|
| H A D | migrate.c | 1533 struct page *new_page = NULL; in alloc_migration_target() local 1563 new_page = __alloc_pages_nodemask(gfp_mask, order, nid, mtc->nmask); in alloc_migration_target() 1565 if (new_page && PageTransHuge(new_page)) in alloc_migration_target() 1566 prep_transhuge_page(new_page); in alloc_migration_target() 1568 return new_page; in alloc_migration_target() 2126 struct page *new_page = NULL; in migrate_misplaced_transhuge_page() local 2130 new_page = alloc_pages_node(node, in migrate_misplaced_transhuge_page() 2133 if (!new_page) in migrate_misplaced_transhuge_page() 2135 prep_transhuge_page(new_page); in migrate_misplaced_transhuge_page() 2139 put_page(new_page); in migrate_misplaced_transhuge_page() [all …]
|
| H A D | memory.c | 842 struct page *new_page; in copy_present_page() local 876 new_page = *prealloc; in copy_present_page() 877 if (!new_page) in copy_present_page() 885 copy_user_highpage(new_page, page, addr, src_vma); in copy_present_page() 886 __SetPageUptodate(new_page); in copy_present_page() 887 page_add_new_anon_rmap(new_page, dst_vma, addr, false); in copy_present_page() 888 lru_cache_add_inactive_or_unevictable(new_page, dst_vma); in copy_present_page() 889 rss[mm_counter(new_page)]++; in copy_present_page() 892 pte = mk_pte(new_page, dst_vma->vm_page_prot); in copy_present_page() 957 struct page *new_page; in page_copy_prealloc() local [all …]
|
| H A D | ksm.c | 2574 struct page *new_page; in ksm_might_need_to_copy() local 2589 new_page = alloc_page_vma(GFP_HIGHUSER_MOVABLE, vma, address); in ksm_might_need_to_copy() 2590 if (new_page && mem_cgroup_charge(new_page, vma->vm_mm, GFP_KERNEL)) { in ksm_might_need_to_copy() 2591 put_page(new_page); in ksm_might_need_to_copy() 2592 new_page = NULL; in ksm_might_need_to_copy() 2594 if (new_page) { in ksm_might_need_to_copy() 2595 copy_user_highpage(new_page, page, address, vma); in ksm_might_need_to_copy() 2597 SetPageDirty(new_page); in ksm_might_need_to_copy() 2598 __SetPageUptodate(new_page); in ksm_might_need_to_copy() 2599 __SetPageLocked(new_page); in ksm_might_need_to_copy() [all …]
|
| H A D | hugetlb.c | 4133 struct page *old_page, *new_page; in hugetlb_cow() local 4171 new_page = alloc_huge_page(vma, haddr, outside_reserve); in hugetlb_cow() 4173 if (IS_ERR(new_page)) { in hugetlb_cow() 4218 ret = vmf_error(PTR_ERR(new_page)); in hugetlb_cow() 4231 copy_user_huge_page(new_page, old_page, address, vma, in hugetlb_cow() 4233 __SetPageUptodate(new_page); in hugetlb_cow() 4246 ClearPagePrivate(new_page); in hugetlb_cow() 4252 make_huge_pte(vma, new_page, 1)); in hugetlb_cow() 4254 hugepage_add_new_anon_rmap(new_page, vma, haddr); in hugetlb_cow() 4255 set_page_huge_active(new_page); in hugetlb_cow() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/mali400/ump/linux/ |
| H A D | ump_kernel_memory_backend_os.c | 133 struct page *new_page; in os_allocate() local 136 new_page = alloc_page(GFP_HIGHUSER | __GFP_ZERO | __GFP_REPEAT | __GFP_NOWARN); in os_allocate() 138 new_page = alloc_page(GFP_HIGHUSER | __GFP_ZERO | __GFP_REPEAT | __GFP_NOWARN | __GFP_COLD); in os_allocate() 140 if (NULL == new_page) { in os_allocate() 146 descriptor->block_array[pages_allocated].addr = page_to_phys(new_page); in os_allocate() 149 …descriptor->block_array[pages_allocated].addr = dma_map_page(NULL, new_page, 0, PAGE_SIZE, DMA_BID… in os_allocate()
|
| /OK3568_Linux_fs/kernel/kernel/events/ |
| H A D | uprobes.c | 155 struct page *old_page, struct page *new_page) in __replace_page() argument 169 if (new_page) { in __replace_page() 170 err = mem_cgroup_charge(new_page, vma->vm_mm, GFP_KERNEL); in __replace_page() 184 if (new_page) { in __replace_page() 185 get_page(new_page); in __replace_page() 186 page_add_new_anon_rmap(new_page, vma, addr, false); in __replace_page() 187 lru_cache_add_inactive_or_unevictable(new_page, vma); in __replace_page() 199 if (new_page) in __replace_page() 201 mk_pte(new_page, vma->vm_page_prot)); in __replace_page() 467 struct page *old_page, *new_page; in uprobe_write_opcode() local [all …]
|
| /OK3568_Linux_fs/kernel/arch/s390/mm/ |
| H A D | vmem.c | 172 void *new_page = vmemmap_alloc_block(PAGE_SIZE, NUMA_NO_NODE); in modify_pte_table() local 174 if (!new_page) in modify_pte_table() 176 pte_val(*pte) = __pa(new_page) | prot; in modify_pte_table() 248 void *new_page; in modify_pmd_table() local 257 new_page = vmemmap_alloc_block(PMD_SIZE, NUMA_NO_NODE); in modify_pmd_table() 258 if (new_page) { in modify_pmd_table() 259 pmd_val(*pmd) = __pa(new_page) | prot; in modify_pmd_table()
|
| /OK3568_Linux_fs/kernel/fs/f2fs/ |
| H A D | namei.c | 926 struct page *old_page, *new_page = NULL; in f2fs_rename() local 1001 &new_page); in f2fs_rename() 1003 if (IS_ERR(new_page)) in f2fs_rename() 1004 err = PTR_ERR(new_page); in f2fs_rename() 1016 f2fs_set_link(new_dir, new_entry, new_page, old_inode); in f2fs_rename() 1017 new_page = NULL; in f2fs_rename() 1097 f2fs_put_page(new_page, 0); in f2fs_rename() 1116 struct page *old_page, *new_page; in f2fs_cross_rename() local 1151 new_entry = f2fs_find_entry(new_dir, &new_dentry->d_name, &new_page); in f2fs_cross_rename() 1153 if (IS_ERR(new_page)) in f2fs_cross_rename() [all …]
|
| /OK3568_Linux_fs/kernel/fs/sysv/ |
| H A D | namei.c | 216 struct page * new_page; in sysv_rename() local 224 new_de = sysv_find_entry(new_dentry, &new_page); in sysv_rename() 227 sysv_set_link(new_de, new_page, old_inode); in sysv_rename()
|
| /OK3568_Linux_fs/kernel/fs/minix/ |
| H A D | namei.c | 211 struct page * new_page; in minix_rename() local 219 new_de = minix_find_entry(new_dentry, &new_page); in minix_rename() 222 minix_set_link(new_de, new_page, old_inode); in minix_rename()
|
| /OK3568_Linux_fs/kernel/fs/ufs/ |
| H A D | namei.c | 270 struct page *new_page; in ufs_rename() local 278 new_de = ufs_find_entry(new_dir, &new_dentry->d_name, &new_page); in ufs_rename() 281 ufs_set_link(new_dir, new_de, new_page, old_inode, 1); in ufs_rename()
|
| /OK3568_Linux_fs/u-boot/fs/ubifs/ |
| H A D | budget.c | 375 znodes = req->new_ino + (req->new_page << UBIFS_BLOCKS_PER_PAGE_SHIFT) + in calc_idx_growth() 392 if (req->new_page) in calc_data_growth() 438 ubifs_assert(req->new_page <= 1); in ubifs_budget_space() 525 ubifs_assert(req->new_page <= 1); in ubifs_release_budget()
|
| /OK3568_Linux_fs/kernel/fs/ubifs/ |
| H A D | budget.c | 367 znodes = req->new_ino + (req->new_page << UBIFS_BLOCKS_PER_PAGE_SHIFT) + in calc_idx_growth() 384 if (req->new_page) in calc_data_growth() 430 ubifs_assert(c, req->new_page <= 1); in ubifs_budget_space() 517 ubifs_assert(c, req->new_page <= 1); in ubifs_release_budget()
|
| H A D | file.c | 198 struct ubifs_budget_req req = { .recalculate = 1, .new_page = 1 }; in release_new_page_budget() 224 struct ubifs_budget_req req = { .new_page = 1 }; in write_begin_slow() 365 req.new_page = 1; in allocate_budget() 1513 struct ubifs_budget_req req = { .new_page = 1 }; in ubifs_vm_page_mkwrite()
|
| /OK3568_Linux_fs/kernel/fs/ext2/ |
| H A D | namei.c | 355 struct page *new_page; in ext2_rename() local 362 new_de = ext2_find_entry(new_dir, &new_dentry->d_name, &new_page); in ext2_rename() 367 ext2_set_link(new_dir, new_de, new_page, old_inode, 1); in ext2_rename()
|
| /OK3568_Linux_fs/kernel/drivers/net/ethernet/ti/ |
| H A D | cpsw.c | 346 struct page *new_page, *page = token; in cpsw_rx_handler() local 378 new_page = page; in cpsw_rx_handler() 387 new_page = page_pool_dev_alloc_pages(pool); in cpsw_rx_handler() 388 if (unlikely(!new_page)) { in cpsw_rx_handler() 389 new_page = page; in cpsw_rx_handler() 449 xmeta = page_address(new_page) + CPSW_XMETA_OFFSET; in cpsw_rx_handler() 453 dma = page_pool_get_dma_addr(new_page) + CPSW_HEADROOM; in cpsw_rx_handler() 454 ret = cpdma_chan_submit_mapped(cpsw->rxv[ch].ch, new_page, dma, in cpsw_rx_handler() 458 page_pool_recycle_direct(pool, new_page); in cpsw_rx_handler()
|
| H A D | cpsw_new.c | 283 struct page *new_page, *page = token; in cpsw_rx_handler() local 321 new_page = page; in cpsw_rx_handler() 330 new_page = page_pool_dev_alloc_pages(pool); in cpsw_rx_handler() 331 if (unlikely(!new_page)) { in cpsw_rx_handler() 332 new_page = page; in cpsw_rx_handler() 392 xmeta = page_address(new_page) + CPSW_XMETA_OFFSET; in cpsw_rx_handler() 396 dma = page_pool_get_dma_addr(new_page) + CPSW_HEADROOM; in cpsw_rx_handler() 397 ret = cpdma_chan_submit_mapped(cpsw->rxv[ch].ch, new_page, dma, in cpsw_rx_handler() 401 page_pool_recycle_direct(pool, new_page); in cpsw_rx_handler()
|
| /OK3568_Linux_fs/kernel/fs/nilfs2/ |
| H A D | namei.c | 374 struct page *new_page; in nilfs_rename() local 382 new_de = nilfs_find_entry(new_dir, &new_dentry->d_name, &new_page); in nilfs_rename() 385 nilfs_set_link(new_dir, new_de, new_page, old_inode); in nilfs_rename()
|
| /OK3568_Linux_fs/kernel/drivers/tty/serial/ |
| H A D | icom.c | 348 unsigned char *new_page = NULL; in load_code() local 421 new_page = dma_alloc_coherent(&dev->dev, 4096, &temp_pci, GFP_KERNEL); in load_code() 423 if (!new_page) { in load_code() 443 new_page[index] = fw->data[index]; in load_code() 500 if (new_page != NULL) in load_code() 501 dma_free_coherent(&dev->dev, 4096, new_page, temp_pci); in load_code()
|
| /OK3568_Linux_fs/kernel/fs/jbd2/ |
| H A D | journal.c | 348 struct page *new_page; in jbd2_journal_write_metadata_buffer() local 377 new_page = virt_to_page(jh_in->b_frozen_data); in jbd2_journal_write_metadata_buffer() 380 new_page = jh2bh(jh_in)->b_page; in jbd2_journal_write_metadata_buffer() 384 mapped_data = kmap_atomic(new_page); in jbd2_journal_write_metadata_buffer() 424 mapped_data = kmap_atomic(new_page); in jbd2_journal_write_metadata_buffer() 428 new_page = virt_to_page(tmp); in jbd2_journal_write_metadata_buffer() 445 mapped_data = kmap_atomic(new_page); in jbd2_journal_write_metadata_buffer() 450 set_bh_page(new_bh, new_page, new_offset); in jbd2_journal_write_metadata_buffer()
|
| /OK3568_Linux_fs/kernel/include/trace/hooks/ |
| H A D | mm.h | 302 TP_PROTO(struct page *old_page, struct page *new_page), 303 TP_ARGS(old_page, new_page));
|