Home
last modified time | relevance | path

Searched refs:kbase_va_region (Results 1 – 25 of 46) sorted by relevance

12

/OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem.h72 struct kbase_va_region *region;
144 struct kbase_va_region *reg;
266 struct kbase_va_region *reg;
443 struct kbase_va_region { struct
709 #define KBASE_RESERVED_REG_JIT_ALLOC ((struct kbase_va_region *)-1) argument
711 static inline bool kbase_is_region_free(struct kbase_va_region *reg) in kbase_is_region_free()
716 static inline bool kbase_is_region_invalid(struct kbase_va_region *reg) in kbase_is_region_invalid()
721 static inline bool kbase_is_region_invalid_or_free(struct kbase_va_region *reg) in kbase_is_region_invalid_or_free()
741 static inline bool kbase_is_region_shrinkable(struct kbase_va_region *reg) in kbase_is_region_shrinkable()
747 struct kbase_va_region *reg);
[all …]
H A Dmali_kbase_mem.c153 static void kbase_region_tracker_insert(struct kbase_va_region *new_reg) in kbase_region_tracker_insert()
165 struct kbase_va_region *old_reg; in kbase_region_tracker_insert()
168 old_reg = rb_entry(parent, struct kbase_va_region, rblink); in kbase_region_tracker_insert()
185 static struct kbase_va_region *find_region_enclosing_range_rbtree( in find_region_enclosing_range_rbtree()
189 struct kbase_va_region *reg; in find_region_enclosing_range_rbtree()
197 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in find_region_enclosing_range_rbtree()
214 struct kbase_va_region *kbase_find_region_enclosing_address( in kbase_find_region_enclosing_address()
219 struct kbase_va_region *reg; in kbase_find_region_enclosing_address()
226 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in kbase_find_region_enclosing_address()
244 struct kbase_va_region *kbase_region_tracker_find_region_enclosing_address( in kbase_region_tracker_find_region_enclosing_address()
[all …]
H A Dmali_kbase_mem_linux.h53 struct kbase_va_region *kbase_mem_alloc(struct kbase_context *kctx, u64 va_pages, u64 commit_pages,
139 struct kbase_va_region *reg, u64 new_pages);
184 struct kbase_va_region *reg, u64 new_pages,
264 int kbase_mem_shrink_gpu_mapping(struct kbase_context *kctx, struct kbase_va_region *reg,
294 void *kbase_vmap_reg(struct kbase_context *kctx, struct kbase_va_region *reg, u64 gpu_addr,
411 struct kbase_va_region *reg,
H A Dmali_kbase_mem_linux.c92 static int kbase_vmap_phy_pages(struct kbase_context *kctx, struct kbase_va_region *reg,
121 static struct kbase_va_region *kbase_find_event_mem_region( in kbase_find_event_mem_region()
126 struct kbase_va_region *reg; in kbase_find_event_mem_region()
183 struct kbase_va_region *reg, size_t vsize, size_t size) in kbase_phy_alloc_mapping_init()
231 struct kbase_va_region *reg; in kbase_phy_alloc_mapping_get()
288 struct kbase_va_region *kbase_mem_alloc(struct kbase_context *kctx, u64 va_pages, u64 commit_pages, in kbase_mem_alloc()
293 struct kbase_va_region *reg; in kbase_mem_alloc()
557 struct kbase_va_region *reg; in kbase_mem_query()
955 struct kbase_va_region *reg; in kbase_mem_flags_change()
1110 struct kbase_va_region *reg, enum kbase_sync_type sync_fn) in kbase_mem_do_sync_imported()
[all …]
H A Dmali_linux_trace.h224 TP_PROTO(struct kbase_va_region *reg, struct kbase_fault *fault,
265 TP_PROTO(struct kbase_va_region *reg, u8 jit_id),
290 TP_PROTO(struct kbase_va_region *reg, u8 jit_id),
303 TP_PROTO(struct kbase_va_region *reg, u8 jit_id),
316 TP_PROTO(struct kbase_jd_atom *katom, struct kbase_va_region *reg,
361 TP_PROTO(struct kbase_va_region *reg, u64 new_used_pages,
496 TP_PROTO(struct kbase_va_region *reg, size_t freed_pages,
H A Dmali_kbase_debug_mem_allocs.c49 struct kbase_va_region *reg; in debug_zone_mem_allocs_show()
63 reg = rb_entry(p, struct kbase_va_region, rblink); in debug_zone_mem_allocs_show()
H A Dmali_kbase_defs.h1916 struct kbase_va_region *pending_regions[BITS_PER_LONG];
1959 struct kbase_va_region *jit_alloc[1 + BASE_JIT_ALLOC_COUNT];
2019 struct kbase_va_region *region;
2047 struct kbase_va_region *reg;
H A Dmali_kbase_gwt.c33 struct kbase_va_region *reg; in kbase_gpu_gwt_setup_page_permission()
36 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in kbase_gpu_gwt_setup_page_permission()
H A Dmali_kbase_debug_mem_view.c197 struct kbase_va_region *reg; in debug_mem_zone_open()
201 reg = rb_entry(p, struct kbase_va_region, rblink); in debug_mem_zone_open()
/OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_mem.h71 struct kbase_va_region *region;
125 struct kbase_va_region *reg;
217 struct kbase_va_region { struct
311 static inline phys_addr_t *kbase_get_cpu_phy_pages(struct kbase_va_region *reg) in kbase_get_cpu_phy_pages() argument
321 static inline phys_addr_t *kbase_get_gpu_phy_pages(struct kbase_va_region *reg) in kbase_get_gpu_phy_pages()
331 static inline size_t kbase_reg_current_backed_size(struct kbase_va_region *reg) in kbase_reg_current_backed_size()
395 static inline int kbase_reg_prepare_native(struct kbase_va_region *reg, in kbase_reg_prepare_native()
622 struct kbase_va_region *kbase_region_tracker_find_region_enclosing_address(struct kbase_context *kc…
629 struct kbase_va_region *kbase_region_tracker_find_region_base_address(struct kbase_context *kctx, u…
631 struct kbase_va_region *kbase_alloc_free_region(struct kbase_context *kctx, u64 start_pfn, size_t n…
[all …]
H A Dmali_kbase_mem.c47 struct kbase_va_region *reg) in kbase_reg_flags_to_rbtree()
99 struct kbase_va_region *new_reg) in kbase_region_tracker_insert()
111 struct kbase_va_region *old_reg; in kbase_region_tracker_insert()
114 old_reg = rb_entry(parent, struct kbase_va_region, rblink); in kbase_region_tracker_insert()
132 static struct kbase_va_region *kbase_region_tracker_find_region_enclosing_range_free( in kbase_region_tracker_find_region_enclosing_range_free()
136 struct kbase_va_region *reg = NULL; in kbase_region_tracker_find_region_enclosing_range_free()
148 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in kbase_region_tracker_find_region_enclosing_range_free()
166 struct kbase_va_region *kbase_region_tracker_find_region_enclosing_address(struct kbase_context *kc… in kbase_region_tracker_find_region_enclosing_address()
169 struct kbase_va_region *reg; in kbase_region_tracker_find_region_enclosing_address()
184 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in kbase_region_tracker_find_region_enclosing_address()
[all …]
H A Dmali_kbase_mem_linux.c64 struct kbase_va_region *reg,
81 struct kbase_va_region *reg,
84 struct kbase_va_region *kbase_mem_alloc(struct kbase_context *kctx, in kbase_mem_alloc()
91 struct kbase_va_region *reg; in kbase_mem_alloc()
261 struct kbase_va_region *reg; in kbase_mem_query()
582 struct kbase_va_region *reg; in kbase_mem_flags_change()
694 static struct kbase_va_region *kbase_mem_from_ump(struct kbase_context *kctx, ump_secure_id id, u64… in kbase_mem_from_ump()
696 struct kbase_va_region *reg; in kbase_mem_from_ump()
812 static struct kbase_va_region *kbase_mem_from_umm(struct kbase_context *kctx, in kbase_mem_from_umm()
815 struct kbase_va_region *reg; in kbase_mem_from_umm()
[all …]
H A Dmali_kbase_mem_linux.h35 struct kbase_va_region *kbase_mem_alloc(struct kbase_context *kctx,
89 struct kbase_va_region *reg,
H A Dmali_kbase_debug_mem_view.c163 struct kbase_va_region *reg; in debug_mem_zone_open()
167 reg = rb_entry(p, struct kbase_va_region, rblink); in debug_mem_zone_open()
H A Dmali_kbase_10969_workaround.c48 struct kbase_va_region *region; in kbasep_10969_workaround_clamp_coordinates()
H A Dmali_kbase_softjobs.c606 struct kbase_va_region *reg; in kbase_debug_copy_prepare()
988 struct kbase_va_region *reg; in kbase_jit_allocate_process()
1037 (struct kbase_va_region *) -1; in kbase_jit_allocate_process()
1132 if (kctx->jit_alloc[id] != (struct kbase_va_region *) -1) in kbase_jit_free_process()
H A Dmali_kbase_defs.h1369 struct kbase_va_region *pending_regions[BITS_PER_LONG];
1487 struct kbase_va_region *jit_alloc[256];
/OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/mmu/
H A Dmali_kbase_mmu.h32 struct kbase_va_region;
157 struct kbase_va_region *reg, bool ignore_page_migration);
161 struct kbase_va_region *reg, bool ignore_page_migration);
166 struct kbase_va_region *reg);
171 struct kbase_va_region *reg);
H A Dmali_kbase_mmu_internal.h56 struct kbase_va_region *reg);
H A Dmali_kbase_mmu.c516 struct kbase_va_region *reg, size_t fault_rel_pfn) in reg_grow_calc_extra_pages()
649 struct kbase_va_region *region; in kbase_gpu_mmu_handle_write_fault()
815 struct kbase_va_region *region, size_t new_pages, in page_fault_try_alloc()
979 struct kbase_va_region *region; in kbase_mmu_page_fault_worker()
2212 struct kbase_va_region *reg, in kbase_mmu_progress_migration_on_insert()
2304 struct kbase_va_region *reg, bool ignore_page_migration) in kbase_mmu_insert_pages_no_flush()
2499 struct kbase_va_region *reg, bool ignore_page_migration) in kbase_mmu_insert_pages()
2524 struct kbase_va_region *reg) in kbase_mmu_insert_imported_pages()
2550 struct kbase_va_region *reg) in kbase_mmu_insert_aliased_pages()
/OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_tiler_heap_def.h85 struct kbase_va_region *region;
130 struct kbase_va_region *buf_desc_reg;
H A Dmali_kbase_csf_mcu_shared_reg.c65 struct kbase_va_region *reg;
258 struct kbase_va_region *reg = csg_reg->reg; in group_bind_csg_reg()
481 struct kbase_va_region *reg = csg_reg->reg; in kbase_csf_mcu_shared_group_update_pmode_map()
516 struct kbase_va_region *reg; in kbase_csf_mcu_shared_clear_evicted_group_csg_reg()
611 struct kbase_va_region *reg; in shared_mcu_csg_reg_init()
696 struct kbase_va_region *reg = csg_reg->reg; in shared_mcu_csg_reg_term()
H A Dmali_kbase_csf_defs.h396 struct kbase_va_region *queue_reg;
655 struct kbase_va_region *region;
H A Dmali_kbase_csf_firmware.h94 struct kbase_va_region *va_reg;
/OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/jm/
H A Dmali_kbase_jm_defs.h522 struct kbase_va_region **extres;

12