Home
last modified time | relevance | path

Searched refs:I915_GTT_PAGE_SIZE (Results 1 – 24 of 24) sorted by relevance

/OK3568_Linux_fs/kernel/drivers/gpu/drm/i915/selftests/
H A Di915_gem_evict.c56 I915_GTT_PAGE_SIZE); in populate_ggtt()
121 I915_GTT_PAGE_SIZE, 0, 0, in igt_evict_something()
136 I915_GTT_PAGE_SIZE, 0, 0, in igt_evict_something()
168 obj = i915_gem_object_create_internal(gt->i915, I915_GTT_PAGE_SIZE); in igt_overcommit()
245 .start = I915_GTT_PAGE_SIZE * 2, in igt_evict_for_cache_color()
246 .size = I915_GTT_PAGE_SIZE, in igt_evict_for_cache_color()
263 obj = i915_gem_object_create_internal(gt->i915, I915_GTT_PAGE_SIZE); in igt_evict_for_cache_color()
272 I915_GTT_PAGE_SIZE | flags); in igt_evict_for_cache_color()
279 obj = i915_gem_object_create_internal(gt->i915, I915_GTT_PAGE_SIZE); in igt_evict_for_cache_color()
289 (I915_GTT_PAGE_SIZE * 2) | flags); in igt_evict_for_cache_color()
H A Di915_gem_gtt.c113 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); in fake_dma_object()
670 obj = i915_gem_object_create_internal(vm->i915, 2 * I915_GTT_PAGE_SIZE); in pot_hole()
682 pot > ilog2(2 * I915_GTT_PAGE_SIZE); in pot_hole()
687 for (addr = round_up(hole_start + I915_GTT_PAGE_SIZE, step) - I915_GTT_PAGE_SIZE; in pot_hole()
688 addr <= round_down(hole_end - 2*I915_GTT_PAGE_SIZE, step) - I915_GTT_PAGE_SIZE; in pot_hole()
1324 total + 2 * I915_GTT_PAGE_SIZE <= ggtt->vm.total; in igt_gtt_reserve()
1325 total += 2 * I915_GTT_PAGE_SIZE) { in igt_gtt_reserve()
1365 vma->node.size != 2*I915_GTT_PAGE_SIZE) { in igt_gtt_reserve()
1368 total, 2*I915_GTT_PAGE_SIZE); in igt_gtt_reserve()
1375 for (total = I915_GTT_PAGE_SIZE; in igt_gtt_reserve()
[all …]
/OK3568_Linux_fs/kernel/drivers/gpu/drm/i915/gt/
H A Dintel_ggtt.c29 *start += I915_GTT_PAGE_SIZE; in i915_ggtt_color_adjust()
39 *end -= I915_GTT_PAGE_SIZE; in i915_ggtt_color_adjust()
199 (gen8_pte_t __iomem *)ggtt->gsm + offset / I915_GTT_PAGE_SIZE; in gen8_ggtt_insert_page()
224 gte += vma->node.start / I915_GTT_PAGE_SIZE; in gen8_ggtt_insert_entries()
225 end = gte + vma->node.size / I915_GTT_PAGE_SIZE; in gen8_ggtt_insert_entries()
250 (gen6_pte_t __iomem *)ggtt->gsm + offset / I915_GTT_PAGE_SIZE; in gen6_ggtt_insert_page()
275 gte += vma->node.start / I915_GTT_PAGE_SIZE; in gen6_ggtt_insert_entries()
276 end = gte + vma->node.size / I915_GTT_PAGE_SIZE; in gen6_ggtt_insert_entries()
302 unsigned int first_entry = start / I915_GTT_PAGE_SIZE; in gen8_ggtt_clear_range()
303 unsigned int num_entries = length / I915_GTT_PAGE_SIZE; in gen8_ggtt_clear_range()
[all …]
H A Dgen6_ppgtt.c83 const unsigned int first_entry = start / I915_GTT_PAGE_SIZE; in gen6_ppgtt_clear_range()
87 unsigned int num_entries = length / I915_GTT_PAGE_SIZE; in gen6_ppgtt_clear_range()
123 unsigned int first_entry = vma->node.start / I915_GTT_PAGE_SIZE; in gen6_ppgtt_insert_entries()
134 GEM_BUG_ON(iter.sg->length < I915_GTT_PAGE_SIZE); in gen6_ppgtt_insert_entries()
137 iter.dma += I915_GTT_PAGE_SIZE; in gen6_ppgtt_insert_entries()
155 vma->page_sizes.gtt = I915_GTT_PAGE_SIZE; in gen6_ppgtt_insert_entries()
309 u32 ggtt_offset = i915_ggtt_offset(vma) / I915_GTT_PAGE_SIZE; in pd_vma_bind()
351 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); in pd_vma_create()
H A Dselftest_context.c111 vaddr += engine->context_size - I915_GTT_PAGE_SIZE; in __live_context_size()
112 memset(vaddr, POISON_INUSE, I915_GTT_PAGE_SIZE); in __live_context_size()
135 if (memchr_inv(vaddr, POISON_INUSE, I915_GTT_PAGE_SIZE)) { in __live_context_size()
177 engine->context_size += I915_GTT_PAGE_SIZE; in live_context_size()
181 engine->context_size -= I915_GTT_PAGE_SIZE; in live_context_size()
H A Dgen8_ppgtt.c372 GEM_BUG_ON(iter->sg->length < I915_GTT_PAGE_SIZE); in gen8_ppgtt_insert_pte()
375 iter->dma += I915_GTT_PAGE_SIZE; in gen8_ppgtt_insert_pte()
443 page_size = I915_GTT_PAGE_SIZE; in gen8_ppgtt_insert_huge()
449 rem >= (I915_PDES - index) * I915_GTT_PAGE_SIZE)) in gen8_ppgtt_insert_huge()
474 rem >= (I915_PDES - index) * I915_GTT_PAGE_SIZE))) in gen8_ppgtt_insert_huge()
536 if (vma->page_sizes.sg > I915_GTT_PAGE_SIZE) { in gen8_ppgtt_insert()
549 vma->page_sizes.gtt = I915_GTT_PAGE_SIZE; in gen8_ppgtt_insert()
H A Dintel_gtt.h46 #define I915_GTT_PAGE_SIZE I915_GTT_PAGE_SIZE_4K macro
49 #define I915_GTT_PAGE_MASK -I915_GTT_PAGE_SIZE
51 #define I915_GTT_MIN_ALIGNMENT I915_GTT_PAGE_SIZE
146 __for_each_sgt_daddr(__dp, __iter, __sgt, I915_GTT_PAGE_SIZE)
H A Dintel_lrc.c3305 memset(vaddr, CONTEXT_REDZONE, I915_GTT_PAGE_SIZE); in set_redzone()
3316 if (memchr_inv(vaddr, CONTEXT_REDZONE, I915_GTT_PAGE_SIZE)) in check_redzone()
3443 GEM_BUG_ON(cs - start > I915_GTT_PAGE_SIZE / sizeof(*cs)); in setup_indirect_ctx_bb()
5368 context_size = round_up(engine->context_size, I915_GTT_PAGE_SIZE); in __execlists_context_alloc()
5371 context_size += I915_GTT_PAGE_SIZE; /* for redzone */ in __execlists_context_alloc()
H A Dintel_engine_cs.c162 BUILD_BUG_ON(I915_GTT_PAGE_SIZE != PAGE_SIZE); in intel_engine_context_size()
/OK3568_Linux_fs/kernel/drivers/gpu/drm/i915/gvt/
H A Daperture_gm.c53 start = ALIGN(gvt_hidden_gmadr_base(gvt), I915_GTT_PAGE_SIZE); in alloc_gm()
54 end = ALIGN(gvt_hidden_gmadr_end(gvt), I915_GTT_PAGE_SIZE); in alloc_gm()
59 start = ALIGN(gvt_aperture_gmadr_base(gvt), I915_GTT_PAGE_SIZE); in alloc_gm()
60 end = ALIGN(gvt_aperture_gmadr_end(gvt), I915_GTT_PAGE_SIZE); in alloc_gm()
67 size, I915_GTT_PAGE_SIZE, in alloc_gm()
262 vgpu_aperture_sz(vgpu) = ALIGN(request, I915_GTT_PAGE_SIZE); in alloc_resource()
273 vgpu_hidden_sz(vgpu) = ALIGN(request, I915_GTT_PAGE_SIZE); in alloc_resource()
H A Dscheduler.c173 I915_GTT_PAGE_SIZE - sizeof(*shadow_ring_context)); in populate_shadow_context()
227 gpa_size += I915_GTT_PAGE_SIZE; in populate_shadow_context()
237 gpa_size = I915_GTT_PAGE_SIZE; in populate_shadow_context()
953 gpa_size += I915_GTT_PAGE_SIZE; in update_guest_context()
963 gpa_size = I915_GTT_PAGE_SIZE; in update_guest_context()
995 I915_GTT_PAGE_SIZE - sizeof(*shadow_ring_context)); in update_guest_context()
H A Dreg.h116 I915_GTT_PAGE_SIZE)
H A Dcmd_parser.c1533 if (guest_gma >= I915_GTT_PAGE_SIZE) { in cmd_address_audit()
1738 offset = gma & (I915_GTT_PAGE_SIZE - 1); in copy_gma_to_hva()
1740 copy_len = (end_gma - gma) >= (I915_GTT_PAGE_SIZE - offset) ? in copy_gma_to_hva()
1741 I915_GTT_PAGE_SIZE - offset : end_gma - gma; in copy_gma_to_hva()
2803 if (WARN_ON(!IS_ALIGNED(workload->rb_start, I915_GTT_PAGE_SIZE))) in scan_workload()
2848 I915_GTT_PAGE_SIZE))) in scan_wa_ctx()
H A Dgtt.c906 (I915_GTT_PAGE_SIZE >> pt_entry_size_shift(spt))
1444 for (index = 0; index < (I915_GTT_PAGE_SIZE >> in sync_oos_page()
1497 oos_page->mem, I915_GTT_PAGE_SIZE); in attach_oos_page()
2375 int page_entry_num = I915_GTT_PAGE_SIZE >> in alloc_scratch_pages()
H A Dhandlers.c1493 !intel_gvt_ggtt_validate_range(vgpu, value, I915_GTT_PAGE_SIZE)) { in hws_pga_write()
/OK3568_Linux_fs/kernel/drivers/gpu/drm/i915/
H A Di915_gem_gtt.c109 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); in i915_gem_gtt_reserve()
204 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); in i915_gem_gtt_insert()
208 GEM_BUG_ON(start > 0 && !IS_ALIGNED(start, I915_GTT_PAGE_SIZE)); in i915_gem_gtt_insert()
209 GEM_BUG_ON(end < U64_MAX && !IS_ALIGNED(end, I915_GTT_PAGE_SIZE)); in i915_gem_gtt_insert()
231 BUILD_BUG_ON(I915_GTT_MIN_ALIGNMENT > I915_GTT_PAGE_SIZE); in i915_gem_gtt_insert()
H A Di915_gem_evict.c263 GEM_BUG_ON(!IS_ALIGNED(start, I915_GTT_PAGE_SIZE)); in i915_gem_evict_for_node()
264 GEM_BUG_ON(!IS_ALIGNED(end, I915_GTT_PAGE_SIZE)); in i915_gem_evict_for_node()
279 start -= I915_GTT_PAGE_SIZE; in i915_gem_evict_for_node()
282 end += I915_GTT_PAGE_SIZE; in i915_gem_evict_for_node()
H A Di915_vma.c160 GEM_BUG_ON(!IS_ALIGNED(vma->size, I915_GTT_PAGE_SIZE)); in vma_create()
643 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); in i915_vma_insert()
648 GEM_BUG_ON(!IS_ALIGNED(start, I915_GTT_PAGE_SIZE)); in i915_vma_insert()
654 end = min_t(u64, end, (1ULL << 32) - I915_GTT_PAGE_SIZE); in i915_vma_insert()
655 GEM_BUG_ON(!IS_ALIGNED(end, I915_GTT_PAGE_SIZE)); in i915_vma_insert()
693 vma->page_sizes.sg > I915_GTT_PAGE_SIZE) { in i915_vma_insert()
H A Di915_perf.c1856 config_length = ALIGN(sizeof(u32) * config_length, I915_GTT_PAGE_SIZE); in alloc_oa_config_buffer()
/OK3568_Linux_fs/kernel/drivers/gpu/drm/i915/gem/selftests/
H A Dhuge_gem_object.c108 GEM_BUG_ON(!IS_ALIGNED(dma_size, I915_GTT_PAGE_SIZE)); in huge_gem_object()
H A Di915_gem_context.c1504 GEM_BUG_ON(offset < I915_GTT_PAGE_SIZE); in write_to_scratch()
1601 GEM_BUG_ON(offset < I915_GTT_PAGE_SIZE); in read_from_scratch()
1849 I915_GTT_PAGE_SIZE, vm_total, in igt_vm_isolation()
H A Dhuge_pages.c307 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); in fake_huge_pages_object()
/OK3568_Linux_fs/kernel/drivers/gpu/drm/i915/gem/
H A Di915_gem_stolen.c709 GEM_WARN_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)) || in i915_gem_object_create_stolen_for_preallocated()
H A Di915_gem_context.c1381 if (!IS_ALIGNED(args->value, I915_GTT_PAGE_SIZE)) in set_ringsize()
1384 if (args->value < I915_GTT_PAGE_SIZE) in set_ringsize()
1387 if (args->value > 128 * I915_GTT_PAGE_SIZE) in set_ringsize()