Lines Matching refs:shadow_start

74 	void *shadow_start, *shadow_end;  in kasan_poison()  local
92 shadow_start = kasan_mem_to_shadow(addr); in kasan_poison()
95 __memset(shadow_start, value, shadow_end - shadow_start); in kasan_poison()
178 unsigned long nr_shadow_pages, start_kaddr, shadow_start; in kasan_mem_notifier() local
183 shadow_start = (unsigned long)kasan_mem_to_shadow((void *)start_kaddr); in kasan_mem_notifier()
185 shadow_end = shadow_start + shadow_size; in kasan_mem_notifier()
200 if (shadow_mapped(shadow_start)) in kasan_mem_notifier()
203 ret = __vmalloc_node_range(shadow_size, PAGE_SIZE, shadow_start, in kasan_mem_notifier()
230 vm = find_vm_area((void *)shadow_start); in kasan_mem_notifier()
232 vfree((void *)shadow_start); in kasan_mem_notifier()
280 unsigned long shadow_start, shadow_end; in kasan_populate_vmalloc() local
286 shadow_start = (unsigned long)kasan_mem_to_shadow((void *)addr); in kasan_populate_vmalloc()
287 shadow_start = ALIGN_DOWN(shadow_start, PAGE_SIZE); in kasan_populate_vmalloc()
291 ret = apply_to_page_range(&init_mm, shadow_start, in kasan_populate_vmalloc()
292 shadow_end - shadow_start, in kasan_populate_vmalloc()
297 flush_cache_vmap(shadow_start, shadow_end); in kasan_populate_vmalloc()
455 void *shadow_start, *shadow_end; in kasan_release_vmalloc() local
474 shadow_start = kasan_mem_to_shadow((void *)region_start); in kasan_release_vmalloc()
477 if (shadow_end > shadow_start) { in kasan_release_vmalloc()
478 size = shadow_end - shadow_start; in kasan_release_vmalloc()
480 (unsigned long)shadow_start, in kasan_release_vmalloc()
483 flush_tlb_kernel_range((unsigned long)shadow_start, in kasan_release_vmalloc()
495 unsigned long shadow_start; in kasan_module_alloc() local
497 shadow_start = (unsigned long)kasan_mem_to_shadow(addr); in kasan_module_alloc()
502 if (WARN_ON(!PAGE_ALIGNED(shadow_start))) in kasan_module_alloc()
505 ret = __vmalloc_node_range(shadow_size, 1, shadow_start, in kasan_module_alloc()
506 shadow_start + shadow_size, in kasan_module_alloc()