Searched refs:L1_CACHE_BYTES (Results 1 – 8 of 8) sorted by relevance
20 #define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) macro26 #define ARCH_DMA_MINALIGN L1_CACHE_BYTES32 #define CONFIG_SYS_CACHELINE_SIZE L1_CACHE_BYTES35 #define L1_CACHE_ALIGN(x) (((x)+(L1_CACHE_BYTES-1))&~(L1_CACHE_BYTES-1))38 #define SMP_CACHE_BYTES L1_CACHE_BYTES41 #define __cacheline_aligned __attribute__((__aligned__(L1_CACHE_BYTES)))44 __attribute__((__aligned__(L1_CACHE_BYTES), \
61 start &= ~(L1_CACHE_BYTES - 1); in flush_dcache_range()62 for (v = start; v < end; v += L1_CACHE_BYTES) { in flush_dcache_range()72 start &= ~(L1_CACHE_BYTES - 1); in invalidate_dcache_range()73 for (v = start; v < end; v += L1_CACHE_BYTES) { in invalidate_dcache_range()
69 li r5,L1_CACHE_BYTES-178 addi r3,r3,L1_CACHE_BYTES93 li r5,L1_CACHE_BYTES-1103 addi r3,r3,L1_CACHE_BYTES
11 #define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) macro13 #define ARCH_DMA_MINALIGN (L1_CACHE_BYTES)
6 #define L1_CACHE_BYTES 32 macro26 #define ARCH_DMA_MINALIGN L1_CACHE_BYTES
26 #ifndef L1_CACHE_BYTES28 #define L1_CACHE_BYTES BIT(L1_CACHE_SHIFT) macro
16 # define ATOMIC_HASH(a) (&(__atomic_hash[ (((unsigned long) a)/L1_CACHE_BYTES) & (ATOMIC_HASH_SIZE…
11 # define CACHE_LINE_SIZE L1_CACHE_BYTES