Viewing file: cacheflush.h (3.76 KB) -rw-r--r-- Select action/file-type: (+) | (+) | (+) | Code (+) | Session (+) | (+) | SDB (+) | (+) | (+) | (+) | (+) | (+) |
/* SPDX-License-Identifier: GPL-2.0-or-later */ /* */ #ifndef _ASM_POWERPC_CACHEFLUSH_H #define _ASM_POWERPC_CACHEFLUSH_H
#include <linux/mm.h> #include <asm/cputable.h> #include <asm/cpu_has_feature.h>
/* * This flag is used to indicate that the page pointed to by a pte is clean * and does not require cleaning before returning it to the user. */ #define PG_dcache_clean PG_arch_1
#ifdef CONFIG_PPC_BOOK3S_64 /* * Book3s has no ptesync after setting a pte, so without this ptesync it's * possible for a kernel virtual mapping access to return a spurious fault * if it's accessed right after the pte is set. The page fault handler does * not expect this type of fault. flush_cache_vmap is not exactly the right * place to put this, but it seems to work well enough. */ static inline void flush_cache_vmap(unsigned long start, unsigned long end) { asm volatile("ptesync" ::: "memory"); } #define flush_cache_vmap flush_cache_vmap #endif /* CONFIG_PPC_BOOK3S_64 */
#define ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE 1 /* * This is called when a page has been modified by the kernel. * It just marks the page as not i-cache clean. We do the i-cache * flush later when the page is given to a user process, if necessary. */ static inline void flush_dcache_page(struct page *page) { if (cpu_has_feature(CPU_FTR_COHERENT_ICACHE)) return; /* avoid an atomic op if possible */ if (test_bit(PG_dcache_clean, &page->flags)) clear_bit(PG_dcache_clean, &page->flags); }
void flush_icache_range(unsigned long start, unsigned long stop); #define flush_icache_range flush_icache_range
void flush_icache_user_page(struct vm_area_struct *vma, struct page *page, unsigned long addr, int len); #define flush_icache_user_page flush_icache_user_page
void flush_dcache_icache_page(struct page *page);
/** * flush_dcache_range(): Write any modified data cache blocks out to memory and * invalidate them. Does not invalidate the corresponding instruction cache * blocks. * * @start: the start address * @stop: the stop address (exclusive) */ static inline void flush_dcache_range(unsigned long start, unsigned long stop) { unsigned long shift = l1_dcache_shift(); unsigned long bytes = l1_dcache_bytes(); void *addr = (void *)(start & ~(bytes - 1)); unsigned long size = stop - (unsigned long)addr + (bytes - 1); unsigned long i;
if (IS_ENABLED(CONFIG_PPC64)) mb(); /* sync */
for (i = 0; i < size >> shift; i++, addr += bytes) dcbf(addr); mb(); /* sync */
}
/* * Write any modified data cache blocks out to memory. * Does not invalidate the corresponding cache lines (especially for * any corresponding instruction cache). */ static inline void clean_dcache_range(unsigned long start, unsigned long stop) { unsigned long shift = l1_dcache_shift(); unsigned long bytes = l1_dcache_bytes(); void *addr = (void *)(start & ~(bytes - 1)); unsigned long size = stop - (unsigned long)addr + (bytes - 1); unsigned long i;
for (i = 0; i < size >> shift; i++, addr += bytes) dcbst(addr); mb(); /* sync */ }
/* * Like above, but invalidate the D-cache. This is used by the 8xx * to invalidate the cache so the PPC core doesn't get stale data * from the CPM (no cache snooping here :-). */ static inline void invalidate_dcache_range(unsigned long start, unsigned long stop) { unsigned long shift = l1_dcache_shift(); unsigned long bytes = l1_dcache_bytes(); void *addr = (void *)(start & ~(bytes - 1)); unsigned long size = stop - (unsigned long)addr + (bytes - 1); unsigned long i;
for (i = 0; i < size >> shift; i++, addr += bytes) dcbi(addr); mb(); /* sync */ }
#ifdef CONFIG_4xx static inline void flush_instruction_cache(void) { iccci((void *)KERNELBASE); isync(); } #else void flush_instruction_cache(void); #endif
#include <asm-generic/cacheflush.h>
#endif /* _ASM_POWERPC_CACHEFLUSH_H */
|