| |
| |
| |
| #include <linux/spinlock.h> |
| #include <linux/smp.h> |
| #include <linux/mm.h> |
| #include <asm/cache.h> |
| #include <asm/barrier.h> |
| |
| |
| #define INS_CACHE (1 << 0) |
| #define DATA_CACHE (1 << 1) |
| #define CACHE_INV (1 << 4) |
| #define CACHE_CLR (1 << 5) |
| #define CACHE_OMS (1 << 6) |
| |
| void local_icache_inv_all(void *priv) |
| { |
| <------>mtcr("cr17", INS_CACHE|CACHE_INV); |
| <------>sync_is(); |
| } |
| |
| #ifdef CONFIG_CPU_HAS_ICACHE_INS |
| void icache_inv_range(unsigned long start, unsigned long end) |
| { |
| <------>unsigned long i = start & ~(L1_CACHE_BYTES - 1); |
| |
| <------>for (; i < end; i += L1_CACHE_BYTES) |
| <------><------>asm volatile("icache.iva %0\n"::"r"(i):"memory"); |
| <------>sync_is(); |
| } |
| #else |
| struct cache_range { |
| <------>unsigned long start; |
| <------>unsigned long end; |
| }; |
| |
| static DEFINE_SPINLOCK(cache_lock); |
| |
| static inline void cache_op_line(unsigned long i, unsigned int val) |
| { |
| <------>mtcr("cr22", i); |
| <------>mtcr("cr17", val); |
| } |
| |
| void local_icache_inv_range(void *priv) |
| { |
| <------>struct cache_range *param = priv; |
| <------>unsigned long i = param->start & ~(L1_CACHE_BYTES - 1); |
| <------>unsigned long flags; |
| |
| <------>spin_lock_irqsave(&cache_lock, flags); |
| |
| <------>for (; i < param->end; i += L1_CACHE_BYTES) |
| <------><------>cache_op_line(i, INS_CACHE | CACHE_INV | CACHE_OMS); |
| |
| <------>spin_unlock_irqrestore(&cache_lock, flags); |
| |
| <------>sync_is(); |
| } |
| |
| void icache_inv_range(unsigned long start, unsigned long end) |
| { |
| <------>struct cache_range param = { start, end }; |
| |
| <------>if (irqs_disabled()) |
| <------><------>local_icache_inv_range(¶m); |
| <------>else |
| <------><------>on_each_cpu(local_icache_inv_range, ¶m, 1); |
| } |
| #endif |
| |
| inline void dcache_wb_line(unsigned long start) |
| { |
| <------>asm volatile("dcache.cval1 %0\n"::"r"(start):"memory"); |
| <------>sync_is(); |
| } |
| |
| void dcache_wb_range(unsigned long start, unsigned long end) |
| { |
| <------>unsigned long i = start & ~(L1_CACHE_BYTES - 1); |
| |
| <------>for (; i < end; i += L1_CACHE_BYTES) |
| <------><------>asm volatile("dcache.cval1 %0\n"::"r"(i):"memory"); |
| <------>sync_is(); |
| } |
| |
| void cache_wbinv_range(unsigned long start, unsigned long end) |
| { |
| <------>dcache_wb_range(start, end); |
| <------>icache_inv_range(start, end); |
| } |
| EXPORT_SYMBOL(cache_wbinv_range); |
| |
| void dma_wbinv_range(unsigned long start, unsigned long end) |
| { |
| <------>unsigned long i = start & ~(L1_CACHE_BYTES - 1); |
| |
| <------>for (; i < end; i += L1_CACHE_BYTES) |
| <------><------>asm volatile("dcache.civa %0\n"::"r"(i):"memory"); |
| <------>sync_is(); |
| } |
| |
| void dma_inv_range(unsigned long start, unsigned long end) |
| { |
| <------>unsigned long i = start & ~(L1_CACHE_BYTES - 1); |
| |
| <------>for (; i < end; i += L1_CACHE_BYTES) |
| <------><------>asm volatile("dcache.iva %0\n"::"r"(i):"memory"); |
| <------>sync_is(); |
| } |
| |
| void dma_wb_range(unsigned long start, unsigned long end) |
| { |
| <------>unsigned long i = start & ~(L1_CACHE_BYTES - 1); |
| |
| <------>for (; i < end; i += L1_CACHE_BYTES) |
| <------><------>asm volatile("dcache.cva %0\n"::"r"(i):"memory"); |
| <------>sync_is(); |
| } |
| |