From: NIIBE Y. <gn...@ch...> - 2000-04-28 07:32:46
|
I think I've fixed issues for cache.c. Here is ChangeLog and the file. (It looks good than the diffs to review). I'll commit this after the import of the 2.3.99.pre6. -------------------------- 2000-04-28 NIIBE Yutaka <gn...@m1...> Bug fixes of cache handling of SH-4. * arch/sh/mm/cache.c (dcache_flush_range, dcache_purge_range, dcache_wback_range): Renamed from cache_flush_area, cache_purge_area, cache_wback_area, and only handle D-cache. (icache_purge_range): New function. (flush_icache_range): Use icache_purge_range. (flush_icache_page): Bug fix. Rewritten. (flush_cache_range): Bug fix. Rewritten. (flush_page_to_ram): Implemented. (cache_wback_all): Assumes it is called P2 area, and make it inline function. (cache_init): Jump to P2-area before calling cache_wback_all. (flush_cache_all): Likewise. (CACHE_IC_ENTRY_SHIFT, CACHE_IC_NUM_ENTRIES, CACHE_OC_ENTRY_PHYS_MASK): New macros. * include/asm-sh/cache.h (cache_flush_area, cache_purge_area, cache_wback_area): Hide them as cache.c static function. 2000-04-27 NIIBE Yutaka <gn...@m1...> Bug fix of cache handling of SH-4. * arch/sh/mm/cache.c (flush_page_to_ram): Replaced from __flush_page_to_ram. * include/asm-sh/pgtable.h (__flush_page_to_ram): Removed. (flush_page_to_ram): Declare directly. * include/asm-sh/system.h (back_to_P1): Added one more nop. -------------------------- /* $Id: cache.c,v 1.14 2000/04/27 13:50:11 gniibe Exp $ * * linux/arch/sh/mm/cache.c * * Copyright (C) 1999, 2000 Niibe Yutaka * */ #include <linux/init.h> #include <linux/mman.h> #include <linux/mm.h> #include <linux/threads.h> #include <asm/page.h> #include <asm/pgtable.h> #include <asm/processor.h> #include <asm/cache.h> #include <asm/io.h> #include <asm/uaccess.h> #if defined(__sh3__) #define CCR 0xffffffec /* Address of Cache Control Register */ #define CCR_CACHE_VAL 0x00000005 /* 8k-byte cache, P1-wb, enable */ #define CCR_CACHE_INIT 0x0000000d /* 8k-byte cache, CF, P1-wb, enable */ #define CCR_CACHE_ENABLE 1 #define CACHE_IC_ADDRESS_ARRAY 0xf0000000 /* SH-3 has unified cache system */ #define CACHE_OC_ADDRESS_ARRAY 0xf0000000 #define CACHE_VALID 1 #define CACHE_UPDATED 2 /* 7709A/7729 has 16K cache (256-entry), while 7702 has only 2K(direct) 7702 is not supported (yet) */ struct _cache_system_info { int way_shift; int entry_mask; int num_entries; }; static struct _cache_system_info cache_system_info; #define CACHE_OC_WAY_SHIFT (cache_system_info.way_shift) #define CACHE_IC_WAY_SHIFT (cache_system_info.way_shift) #define CACHE_OC_ENTRY_SHIFT 4 #define CACHE_OC_ENTRY_MASK (cache_system_info.entry_mask) #define CACHE_IC_ENTRY_MASK (cache_system_info.entry_mask) #define CACHE_OC_NUM_ENTRIES (cache_system_info.num_entries) #define CACHE_OC_NUM_WAYS 4 #define CACHE_IC_NUM_WAYS 4 #elif defined(__SH4__) #define CCR 0xff00001c /* Address of Cache Control Register */ #define CCR_CACHE_VAL 0x00000105 /* 8k+16k-byte cache,P1-wb,enable */ #define CCR_CACHE_INIT 0x0000090d /* 8k+16k-byte cache,ICI,OCI,P1-wb,enable */ #define CCR_CACHE_ENABLE 0x00000101 #define CACHE_IC_ADDRESS_ARRAY 0xf0000000 #define CACHE_OC_ADDRESS_ARRAY 0xf4000000 #define CACHE_VALID 1 #define CACHE_UPDATED 2 #define CACHE_OC_WAY_SHIFT 13 #define CACHE_IC_WAY_SHIFT 13 #define CACHE_OC_ENTRY_SHIFT 5 #define CACHE_IC_ENTRY_SHIFT 5 #define CACHE_OC_ENTRY_MASK 0x3fe0 #define CACHE_OC_ENTRY_PHYS_MASK 0x0fe0 #define CACHE_IC_ENTRY_MASK 0x1fe0 #define CACHE_IC_NUM_ENTRIES 256 #define CACHE_OC_NUM_ENTRIES 512 #define CACHE_OC_NUM_WAYS 1 #define CACHE_IC_NUM_WAYS 1 #endif /* * Write back all the cache. * * For SH-4, we only need to flush (write back) Operand Cache, * as Instruction Cache doesn't have "updated" data. * * Assumes that this is called in interrupt disabled context, and P2. * Shuld be INLINE function. */ static inline void cache_wback_all(void) { unsigned long addr, data, i, j; for (i=0; i<CACHE_OC_NUM_ENTRIES; i++) { for (j=0; j<CACHE_OC_NUM_WAYS; j++) { addr = CACHE_OC_ADDRESS_ARRAY|(j<<CACHE_OC_WAY_SHIFT)| (i<<CACHE_OC_ENTRY_SHIFT); data = ctrl_inl(addr); if (data & CACHE_UPDATED) { data &= ~CACHE_UPDATED; ctrl_outl(data, addr); } } } } static void detect_cpu_and_cache_system(void) { #if defined(__sh3__) unsigned long addr0, addr1, data0, data1, data2, data3; jump_to_P2(); /* * Check if the entry shadows or not. * When shadowed, it's 128-entry system. * Otherwise, it's 256-entry system. */ addr0 = CACHE_OC_ADDRESS_ARRAY + (3 << 12); addr1 = CACHE_OC_ADDRESS_ARRAY + (1 << 12); data0 = ctrl_inl(addr0); data0 ^= 0x00000001; ctrl_outl(data0,addr0); data1 = ctrl_inl(addr1); data2 = data1 ^ 0x00000001; ctrl_outl(data2,addr1); data3 = ctrl_inl(addr0); /* Invaliate them, in case the cache has been enabled already. */ ctrl_outl(data0&~0x00000001, addr0); ctrl_outl(data2&~0x00000001, addr1); back_to_P1(); if (data0 == data1 && data2 == data3) { /* Shadow */ cache_system_info.way_shift = 11; cache_system_info.entry_mask = 0x7f0; cache_system_info.num_entries = 128; cpu_data->type = CPU_SH7708; } else { /* 7709A or 7729 */ cache_system_info.way_shift = 12; cache_system_info.entry_mask = 0xff0; cache_system_info.num_entries = 256; cpu_data->type = CPU_SH7729; } #elif defined(__SH4__) cpu_data->type = CPU_SH7750; #endif } void __init cache_init(void) { unsigned long ccr; detect_cpu_and_cache_system(); ccr = ctrl_inl(CCR); if (ccr == CCR_CACHE_VAL) return; jump_to_P2(); if (ccr & CCR_CACHE_ENABLE) /* * XXX: Should check RA here. * If RA was 1, we only need to flush the half of the caches. */ cache_wback_all(); ctrl_outl(CCR_CACHE_INIT, CCR); back_to_P1(); } #if defined(__SH4__) /* * SH-4 has virtually indexed and physically tagged cache. */ /* * Write back D-caches to memory (if needed) and invalidate then. * * START, END: Virtual Address */ static void dcache_flush_range(unsigned long start, unsigned long end) { unsigned long v; start &= ~(L1_CACHE_BYTES-1); for (v = start; v < end; v+=L1_CACHE_BYTES) { asm volatile("ocbp %0" : /* no output */ : "m" (__m(v))); } } /* * Purge (just invalidate, no write back) the D-caches. * * START, END: Virtual Address * * FYI: * On SH-3, it seems that we have no way to just purge (with no write * back action) the cache line. */ static void dcache_purge_range(unsigned long start, unsigned long end) { unsigned long v; start &= ~(L1_CACHE_BYTES-1); for (v = start; v < end; v+=L1_CACHE_BYTES) { asm volatile("ocbi %0" : /* no output */ : "m" (__m(v))); } } /* * Write back the dirty D-caches, but not invalidate them. * * START, END: Virtual Address */ static void dcache_wback_range(unsigned long start, unsigned long end) { unsigned long v; start &= ~(L1_CACHE_BYTES-1); for (v = start; v < end; v+=L1_CACHE_BYTES) { asm volatile("ocbwb %0" : /* no output */ : "m" (__m(v))); } } /* * Invalidate I-caches. * * START, END: Virtual Address * */ static void icache_purge_range(unsigned long start, unsigned long end) { unsigned long addr, data, v; start &= ~(L1_CACHE_BYTES-1); jump_to_P2(); /* * To handle the cache-line, we calculate the entry with virtual * address: entry = vaddr & CACHE_IC_ENTRY_MASK. * * With A-bit "on", data written to is translated by MMU and * compared the tag of cache and if it's not matched, nothing * will be occurred. (We can avoid flushing other caches.) * */ for (v = start; v < end; v+=L1_CACHE_BYTES) { addr = CACHE_IC_ADDRESS_ARRAY | (v&CACHE_IC_ENTRY_MASK) | 0x8 /* A-bit */; data = (v&0xfffffc00); /* Valid=0 */ ctrl_outl(data, addr); } back_to_P1(); } /* * Write back the range of D-cache, and purge the I-cache. * * Called from sh/kernel/signal.c. */ void flush_icache_range(unsigned long start, unsigned long end) { dcache_wback_range(start, end); icache_purge_range(start, end); } /* * Invalidate the I-cache of the page (don't need to write back D-cache). * * Called from kernel/ptrace.c, mm/memory.c after flush_page_to_ram is called. */ void flush_icache_page(struct vm_area_struct *vma, struct page *pg) { unsigned long phys, addr, data, i; /* * Alas, we don't know where the virtual address is, * So, we can't use icache_purge_range(). */ /* Physical address of this page */ phys = (pg - mem_map)*PAGE_SIZE + __MEMORY_START; jump_to_P2(); /* Loop all the I-cache */ for (i=0; i<CACHE_IC_NUM_ENTRIES; i++) { addr = CACHE_IC_ADDRESS_ARRAY| (i<<CACHE_IC_ENTRY_SHIFT); data = ctrl_inl(addr); if ((data & CACHE_VALID) && (data&PAGE_MASK) == phys) { data &= ~CACHE_VALID; ctrl_outl(data, addr); } } back_to_P1(); } void flush_cache_all(void) { unsigned long flags; save_and_cli(flags); jump_to_P2(); /* Write back Operand Cache */ cache_wback_all(); /* Then, invalidate Instruction Cache and Operand Cache */ ctrl_outl(CCR_CACHE_INIT, CCR); back_to_P1(); restore_flags(flags); } void flush_cache_mm(struct mm_struct *mm) { /* Is there any good way? */ /* XXX: possibly call flush_cache_range for each vm area */ flush_cache_all(); } void flush_cache_range(struct mm_struct *mm, unsigned long start, unsigned long end) { /* * Calling * dcache_flush_range(start, end); * is not good for the purpose of this function. That is, * flushing cache lines indexed by the virtual address is not * sufficient. * * Instead, we need to flush the relevant cache lines which * hold the data of the corresponding physical memory, as we * have "alias" issues. * * This is needed because, kernel accesses the memory through * P1-area (and/or U0-area) and user-space accesses through U0-area. * And P1-area and U0-area may use different cache lines for * same physical memory. * * If we would call dcache_flush_range(), the line of P1-area * could remain in the cache, unflushed. */ unsigned long addr, data, v; start &= ~(L1_CACHE_BYTES-1); jump_to_P2(); for (v = start; v < end; v+=L1_CACHE_BYTES) { addr = CACHE_OC_ADDRESS_ARRAY | (v&CACHE_OC_ENTRY_PHYS_MASK) | 0x8 /* A-bit */; data = (v&0xfffffc00); /* Update=0, Valid=0 */ /* Try all the cases for aliases */ ctrl_outl(data, addr); ctrl_outl(data, addr | 0x1000); ctrl_outl(data, addr | 0x2000); ctrl_outl(data, addr | 0x3000); } back_to_P1(); #if 1 icache_purge_range(start, end); /* Do we needed this? */ #endif } void flush_cache_page(struct vm_area_struct *vma, unsigned long addr) { flush_cache_range(vma->vm_mm, addr, addr+PAGE_SIZE); } void flush_page_to_ram(struct page *pg) { unsigned long phys, addr, data, i; /* Physical address of this page */ phys = (pg - mem_map)*PAGE_SIZE + __MEMORY_START; jump_to_P2(); /* Loop all the D-cache */ for (i=0; i<CACHE_OC_NUM_ENTRIES; i++) { addr = CACHE_OC_ADDRESS_ARRAY| (i<<CACHE_OC_ENTRY_SHIFT); data = ctrl_inl(addr); if ((data & CACHE_UPDATED) && (data&PAGE_MASK) == phys) { data &= ~(CACHE_UPDATED|CACHE_VALID); ctrl_outl(data, addr); } } back_to_P1(); } #endif |