233 lines
6.0 KiB
C
233 lines
6.0 KiB
C
/*
|
|
* Cache operations for the cache instruction.
|
|
*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*
|
|
* (C) Copyright 1996, 97, 99, 2002, 03 Ralf Baechle
|
|
* (C) Copyright 1999 Silicon Graphics, Inc.
|
|
*/
|
|
#ifndef __CACHE_H__
|
|
#define __CACHE_H__
|
|
|
|
|
|
#define KUSEG 0x00000000
|
|
#define KSEG0 0x80000000
|
|
#define KSEG1 0xa0000000
|
|
#define KSEG2 0xc0000000
|
|
#define KSEG3 0xe0000000
|
|
|
|
/*
|
|
* Cache Operations available on all MIPS processors with R4000-style caches
|
|
*/
|
|
#define Index_Invalidate_I 0x00
|
|
#define Index_Writeback_Inv_D 0x01
|
|
#define Index_Load_Tag_I 0x04
|
|
#define Index_Load_Tag_D 0x05
|
|
#define Index_Store_Tag_I 0x08
|
|
#define Index_Store_Tag_D 0x09
|
|
|
|
#define Hit_Invalidate_I 0x10
|
|
#define Hit_Invalidate_D 0x11
|
|
#define Hit_Writeback_Inv_D 0x15
|
|
#define Hit_Writeback_I 0x18
|
|
#define Hit_Writeback_D 0x19
|
|
|
|
/*
|
|
*The lock state is cleared by executing an Index
|
|
Invalidate, Index Writeback Invalidate, Hit
|
|
Invalidate, or Hit Writeback Invalidate
|
|
operation to the locked line, or via an Index
|
|
Store Tag operation with the lock bit reset in
|
|
the TagLo register.
|
|
*/
|
|
#define Fetch_And_Lock_I 0x1c
|
|
#define Fetch_And_Lock_D 0x1d
|
|
/*
|
|
* R4000-specific cacheops
|
|
*/
|
|
#define Create_Dirty_Excl_D 0x0d
|
|
#define Fill 0x14
|
|
|
|
/*
|
|
* R4000SC and R4400SC-specific cacheops
|
|
*/
|
|
#define Index_Invalidate_SI 0x02
|
|
#define Index_Writeback_Inv_SD 0x03
|
|
#define Index_Load_Tag_SI 0x06
|
|
#define Index_Load_Tag_SD 0x07
|
|
#define Index_Store_Tag_SI 0x0A
|
|
#define Index_Store_Tag_SD 0x0B
|
|
#define Create_Dirty_Excl_SD 0x0f
|
|
#define Hit_Invalidate_SI 0x12
|
|
#define Hit_Invalidate_SD 0x13
|
|
#define Hit_Writeback_Inv_SD 0x17
|
|
#define Hit_Writeback_SD 0x1b
|
|
#define Hit_Set_Virtual_SI 0x1e
|
|
#define Hit_Set_Virtual_SD 0x1f
|
|
|
|
/*
|
|
* R5000-specific cacheops
|
|
*/
|
|
#define R5K_Page_Invalidate_S 0x17
|
|
|
|
/*
|
|
* RM7000-specific cacheops
|
|
*/
|
|
#define Page_Invalidate_T 0x16
|
|
|
|
/*
|
|
* R1000-specific cacheops
|
|
*
|
|
* Cacheops 0x02, 0x06, 0x0a, 0x0c-0x0e, 0x16, 0x1a and 0x1e are unused.
|
|
* Most of the _S cacheops are identical to the R4000SC _SD cacheops.
|
|
*/
|
|
#define Index_Writeback_Inv_S 0x03
|
|
#define Index_Load_Tag_S 0x07
|
|
#define Index_Store_Tag_S 0x0B
|
|
#define Hit_Invalidate_S 0x13
|
|
#define Cache_Barrier 0x14
|
|
#define Hit_Writeback_Inv_S 0x17
|
|
#define Index_Load_Data_I 0x18
|
|
#define Index_Load_Data_D 0x19
|
|
#define Index_Load_Data_S 0x1b
|
|
#define Index_Store_Data_I 0x1c
|
|
#define Index_Store_Data_D 0x1d
|
|
#define Index_Store_Data_S 0x1f
|
|
|
|
|
|
#ifndef __ASSEMBLER__
|
|
|
|
#ifndef dcache_size
|
|
#define dcache_size (g_mips_core.dcache_ways * g_mips_core.dcache_lines_per_way * g_mips_core.dcache_line_size)
|
|
#endif
|
|
|
|
#ifndef icache_size
|
|
#define icache_size (g_mips_core.dcache_ways * g_mips_core.dcache_lines_per_way * g_mips_core.dcache_line_size)
|
|
#endif
|
|
|
|
#ifndef cpu_dcache_line_size
|
|
#define cpu_dcache_line_size() g_mips_core.icache_line_size
|
|
#endif
|
|
|
|
#ifndef cpu_icache_line_size
|
|
#define cpu_icache_line_size() g_mips_core.icache_line_size
|
|
#endif
|
|
|
|
#define cache_op(op, addr) \
|
|
__asm__ __volatile__( \
|
|
" .set noreorder \n" \
|
|
" .set mips3\n\t \n" \
|
|
" cache %0, %1 \n" \
|
|
" .set mips0 \n" \
|
|
" .set reorder" \
|
|
: \
|
|
: "i" (op), "m" (*(unsigned char *)(addr)))
|
|
|
|
#define cache16_unroll32(base, op) \
|
|
__asm__ __volatile__( \
|
|
" .set noreorder \n" \
|
|
" .set mips3 \n" \
|
|
" cache %1, 0x000(%0); cache %1, 0x010(%0) \n" \
|
|
" cache %1, 0x020(%0); cache %1, 0x030(%0) \n" \
|
|
" cache %1, 0x040(%0); cache %1, 0x050(%0) \n" \
|
|
" cache %1, 0x060(%0); cache %1, 0x070(%0) \n" \
|
|
" cache %1, 0x080(%0); cache %1, 0x090(%0) \n" \
|
|
" cache %1, 0x0a0(%0); cache %1, 0x0b0(%0) \n" \
|
|
" cache %1, 0x0c0(%0); cache %1, 0x0d0(%0) \n" \
|
|
" cache %1, 0x0e0(%0); cache %1, 0x0f0(%0) \n" \
|
|
" cache %1, 0x100(%0); cache %1, 0x110(%0) \n" \
|
|
" cache %1, 0x120(%0); cache %1, 0x130(%0) \n" \
|
|
" cache %1, 0x140(%0); cache %1, 0x150(%0) \n" \
|
|
" cache %1, 0x160(%0); cache %1, 0x170(%0) \n" \
|
|
" cache %1, 0x180(%0); cache %1, 0x190(%0) \n" \
|
|
" cache %1, 0x1a0(%0); cache %1, 0x1b0(%0) \n" \
|
|
" cache %1, 0x1c0(%0); cache %1, 0x1d0(%0) \n" \
|
|
" cache %1, 0x1e0(%0); cache %1, 0x1f0(%0) \n" \
|
|
" .set mips0 \n" \
|
|
" .set reorder \n" \
|
|
: \
|
|
: "r" (base), \
|
|
"i" (op));
|
|
|
|
|
|
static inline void flush_icache_line_indexed(rt_ubase_t addr)
|
|
{
|
|
cache_op(Index_Invalidate_I, addr);
|
|
}
|
|
|
|
static inline void flush_dcache_line_indexed(rt_ubase_t addr)
|
|
{
|
|
cache_op(Index_Writeback_Inv_D, addr);
|
|
}
|
|
|
|
static inline void flush_icache_line(rt_ubase_t addr)
|
|
{
|
|
cache_op(Hit_Invalidate_I, addr);
|
|
}
|
|
|
|
static inline void lock_icache_line(rt_ubase_t addr)
|
|
{
|
|
cache_op(Fetch_And_Lock_I, addr);
|
|
}
|
|
|
|
static inline void lock_dcache_line(rt_ubase_t addr)
|
|
{
|
|
cache_op(Fetch_And_Lock_D, addr);
|
|
}
|
|
|
|
static inline void flush_dcache_line(rt_ubase_t addr)
|
|
{
|
|
cache_op(Hit_Writeback_Inv_D, addr);
|
|
}
|
|
|
|
static inline void invalidate_dcache_line(rt_ubase_t addr)
|
|
{
|
|
cache_op(Hit_Invalidate_D, addr);
|
|
}
|
|
|
|
static inline void blast_dcache16(void)
|
|
{
|
|
rt_ubase_t start = KSEG0;
|
|
rt_ubase_t end = start + dcache_size;
|
|
rt_ubase_t addr;
|
|
|
|
for (addr = start; addr < end; addr += 0x200)
|
|
cache16_unroll32(addr, Index_Writeback_Inv_D);
|
|
}
|
|
|
|
static inline void inv_dcache16(void)
|
|
{
|
|
rt_ubase_t start = KSEG0;
|
|
rt_ubase_t end = start + dcache_size;
|
|
rt_ubase_t addr;
|
|
|
|
for (addr = start; addr < end; addr += 0x200)
|
|
cache16_unroll32(addr, Hit_Invalidate_D);
|
|
}
|
|
|
|
static inline void blast_icache16(void)
|
|
{
|
|
rt_ubase_t start = KSEG0;
|
|
rt_ubase_t end = start + icache_size;
|
|
rt_ubase_t addr;
|
|
|
|
for (addr = start; addr < end; addr += 0x200)
|
|
cache16_unroll32(addr, Index_Invalidate_I);
|
|
}
|
|
|
|
|
|
|
|
void r4k_cache_init(void);
|
|
void r4k_cache_flush_all(void);
|
|
void r4k_icache_flush_all(void);
|
|
void r4k_icache_flush_range(rt_ubase_t addr, rt_ubase_t size);
|
|
void r4k_icache_lock_range(rt_ubase_t addr, rt_ubase_t size);
|
|
void r4k_dcache_inv(rt_ubase_t addr, rt_ubase_t size);
|
|
void r4k_dcache_wback_inv(rt_ubase_t addr, rt_ubase_t size);
|
|
|
|
#endif /*end of __ASSEMBLER__ */
|
|
|
|
#endif /* end of __CACHE_H__ */
|