/* $OpenBSD: cache.h,v 1.10 2007/01/22 19:39:33 miod Exp $ */ /* $NetBSD: cache.h,v 1.16 1997/07/06 21:15:14 pk Exp $ */ /* * Copyright (c) 1996 * The President and Fellows of Harvard College. All rights reserved. * Copyright (c) 1992, 1993 * The Regents of the University of California. All rights reserved. * * This software was developed by the Computer Systems Engineering group * at Lawrence Berkeley Laboratory under DARPA contract BG 91-66 and * contributed to Berkeley. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. All advertising materials mentioning features or use of this software * must display the following acknowledgement: * This product includes software developed by Aaron Brown and * Harvard University. * This product includes software developed by the University of * California, Berkeley and its contributors. * 4. Neither the name of the University nor the names of its contributors * may be used to endorse or promote products derived from this software * without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * * @(#)cache.h 8.1 (Berkeley) 6/11/93 */ #ifndef SPARC_CACHE_H #define SPARC_CACHE_H /* * Sun-4 and Sun-4c virtual address cache. * * Sun-4 virtual caches come in two flavors, write-through (Sun-4c) * and write-back (Sun-4). The write-back caches are much faster * but require a bit more care. * */ enum vactype { VAC_UNKNOWN, VAC_NONE, VAC_WRITETHROUGH, VAC_WRITEBACK }; /* * Cache tags can be written in control space, and must be set to 0 * (or invalid anyway) before turning on the cache. The tags are * addressed as an array of 32-bit structures of the form: * * struct cache_tag { * u_int :7, (unused; must be zero) * ct_cid:3, (context ID) * ct_w:1, (write flag from PTE) * ct_s:1, (supervisor flag from PTE) * ct_v:1, (set => cache entry is valid) * :3, (unused; must be zero) * ct_tid:14, (cache tag ID) * :2; (unused; must be zero) * }; * * The SPARCstation 1 cache sees virtual addresses as: * * struct cache_va { * u_int :2, (unused; probably copies of va_tid<13>) * cva_tid:14, (tag ID) * cva_line:12, (cache line number) * cva_byte:4; (byte in cache line) * }; * * (The SS2 cache is similar but has half as many lines, each twice as long.) * * Note that, because the 12-bit line ID is `wider' than the page offset, * it is possible to have one page map to two different cache lines. * This can happen whenever two different physical pages have the same bits * in the part of the virtual address that overlaps the cache line ID, i.e., * bits <15:12>. In order to prevent cache duplication, we have to * make sure that no one page has more than one virtual address where * (va1 & 0xf000) != (va2 & 0xf000). (The cache hardware turns off ct_v * when a cache miss occurs on a write, i.e., if va1 is in the cache and * va2 is not, and you write to va2, va1 goes out of the cache. If va1 * is in the cache and va2 is not, reading va2 also causes va1 to become * uncached, and the [same] data is then read from main memory into the * cache.) * * The other alternative, of course, is to disable caching of aliased * pages. (In a few cases this might be faster anyway, but we do it * only when forced.) * * The Sun4, since it has an 8K pagesize instead of 4K, needs to check * bits that are one position higher. */ /* Some more well-known values: */ #define CACHE_ALIAS_DIST_SUN4 0x20000 #define CACHE_ALIAS_DIST_SUN4C 0x10000 #define CACHE_ALIAS_BITS_SUN4 0x1e000 #define CACHE_ALIAS_BITS_SUN4C 0xf000 #define CACHE_ALIAS_DIST_HS128k 0x20000 #define CACHE_ALIAS_BITS_HS128k 0x1f000 #define CACHE_ALIAS_DIST_HS256k 0x40000 #define CACHE_ALIAS_BITS_HS256k 0x3f000 extern int cache_alias_dist; extern int cache_alias_bits; /* Optimize cache alias macros on single architecture kernels */ #if defined(SUN4) && !defined(SUN4C) && !defined(SUN4M) #define CACHE_ALIAS_DIST CACHE_ALIAS_DIST_SUN4 #define CACHE_ALIAS_BITS CACHE_ALIAS_BITS_SUN4 #elif !defined(SUN4) && defined(SUN4C) && !defined(SUN4M) #define CACHE_ALIAS_DIST CACHE_ALIAS_DIST_SUN4C #define CACHE_ALIAS_BITS CACHE_ALIAS_BITS_SUN4C #else #define CACHE_ALIAS_DIST cache_alias_dist #define CACHE_ALIAS_BITS cache_alias_bits #endif /* * True iff a1 and a2 are `bad' aliases (will cause cache duplication). */ #define BADALIAS(a1, a2) (((int)(a1) ^ (int)(a2)) & CACHE_ALIAS_BITS) /* * Routines for dealing with the cache. */ void sun4_cache_enable(void); /* turn it on */ void ms1_cache_enable(void); /* turn it on */ void viking_cache_enable(void); /* turn it on */ void hypersparc_cache_enable(void); /* turn it on */ void swift_cache_enable(void); /* turn it on */ void cypress_cache_enable(void); /* turn it on */ void turbosparc_cache_enable(void); /* turn it on */ void kap_cache_enable(void); /* turn it on */ void sun4_vcache_flush_context(void); /* flush current context */ void sun4_vcache_flush_region(int); /* flush region in cur ctx */ void sun4_vcache_flush_segment(int, int); /* flush seg in cur ctx */ void sun4_vcache_flush_page(int va); /* flush page in cur ctx */ void sun4_cache_flush(caddr_t, u_int); /* flush region */ void srmmu_vcache_flush_context(void); /* flush current context */ void srmmu_vcache_flush_region(int); /* flush region in cur ctx */ void srmmu_vcache_flush_segment(int, int); /* flush seg in cur ctx */ void srmmu_vcache_flush_page(int va); /* flush page in cur ctx */ void srmmu_cache_flush(caddr_t, u_int); /* flush region */ void hypersparc_pure_vcache_flush(void); void ms1_cache_flush_all(void); void srmmu_cache_flush_all(void); void cypress_cache_flush_all(void); void hypersparc_cache_flush_all(void); void ms1_cache_flush(caddr_t, u_int); void viking_cache_flush(caddr_t, u_int); void viking_pcache_flush_line(int, int); void srmmu_pcache_flush_line(int, int); void kap_vcache_flush_context(void); /* flush current context */ void kap_vcache_flush_page(int va); /* flush page in cur ctx */ void kap_cache_flush(caddr_t, u_int); /* flush region */ extern void sparc_noop(void); #define noop_vcache_flush_context \ (void (*)(void)) sparc_noop #define noop_vcache_flush_region \ (void (*)(int)) sparc_noop #define noop_vcache_flush_segment \ (void (*)(int,int)) sparc_noop #define noop_vcache_flush_page \ (void (*)(int)) sparc_noop #define noop_cache_flush \ (void (*)(caddr_t, u_int)) sparc_noop #define noop_pcache_flush_line \ (void (*)(int, int)) sparc_noop #define noop_pure_vcache_flush \ (void (*)(void)) sparc_noop #define noop_cache_flush_all \ (void (*)(void)) sparc_noop #define cache_flush_page(va) cpuinfo.vcache_flush_page(va) #define cache_flush_segment(vr,vs) cpuinfo.vcache_flush_segment(vr,vs) #define cache_flush_region(vr) cpuinfo.vcache_flush_region(vr) #define cache_flush_context() cpuinfo.vcache_flush_context() /* * Cache control information. */ struct cacheinfo { int c_totalsize; /* total size, in bytes */ /* if split, MAX(icache,dcache) */ int c_enabled; /* true => cache is enabled */ int c_hwflush; /* true => have hardware flush */ int c_linesize; /* line size, in bytes */ int c_l2linesize; /* log2(linesize) */ int c_nlines; /* number of cache lines */ int c_physical; /* true => cache has physical address tags */ int c_associativity; /* # of "buckets" in cache line */ int c_split; /* true => cache is split */ int ic_totalsize; /* instruction cache */ int ic_enabled; int ic_linesize; int ic_l2linesize; int ic_nlines; int ic_associativity; int dc_totalsize; /* data cache */ int dc_enabled; int dc_linesize; int dc_l2linesize; int dc_nlines; int dc_associativity; int ec_totalsize; /* external cache info */ int ec_enabled; int ec_linesize; int ec_l2linesize; int ec_nlines; int ec_associativity; enum vactype c_vactype; }; #define CACHEINFO cpuinfo.cacheinfo /* * Cache control statistics. */ struct cachestats { int cs_npgflush; /* # page flushes */ int cs_nsgflush; /* # seg flushes */ int cs_nrgflush; /* # seg flushes */ int cs_ncxflush; /* # context flushes */ int cs_nraflush; /* # range flushes */ #ifdef notyet int cs_ra[65]; /* pages/range */ #endif }; #endif /* SPARC_CACHE_H */