URL
https://opencores.org/ocsvn/or1k_old/or1k_old/trunk
Subversion Repositories or1k_old
[/] [or1k_old/] [trunk/] [or1ksim/] [mmu/] [dmmu.c] - Rev 1532
Go to most recent revision | Compare with Previous | Blame | View Log
/* dmmu.c -- Data MMU simulation Copyright (C) 1999 Damjan Lampret, lampret@opencores.org This file is part of OpenRISC 1000 Architectural Simulator. This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program; if not, write to the Free Software Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. */ /* DMMU model (not functional yet, currently just copy of data cache). */ #include "config.h" #ifdef HAVE_INTTYPES_H #include <inttypes.h> #endif #include "port.h" #include "arch.h" #include "dmmu.h" #include "abstract.h" #include "opcode/or32.h" #include "spr_defs.h" #include "execute.h" #include "stats.h" #include "sprs.h" #include "except.h" #include "sim-config.h" #include "debug.h" DEFAULT_DEBUG_CHANNEL(dmmu); extern int cont_run; /* Data MMU */ inline oraddr_t dmmu_simulate_tlb(oraddr_t virtaddr, int write_access) { int set, way = -1; int i; oraddr_t tagaddr; oraddr_t vpn, ppn; if (!(cpu_state.sprs[SPR_SR] & SPR_SR_DME) || !(cpu_state.sprs[SPR_UPR] & SPR_UPR_DMP)) { data_ci = (virtaddr >= 0x80000000); return virtaddr; } /* Which set to check out? */ set = (virtaddr / config.dmmu.pagesize) % config.dmmu.nsets; tagaddr = (virtaddr / config.dmmu.pagesize) / config.dmmu.nsets; vpn = virtaddr / (config.dmmu.pagesize * config.dmmu.nsets); /* Scan all ways and try to find a matching way. */ for (i = 0; i < config.dmmu.nways; i++) if (((cpu_state.sprs[SPR_DTLBMR_BASE(i) + set] / (config.dmmu.pagesize * config.dmmu.nsets)) == vpn) && (cpu_state.sprs[SPR_DTLBMR_BASE(i) + set] & SPR_DTLBMR_V)) way = i; /* Did we find our tlb entry? */ if (way >= 0) { /* Yes, we did. */ dmmu_stats.loads_tlbhit++; TRACE("DTLB hit (virtaddr=%"PRIxADDR") at %lli.\n", virtaddr, runtime.sim.cycles); /* Set LRUs */ for (i = 0; i < config.dmmu.nways; i++) { uorreg_t lru = cpu_state.sprs[SPR_DTLBMR_BASE(i) + set]; if (lru & SPR_DTLBMR_LRU) { lru = (lru & ~SPR_DTLBMR_LRU) | ((lru & SPR_DTLBMR_LRU) - 0x40); cpu_state.sprs[SPR_DTLBMR_BASE(i) + set] = lru; } } cpu_state.sprs[SPR_DTLBMR_BASE(way) + set] &= ~SPR_DTLBMR_LRU; cpu_state.sprs[SPR_DTLBMR_BASE(way) + set] |= (config.dmmu.nsets - 1) << 6; /* Check if page is cache inhibited */ data_ci = (cpu_state.sprs[SPR_DTLBTR_BASE(way) + set] & SPR_DTLBTR_CI) == SPR_DTLBTR_CI; runtime.sim.mem_cycles += config.dmmu.hitdelay; ppn = cpu_state.sprs[SPR_DTLBTR_BASE(way) + set] / config.dmmu.pagesize; /* Test for page fault */ if (cpu_state.sprs[SPR_SR] & SPR_SR_SM) { if ( write_access && !(cpu_state.sprs[SPR_DTLBTR_BASE(way) + set] & SPR_DTLBTR_SWE) || !write_access && !(cpu_state.sprs[SPR_DTLBTR_BASE(way) + set] & SPR_DTLBTR_SRE)) except_handle(EXCEPT_DPF, virtaddr); } else { if ( write_access && !(cpu_state.sprs[SPR_DTLBTR_BASE(way) + set] & SPR_DTLBTR_UWE) || !write_access && !(cpu_state.sprs[SPR_DTLBTR_BASE(way) + set] & SPR_DTLBTR_URE)) except_handle(EXCEPT_DPF, virtaddr); } return (ppn * config.dmmu.pagesize) + (virtaddr % config.dmmu.pagesize); } else { /* No, we didn't. */ dmmu_stats.loads_tlbmiss++; #if 0 for (i = 0; i < config.dmmu.nways; i++) if (((cpu_state.sprs[SPR_DTLBMR_BASE(i) + set] & SPR_DTLBMR_LRU) >> 6) < minlru) minway = i; cpu_state.sprs[SPR_DTLBMR_BASE(minway) + set] &= ~SPR_DTLBMR_VPN; cpu_state.sprs[SPR_DTLBMR_BASE(minway) + set] |= vpn << 12; for (i = 0; i < config.dmmu.nways; i++) { uorreg_t lru = cpu_state.sprs[SPR_DTLBMR_BASE(i) + set]; if (lru & SPR_DTLBMR_LRU) { lru = (lru & ~SPR_DTLBMR_LRU) | ((lru & SPR_DTLBMR_LRU) - 0x40); cpu_state.sprs[SPR_DTLBMR_BASE(i) + set] = lru; } } cpu_state.sprs[SPR_DTLBMR_BASE(way) + set] &= ~SPR_DTLBMR_LRU; cpu_state.sprs[SPR_DTLBMR_BASE(way) + set] |= (config.dmmu.nsets - 1) << 6; /* 1 to 1 mapping */ cpu_state.sprs[SPR_DTLBTR_BASE(minway) + set] &= ~SPR_DTLBTR_PPN; cpu_state.sprs[SPR_DTLBTR_BASE(minway) + set] |= vpn << 12; cpu_state.sprs[SPR_DTLBMR_BASE(minway) + set] |= SPR_DTLBMR_V; #endif TRACE("DTLB miss (virtaddr=%"PRIxADDR") at %lli.\n", virtaddr, runtime.sim.cycles); runtime.sim.mem_cycles += config.dmmu.missdelay; /* if tlb refill implemented in HW */ /* return ((cpu_state.sprs[SPR_DTLBTR_BASE(minway) + set] & SPR_DTLBTR_PPN) >> 12) * config.dmmu.pagesize + (virtaddr % config.dmmu.pagesize); */ except_handle(EXCEPT_DTLBMISS, virtaddr); return 0; } } /* DESC: try to find EA -> PA transaltion without changing * any of precessor states. if this is not passible gives up * (without triggering exceptions) * * PRMS: virtaddr - EA for which to find translation * * write_access - 0 ignore testing for write access * 1 test for write access, if fails * do not return translation * * through_dc - 1 go through data cache * 0 ignore data cache * * RTRN: 0 - no DMMU, DMMU disabled or ITLB miss * else - appropriate PA (note it DMMU is not present * PA === EA) */ oraddr_t peek_into_dtlb(oraddr_t virtaddr, int write_access, int through_dc) { int set, way = -1; int i; oraddr_t tagaddr; oraddr_t vpn, ppn; if (!(cpu_state.sprs[SPR_SR] & SPR_SR_DME) || !(cpu_state.sprs[SPR_UPR] & SPR_UPR_DMP)) { if (through_dc) data_ci = (virtaddr >= 0x80000000); return virtaddr; } /* Which set to check out? */ set = (virtaddr / config.dmmu.pagesize) % config.dmmu.nsets; tagaddr = (virtaddr / config.dmmu.pagesize) / config.dmmu.nsets; vpn = virtaddr / (config.dmmu.pagesize * config.dmmu.nsets); /* Scan all ways and try to find a matching way. */ for (i = 0; i < config.dmmu.nways; i++) if (((cpu_state.sprs[SPR_DTLBMR_BASE(i) + set] / (config.dmmu.pagesize * config.dmmu.nsets)) == vpn) && (cpu_state.sprs[SPR_DTLBMR_BASE(i) + set] & SPR_DTLBMR_V)) way = i; /* Did we find our tlb entry? */ if (way >= 0) { /* Yes, we did. */ dmmu_stats.loads_tlbhit++; TRACE("DTLB hit (virtaddr=%"PRIxADDR") at %lli.\n", virtaddr, runtime.sim.cycles); /* Test for page fault */ if (cpu_state.sprs[SPR_SR] & SPR_SR_SM) { if ( write_access && !(cpu_state.sprs[SPR_DTLBTR_BASE(way) + set] & SPR_DTLBTR_SWE) || !write_access && !(cpu_state.sprs[SPR_DTLBTR_BASE(way) + set] & SPR_DTLBTR_SRE)) /* otherwise exception DPF would be raised */ return(0); } else { if ( write_access && !(cpu_state.sprs[SPR_DTLBTR_BASE(way) + set] & SPR_DTLBTR_UWE) || !write_access && !(cpu_state.sprs[SPR_DTLBTR_BASE(way) + set] & SPR_DTLBTR_URE)) /* otherwise exception DPF would be raised */ return(0); } if (through_dc) { /* Check if page is cache inhibited */ data_ci = (cpu_state.sprs[SPR_DTLBTR_BASE(way) + set] & SPR_DTLBTR_CI) == SPR_DTLBTR_CI; } ppn = cpu_state.sprs[SPR_DTLBTR_BASE(way) + set] / config.dmmu.pagesize; return (ppn * config.dmmu.pagesize) + (virtaddr % config.dmmu.pagesize); } else { /* No, we didn't. */ return(0); } ERR("ERR, should never have happened\n"); return(0); } oraddr_t dmmu_translate(oraddr_t virtaddr, int write_access) { oraddr_t phyaddr = dmmu_simulate_tlb(virtaddr, write_access); /* PRINTF("DMMU translate(%"PRIxADDR") = %"PRIxADDR"\n", virtaddr, phyaddr);*/ return phyaddr; } void dtlb_info(void) { if (!(cpu_state.sprs[SPR_UPR] & SPR_UPR_DMP)) { PRINTF("DMMU not implemented. Set UPR[DMP].\n"); return; } PRINTF("Data MMU %dKB: ", config.dmmu.nsets * config.dmmu.entrysize * config.dmmu.nways / 1024); PRINTF("%d ways, %d sets, entry size %d bytes\n", config.dmmu.nways, config.dmmu.nsets, config.dmmu.entrysize); } /* First check if virtual address is covered by DTLB and if it is: - increment DTLB read hit stats, - set 'lru' at this way to config.dmmu.ustates - 1 and decrement 'lru' of other ways unless they have reached 0, - check page access attributes and invoke DMMU page fault exception handler if necessary and if not: - increment DTLB read miss stats - find lru way and entry and invoke DTLB miss exception handler - set 'lru' with config.dmmu.ustates - 1 and decrement 'lru' of other ways unless they have reached 0 */ void dtlb_status(int start_set) { int set; int way; int end_set = config.dmmu.nsets; if (!(cpu_state.sprs[SPR_UPR] & SPR_UPR_DMP)) { PRINTF("DMMU not implemented. Set UPR[DMP].\n"); return; } if ((start_set >= 0) && (start_set < end_set)) end_set = start_set + 1; else start_set = 0; if (start_set < end_set) PRINTF("\nDMMU: "); /* Scan set(s) and way(s). */ for (set = start_set; set < end_set; set++) { PRINTF("\nSet %x: ", set); for (way = 0; way < config.dmmu.nways; way++) { PRINTF(" way %d: ", way); PRINTF("%s\n", dump_spr(SPR_DTLBMR_BASE(way) + set, cpu_state.sprs[SPR_DTLBMR_BASE(way) + set])); PRINTF("%s\n", dump_spr(SPR_DTLBTR_BASE(way) + set, cpu_state.sprs[SPR_DTLBTR_BASE(way) + set])); } } if (start_set < end_set) PRINTF("\n"); } /*---------------------------------------------------[ DMMU configuration ]---*/ void dmmu_enabled(union param_val val, void *dat) { if(val.int_val) cpu_state.sprs[SPR_UPR] |= SPR_UPR_DMP; else cpu_state.sprs[SPR_UPR] &= ~SPR_UPR_DMP; config.dmmu.enabled = val.int_val; } void dmmu_nsets(union param_val val, void *dat) { if (is_power2(val.int_val) && val.int_val <= 256) { config.dmmu.nsets = val.int_val; cpu_state.sprs[SPR_DMMUCFGR] &= ~SPR_DMMUCFGR_NTS; cpu_state.sprs[SPR_DMMUCFGR] |= log2(val.int_val) << 3; } else CONFIG_ERROR("value of power of two and lower or equal than 256 expected."); } void dmmu_nways(union param_val val, void *dat) { if (val.int_val >= 1 && val.int_val <= 4) { config.dmmu.nways = val.int_val; cpu_state.sprs[SPR_DMMUCFGR] &= ~SPR_DMMUCFGR_NTW; cpu_state.sprs[SPR_DMMUCFGR] |= val.int_val - 1; } else CONFIG_ERROR("value 1, 2, 3 or 4 expected."); } void dmmu_pagesize(union param_val val, void *dat) { if (is_power2(val.int_val)) config.dmmu.pagesize = val.int_val; else CONFIG_ERROR("value of power of two expected."); } void dmmu_entrysize(union param_val val, void *dat) { if (is_power2(val.int_val)) config.dmmu.entrysize = val.int_val; else CONFIG_ERROR("value of power of two expected."); } void dmmu_ustates(union param_val val, void *dat) { if (val.int_val >= 2 && val.int_val <= 4) config.dmmu.ustates = val.int_val; else CONFIG_ERROR("invalid USTATE."); } void dmmu_missdelay(union param_val val, void *dat) { config.dmmu.missdelay = val.int_val; } void dmmu_hitdelay(union param_val val, void *dat) { config.immu.hitdelay = val.int_val; } void reg_dmmu_sec(void) { struct config_section *sec = reg_config_sec("dmmu", NULL, NULL); reg_config_param(sec, "enabled", paramt_int, dmmu_enabled); reg_config_param(sec, "nsets", paramt_int, dmmu_nsets); reg_config_param(sec, "nways", paramt_int, dmmu_nways); reg_config_param(sec, "pagesize", paramt_int, dmmu_pagesize); reg_config_param(sec, "entrysize", paramt_int, dmmu_entrysize); reg_config_param(sec, "ustates", paramt_int, dmmu_ustates); reg_config_param(sec, "missdelay", paramt_int, dmmu_missdelay); reg_config_param(sec, "hitdelay", paramt_int, dmmu_hitdelay); }
Go to most recent revision | Compare with Previous | Blame | View Log