42 #include "debug/IPR.hh" 43 #include "debug/TLB.hh" 56 :
BaseTLB(p), size(p->size), usedEntries(0), lastReplaced(0),
57 cacheState(0), cacheValid(false)
61 fatal(
"SPARC T1 TLB registers don't support more than 64 TLB entries");
105 va &= ~(PTE.
size()-1);
108 "TLB: Inserting Entry; va=%#x pa=%#x pid=%d cid=%d r=%d entryid=%d\n",
109 va, PTE.
paddr(), partition_id, context_id, (int)real, entry);
112 for (x = 0; x <
size; x++) {
113 if (
tlb[x].range.real == real &&
121 DPRINTF(
TLB,
"TLB: Conflicting entry %#X , deleting it\n", x);
134 assert(entry < size && entry >= 0);
135 new_entry = &
tlb[entry];
146 goto insertAllLocked;
147 }
while (
tlb[x].pte.locked());
156 new_entry = &
tlb[size-1];
162 if (new_entry->
valid)
172 new_entry->
pte = PTE;
173 new_entry->
used =
true;;
174 new_entry->
valid =
true;
184 new_entry->
used =
true;
198 DPRINTF(
TLB,
"TLB: Looking up entry va=%#x pid=%d cid=%d r=%d\n",
199 va, partition_id, context_id, real);
221 if (!t->
used && update_used) {
238 for (
int x = 0;
x <
size;
x++) {
240 DPRINTFN(
"%4d: %#2x:%#2x %c %#4x %#8x %#8x %#16x\n",
254 DPRINTF(IPR,
"TLB: Demapping Page va=%#x pid=%#d cid=%d r=%d\n",
255 va, partition_id, context_id, real);
269 DPRINTF(IPR,
"TLB: Demapped page\n");
270 i->second->valid =
false;
271 if (i->second->used) {
272 i->second->used =
false;
283 DPRINTF(IPR,
"TLB: Demapping Context pid=%#d cid=%d\n",
284 partition_id, context_id);
286 for (
int x = 0;
x <
size;
x++) {
287 if (
tlb[
x].range.contextId == context_id &&
305 DPRINTF(
TLB,
"TLB: Demapping All pid=%#d\n", partition_id);
307 for (
int x = 0;
x <
size;
x++) {
308 if (
tlb[
x].valid && !
tlb[
x].pte.locked() &&
327 for (
int x = 0;
x <
size;
x++) {
340 panic(
"entry: %d\n", entry);
342 assert(entry <
size);
343 if (
tlb[entry].valid)
346 return (uint64_t)-1ll;
352 assert(entry <
size);
354 if (!
tlb[entry].valid)
355 return (uint64_t)-1ll;
359 tag |= (uint64_t)
tlb[entry].range.partitionId << 61;
361 tag |= (uint64_t)~
tlb[entry].pte._size() << 56;
395 DPRINTF(
TLB,
"TLB: Writing Tag Access: va: %#X ctx: %#X value: %#X\n",
396 va, context,
mbits(va, 63,13) |
mbits(context,12,0));
405 DPRINTF(
TLB,
"TLB: Fault: A=%#x w=%d ct=%d ft=%d asi=%d\n",
406 a, (
int)write, ct, ft, asi);
421 DPRINTF(
TLB,
"TLB: ITB Request to translate va=%#x size=%d\n",
422 vaddr, req->getSize());
429 req->setPaddr(
cacheEntry[0]->pte.translate(vaddr));
441 bool addr_mask =
bits(tlbdata,3,3);
442 bool lsu_im =
bits(tlbdata,4,4);
444 int part_id =
bits(tlbdata,15,8);
445 int tl =
bits(tlbdata,18,16);
446 int pri_context =
bits(tlbdata,47,32);
452 DPRINTF(
TLB,
"TLB: priv:%d hpriv:%d red:%d lsuim:%d part_id: %#X\n",
453 priv, hpriv, red, lsu_im, part_id);
462 context = pri_context;
465 if ( hpriv || red ) {
476 return std::make_shared<MemAddressNotAligned>();
484 return std::make_shared<InstructionAccessException>();
488 e =
lookup(vaddr, part_id,
true);
492 e =
lookup(vaddr, part_id,
false, context);
495 if (e == NULL || !e->
valid) {
498 return std::make_shared<InstructionRealTranslationMiss>();
501 return std::make_shared<FastInstructionAccessMMUMiss>();
503 return std::make_shared<FastInstructionAccessMMUMiss>(
512 return std::make_shared<InstructionAccessException>();
521 DPRINTF(
TLB,
"TLB: %#X -> %#X\n", vaddr, req->getPaddr());
536 asi = (
ASI)req->getArchFlags();
537 bool implicit =
false;
539 bool unaligned = vaddr & (size - 1);
541 DPRINTF(
TLB,
"TLB: DTB Request to translate va=%#x size=%d asi=%#x\n",
553 if (hpriv && implicit) {
567 ce_va < vaddr + size && ce_va + ce->range.size > vaddr &&
574 DPRINTF(
TLB,
"TLB: %#X -> %#X\n", vaddr, req->getPaddr());
582 ce_va < vaddr + size && ce_va + ce->range.size > vaddr &&
589 DPRINTF(
TLB,
"TLB: %#X -> %#X\n", vaddr, req->getPaddr());
598 bool addr_mask =
bits(tlbdata,3,3);
599 bool lsu_dm =
bits(tlbdata,5,5);
601 int part_id =
bits(tlbdata,15,8);
602 int tl =
bits(tlbdata,18,16);
603 int pri_context =
bits(tlbdata,47,32);
604 int sec_context =
bits(tlbdata,63,48);
612 DPRINTF(
TLB,
"TLB: priv:%d hpriv:%d red:%d lsudm:%d part_id: %#X\n",
613 priv, hpriv, red, lsu_dm, part_id);
623 context = pri_context;
630 return std::make_shared<PrivilegedAction>();
635 return std::make_shared<DataAccessException>();
639 context = pri_context;
642 context = sec_context;
649 context = pri_context;
653 if (!implicit && asi !=
ASI_P && asi !=
ASI_S) {
655 panic(
"Little Endian ASIs not supported\n");
658 panic(
"Partial Store ASIs not supported\n");
661 panic(
"Cmt ASI registers not implmented\n");
664 goto handleIntRegAccess;
666 goto handleMmuRegAccess;
668 goto handleScratchRegAccess;
670 goto handleQueueRegAccess;
672 goto handleSparcErrorRegAccess;
676 panic(
"Accessing ASI %#X. Should we?\n", asi);
682 return std::make_shared<MemAddressNotAligned>();
690 return std::make_shared<DataAccessException>();
693 if ((!lsu_dm && !hpriv && !red) ||
asiIsReal(asi)) {
703 e =
lookup(vaddr, part_id, real, context);
705 if (e == NULL || !e->
valid) {
707 DPRINTF(
TLB,
"TLB: DTB Failed to find matching TLB entry\n");
709 return std::make_shared<DataRealTranslationMiss>();
712 return std::make_shared<FastDataAccessMMUMiss>();
714 return std::make_shared<FastDataAccessMMUMiss>(
723 return std::make_shared<DataAccessException>();
729 return std::make_shared<FastDataAccessProtection>();
735 return std::make_shared<DataAccessException>();
741 return std::make_shared<DataAccessException>();
764 DPRINTF(
TLB,
"TLB: %#X -> %#X\n", vaddr, req->getPaddr());
772 return std::make_shared<DataAccessException>();
774 return std::make_shared<PrivilegedAction>();
780 return std::make_shared<DataAccessException>();
786 handleScratchRegAccess:
787 if (vaddr > 0x38 || (vaddr >= 0x20 && vaddr < 0x30 && !hpriv)) {
789 return std::make_shared<DataAccessException>();
793 handleQueueRegAccess:
794 if (!priv && !hpriv) {
796 return std::make_shared<PrivilegedAction>();
798 if ((!hpriv && vaddr & 0xF) || vaddr > 0x3f8 || vaddr < 0x3c0) {
800 return std::make_shared<DataAccessException>();
804 handleSparcErrorRegAccess:
808 return std::make_shared<DataAccessException>();
810 return std::make_shared<PrivilegedAction>();
817 DPRINTF(
TLB,
"TLB: DTB Translating local access\n");
818 req->setLocalAccessor(
824 req->setPaddr(req->getVaddr());
856 bool addr_mask =
bits(tlbdata,3,3);
857 bool data_real = !
bits(tlbdata,5,5);
858 bool inst_real = !
bits(tlbdata,4,4);
859 bool ctx_zero =
bits(tlbdata,18,16) > 0;
860 int part_id =
bits(tlbdata,15,8);
861 int pri_context =
bits(tlbdata,47,32);
864 bool real = (mode ==
Execute) ? inst_real : data_real;
873 req->setPaddr(vaddr);
882 return std::make_shared<InstructionAccessException>();
884 return std::make_shared<DataAccessException>();
887 tbe =
lookup(vaddr, part_id, real, ctx_zero ? 0 : pri_context,
false);
890 DPRINTF(
TLB,
"Virtual(%#x)->Physical(%#x) found in TLB\n", vaddr,
901 GetTsbPtr(tc, vaddr, ctx_zero ? 0 : pri_context, tsbs);
902 va_tag =
bits(vaddr, 63, 22);
903 for (
int x = 0;
x < 4;
x++) {
905 if (ttetag.
valid() && ttetag.
va() == va_tag) {
906 uint64_t entry = mem.
read<uint64_t>(tsbs[
x]) +
sizeof(uint64_t);
909 DPRINTF(
TLB,
"Virtual(%#x)->Physical(%#x) found in TTE\n",
918 return std::make_shared<InstructionRealTranslationMiss>();
920 return std::make_shared<FastInstructionAccessMMUMiss>();
922 return std::make_shared<FastInstructionAccessMMUMiss>(
vaddr);
925 return std::make_shared<DataRealTranslationMiss>();
927 return std::make_shared<FastDataAccessMMUMiss>();
929 return std::make_shared<FastDataAccessMMUMiss>(
vaddr);
955 DPRINTF(IPR,
"Memory Mapped IPR Read: asi=%#X a=%#x\n",
956 (uint32_t)pkt->
req->getArchFlags(), pkt->
getAddr());
995 pkt->
setBE(itb->c0_tsb_ps0);
999 pkt->
setBE(itb->c0_tsb_ps1);
1003 pkt->
setBE(itb->c0_config);
1019 pkt->
setBE(itb->cx_tsb_ps0);
1023 pkt->
setBE(itb->cx_tsb_ps1);
1027 pkt->
setBE(itb->cx_config);
1030 pkt->
setBE((uint64_t)0);
1039 temp = itb->tag_access;
1043 pkt->
setBE(itb->sfsr);
1046 pkt->
setBE(itb->tag_access);
1049 goto doMmuReadError;
1071 goto doMmuReadError;
1126 panic(
"need to impl DTB::doMmuRegRead() got asi=%#x, va=%#x\n",
1138 ASI asi = (
ASI)pkt->
req->getArchFlags();
1144 int entry_insert = -1;
1151 DPRINTF(IPR,
"Memory Mapped IPR Write: asi=%#X a=%#x d=%#X\n",
1152 (uint32_t)asi, va, data);
1170 goto doMmuWriteError;
1174 assert(
mbits(data,13,6) == data);
1176 (va >> 4) - 0x3c, data);
1228 inform(
"Ignoring write to SPARC ERROR regsiter\n");
1240 sext<59>(
bits(data, 59,0));
1244 goto doMmuWriteError;
1248 entry_insert =
bits(va, 8,3);
1251 assert(entry_insert != -1 ||
mbits(va,10,9) == va);
1253 va_insert =
mbits(ta_insert, 63,13);
1254 ct_insert =
mbits(ta_insert, 12,0);
1256 real_insert =
bits(va, 9,9);
1259 itb->
insert(va_insert, part_insert, ct_insert, real_insert,
1263 entry_insert =
bits(va, 8,3);
1266 assert(entry_insert != -1 ||
mbits(va,10,9) == va);
1268 va_insert =
mbits(ta_insert, 63,13);
1269 ct_insert =
mbits(ta_insert, 12,0);
1271 real_insert =
bits(va, 9,9);
1273 PageTableEntry::sun4u);
1274 insert(va_insert, part_insert, ct_insert, real_insert, pte,
1281 switch (
bits(va,5,4)) {
1295 switch (
bits(va,7,6)) {
1308 panic(
"Invalid type for IMMU demap\n");
1317 sext<59>(
bits(data, 59,0));
1324 goto doMmuWriteError;
1331 switch (
bits(va,5,4)) {
1345 switch (
bits(va,7,6)) {
1358 panic(
"Invalid type for IMMU demap\n");
1376 postInterrupt(0,
bits(data, 5, 0), 0);
1380 panic(
"need to impl DTB::doMmuRegWrite() got asi=%#x, va=%#x d=%#x\n",
1416 uint64_t
c0_config, uint64_t cX_tsb, uint64_t cX_config)
1421 if (
bits(tag_access, 12,0) == 0) {
1429 uint64_t ptr =
mbits(tsb,63,13);
1430 bool split =
bits(tsb,12,12);
1431 int tsb_size =
bits(tsb,3,0);
1432 int page_size = (ps ==
Ps0) ?
bits(config, 2,0) :
bits(config,10,8);
1434 if (ps ==
Ps1 && split)
1435 ptr |=
ULL(1) << (13 + tsb_size);
1436 ptr |= (tag_access >> (9 + page_size * 3)) &
mask(12+tsb_size, 4);
1451 free_list.push_back(entry -
tlb);
1465 for (
int x = 0;
x <
size;
x++) {
1477 if (oldSize !=
size)
1478 panic(
"Don't support unserializing different sized TLBs\n");
1485 for (
int idx : free_list)
1498 for (
int x = 0;
x <
size;
x++) {
1511 SparcTLBParams::create()
void demapContext(int partition_id, int context_id)
Remove all entries that match a given context/partition id.
Addr translate(Addr vaddr) const
static void ignore(const char *expr)
#define panic(...)
This implements a cprintf based panic() function.
bool asiIsSparcError(ASI asi)
virtual void setMiscReg(RegIndex misc_reg, RegVal val)=0
void writeSfsr(bool write, ContextType ct, bool se, FaultTypes ft, int asi)
virtual System * getSystemPtr()=0
#define UNSERIALIZE_CONTAINER(member)
decltype(nullptr) constexpr NoFault
Cycles is a wrapper class for representing cycle counts, i.e.
iterator insert(TlbRange &r, TlbEntry *d)
#define fatal(...)
This implements a cprintf based fatal() function.
Declaration of a request, the overall memory request consisting of the parts of the request that are ...
std::shared_ptr< Request > RequestPtr
Fault translateData(const RequestPtr &req, ThreadContext *tc, bool write)
void clearInterrupt(ThreadID tid, int int_num, int index)
T read(Addr address) const
Read sizeof(T) bytes from address and return as object T.
bool FullSystem
The FullSystem variable can be used to determine the current mode of simulation.
virtual Process * getProcessPtr()=0
bool validVirtualAddress(Addr va, bool am)
Checks if the virtual address provided is a valid one.
uint64_t MakeTsbPtr(TsbPageSize ps, uint64_t tag_access, uint64_t c0_tsb, uint64_t c0_config, uint64_t cX_tsb, uint64_t cX_config)
virtual BaseCPU * getCpuPtr()=0
TlbEntry * lookup(Addr va, int partition_id, bool real, int context_id=0, bool update_used=true)
lookup an entry in the TLB based on the partition id, and real bit if real is true or the partition i...
void setBE(T v)
Set the value in the data pointer to v as big endian.
ThreadContext is the external interface to all thread state for anything outside of the CPU...
Fault translateAtomic(const RequestPtr &req, ThreadContext *tc, Mode mode) override
Cycles doMmuRegRead(ThreadContext *tc, Packet *pkt)
The request is to an uncacheable address.
void unserialize(CheckpointIn &cp) override
Unserialize an object.
void populate(uint64_t e, EntryType t=sun4u)
RequestPtr req
A pointer to the original request.
#define UNSERIALIZE_SCALAR(scalar)
Cycles doMmuRegWrite(ThreadContext *tc, Packet *pkt)
virtual PortProxy & getPhysProxy()=0
std::list< TlbEntry * > freeList
std::string csprintf(const char *format, const Args &...args)
void demapPage(Addr va, int partition_id, bool real, int context_id)
Remve all entries that match a certain partition id, (contextid), and va).
bool translate(Addr vaddr, Addr &paddr)
Translate function.
const Addr StartVAddrHole
void makeAtomicResponse()
uint64_t get_vec(int int_num)
void flushAll() override
Remove all entries from the TLB.
void translateTiming(const RequestPtr &req, ThreadContext *tc, Translation *translation, Mode mode) override
std::vector< ThreadContext * > threadContexts
bool asiIsPrimary(ASI asi)
virtual BaseTLB * getITBPtr()=0
Fault translateFunctional(const RequestPtr &req, ThreadContext *tc, Mode mode) override
virtual void finish(const Fault &fault, const RequestPtr &req, ThreadContext *tc, Mode mode)=0
uint64_t Addr
Address type This will probably be moved somewhere else in the near future.
#define ULL(N)
uint64_t constant
Fault finalizePhysical(const RequestPtr &req, ThreadContext *tc, Mode mode) const override
Do post-translation physical address finalization.
A Packet is used to encapsulate a transfer between two objects in the memory system (e...
void unserialize(CheckpointIn &cp)
void writeTagAccess(Addr va, int context)
bool asiIsScratchPad(ASI asi)
#define SERIALIZE_SCALAR(scalar)
This object is a proxy for a port or other object which implements the functional response protocol...
void insert(Addr vpn, int partition_id, int context_id, bool real, const PageTableEntry &PTE, int entry=-1)
Insert a PTE into the TLB.
bool asiIsNoFault(ASI asi)
EmulationPageTable * pTable
virtual RegVal readMiscRegNoEffect(RegIndex misc_reg) const =0
Declarations of a non-full system Page Table.
#define SERIALIZE_CONTAINER(member)
void GetTsbPtr(ThreadContext *tc, Addr addr, int ctx, Addr *ptrs)
std::ostream CheckpointOut
uint64_t TagRead(int entry)
Given an entry id, read that tlb entries' tag.
void serialize(CheckpointOut &cp) const override
Serialize an object.
bool asiIsUnPriv(ASI asi)
iterator find(const TlbRange &r)
uint64_t TteRead(int entry)
Give an entry id, read that tlb entries' tte.
BaseInterrupts * getInterruptController(ThreadID tid)
int findMsbSet(uint64_t val)
Returns the bit position of the MSB that is set in the input.
bool asiIsInterrupt(ASI asi)
void paramIn(CheckpointIn &cp, const string &name, ExtMachInst &machInst)
bool asiIsAsIfUser(ASI asi)
The request is required to be strictly ordered by CPU models and is non-speculative.
bool asiIsLittle(ASI asi)
T mbits(T val, int first, int last)
Mask off the given bits in place like bits() but without shifting.
T bits(T val, int first, int last)
Extract the bitfield from position 'first' to 'last' (inclusive) from 'val' and right justify it...
void demapAll(int partition_id)
Remove all non-locked entries from the tlb that match partition id.
Scoped checkpoint section helper class.
bool asiIsNucleus(ASI asi)
bool asiIsSecondary(ASI asi)
Fault translateInst(const RequestPtr &req, ThreadContext *tc)
virtual RegVal readMiscReg(RegIndex misc_reg)=0
std::shared_ptr< FaultBase > Fault
T getBE() const
Get the data in the packet byte swapped from big endian to host endian.
void serialize(CheckpointOut &cp) const
bool asiIsPartialStore(ASI asi)