Go to the documentation of this file.
48 #include "debug/TLB.hh"
49 #include "debug/TLBVerbose.hh"
57 using namespace ArmISA;
61 itbStage2(
p.stage2_itb), dtbStage2(
p.stage2_dtb),
62 itbWalker(
p.itb_walker), dtbWalker(
p.dtb_walker),
63 itbStage2Walker(
p.stage2_itb_walker),
64 dtbStage2Walker(
p.stage2_dtb_walker),
67 s1State(this, false), s2State(this, true),
114 if (
static_cast<TLB*
>(
tlb)->walkCache())
118 if (
static_cast<TLB*
>(
tlb)->walkCache())
122 if (
static_cast<TLB*
>(
tlb)->walkCache())
185 lookup_data.
inHost =
false;
206 const Addr paddr = req->getPaddr();
211 req->setLocalAccessor(
216 pseudo_inst::pseudoInst<RegABI64>(tc, func, ret);
218 pseudo_inst::pseudoInst<RegABI32>(tc, func, ret);
234 Translation *translation,
bool &delay,
bool timing,
238 Addr vaddr_tainted = req->getVaddr();
244 vaddr = vaddr_tainted;
254 return std::make_shared<DataAbort>(
266 if (!
p->pTable->translate(
vaddr, paddr))
267 return std::make_shared<GenericPageTableFault>(vaddr_tainted);
268 req->setPaddr(paddr);
286 if (req->isCacheMaintenance()) {
304 if (state.
isStage2 && req->isPTWalk() && state.
hcr.ptw &&
306 return std::make_shared<DataAbort>(
318 return std::make_shared<DataAbort>(
326 if (
te->nonCacheable) {
328 if (req->isPrefetch()) {
331 return std::make_shared<PrefetchAbort>(
337 if (!
te->longDescFormat) {
338 switch ((state.
dacr >> (
static_cast<uint8_t
>(
te->domain) * 2)) & 0x3) {
341 DPRINTF(
TLB,
"TLB Fault: Data abort on domain. DACR: %#x"
342 " domain: %#x write:%d\n", state.
dacr,
343 static_cast<uint8_t
>(
te->domain), is_write);
348 return std::make_shared<PrefetchAbort>(
353 return std::make_shared<DataAbort>(
361 panic(
"UNPRED domain\n");
368 uint8_t ap =
te->longDescFormat ?
te->ap << 1 :
te->ap;
369 uint8_t hap =
te->hap;
371 if (state.
sctlr.afe == 1 ||
te->longDescFormat)
375 bool isWritable =
true;
384 DPRINTF(
TLB,
"Access permissions 0, checking rs:%#x\n",
385 (
int)state.
sctlr.rs);
386 if (!state.
sctlr.xp) {
387 switch ((
int)state.
sctlr.rs) {
392 abt = is_write || !is_priv;
408 abt = !is_priv && is_write;
409 isWritable = is_priv;
415 panic(
"UNPRED premissions\n");
417 abt = !is_priv || is_write;
426 panic(
"Unknown permissions %#x\n", ap);
430 bool hapAbt = is_write ? !(hap & 2) : !(hap & 1);
431 bool xn =
te->xn || (isWritable && state.
sctlr.wxn) ||
432 (ap == 3 && state.
sctlr.uwxn && is_priv);
433 if (is_fetch && (abt || xn ||
434 (
te->longDescFormat &&
te->pxn && is_priv) ||
437 DPRINTF(
TLB,
"TLB Fault: Prefetch abort on permission check. AP:%d "
438 "priv:%d write:%d ns:%d sif:%d sctlr.afe: %d \n",
439 ap, is_priv, is_write,
te->ns,
443 return std::make_shared<PrefetchAbort>(
447 }
else if (abt | hapAbt) {
449 DPRINTF(
TLB,
"TLB Fault: Data abort on permission check. AP:%d priv:%d"
450 " write:%d\n", ap, is_priv, is_write);
451 return std::make_shared<DataAbort>(
481 Addr vaddr_tainted = req->getVaddr();
488 bool is_write = !req->isCacheClean() &&
mode ==
Write;
489 bool is_atomic = req->isAtomic();
497 if (state.
isStage2 && req->isPTWalk() && state.
hcr.ptw &&
499 return std::make_shared<DataAbort>(
500 vaddr_tainted,
te->domain, is_write,
511 return std::make_shared<DataAbort>(
514 is_atomic ?
false : is_write,
521 if (
te->nonCacheable) {
523 if (req->isPrefetch()) {
526 return std::make_shared<PrefetchAbort>(
538 bool grant_read =
true;
542 (!is_write && !is_fetch), is_write, is_fetch);
545 (!is_write && !is_fetch), is_write, is_fetch);
551 DPRINTF(
TLB,
"TLB Fault: Prefetch abort on permission check. "
552 "ns:%d scr.sif:%d sctlr.afe: %d\n",
556 return std::make_shared<PrefetchAbort>(
562 DPRINTF(
TLB,
"TLB Fault: Data abort on permission check."
564 return std::make_shared<DataAbort>(
565 vaddr_tainted,
te->domain,
566 (is_atomic && !grant_read) ?
false : is_write,
585 bool grant_read =
te->hap & 0b01;
586 bool grant_write =
te->hap & 0
b10;
589 uint8_t pxn =
te->pxn;
592 te->ns && state.
scr.sif) {
597 "Checking S2 permissions: hap:%d, xn:%d, pxn:%d, r:%d, "
598 "w:%d, x:%d\n",
te->hap, xn, pxn,
r,
w,
x);
601 grant = grant_read && !xn;
602 }
else if (req->isAtomic()) {
603 grant = grant_read || grant_write;
609 panic(
"Invalid Operation\n");
612 return std::make_pair(grant, grant_read);
619 bool grant =
false, grant_read =
true;
621 const uint8_t ap =
te->ap & 0
b11;
622 const bool is_priv = state.
isPriv && !(req->getFlags() &
UserMode);
626 uint8_t pxn =
te->pxn;
629 te->ns && state.
scr.sif) {
633 DPRINTF(TLBVerbose,
"Checking S1 permissions: ap:%d, xn:%d, pxn:%d, r:%d, "
634 "w:%d, x:%d, is_priv: %d, wxn: %d\n", ap, xn,
635 pxn,
r,
w,
x, is_priv,
wxn);
638 return std::make_pair(
false,
false);
645 grant_read = ap & 0x1;
646 uint8_t
perm = (ap << 2) | (xn << 1) | pxn;
656 grant =
r ||
w || (
x && !
wxn);
677 uint8_t
perm = (ap << 2) | (xn << 1) | pxn;
681 grant =
r ||
w || (
x && !
wxn);
713 uint8_t
perm = (ap & 0x2) | xn;
716 grant =
r ||
w || (
x && !
wxn);
734 return std::make_pair(grant, grant_read);
741 bool exception =
false;
775 if (mmfr1.pan && state.
cpsr.pan && (ap & 0x1) &&
778 if (req->isCacheMaintenance() &&
782 }
else if (!is_priv && !(state.
hcr.e2h && !state.
hcr.tge)) {
798 bool is_atomic = req->isAtomic();
799 req->setPaddr(
vaddr);
813 f = std::make_shared<PrefetchAbort>(
vaddr,
817 f = std::make_shared<DataAbort>(
vaddr,
827 if (long_desc_format || state.
sctlr.tre == 0 || state.
nmrr.ir0 == 0 ||
828 state.
nmrr.or0 == 0 || state.
prrr.tr0 != 0x2) {
829 if (!req->isCacheMaintenance()) {
839 && state.
hcr.e2h == 1 && state.
hcr.tge == 1) ? 0: state.
hcr.dc;
840 bool i_cacheability = state.
sctlr.i && !state.
sctlr.m;
846 temp_te.
innerAttrs = i_cacheability? 0x2: 0x0;
847 temp_te.
outerAttrs = i_cacheability? 0x2: 0x0;
858 DPRINTF(TLBVerbose,
"(No MMU) setting memory attributes: shareable: "
859 "%d, innerAttrs: %d, outerAttrs: %d, stage2: %d\n",
869 Translation *translation,
bool &delay,
bool timing,
878 Addr vaddr_tainted = req->getVaddr();
881 functional, &mergeTe, state);
890 "Setting memory attributes: shareable: %d, innerAttrs: %d, "
891 "outerAttrs: %d, mtype: %d, stage2: %d\n",
892 te->shareable,
te->innerAttrs,
te->outerAttrs,
893 static_cast<uint8_t
>(
te->mtype), state.
isStage2);
896 if (
te->nonCacheable && !req->isCacheMaintenance())
911 if (!is_fetch && fault ==
NoFault &&
918 return std::make_shared<DataAbort>(
941 Translation *translation,
bool &delay,
bool timing,
946 assert(!(timing && functional));
948 Addr vaddr_tainted = req->getVaddr();
954 vaddr = vaddr_tainted;
964 "CPSR is priv:%d UserMode:%d secure:%d S1S2NsTran:%d\n",
968 DPRINTF(
TLB,
"translateFs addr %#x, mode %d, st2 %d, scr %#x sctlr %#x "
969 "flags %#lx tranType 0x%x\n", vaddr_tainted,
mode,
973 if ((req->isInstFetch() && (!state.
sctlr.i)) ||
974 ((!req->isInstFetch()) && (!state.
sctlr.c))){
975 if (!req->isCacheMaintenance()) {
985 return std::make_shared<DataAbort>(
994 bool vm = state.
hcr.vm;
997 else if (state.
hcr.dc == 1)
1004 long_desc_format, state);
1006 DPRINTF(TLBVerbose,
"Translating %s=%#x context=%d\n",
1007 state.
isStage2 ?
"IPA" :
"VA", vaddr_tainted, state.
asid);
1010 functional,
vaddr, tranMethod, state);
1016 if (
sd->enabled() && fault ==
NoFault) {
1017 fault =
sd->testDebug(tc, req,
mode);
1040 tran_type,
false, state);
1070 tran_type,
true, state);
1084 assert(translation);
1107 fault =
translateFs(req, tc,
mode, translation, delay,
true, tran_type,
1112 DPRINTF(TLBVerbose,
"Translation returning delay=%d fault=%d\n", delay,
1121 if (translation && (call_from_s2 || !state.
stage2Req || req->hasPaddr() ||
1138 switch (mmfr1.vmidbits) {
1153 panic(
"Reserved ID_AA64MMFR1_EL1.VMIDBits value: %#x",
1170 ((tran_type == state.curTranType) || stage2)) {
1173 DPRINTF(TLBVerbose,
"TLB variables changed!\n");
1174 state.updateMiscReg(tc, tran_type);
1180 static_cast<TLB*
>(
tlb)->setVMID(state.vmid);
1183 static_cast<TLB*
>(
tlb)->setVMID(state.vmid);
1186 static_cast<TLB*
>(
tlb)->setVMID(state.vmid);
1192 if (state.directToStage2) {
1218 switch (aarch64EL) {
1224 uint64_t ttbr_asid = ttbcr.a1 ?
1228 (mmu->haveLargeAsid64 && ttbcr.as) ? 63 : 55, 48);
1233 uint64_t ttbr_asid = ttbcr.a1 ?
1237 (mmu->haveLargeAsid64 && ttbcr.as) ? 63 : 55, 48);
1245 uint64_t ttbr_asid = ttbcr.a1 ?
1249 (mmu->haveLargeAsid64 && ttbcr.as) ? 63 : 55, 48);
1257 uint64_t ttbr_asid = ttbcr.a1 ?
1261 (mmu->haveLargeAsid64 && ttbcr.as) ? 63 : 55, 48);
1274 isPriv = aarch64EL !=
EL0;
1275 if (mmu->release()->has(ArmExtension::VIRTUALIZATION)) {
1277 isHyp = aarch64EL ==
EL2;
1280 isHyp &= (tran_type &
S1CTran) == 0;
1286 if (hcr.e2h == 1 && (aarch64EL ==
EL2
1287 || (hcr.tge ==1 && aarch64EL ==
EL0))) {
1289 directToStage2 =
false;
1291 stage2DescReq =
false;
1297 stage2Req = isStage2 ||
1298 (
vm && !isHyp && sec &&
1299 !(tran_type &
S1CTran) && (aarch64EL <
EL2) &&
1301 stage2DescReq = isStage2 || (
vm && !isHyp && sec &&
1303 directToStage2 = !isStage2 && stage2Req && !sctlr.m;
1308 directToStage2 =
false;
1310 stage2DescReq =
false;
1328 asid = context_id.asid;
1338 if (mmu->release()->has(ArmExtension::VIRTUALIZATION)) {
1343 isHyp &= (tran_type &
S1CTran) == 0;
1351 stage2Req = hcr.vm && !isStage2 && !isHyp && sec &&
1353 stage2DescReq = hcr.vm && !isStage2 && !isHyp && sec;
1354 directToStage2 = stage2Req && !sctlr.m;
1359 directToStage2 =
false;
1360 stage2DescReq =
false;
1363 miscRegValid =
true;
1364 curTranType = tran_type;
1392 panic(
"Unknown translation mode!\n");
1398 Translation *translation,
bool timing,
bool functional,
1402 return getTE(
te, req, tc,
mode, translation, timing, functional,
1415 lookup_data.
va =
va;
1418 lookup_data.
vmid = vmid;
1419 lookup_data.
hyp = hyp;
1420 lookup_data.
secure = secure;
1423 lookup_data.
inHost = in_host;
1426 return tlb->multiLookup(lookup_data);
1431 Translation *translation,
bool timing,
bool functional,
1441 Addr vaddr_tainted = req->getVaddr();
1448 vaddr = vaddr_tainted;
1455 if (req->isPrefetch()) {
1460 return std::make_shared<PrefetchAbort>(
1467 "TLB Miss: Starting hardware table walker for %#x(%d:%d)\n",
1468 vaddr_tainted, state.
asid, state.
vmid);
1473 translation, timing, functional, is_secure,
1477 if (timing || fault !=
NoFault) {
1491 Translation *translation,
bool timing,
bool functional,
1501 fault =
getTE(&s2_te, req, tc,
mode, translation, timing, functional,
1516 Addr vaddr_tainted = req->getVaddr();
1519 fault =
getTE(&s1_te, req, tc,
mode, translation, timing, functional,
1530 req, translation,
mode, timing, functional, state.
isSecure,
1532 fault = s2_lookup->
getTe(tc, mergeTe);
1547 DPRINTF(TLBVerbose,
"s2TLB: reqVa %#x, reqPa %#x, fault %p\n",
1548 vaddr_tainted, req->hasPaddr() ? req->getPaddr() : ~0,
1551 auto arm_fault =
reinterpret_cast<ArmFault*
>(fault.get());
1565 return entry && !entry->
partial;
1573 auto *ommu =
dynamic_cast<MMU*
>(old_mmu);
1576 _attr = ommu->_attr;
1589 fatal_if(!
ti,
"%s is not a valid ARM TLB tester\n", _ti->
name());
1598 if (!
test || !req->hasSize() || req->getSize() == 0 ||
1599 req->isCacheMaintenance()) {
1629 : statistics::
Group(parent),
1630 ADD_STAT(alignFaults, statistics::units::Count::get(),
1631 "Number of MMU faults due to alignment restrictions"),
1632 ADD_STAT(prefetchFaults, statistics::units::Count::get(),
1633 "Number of MMU faults due to prefetch"),
1634 ADD_STAT(domainFaults, statistics::units::Count::get(),
1635 "Number of MMU faults due to domain restrictions"),
1636 ADD_STAT(permsFaults, statistics::units::Count::get(),
1637 "Number of MMU faults due to permissions restrictions")
uint8_t physAddrRange() const
Returns the supported physical address range in bits.
Fault finalizePhysical(const RequestPtr &req, ThreadContext *tc, Mode mode) const override
virtual RegVal readMiscReg(RegIndex misc_reg)=0
constexpr decltype(nullptr) NoFault
bool ELIs64(ThreadContext *tc, ExceptionLevel el)
Addr start() const
Get the start address of the range.
enums::ArmLookupLevel LookupLevel
std::pair< bool, bool > s2PermBits64(TlbEntry *te, const RequestPtr &req, Mode mode, ThreadContext *tc, CachedState &state, bool r, bool w, bool x)
statistics::Scalar domainFaults
Fault getTE(TlbEntry **te, const RequestPtr &req, ThreadContext *tc, Mode mode, Translation *translation, bool timing, bool functional, bool is_secure, ArmTranslationType tran_type, bool stage2)
bool contains(const Addr &a) const
Determine if the range contains an address.
TableWalker * dtbStage2Walker
virtual void markDelayed()=0
Signal that the translation has been delayed due to a hw page table walk.
void updateMiscReg(ThreadContext *tc, ArmTranslationType tran_type)
@ SECURE
The request targets the secure memory space.
bool haveLargeAsid64() const
Returns true if ASID is 16 bits in AArch64 (ARMv8)
virtual ContextID contextId() const =0
Fault getResultTe(TlbEntry **te, const RequestPtr &req, ThreadContext *tc, Mode mode, Translation *translation, bool timing, bool functional, TlbEntry *mergeTe, CachedState &state)
void init() override
Called at init time, this method is traversing the TLB hierarchy and pupulating the instruction/data/...
void setAttributes(bool lpae)
bool inAArch64(ThreadContext *tc)
ArmISA::TLB * getDTBPtr() const
Fault walk(const RequestPtr &req, ThreadContext *tc, uint16_t asid, vmid_t _vmid, bool hyp, BaseMMU::Mode mode, BaseMMU::Translation *_trans, bool timing, bool functional, bool secure, MMU::ArmTranslationType tran_type, bool stage2, const TlbEntry *walk_entry)
Fault testWalk(Addr pa, Addr size, Addr va, bool is_secure, Mode mode, TlbEntry::DomainType domain, LookupLevel lookup_level, bool stage2)
MMU(const ArmMMUParams &p)
void init() override
init() is called after all C++ SimObjects have been created and all ports are connected.
statistics::Scalar permsFaults
ArmTranslationType curTranType
TableWalker * itbStage2Walker
Fault translateFs(const RequestPtr &req, ThreadContext *tc, Mode mode, Translation *translation, bool &delay, bool timing, ArmTranslationType tran_type, bool functional, CachedState &state)
Cycles is a wrapper class for representing cycle counts, i.e.
Addr purifyTaggedAddr(Addr addr, ThreadContext *tc, ExceptionLevel el, TCR tcr, bool is_instr)
Removes the tag from tagged addresses if that mode is enabled.
Fault translateComplete(const RequestPtr &req, ThreadContext *tc, Translation *translation, Mode mode, ArmTranslationType tran_type, bool call_from_s2)
TableWalker * getTableWalker(BaseMMU::Mode mode, bool stage2) const
statistics::Scalar prefetchFaults
ThreadContext is the external interface to all thread state for anything outside of the CPU.
void setTestInterface(SimObject *ti)
virtual std::string name() const
std::shared_ptr< FaultBase > Fault
bool HaveVirtHostExt(ThreadContext *tc)
std::set< BaseTLB * > unified
#define ADD_STAT(n,...)
Convenience macro to add a stat to a statistics group.
A Packet is used to encapsulate a transfer between two objects in the memory system (e....
static ExceptionLevel tranTypeEL(CPSR cpsr, ArmTranslationType type)
Determine the EL to use for the purpose of a translation given a specific translation type.
Fault translateSe(const RequestPtr &req, ThreadContext *tc, Mode mode, Translation *translation, bool &delay, bool timing, CachedState &state)
std::shared_ptr< Request > RequestPtr
Fault translateAtomic(const RequestPtr &req, ThreadContext *tc, Mode mode) override
void drainResume() override
Resume execution after a successful drain.
constexpr T bits(T val, unsigned first, unsigned last)
Extract the bitfield from position 'first' to 'last' (inclusive) from 'val' and right justify it.
Abstract superclass for simulation objects.
static void decodeAddrOffset(Addr offset, uint8_t &func)
@ MISCREG_ID_AA64MMFR1_EL1
Fault translateMmuOff(ThreadContext *tc, const RequestPtr &req, Mode mode, ArmTranslationType tran_type, Addr vaddr, bool long_desc_format, CachedState &state)
uint64_t Addr
Address type This will probably be moved somewhere else in the near future.
void setVMID(vmid_t _vmid)
bool isSecure(ThreadContext *tc)
bool checkWalkCache() const
std::set< BaseTLB * > instruction
It is possible from the MMU to traverse the entire hierarchy of TLBs, starting from the DTB and ITB (...
@ CACHE_BLOCK_ZERO
This is a write that is targeted and zeroing an entire cache block.
ExceptionLevel currEL(const ThreadContext *tc)
Returns the current Exception Level (EL) of the provided ThreadContext.
bool faultPAN(ThreadContext *tc, uint8_t ap, const RequestPtr &req, Mode mode, const bool is_priv, CachedState &state)
virtual Process * getProcessPtr()=0
ArmISA::TLB * getITBPtr() const
bool FullSystem
The FullSystem variable can be used to determine the current mode of simulation.
TLB * getTlb(BaseMMU::Mode mode, bool stage2) const
void translateTiming(const RequestPtr &req, ThreadContext *tc, Translation *translation, Mode mode) override
bool IsSecureEL2Enabled(ThreadContext *tc)
Fault getTe(ThreadContext *tc, TlbEntry *destTe)
@ STRICT_ORDER
The request is required to be strictly ordered by CPU models and is non-speculative.
Fault testTranslation(const RequestPtr &req, Mode mode, TlbEntry::DomainType domain, CachedState &state)
TranslationGenPtr translateFunctional(Addr start, Addr size, ThreadContext *tc, Mode mode, Request::Flags flags) override
Returns a translation generator for a region of virtual addresses, instead of directly translating a ...
void setTableWalker(TableWalker *table_walker)
bool checkPAN(ThreadContext *tc, uint8_t ap, const RequestPtr &req, Mode mode, const bool is_priv, CachedState &state)
SelfDebug * getSelfDebug() const
vmid_t getVMID(ThreadContext *tc) const
Returns the current VMID (information stored in the VTTBR_EL2 register)
Fault checkPermissions(TlbEntry *te, const RequestPtr &req, Mode mode, bool stage2)
bool longDescFormatInUse(ThreadContext *tc)
virtual void annotate(AnnotationIDs id, uint64_t val)
@ UNCACHEABLE
The request is to an uncacheable address.
CachedState & updateMiscReg(ThreadContext *tc, ArmTranslationType tran_type, bool stage2)
const ArmRelease * releaseFS() const
void setAttr(uint64_t attr)
Accessor functions for memory attributes for last accessed TLB entry.
int snsBankedIndex(MiscRegIndex reg, ThreadContext *tc)
void setLE(T v)
Set the value in the data pointer to v as little endian.
void takeOverFrom(BaseMMU *old_mmu) override
static bool haveEL(ThreadContext *tc, ArmISA::ExceptionLevel el)
Return true if the system implements a specific exception level.
#define fatal_if(cond,...)
Conditional fatal macro that checks the supplied condition and only causes a fatal error if the condi...
Fault translateMmuOn(ThreadContext *tc, const RequestPtr &req, Mode mode, Translation *translation, bool &delay, bool timing, bool functional, Addr vaddr, ArmFault::TranMethod tranMethod, CachedState &state)
Reference material can be found at the JEDEC website: UFS standard http://www.jedec....
int computeAddrTop(ThreadContext *tc, bool selbit, bool is_instr, TCR tcr, ExceptionLevel el)
const ArmRelease * _release
virtual void takeOverFrom(BaseMMU *old_mmu)
gem5::ArmISA::MMU::Stats stats
bool isCompleteTranslation(TlbEntry *te) const
virtual void finish(const Fault &fault, const RequestPtr &req, ThreadContext *tc, BaseMMU::Mode mode)=0
Stats(statistics::Group *parent)
statistics::Scalar alignFaults
TlbEntry * lookup(Addr vpn, uint16_t asn, vmid_t vmid, bool hyp, bool secure, bool functional, bool ignore_asn, ExceptionLevel target_el, bool in_host, bool stage2, BaseMMU::Mode mode)
Lookup an entry in the TLB.
std::set< BaseTLB * > data
#define panic(...)
This implements a cprintf based panic() function.
Fault checkPermissions64(TlbEntry *te, const RequestPtr &req, Mode mode, ThreadContext *tc, bool stage2)
std::pair< bool, bool > s1PermBits64(TlbEntry *te, const RequestPtr &req, Mode mode, ThreadContext *tc, CachedState &state, bool r, bool w, bool x)
Generated on Wed May 4 2022 12:13:48 for gem5 by doxygen 1.8.17