51#include "debug/Checkpoint.hh"
52#include "debug/Drain.hh"
53#include "debug/PageTableWalker.hh"
54#include "debug/TLB.hh"
55#include "debug/TLBVerbose.hh"
75 doL2DescEvent([
this]{ doL2DescriptorWrapper(); },
name()),
76 doL0LongDescEvent([
this]{ doL0LongDescriptorWrapper(); },
name()),
77 doL1LongDescEvent([
this]{ doL1LongDescriptorWrapper(); },
name()),
78 doL2LongDescEvent([
this]{ doL2LongDescriptorWrapper(); },
name()),
79 doL3LongDescEvent([
this]{ doL3LongDescriptorWrapper(); },
name()),
80 LongDescEventByLevel { &doL0LongDescEvent, &doL1LongDescEvent,
81 &doL2LongDescEvent, &doL3LongDescEvent },
82 doProcessEvent([
this]{ processWalkWrapper(); },
name()),
89 ArmSystem *arm_sys =
dynamic_cast<ArmSystem *
>(
p.sys);
91 _physAddrRange = arm_sys->physAddrRange();
92 _haveLargeAsid64 = arm_sys->haveLargeAsid64();
94 _haveLargeAsid64 =
false;
114 if (if_name ==
"port") {
162 state->delay = delay;
206 assert(pkt->
req->isUncacheable() ||
249 DPRINTF(Drain,
"TableWalker done draining, processing drain event\n");
257 bool state_queues_not_empty =
false;
259 for (
int i = 0;
i < LookupLevel::Num_ArmLookupLevel; ++
i) {
261 state_queues_not_empty =
true;
267 DPRINTF(Drain,
"TableWalker not drained\n");
270 DPRINTF(Drain,
"TableWalker free, no need to drain\n");
288 bool disable_cacheability =
isStage2 ?
291 return disable_cacheability ||
currState->isUncacheable;
300 bool _stage2Req,
const TlbEntry *walk_entry)
302 assert(!(_functional && _timing));
311 DPRINTF(PageTableWalker,
"creating new instance of WalkerState\n");
315 }
else if (_functional) {
320 "creating functional instance of WalkerState\n");
324 }
else if (_timing) {
331 if (
currState->vaddr_tainted == _req->getVaddr()) {
332 ++
stats.squashedBefore;
333 return std::make_shared<ReExec>();
402 assert(
release->has(ArmExtension::VIRTUALIZATION));
407 assert(
release->has(ArmExtension::SECURITY));
412 panic(
"Invalid translation regime");
439 if (long_desc_format) {
443 currState->longDescData->userTable =
true;
444 currState->longDescData->xnTable =
false;
445 currState->longDescData->pxnTable =
false;
446 ++
stats.walksLongDescriptor;
449 ++
stats.walksShortDescriptor;
466 }
else if (long_desc_format) {
488 currState->longDesc.lookupLevel : LookupLevel::L1;
520 if (!
currState->transState->squashed() && (!
te ||
te->partial)) {
528 if (
te &&
te->partial) {
534 }
else if (long_desc_format) {
551 currState->longDesc.lookupLevel : LookupLevel::L1;
561 unsigned num_squashed = 0;
565 (
te && !
te->partial))) {
568 stats.squashedBefore++;
570 DPRINTF(
TLB,
"Squashing table walk for address %#x\n",
576 std::make_shared<UnimpFault>(
"Squashed Inst"),
618 const auto irgn0_mask = 0x1;
619 const auto irgn1_mask = 0x40;
620 currState->isUncacheable = (ttbr1 & (irgn0_mask | irgn1_mask)) == 0;
624 const bool is_atomic =
currState->req->isAtomic();
625 const bool have_security =
release->has(ArmExtension::SECURITY);
627 DPRINTF(
TLB,
"Beginning table walk for address %#x, TTBCR: %#x, bits:%#x\n",
637 if (have_security &&
currState->ttbcr.pd0) {
639 return std::make_shared<PrefetchAbort>(
645 return std::make_shared<DataAbort>(
658 if (have_security &&
currState->ttbcr.pd1) {
660 return std::make_shared<PrefetchAbort>(
666 return std::make_shared<DataAbort>(
679 DPRINTF(
TLB,
" - Descriptor at address %#x (%s)\n", l1desc_addr,
693 sizeof(uint32_t), flag, LookupLevel::L1,
703 Addr ttbr, ttbr0_max, ttbr1_min, desc_addr;
707 DPRINTF(
TLB,
"Beginning table walk for address %#x, TTBCR: %#x\n",
719 DPRINTF(
TLB,
" - Selecting VTTBR (long-desc.)\n");
722 start_lookup_level =
currState->vtcr.sl0 ?
723 LookupLevel::L1 : LookupLevel::L2;
726 DPRINTF(
TLB,
" - Selecting HTTBR (long-desc.)\n");
735 ttbr0_max = (1ULL << (32 -
currState->ttbcr.t0sz)) - 1;
737 ttbr0_max = (1ULL << 32) -
738 (1ULL << (32 -
currState->ttbcr.t1sz)) - 1;
740 ttbr0_max = (1ULL << 32) - 1;
742 ttbr1_min = (1ULL << 32) - (1ULL << (32 -
currState->ttbcr.t1sz));
744 ttbr1_min = (1ULL << (32 -
currState->ttbcr.t0sz));
746 const bool is_atomic =
currState->req->isAtomic();
753 DPRINTF(
TLB,
" - Selecting TTBR0 (long-desc.)\n");
757 return std::make_shared<PrefetchAbort>(
763 return std::make_shared<DataAbort>(
776 if (ttbr0_max < (1ULL << 30))
777 start_lookup_level = LookupLevel::L2;
778 }
else if (
currState->vaddr >= ttbr1_min) {
779 DPRINTF(
TLB,
" - Selecting TTBR1 (long-desc.)\n");
783 return std::make_shared<PrefetchAbort>(
789 return std::make_shared<DataAbort>(
803 if (ttbr1_min >= (1ULL << 31) + (1ULL << 30))
804 start_lookup_level = LookupLevel::L2;
808 return std::make_shared<PrefetchAbort>(
814 return std::make_shared<DataAbort>(
825 if (start_lookup_level == LookupLevel::L1) {
827 desc_addr =
mbits(ttbr, 39,
n) |
829 DPRINTF(
TLB,
" - Descriptor at address %#x (%s) (long-desc.)\n",
834 n = (tsz >= 2 ? 14 - tsz : 12);
835 desc_addr =
mbits(ttbr, 39,
n) |
837 DPRINTF(
TLB,
" - Descriptor at address %#x (%s) (long-desc.)\n",
846 currState->longDesc.lookupLevel = start_lookup_level;
853 sizeof(uint64_t), flag, start_lookup_level,
901 return tsz > max_txsz || tsz < min_txsz;
924 DPRINTF(
TLB,
"Beginning table walk for address %#llx, TCR: %#llx\n",
941 bool vaddr_fault =
false;
948 DPRINTF(
TLB,
" - Selecting VSTTBR_EL2 (AArch64 stage 2)\n");
953 DPRINTF(
TLB,
" - Selecting VTTBR_EL2 (AArch64 stage 2)\n");
969 DPRINTF(
TLB,
" - Selecting TTBR0_EL1 (AArch64)\n");
984 DPRINTF(
TLB,
" - Selecting TTBR1_EL1 (AArch64)\n");
1009 DPRINTF(
TLB,
" - Selecting TTBR0_EL2 (AArch64)\n");
1026 DPRINTF(
TLB,
" - Selecting TTBR1_EL2 (AArch64)\n");
1050 DPRINTF(
TLB,
" - Selecting TTBR0_EL3 (AArch64)\n");
1075 const bool is_atomic =
currState->req->isAtomic();
1079 return std::make_shared<PrefetchAbort>(
1084 return std::make_shared<DataAbort>(
1094 warn_once(
"Reserved granule size requested; gem5's IMPLEMENTATION "
1095 "DEFINED behavior takes this to mean 4KB granules\n");
1107 auto [table_addr, desc_addr, start_lookup_level] =
walkAddresses(
1108 ttbr, tg, tsz, pa_range);
1113 DPRINTF(
TLB,
"Address size fault before any lookup\n");
1115 return std::make_shared<PrefetchAbort>(
1121 return std::make_shared<DataAbort>(
1139 currState->longDesc.lookupLevel = start_lookup_level;
1146 sizeof(uint64_t), flag, start_lookup_level,
1153std::tuple<Addr, Addr, TableWalker::LookupLevel>
1158 LookupLevel first_level = LookupLevel::Num_ArmLookupLevel;
1159 Addr table_addr = 0;
1166 "Walk Cache hit: va=%#x, level=%d, table address=%#x\n",
1169 if (
currState->longDescData.has_value()) {
1176 table_addr = entry->
pfn;
1181 ptops->firstS2Level(
currState->vtcr.sl0) :
1182 ptops->firstLevel(64 - tsz);
1183 panic_if(first_level == LookupLevel::Num_ArmLookupLevel,
1184 "Table walker couldn't find lookup level\n");
1187 int base_addr_lo = 3 + tsz -
stride * (3 - first_level) - tg;
1189 if (pa_range == 52) {
1190 int z = (base_addr_lo < 6) ? 6 : base_addr_lo;
1191 table_addr =
mbits(ttbr, 47,
z);
1192 table_addr |= (
bits(ttbr, 5, 2) << 48);
1194 table_addr =
mbits(ttbr, 47, base_addr_lo);
1198 desc_addr = table_addr + ptops->index(
currState->vaddr, first_level, tsz);
1200 return std::make_tuple(table_addr, desc_addr, first_level);
1205 uint8_t texcb,
bool s)
1209 DPRINTF(TLBVerbose,
"memAttrs texcb:%d s:%d\n", texcb,
s);
1210 te.shareable =
false;
1211 te.nonCacheable =
false;
1212 te.outerShareable =
false;
1216 te.nonCacheable =
true;
1218 te.shareable =
true;
1223 te.nonCacheable =
true;
1225 te.shareable =
true;
1233 te.outerAttrs =
bits(texcb, 1, 0);
1239 te.outerAttrs =
bits(texcb, 1, 0);
1242 te.nonCacheable =
true;
1246 te.outerAttrs =
bits(texcb, 1, 0);
1249 panic(
"Reserved texcb value!\n");
1252 panic(
"Implementation-defined texcb value!\n");
1261 te.nonCacheable =
true;
1263 te.shareable =
false;
1268 panic(
"Reserved texcb value!\n");
1273 if (
bits(texcb, 1,0) == 0 ||
bits(texcb, 3,2) == 0)
1274 te.nonCacheable =
true;
1275 te.innerAttrs =
bits(texcb, 1, 0);
1276 te.outerAttrs =
bits(texcb, 3, 2);
1279 panic(
"More than 32 states for 5 bits?\n");
1287 DPRINTF(TLBVerbose,
"memAttrs PRRR:%08x NMRR:%08x\n", prrr, nmrr);
1288 uint8_t curr_tr = 0, curr_ir = 0, curr_or = 0;
1289 switch(
bits(texcb, 2,0)) {
1294 te.outerShareable = (prrr.nos0 == 0);
1300 te.outerShareable = (prrr.nos1 == 0);
1306 te.outerShareable = (prrr.nos2 == 0);
1312 te.outerShareable = (prrr.nos3 == 0);
1318 te.outerShareable = (prrr.nos4 == 0);
1324 te.outerShareable = (prrr.nos5 == 0);
1327 panic(
"Imp defined type\n");
1332 te.outerShareable = (prrr.nos7 == 0);
1338 DPRINTF(TLBVerbose,
"StronglyOrdered\n");
1340 te.nonCacheable =
true;
1343 te.shareable =
true;
1346 DPRINTF(TLBVerbose,
"Device ds1:%d ds0:%d s:%d\n",
1347 prrr.ds1, prrr.ds0,
s);
1349 te.nonCacheable =
true;
1353 te.shareable =
true;
1355 te.shareable =
true;
1358 DPRINTF(TLBVerbose,
"Normal ns1:%d ns0:%d s:%d\n",
1359 prrr.ns1, prrr.ns0,
s);
1362 te.shareable =
true;
1364 te.shareable =
true;
1367 panic(
"Reserved type");
1373 te.nonCacheable =
true;
1389 te.nonCacheable =
true;
1404 DPRINTF(TLBVerbose,
"memAttrs: shareable: %d, innerAttrs: %d, "
1406 te.shareable,
te.innerAttrs,
te.outerAttrs);
1407 te.setAttributes(
false);
1414 assert(
release->has(ArmExtension::LPAE));
1417 uint8_t
sh = l_descriptor.
sh();
1422 uint8_t attr_3_2 = (
attr >> 2) & 0x3;
1423 uint8_t attr_1_0 =
attr & 0x3;
1425 DPRINTF(TLBVerbose,
"memAttrsLPAE MemAttr:%#x sh:%#x\n",
attr,
sh);
1427 if (attr_3_2 == 0) {
1431 te.innerAttrs = attr_1_0 == 0 ? 1 : 3;
1432 te.nonCacheable =
true;
1435 te.outerAttrs = attr_3_2 == 1 ? 0 :
1436 attr_3_2 == 2 ? 2 : 1;
1437 te.innerAttrs = attr_1_0 == 1 ? 0 :
1438 attr_1_0 == 2 ? 6 : 5;
1439 te.nonCacheable = (attr_3_2 == 1) || (attr_1_0 == 1);
1442 uint8_t attrIndx = l_descriptor.
attrIndx();
1449 uint32_t mair =
currState->tc->readMiscReg(reg_as_int);
1450 attr = (mair >> (8 * (attrIndx % 4))) & 0xff;
1451 uint8_t attr_7_4 =
bits(
attr, 7, 4);
1452 uint8_t attr_3_0 =
bits(
attr, 3, 0);
1453 DPRINTF(TLBVerbose,
"memAttrsLPAE AttrIndx:%#x sh:%#x, attr %#x\n", attrIndx,
sh,
attr);
1458 te.nonCacheable =
false;
1463 if (attr_3_0 == 0x0)
1465 else if (attr_3_0 == 0x4)
1468 panic(
"Unpredictable behavior\n");
1469 te.nonCacheable =
true;
1476 if (attr_3_0 == 0x4)
1478 te.nonCacheable =
true;
1479 else if (attr_3_0 < 0x8)
1480 panic(
"Unpredictable behavior\n");
1490 if (attr_7_4 & 0x4) {
1491 te.outerAttrs = (attr_7_4 & 1) ? 1 : 3;
1493 te.outerAttrs = 0x2;
1497 if (attr_3_0 != 0x4 && attr_3_0 < 0x8)
1498 panic(
"Unpredictable behavior\n");
1501 panic(
"Unpredictable behavior\n");
1507 te.innerAttrs = 0x1;
1510 te.innerAttrs = attr_7_4 == 0 ? 0x3 : 0;
1522 te.innerAttrs = attr_3_0 & 1 ? 0x5 : 0x7;
1525 panic(
"Unpredictable behavior\n");
1530 te.outerShareable =
sh == 2;
1531 te.shareable = (
sh & 0x2) ?
true :
false;
1532 te.setAttributes(
true);
1533 te.attributes |= (uint64_t)
attr << 56;
1539 return !
bits(attrs, 2) ||
1550 uint8_t
sh = l_descriptor.
sh();
1554 uint8_t attr_hi = (
attr >> 2) & 0x3;
1555 uint8_t attr_lo =
attr & 0x3;
1557 DPRINTF(TLBVerbose,
"memAttrsAArch64 MemAttr:%#x sh:%#x\n",
attr,
sh);
1563 te.innerAttrs = attr_lo == 0 ? 1 : 3;
1564 te.nonCacheable =
true;
1567 te.outerAttrs = attr_hi == 1 ? 0 :
1568 attr_hi == 2 ? 2 : 1;
1569 te.innerAttrs = attr_lo == 1 ? 0 :
1570 attr_lo == 2 ? 6 : 5;
1573 te.nonCacheable = (attr_hi == 1) || (attr_hi == 2) ||
1574 (attr_lo == 1) || (attr_lo == 2);
1577 te.xs = !l_descriptor.
fnxs();
1580 uint8_t attrIndx = l_descriptor.
attrIndx();
1582 DPRINTF(TLBVerbose,
"memAttrsAArch64 AttrIndx:%#x sh:%#x\n", attrIndx,
sh);
1598 panic(
"Invalid exception level");
1603 attr =
bits(mair, 8 * attrIndx + 7, 8 * attrIndx);
1610 case 0b00000000 ... 0b00001111:
1612 te.nonCacheable =
true;
1617 te.nonCacheable =
true;
1622 te.nonCacheable =
true;
1630 te.xs =
te.nonCacheable;
1634 te.shareable =
sh == 2;
1635 te.outerShareable = (
sh & 0x2) ?
true :
false;
1637 te.attributes = ((uint64_t)
attr << 56) |
1652 te.nonCacheable =
true;
1657 te.nonCacheable = (
te.outerAttrs == 0 ||
te.outerAttrs == 2) &&
1658 (
te.innerAttrs == 0 ||
te.innerAttrs == 2);
1662 te.xs =
te.nonCacheable;
1675 DPRINTF(
TLB,
"L1 descriptor for %#x is %#x\n",
1679 const bool is_atomic =
currState->req->isAtomic();
1688 DPRINTF(
TLB,
"L1 Descriptor Reserved/Ignore, causing fault\n");
1691 std::make_shared<PrefetchAbort>(
1698 std::make_shared<DataAbort>(
1712 currState->fault = std::make_shared<DataAbort>(
1721 panic(
"Haven't implemented supersections\n");
1728 l2desc_addr =
currState->l1Desc.l2Addr() |
1730 DPRINTF(
TLB,
"L1 descriptor points to page table at: %#x (%s)\n",
1745 sizeof(uint32_t), flag, LookupLevel::L2,
1754 panic(
"A new type in a 2 bit field?\n");
1762 return std::make_shared<PrefetchAbort>(
1768 return std::make_shared<DataAbort>(
1788 DPRINTF(
TLB,
"L%d descriptor for %#llx is %#llx (%s)\n",
1791 currState->aarch64 ?
"AArch64" :
"long-desc.");
1795 DPRINTF(PageTableWalker,
"Analyzing L%d descriptor: %#llx, pxn: %d, "
1796 "xn: %d, ap: %d, piindex: %d, af: %d, type: %d\n",
1806 DPRINTF(PageTableWalker,
"Analyzing L%d descriptor: %#llx, type: %d\n",
1816 DPRINTF(
TLB,
"L%d descriptor Invalid, causing fault type %d\n",
1835 DPRINTF(
TLB,
"L%d descriptor causing Address Size Fault\n",
1840 }
else if (
currState->longDesc.af() == 0) {
1842 DPRINTF(
TLB,
"L%d descriptor causing Access Fault\n",
1878 DPRINTF(
TLB,
"L%d descriptor points to L%d descriptor at: %#x (%s)\n",
1886 next_desc_addr,
currState->physAddrRange)) {
1887 DPRINTF(
TLB,
"L%d descriptor causing Address Size Fault\n",
1895 if (
mmu->hasWalkCache()) {
1909 Event *
event = NULL;
1911 case LookupLevel::L1:
1913 case LookupLevel::L2:
1914 case LookupLevel::L3:
1918 panic(
"Wrong lookup level in table walk\n");
1924 sizeof(uint64_t), flag,
L,
event,
1931 panic(
"A new type in a 2 bit field?\n");
1945 DPRINTF(
TLB,
"L2 descriptor for %#x is %#x\n",
1949 const bool is_atomic =
currState->req->isAtomic();
1952 DPRINTF(
TLB,
"L2 descriptor invalid, causing fault\n");
1958 currState->fault = std::make_shared<PrefetchAbort>(
1964 currState->fault = std::make_shared<DataAbort>(
1977 DPRINTF(
TLB,
"Generating access fault at L2, afe: %d, ap: %d\n",
1980 currState->fault = std::make_shared<DataAbort>(
2003 DPRINTF(PageTableWalker,
"L1 Desc object host addr: %p\n",
2005 DPRINTF(PageTableWalker,
"L1 Desc object data: %08x\n",
2008 DPRINTF(PageTableWalker,
"calling doL1Descriptor for vaddr:%#x\n",
2017 stats.walksShortTerminatedAtLevel[0]++;
2031 DPRINTF(PageTableWalker,
"calling translateTiming again\n");
2037 stats.walksShortTerminatedAtLevel[0]++;
2064 DPRINTF(PageTableWalker,
"calling doL2Descriptor for vaddr:%#x\n",
2072 stats.walksShortTerminatedAtLevel[1]++;
2075 DPRINTF(PageTableWalker,
"calling translateTiming again\n");
2081 stats.walksShortTerminatedAtLevel[1]++;
2125 assert(curr_lookup_level ==
currState->longDesc.lookupLevel);
2134 DPRINTF(PageTableWalker,
"calling doLongDescriptor for vaddr:%#x\n",
2154 DPRINTF(PageTableWalker,
"calling translateTiming again\n");
2161 stats.walksLongTerminatedAtLevel[(unsigned) curr_lookup_level]++;
2171 if (curr_lookup_level >= LookupLevel::Num_ArmLookupLevel - 1)
2172 panic(
"Max. number of lookups already reached in table walk\n");
2198 "Fetching descriptor at address: 0x%x stage2Req: %d\n",
2213 fault = tran->fault;
2229 (this->*doDescriptor)();
2252 currState->tc->getCpuPtr()->clockPeriod());
2254 (this->*doDescriptor)();
2256 port->sendFunctionalReq(req,
data);
2257 (this->*doDescriptor)();
2265 DPRINTF(PageTableWalker,
"Adding to walker fifo: "
2266 "queue size before adding: %d\n",
2275 const bool have_security =
release->has(ArmExtension::SECURITY);
2280 te.longDescFormat =
true;
2290 te.size = (1ULL <<
te.N) - 1;
2297 te.type = TypeTLB::unified;
2310 DPRINTF(
TLB,
" - N:%d pfn:%#x size:%#x global:%d valid:%d\n",
2312 DPRINTF(
TLB,
" - vpn:%#x xn:%d pxn:%d ap:%d domain:%d asid:%d "
2313 "vmid:%d nc:%d ns:%d\n",
te.vpn,
te.xn,
te.pxn,
2314 te.ap,
static_cast<uint8_t
>(
te.domain),
te.asid,
te.vmid,
2315 te.nonCacheable,
te.ns);
2316 DPRINTF(
TLB,
" - domain from L%d desc:%d data:%#x\n",
2327 const bool have_security =
release->has(ArmExtension::SECURITY);
2332 te.longDescFormat = long_descriptor;
2337 te.size = (1<<
te.N) - 1;
2338 te.pfn = descriptor.
pfn();
2344 te.xn = descriptor.
xn();
2346 TypeTLB::instruction : TypeTLB::data;
2356 if (long_descriptor) {
2362 te.pxn =
currState->longDescData->pxnTable || l_descriptor.
pxn();
2366 te.hap = l_descriptor.
ap();
2369 descriptor.
ap() >> 1) << 1) |
2370 (
currState->longDescData->userTable && (descriptor.
ap() & 0x1));
2379 te.ap = descriptor.
ap();
2386 DPRINTF(
TLB,
" - N:%d pfn:%#x size:%#x global:%d valid:%d\n",
2388 DPRINTF(
TLB,
" - vpn:%#x xn:%d pxn:%d ap:%d piindex:%d domain:%d asid:%d "
2389 "vmid:%d nc:%d ns:%d\n",
te.vpn,
te.xn,
te.pxn,
2391 static_cast<uint8_t
>(
te.domain),
te.asid,
te.vmid,
2392 te.nonCacheable,
te.ns);
2393 DPRINTF(
TLB,
" - domain from L%d desc:%d data:%#x\n",
2408 switch (lookup_level_as_int) {
2409 case LookupLevel::L1:
2410 return LookupLevel::L1;
2411 case LookupLevel::L2:
2412 return LookupLevel::L2;
2413 case LookupLevel::L3:
2414 return LookupLevel::L3;
2416 panic(
"Invalid lookup level conversion");
2474 panic(
"unknown page size");
2487 auto req = std::make_shared<Request>();
2504 port->sendFunctional(&pkt);
2506 port->sendAtomic(&pkt);
2544 req = std::make_shared<Request>();
2581 "Table walker walks requested"),
2583 "Table walker walks initiated with short descriptors"),
2585 "Table walker walks initiated with long descriptors"),
2587 "Level at which table walker walks with short descriptors "
2590 "Level at which table walker walks with long descriptors "
2593 "Table walks squashed before starting"),
2595 "Table walks squashed after completion"),
2597 "Table walker wait (enqueue to first request) latency"),
2599 "Table walker service (enqueue to completion) latency"),
2601 "Table walker pending requests distribution"),
2603 "Table walker page sizes translated"),
2605 "Table walker requests started/completed, data/inst")
FaultSource
Generic fault source enums used to index into {short/long/aarch64}DescFaultSources[] to get the actua...
virtual void annotate(AnnotationIDs id, uint64_t val)
static ExceptionLevel tranTypeEL(CPSR cpsr, SCR scr, ArmTranslationType type)
Determine the EL to use for the purpose of a translation given a specific translation type.
virtual bool global(WalkerState *currState) const =0
virtual bool xn() const =0
virtual uint8_t * getRawPtr()=0
virtual uint64_t getRawData() const =0
virtual DomainType domain() const =0
virtual uint8_t texcb() const
virtual std::string dbgHeader() const =0
virtual uint8_t ap() const =0
virtual bool shareable() const
virtual uint8_t offsetBits() const =0
virtual Addr pfn() const =0
LookupLevel lookupLevel
Current lookup level for this descriptor.
virtual bool secure(bool have_security, WalkerState *currState) const =0
Long-descriptor format (LPAE)
uint8_t sh() const
2-bit shareability field
uint8_t memAttr() const
Memory attributes, only used by stage 2 translations.
uint8_t offsetBits() const override
Return the bit width of the page/block offset.
bool pxn() const
Is privileged execution allowed on this mapping?
DomainType domain() const override
uint8_t piindex() const
Stage 1 Indirect permissions.
bool secure(bool have_security, WalkerState *currState) const override
Returns true if this entry targets the secure physical address map.
std::string dbgHeader() const override
Addr nextTableAddr() const
Return the address of the next page table.
GrainSize grainSize
Width of the granule size in bits.
uint8_t attrIndx() const
Attribute index.
uint8_t ap() const override
2-bit access protection flags
bool fnxs() const
FNXS for FEAT_XS only.
uint64_t getRawData() const override
SnoopRespPacketQueue snoopRespQueue
Packet queue used to store outgoing snoop responses.
ReqPacketQueue reqQueue
Packet queue used to store outgoing requests.
Port(TableWalker &_walker)
void sendAtomicReq(const RequestPtr &req, uint8_t *data, Tick delay)
void handleResp(TableWalkerState *state, Addr addr, Addr size, Tick delay=0)
void sendFunctionalReq(const RequestPtr &req, uint8_t *data)
void handleRespPacket(PacketPtr pkt, Tick delay=0)
void sendTimingReq(const RequestPtr &req, uint8_t *data, Tick delay, Event *event)
PacketPtr createPacket(const RequestPtr &req, uint8_t *data, Tick delay, Event *event)
bool recvTimingResp(PacketPtr pkt) override
Receive a timing response from the peer.
This translation class is used to trigger the data fetch once a timing translation returns the transl...
void finish(const Fault &fault, const RequestPtr &req, ThreadContext *tc, BaseMMU::Mode mode)
void setVirt(Addr vaddr, int size, Request::Flags flags, int requestorId)
MMU::ArmTranslationType tranType
Stage2Walk(TableWalker &_parent, uint8_t *_data, Event *_event, Addr vaddr, BaseMMU::Mode mode, MMU::ArmTranslationType tran_type)
void translateTiming(ThreadContext *tc)
bool isWrite
If the access is a write.
CPSR cpsr
Cached copy of the cpsr as it existed when translation began.
Addr vaddr_tainted
The virtual address that is being translated.
RequestPtr req
Request that is currently being serviced.
VTCR_t vtcr
Cached copy of the vtcr as it existed when translation began.
HCR hcr
Cached copy of the htcr as it existed when translation began.
Addr vaddr
The virtual address that is being translated with tagging removed.
bool functional
If the atomic mode should be functional.
bool isUncacheable
True if table walks are uncacheable (for table descriptors)
ThreadContext * tc
Thread context that we're doing the walk for.
bool hpd
Hierarchical access permission disable.
BaseMMU::Translation * transState
Translation state for delayed requests.
SecurityState ss
Security State of the access.
std::optional< LongDescData > longDescData
BaseMMU::Mode mode
Save mode for use in delayed response.
HTCR htcr
Cached copy of the htcr as it existed when translation began.
SCR scr
Cached copy of the scr as it existed when translation began.
MMU::ArmTranslationType tranType
The translation type that has been requested.
SCTLR sctlr
Cached copy of the sctlr as it existed when translation began.
TableWalker * tableWalker
bool isFetch
If the access is a fetch (for execution, and no-exec) must be checked?
bool stage2Req
Flag indicating if a second stage of lookup is required.
TranslationRegime regime
Current translation regime.
bool timing
If the mode is timing or atomic.
int physAddrRange
Current physical address range in bits.
bool delayed
Whether the response is delayed in timing mode due to additional lookups.
uint16_t asid
ASID that we're servicing the request under.
L1Descriptor l1Desc
Short-format descriptors.
bool aarch64
If the access is performed in AArch64 state.
BaseMMU::Translation * stage2Tran
A pointer to the stage 2 translation that's in progress.
void doL2DescriptorWrapper()
static LookupLevel toLookupLevel(uint8_t lookup_level_as_int)
enums::ArmLookupLevel LookupLevel
void doL2LongDescriptorWrapper()
void memAttrs(ThreadContext *tc, TlbEntry &te, SCTLR sctlr, uint8_t texcb, bool s)
void doL3LongDescriptorWrapper()
const ArmRelease * release
Cached copies of system-level properties.
void doL1DescriptorWrapper()
bool checkVAOutOfRange(Addr addr, int top_bit, int tsz, bool low_range)
Fault generateLongDescFault(ArmFault::FaultSource src)
EventFunctionWrapper doL1DescEvent
EventFunctionWrapper doProcessEvent
static const unsigned REQUESTED
static const unsigned COMPLETED
bool uncacheableWalk() const
Returns true if the table walk should be uncacheable.
Event * LongDescEventByLevel[4]
void memAttrsAArch64(ThreadContext *tc, TlbEntry &te, LongDescriptor &lDescriptor)
void insertPartialTableEntry(LongDescriptor &descriptor)
void doL1LongDescriptorWrapper()
void fetchDescriptor(Addr desc_addr, DescriptorBase &descriptor, int num_bytes, Request::Flags flags, LookupLevel lookup_lvl, Event *event, void(TableWalker::*doDescriptor)())
void drainResume() override
Resume execution after a successful drain.
void doLongDescriptorWrapper(LookupLevel curr_lookup_level)
void doL0LongDescriptorWrapper()
bool pending
If a timing translation is currently in progress.
Port * port
Port shared by the two table walkers.
Fault testWalk(const RequestPtr &walk_req, DomainType domain, LookupLevel lookup_level)
Addr s1MinTxSz(GrainSize tg) const
std::tuple< Addr, Addr, LookupLevel > walkAddresses(Addr ttbr, GrainSize tg, int tsz, int pa_range)
Returns a tuple made of: 1) The address of the first page table 2) The address of the first descripto...
Fault readDataUntimed(ThreadContext *tc, Addr vaddr, Addr desc_addr, uint8_t *data, int num_bytes, Request::Flags flags, BaseMMU::Mode mode, MMU::ArmTranslationType tran_type, bool functional)
TableWalker(const Params &p)
void nextWalk(ThreadContext *tc)
void readDataTimed(ThreadContext *tc, Addr desc_addr, Stage2Walk *translation, int num_bytes, Request::Flags flags)
EventFunctionWrapper doL2DescEvent
gem5::Port & getPort(const std::string &if_name, PortID idx=InvalidPortID) override
Get a port with a given name and index.
std::list< WalkerState * > pendingQueue
Queue of requests that have passed are waiting because the walker is currently busy.
bool s1TxSzFault(GrainSize tg, int tsz) const
MMU * mmu
The MMU to forward second stage look upts to.
RequestorID requestorId
Requestor id assigned by the MMU.
gem5::ArmISA::TableWalker::TableWalkerStats stats
Port & getTableWalkerPort()
Fault walk(const RequestPtr &req, ThreadContext *tc, uint16_t asid, vmid_t _vmid, BaseMMU::Mode mode, BaseMMU::Translation *_trans, bool timing, bool functional, SecurityState ss, PASpace ipaspace, MMU::ArmTranslationType tran_type, bool stage2, const TlbEntry *walk_entry)
Fault processWalkAArch64()
const bool isStage2
Indicates whether this table walker is part of the stage 2 mmu.
bool uncacheableFromAttrs(uint8_t attrs)
bool checkAddrSizeFaultAArch64(Addr addr, int pa_range)
Returns true if the address exceeds the range permitted by the system-wide setting or by the TCR_ELx ...
void mpamTagTableWalk(RequestPtr &req) const
static uint8_t pageSizeNtoStatBin(uint8_t N)
void completeDrain()
Checks if all state is cleared and if so, completes drain.
void insertTableEntry(DescriptorBase &descriptor, bool longDescriptor)
DrainState drain() override
Provide a default implementation of the drain interface for objects that don't need draining.
std::list< WalkerState * > stateQueues[LookupLevel::Num_ArmLookupLevel]
Queues of requests for all the different lookup levels.
unsigned numSquashable
The number of walks belonging to squashed instructions that can be removed from the pendingQueue per ...
TLB * tlb
TLB that is initiating these table walks.
void memAttrsLPAE(ThreadContext *tc, TlbEntry &te, LongDescriptor &lDescriptor)
void setTestInterface(TlbTestInterface *ti)
void processWalkWrapper()
void memAttrsWalkAArch64(TlbEntry &te)
Addr maxTxSz(GrainSize tg) const
SCTLR sctlr
Cached copy of the sctlr as it existed when translation began.
void stashCurrState(int queue_idx)
Timing mode: saves the currState into the stateQueues.
ClockedObject(const ClockedObjectParams &p)
ClockedObjectParams Params
Parameters of ClockedObject.
Tick clockEdge(Cycles cycles=Cycles(0)) const
Determine the tick when a cycle begins, by default the current one, but the argument also enables the...
Cycles is a wrapper class for representing cycle counts, i.e.
A Packet is used to encapsulate a transfer between two objects in the memory system (e....
void dataStatic(T *p)
Set the data pointer to the following value that should not be freed.
SenderState * senderState
This packet's sender state.
RequestPtr req
A pointer to the original request.
bool cacheResponding() const
Ports are used to interface objects to each other.
const std::string name() const
Return port name (for DPRINTF).
QueuedRequestPort(const std::string &name, ReqPacketQueue &req_queue, SnoopRespPacketQueue &snoop_resp_queue, PortID id=InvalidPortID)
Create a QueuedPort with a given name, and a supplied implementation of two packet queues.
void schedTimingReq(PacketPtr pkt, Tick when)
Schedule the sending of a timing request.
Tick sendAtomic(PacketPtr pkt)
Send an atomic request packet, where the data is moved and the state is updated in zero time,...
void sendFunctional(PacketPtr pkt) const
Send a functional request packet, where the data is instantly updated everywhere in the memory system...
@ PT_WALK
The request is a page table walk.
@ SECURE
The request targets the secure memory space.
@ UNCACHEABLE
The request is to an uncacheable address.
@ NO_ACCESS
The request should not cause a memory access.
gem5::Flags< FlagsType > Flags
ThreadContext is the external interface to all thread state for anything outside of the CPU.
virtual RegVal readMiscReg(RegIndex misc_reg)=0
virtual BaseCPU * getCpuPtr()=0
#define ADD_STAT(n,...)
Convenience macro to add a stat to a statistics group.
constexpr T bits(T val, unsigned first, unsigned last)
Extract the bitfield from position 'first' to 'last' (inclusive) from 'val' and right justify it.
constexpr T mbits(T val, unsigned first, unsigned last)
Mask off the given bits in place like bits() but without shifting.
constexpr uint64_t sext(uint64_t val)
Sign-extend an N-bit value to 64 bits.
void signalDrainDone() const
Signal that an object is drained.
DrainState drainState() const
Return the current drain state of an object.
DrainState
Object drain/handover states.
@ Draining
Draining buffers pending serialization/handover.
@ Drained
Buffers drained, ready for serialization/handover.
void schedule(Event &event, Tick when)
void set(Type mask)
Set all flag's bits matching the given mask.
#define panic(...)
This implements a cprintf based panic() function.
#define panic_if(cond,...)
Conditional panic macro that checks the supplied condition and only panics if the condition is true a...
const Params & params() const
virtual Port & getPort(const std::string &if_name, PortID idx=InvalidPortID)
Get a port with a given name and index.
void tagRequest(ThreadContext *tc, const RequestPtr &req, bool ind)
Tag a memory request with MPAM information.
ByteOrder byteOrder(const ThreadContext *tc)
const PageTableOps * getPageTableOps(GrainSize trans_granule)
bool longDescFormatInUse(ThreadContext *tc)
bool ELIs64(ThreadContext *tc, ExceptionLevel el)
const GrainSize GrainMap_tg1[]
TranslationRegime translationRegime(ThreadContext *tc, ExceptionLevel el)
ExceptionLevel translationEl(TranslationRegime regime)
SecurityState
Security State.
int computeAddrTop(ThreadContext *tc, bool selbit, bool is_instr, TCR tcr, ExceptionLevel el)
Bitfield< 21, 20 > stride
int decodePhysAddrRange64(uint8_t pa_enc)
Returns the n.
const GrainSize GrainMap_tg0[]
int snsBankedIndex(MiscRegIndex reg, ThreadContext *tc)
PASpace
Physical Address Space.
bool HaveExt(ThreadContext *tc, ArmExtension ext)
Returns true if the provided ThreadContext supports the ArmExtension passed as a second argument.
Addr purifyTaggedAddr(Addr addr, ThreadContext *tc, ExceptionLevel el, TCR tcr, bool is_instr)
Removes the tag from tagged addresses if that mode is enabled.
const FlagsType pdf
Print the percent of the total that this entry represents.
const FlagsType nonan
Don't print if this is NAN.
const FlagsType nozero
Don't print if this is zero.
const FlagsType total
Print the total.
const FlagsType dist
Print the distribution.
Copyright (c) 2024 Arm Limited All rights reserved.
std::shared_ptr< FaultBase > Fault
std::shared_ptr< Request > RequestPtr
Tick curTick()
The universal simulation clock.
uint64_t Addr
Address type This will probably be moved somewhere else in the near future.
int16_t PortID
Port index/ID type, and a symbolic name for an invalid port id.
bool FullSystem
The FullSystem variable can be used to determine the current mode of simulation.
uint64_t Tick
Tick count type.
T htog(T value, ByteOrder guest_byte_order)
constexpr decltype(nullptr) NoFault
Overload hash function for BasicBlockRange type.
statistics::Scalar squashedBefore
statistics::Vector2d requestOrigin
TableWalkerStats(statistics::Group *parent)
statistics::Scalar walksLongDescriptor
statistics::Scalar walksShortDescriptor
statistics::Histogram walkWaitTime
statistics::Vector walksShortTerminatedAtLevel
statistics::Vector pageSizes
statistics::Vector walksLongTerminatedAtLevel
statistics::Histogram walkServiceTime
statistics::Histogram pendingWalks
statistics::Scalar squashedAfter
Helper variables used to implement hierarchical access permissions when the long-desc.
TLBTypes::KeyType KeyType
const std::string & name()