Go to the documentation of this file.
44 #include "debug/DRAM.hh"
45 #include "debug/Drain.hh"
46 #include "debug/MemCtrl.hh"
47 #include "debug/NVM.hh"
48 #include "debug/QOS.hh"
60 port(
name() +
".port", *this), isTimingMode(false),
61 retryRdReq(false), retryWrReq(false),
63 respondEvent([
this]{ processRespondEvent(); },
name()),
64 dram(
p.dram), nvm(
p.nvm),
65 readBufferSize((dram ? dram->readBufferSize : 0) +
66 (nvm ? nvm->readBufferSize : 0)),
67 writeBufferSize((dram ? dram->writeBufferSize : 0) +
68 (nvm ? nvm->writeBufferSize : 0)),
69 writeHighThreshold(writeBufferSize *
p.write_high_thresh_perc / 100.0),
70 writeLowThreshold(writeBufferSize *
p.write_low_thresh_perc / 100.0),
71 minWritesPerSwitch(
p.min_writes_per_switch),
72 writesThisTime(0), readsThisTime(0),
73 memSchedPolicy(
p.mem_sched_policy),
74 frontendLatency(
p.static_frontend_latency),
75 backendLatency(
p.static_backend_latency),
76 commandWindow(
p.command_window),
77 nextBurstAt(0), prevArrival(0),
81 DPRINTF(MemCtrl,
"Setting up controller\n");
82 readQueue.resize(
p.qos_priorities);
83 writeQueue.resize(
p.qos_priorities);
87 dram->setCtrl(
this, commandWindow);
89 nvm->setCtrl(
this, commandWindow);
91 fatal_if(!dram && !nvm,
"Memory controller must have an interface");
94 if (
p.write_low_thresh_perc >=
p.write_high_thresh_perc)
95 fatal(
"Write buffer low threshold %d must be smaller than the "
96 "high threshold %d\n",
p.write_low_thresh_perc,
97 p.write_high_thresh_perc);
104 fatal(
"MemCtrl %s is unconnected!\n",
name());
154 panic(
"Can't handle address range for packet %s\n",
177 "Read queue limit %d, current size %d, entries needed %d\n",
189 "Write queue limit %d, current size %d, entries needed %d\n",
203 assert(pkt_count != 0);
213 unsigned pktsServicedByWrQ = 0;
218 for (
int cnt = 0; cnt < pkt_count; ++cnt) {
219 unsigned size = std::min((
addr | (burst_size - 1)) + 1,
227 bool foundInWrQ =
false;
233 for (
const auto&
p :
vec) {
236 if (
p->addr <=
addr &&
237 ((
addr + size) <= (
p->addr +
p->size))) {
243 "Read to addr %#x with size %d serviced by "
258 if (pkt_count > 1 && burst_helper == NULL) {
260 "memory requests\n", pkt->
getAddr(), pkt_count);
294 addr = (
addr | (burst_size - 1)) + 1;
298 if (pktsServicedByWrQ == pkt_count) {
304 if (burst_helper != NULL)
328 for (
int cnt = 0; cnt < pkt_count; ++cnt) {
329 unsigned size = std::min((
addr | (burst_size - 1)) + 1,
370 "Merging write burst with existing queue entry\n");
378 addr = (
addr | (burst_size - 1)) + 1;
402 for (
const auto& packet : queue) {
414 for (
const auto& packet : queue) {
432 "Should only see read and writes at memory controller\n");
447 panic(
"Can't handle address range for packet %s\n",
456 unsigned size = pkt->
getSize();
502 "processRespondEvent(): Some req has reached its readyTime\n");
541 DPRINTF(Drain,
"Controller done draining\n");
543 }
else if (mem_pkt->
isDram()) {
561 MemPacketQueue::iterator
566 MemPacketQueue::iterator ret = queue.end();
568 if (!queue.empty()) {
569 if (queue.size() == 1) {
580 for (
auto i = queue.begin();
i != queue.end(); ++
i) {
590 panic(
"No scheduling policy chosen\n");
596 MemPacketQueue::iterator
599 auto selected_pkt_it = queue.end();
608 auto nvm_pkt_it = queue.end();
613 std::tie(selected_pkt_it, col_allowed_at) =
615 std::tie(nvm_pkt_it, nvm_col_at) =
620 if (col_allowed_at > nvm_col_at) {
621 selected_pkt_it = nvm_pkt_it;
624 std::tie(selected_pkt_it, col_allowed_at) =
627 std::tie(selected_pkt_it, col_allowed_at) =
631 if (selected_pkt_it == queue.end()) {
635 return selected_pkt_it;
651 panic(
"Can't handle address range for packet %s\n",
687 auto current_it = it++;
700 return (cmd_tick - burst_offset);
707 Tick cmd_at = cmd_tick;
714 while (
burstTicks.count(burst_tick) >= max_cmds_per_burst) {
728 Tick max_multi_cmd_split)
731 Tick cmd_at = cmd_tick;
741 Tick burst_offset = 0;
743 while (max_multi_cmd_split > (first_cmd_offset + burst_offset)) {
748 Tick first_cmd_tick = burst_tick - std::min(burst_offset, burst_tick);
751 bool first_can_issue =
false;
752 bool second_can_issue =
false;
754 while (!first_can_issue || !second_can_issue) {
755 bool same_burst = (burst_tick == first_cmd_tick);
756 auto first_cmd_count =
burstTicks.count(first_cmd_tick);
757 auto second_cmd_count = same_burst ? first_cmd_count + 1 :
760 first_can_issue = first_cmd_count < max_cmds_per_burst;
761 second_can_issue = second_cmd_count < max_cmds_per_burst;
763 if (!second_can_issue) {
764 DPRINTF(
MemCtrl,
"Contention (cmd2) found on command bus at %d\n",
773 bool gap_violated = !same_burst &&
774 ((burst_tick - first_cmd_tick) > max_multi_cmd_split);
776 if (!first_can_issue || (!second_can_issue && gap_violated)) {
777 DPRINTF(
MemCtrl,
"Contention (cmd1) found on command bus at %d\n",
847 DPRINTF(
MemCtrl,
"Access to %#x, ready at %lld next burst at %lld.\n",
889 switched_cmd_type?
"[turnaround triggered]":
"");
891 if (switched_cmd_type) {
894 "Switching to writes after %d reads with %d reads "
900 "Switching to reads after %d writes with %d writes "
927 bool nvm_busy =
true;
928 bool all_writes_nvm =
false;
932 nvm_busy =
nvm->
isBusy(read_queue_empty, all_writes_nvm);
936 if (dram_busy && nvm_busy) {
947 bool switch_to_writes =
false;
958 "Switching to writes due to read queue empty\n");
959 switch_to_writes =
true;
968 DPRINTF(Drain,
"MemCtrl controller done draining\n");
978 bool read_found =
false;
979 MemPacketQueue::iterator to_read;
988 "Checking READ queue [%d] priority [%d elements]\n",
989 prio, queue->size());
994 to_read =
chooseNext((*queue), switched_cmd_type ?
997 if (to_read != queue->end()) {
1014 auto mem_pkt = *to_read;
1019 assert(mem_pkt->size <= (mem_pkt->isDram() ?
1022 assert(mem_pkt->readyTime >=
curTick());
1026 mem_pkt->qosValue(), mem_pkt->getAddr(), 1,
1027 mem_pkt->readyTime - mem_pkt->entryTime);
1036 assert(
respQueue.back()->readyTime <= mem_pkt->readyTime);
1047 switch_to_writes =
true;
1052 readQueue[mem_pkt->qosValue()].erase(to_read);
1058 if (switch_to_writes) {
1064 bool write_found =
false;
1065 MemPacketQueue::iterator to_write;
1074 "Checking WRITE queue [%d] priority [%d elements]\n",
1075 prio, queue->size());
1082 if (to_write != queue->end()) {
1098 auto mem_pkt = *to_write;
1101 assert(mem_pkt->size <= (mem_pkt->isDram() ?
1111 mem_pkt->qosValue(), mem_pkt->getAddr(), 1,
1112 mem_pkt->readyTime - mem_pkt->entryTime);
1116 writeQueue[mem_pkt->qosValue()].erase(to_write);
1126 bool below_threshold =
1171 return std::min(dram_min, nvm_min);
1179 return std::min(dram_min, nvm_min);
1192 : statistics::
Group(&_ctrl),
1195 ADD_STAT(readReqs, statistics::units::Count::get(),
1196 "Number of read requests accepted"),
1197 ADD_STAT(writeReqs, statistics::units::Count::get(),
1198 "Number of write requests accepted"),
1200 ADD_STAT(readBursts, statistics::units::Count::get(),
1201 "Number of controller read bursts, including those serviced by "
1203 ADD_STAT(writeBursts, statistics::units::Count::get(),
1204 "Number of controller write bursts, including those merged in "
1206 ADD_STAT(servicedByWrQ, statistics::units::Count::get(),
1207 "Number of controller read bursts serviced by the write queue"),
1208 ADD_STAT(mergedWrBursts, statistics::units::Count::get(),
1209 "Number of controller write bursts merged with an existing one"),
1211 ADD_STAT(neitherReadNorWriteReqs, statistics::units::Count::get(),
1212 "Number of requests that are neither read nor write"),
1214 ADD_STAT(avgRdQLen, statistics::units::Rate<
1215 statistics::units::Count, statistics::units::
Tick>::get(),
1216 "Average read queue length when enqueuing"),
1217 ADD_STAT(avgWrQLen, statistics::units::Rate<
1218 statistics::units::Count, statistics::units::
Tick>::get(),
1219 "Average write queue length when enqueuing"),
1221 ADD_STAT(numRdRetry, statistics::units::Count::get(),
1222 "Number of times read queue was full causing retry"),
1223 ADD_STAT(numWrRetry, statistics::units::Count::get(),
1224 "Number of times write queue was full causing retry"),
1226 ADD_STAT(readPktSize, statistics::units::Count::get(),
1227 "Read request sizes (log2)"),
1228 ADD_STAT(writePktSize, statistics::units::Count::get(),
1229 "Write request sizes (log2)"),
1231 ADD_STAT(rdQLenPdf, statistics::units::Count::get(),
1232 "What read queue length does an incoming req see"),
1233 ADD_STAT(wrQLenPdf, statistics::units::Count::get(),
1234 "What write queue length does an incoming req see"),
1236 ADD_STAT(rdPerTurnAround, statistics::units::Count::get(),
1237 "Reads before turning the bus around for writes"),
1238 ADD_STAT(wrPerTurnAround, statistics::units::Count::get(),
1239 "Writes before turning the bus around for reads"),
1241 ADD_STAT(bytesReadWrQ, statistics::units::Byte::get(),
1242 "Total number of bytes read from write queue"),
1243 ADD_STAT(bytesReadSys, statistics::units::Byte::get(),
1244 "Total read bytes from the system interface side"),
1245 ADD_STAT(bytesWrittenSys, statistics::units::Byte::get(),
1246 "Total written bytes from the system interface side"),
1248 ADD_STAT(avgRdBWSys, statistics::units::Rate<
1249 statistics::units::Byte, statistics::units::Second>::get(),
1250 "Average system read bandwidth in Byte/s"),
1251 ADD_STAT(avgWrBWSys, statistics::units::Rate<
1252 statistics::units::Byte, statistics::units::Second>::get(),
1253 "Average system write bandwidth in Byte/s"),
1255 ADD_STAT(totGap, statistics::units::
Tick::get(),
"Total gap between requests"),
1256 ADD_STAT(avgGap, statistics::units::Rate<
1257 statistics::units::
Tick, statistics::units::Count>::get(),
1258 "Average gap between requests"),
1260 ADD_STAT(requestorReadBytes, statistics::units::Byte::get(),
1261 "Per-requestor bytes read from memory"),
1262 ADD_STAT(requestorWriteBytes, statistics::units::Byte::get(),
1263 "Per-requestor bytes write to memory"),
1264 ADD_STAT(requestorReadRate, statistics::units::Rate<
1265 statistics::units::Byte, statistics::units::Second>::get(),
1266 "Per-requestor bytes read from memory rate"),
1267 ADD_STAT(requestorWriteRate, statistics::units::Rate<
1268 statistics::units::Byte, statistics::units::Second>::get(),
1269 "Per-requestor bytes write to memory rate"),
1270 ADD_STAT(requestorReadAccesses, statistics::units::Count::get(),
1271 "Per-requestor read serviced memory accesses"),
1272 ADD_STAT(requestorWriteAccesses, statistics::units::Count::get(),
1273 "Per-requestor write serviced memory accesses"),
1274 ADD_STAT(requestorReadTotalLat, statistics::units::
Tick::get(),
1275 "Per-requestor read total memory access latency"),
1276 ADD_STAT(requestorWriteTotalLat, statistics::units::
Tick::get(),
1277 "Per-requestor write total memory access latency"),
1278 ADD_STAT(requestorReadAvgLat, statistics::units::Rate<
1279 statistics::units::
Tick, statistics::units::Count>::get(),
1280 "Per-requestor read average memory access latency"),
1281 ADD_STAT(requestorWriteAvgLat, statistics::units::Rate<
1282 statistics::units::
Tick, statistics::units::Count>::get(),
1283 "Per-requestor write average memory access latency")
1290 using namespace statistics;
1292 assert(ctrl.system());
1293 const auto max_requestors = ctrl.system()->maxRequestors();
1295 avgRdQLen.precision(2);
1296 avgWrQLen.precision(2);
1298 readPktSize.init(
ceilLog2(ctrl.system()->cacheLineSize()) + 1);
1299 writePktSize.init(
ceilLog2(ctrl.system()->cacheLineSize()) + 1);
1301 rdQLenPdf.init(ctrl.readBufferSize);
1302 wrQLenPdf.init(ctrl.writeBufferSize);
1305 .init(ctrl.readBufferSize)
1308 .init(ctrl.writeBufferSize)
1311 avgRdBWSys.precision(8);
1312 avgWrBWSys.precision(8);
1313 avgGap.precision(2);
1317 .init(max_requestors)
1321 .init(max_requestors)
1329 requestorReadAccesses
1330 .init(max_requestors)
1333 requestorWriteAccesses
1334 .init(max_requestors)
1337 requestorReadTotalLat
1338 .init(max_requestors)
1349 requestorWriteTotalLat
1350 .init(max_requestors)
1353 requestorWriteAvgLat
1357 for (
int i = 0;
i < max_requestors;
i++) {
1358 const std::string requestor = ctrl.system()->getRequestorName(
i);
1359 requestorReadBytes.subname(
i, requestor);
1360 requestorReadRate.subname(
i, requestor);
1361 requestorWriteBytes.subname(
i, requestor);
1362 requestorWriteRate.subname(
i, requestor);
1363 requestorReadAccesses.subname(
i, requestor);
1364 requestorWriteAccesses.subname(
i, requestor);
1365 requestorReadTotalLat.subname(
i, requestor);
1366 requestorReadAvgLat.subname(
i, requestor);
1367 requestorWriteTotalLat.subname(
i, requestor);
1368 requestorWriteAvgLat.subname(
i, requestor);
1375 avgGap = totGap / (readReqs + writeReqs);
1377 requestorReadRate = requestorReadBytes /
simSeconds;
1378 requestorWriteRate = requestorWriteBytes /
simSeconds;
1379 requestorReadAvgLat = requestorReadTotalLat / requestorReadAccesses;
1380 requestorWriteAvgLat = requestorWriteTotalLat / requestorWriteAccesses;
1393 panic(
"Can't handle address range for packet %s\n",
1401 if (if_name !=
"port") {
1416 return (dram_drained && nvm_drained);
1427 DPRINTF(Drain,
"Memory controller not drained, write: %d, read: %d,"
1475 DPRINTF(DRAM,
"Pushing DRAM ranges to port\n");
1476 ranges.push_back(ctrl.dram->getAddrRange());
1479 DPRINTF(NVM,
"Pushing NVM ranges to port\n");
1480 ranges.push_back(ctrl.nvm->getAddrRange());
1490 if (!queue.trySatisfyFunctional(pkt)) {
1494 ctrl.recvFunctional(pkt);
1503 return ctrl.recvAtomic(pkt);
1510 return ctrl.recvAtomicBackdoor(pkt, backdoor);
1517 return ctrl.recvTimingReq(pkt);
const uint32_t readBufferSize
The following are basic design parameters of the memory controller, and are initialized based on para...
Tick curTick()
The universal simulation clock.
#define fatal(...)
This implements a cprintf based fatal() function.
int16_t PortID
Port index/ID type, and a symbolic name for an invalid port id.
const std::string & cmdString() const
Return the string name of the cmd field (for debugging and tracing).
uint64_t totalReadQueueSize
Total read request packets queue length in #packets.
std::pair< MemPacketQueue::iterator, Tick > chooseNextFRFCFS(MemPacketQueue &queue, Tick min_col_at) const override
For FR-FCFS policy, find first NVM command that can issue default to first command to prepped region.
virtual Port & getPort(const std::string &if_name, PortID idx=InvalidPortID)
Get a port with a given name and index.
bool burstReady(MemPacket *pkt) const override
Check if a burst operation can be issued to the DRAM.
const Tick backendLatency
Pipeline latency of the backend and PHY.
void respondEvent(uint8_t rank)
Complete response process for DRAM when read burst is complete This will update the counters and chec...
DRAMInterface *const dram
Create pointer to interface of the actual dram media when connected.
statistics::Scalar bytesReadWrQ
EventFunctionWrapper nextReqEvent
void logResponse(BusState dir, RequestorID id, uint8_t _qos, Addr addr, uint64_t entries, double delay)
Called upon receiving a response, updates statistics and updates queues status.
Tick accessLatency() const override
bool recvTimingReq(PacketPtr) override
Receive a timing request from the peer.
BurstHelper * burstHelper
A pointer to the BurstHelper if this MemPacket is a split packet If not a split packet (common case),...
void sendRetryReq()
Send a retry to the request port that previously attempted a sendTimingReq to this response port and ...
Tick readyTime
When will request leave the controller.
unsigned int size
The size of this dram packet in bytes It is always equal or smaller than the burst size.
std::unordered_set< Addr > isInWriteQueue
To avoid iterating over the write queue to check for overlapping transactions, maintain a set of burs...
DrainState drainState() const
Return the current drain state of an object.
Tick recvAtomicBackdoor(PacketPtr pkt, MemBackdoorPtr &backdoor)
bool isBusy(bool read_queue_empty, bool all_writes_nvm)
This function checks if ranks are busy.
statistics::Vector requestorWriteAccesses
MemoryPort(const std::string &name, MemCtrl &_ctrl)
void addRankToRankDelay(Tick cmd_at) override
Add rank to rank delay to bus timing to all DRAM banks in alli ranks when access to an alternate inte...
void addToWriteQueue(PacketPtr pkt, unsigned int pkt_count, bool is_dram)
Decode the incoming pkt, create a mem_pkt and push to the back of the write queue.
AddrRange getAddrRange() const
Get the address range.
bool contains(const Addr &a) const
Determine if the range contains an address.
void processRespondEvent()
const FlagsType nozero
Don't print if this is zero.
Tick verifySingleCmd(Tick cmd_tick, Tick max_cmds_per_burst)
Check for command bus contention for single cycle command.
statistics::Vector requestorReadTotalLat
std::pair< MemPacketQueue::iterator, Tick > chooseNextFRFCFS(MemPacketQueue &queue, Tick min_col_at) const override
For FR-FCFS policy, find first DRAM command that can issue.
const PacketPtr pkt
This comes from the outside world.
uint8_t qosValue() const
QoS Value getter Returns 0 if QoS value was never set (constructor default).
The memory controller is a single-channel memory controller capturing the most important timing const...
void addRankToRankDelay(Tick cmd_at) override
Add rank to rank delay to bus timing to all NVM banks in alli ranks when access to an alternate inter...
bool cacheResponding() const
bool readsWaitingToIssue() const
bool allRanksDrained() const override
Check drain state of NVM interface.
bool allRanksDrained() const override
Return true once refresh is complete for all ranks and there are no additional commands enqueued.
const uint32_t writeBufferSize
void setupRank(const uint8_t rank, const bool is_read) override
Setup the rank based on packet received.
EventFunctionWrapper respondEvent
const FlagsType nonan
Don't print if this is NAN.
virtual void startup() override
startup() is the final initialization call before simulation.
statistics::Vector requestorWriteTotalLat
std::vector< MemPacketQueue > & selQueue(bool is_read)
Select either the read or write queue.
enums::MemSched memSchedPolicy
Memory controller configuration initialized based on parameter values.
bool packetReady(MemPacket *pkt)
Determine if there is a packet that can issue.
bool writeRespQueueFull() const
Check if the write response queue has reached defined threshold.
statistics::Vector readPktSize
uint32_t headerDelay
The extra delay from seeing the packet until the header is transmitted.
MemCtrl(const MemCtrlParams &p)
void drainRanks()
Iterate through dram ranks to exit self-refresh in order to drain.
Tick accessLatency() const override
BusState busStateNext
bus state for next request event triggered
BusState busState
Bus state used to control the read/write switching and drive the scheduling of the next request.
uint8_t qosSchedule(std::initializer_list< Queues * > queues_ptr, uint64_t queue_entry_size, const PacketPtr pkt)
Assign priority to a packet by executing the configured QoS policy.
void sample(const U &v, int n=1)
Add a value to the distribtion n times.
statistics::Scalar bytesReadSys
void setupRank(const uint8_t rank, const bool is_read) override
Setup the rank based on packet received.
void suspend()
Iterate through DRAM ranks and suspend them.
bool inReadBusState(bool next_state) const
Check the current direction of the memory channel.
bool retryRdReq
Remember if we have to retry a request when available.
statistics::Vector requestorReadAccesses
const Tick entryTime
When did request enter the controller.
Tick minWriteToReadDataGap() const
const std::unique_ptr< TurnaroundPolicy > turnPolicy
QoS Bus Turnaround Policy: selects the bus direction (READ/WRITE)
NVMInterface *const nvm
Create pointer to interface of the actual nvm media when connected.
statistics::Average avgWrQLen
uint32_t payloadDelay
The extra pipelining delay from seeing the packet until the end of payload is transmitted by the comp...
Tick recvAtomic(PacketPtr pkt)
statistics::Vector writePktSize
std::vector< MemPacketQueue > writeQueue
const uint32_t writeHighThreshold
DrainState
Object drain/handover states.
bool inWriteBusState(bool next_state) const
Check the current direction of the memory channel.
uint8_t schedule(RequestorID id, uint64_t data)
void accessAndRespond(PacketPtr pkt, Tick static_latency)
When a packet reaches its "readyTime" in the response Q, use the "access()" method in AbstractMemory ...
Addr addr
The starting address of the packet.
Tick minWriteToReadDataGap()
Calculate the minimum delay used when scheduling a write-to-read transision.
bool isBusy()
This function checks if ranks are actively refreshing and therefore busy.
MemoryPort port
Our incoming port, for a multi-ported controller add a crossbar in front of it.
Tick nextBurstAt
Till when must we wait before issuing next RD/WR burst?
void processNextReqEvent()
Bunch of things requires to setup "events" in gem5 When event "respondEvent" occurs for example,...
statistics::Vector requestorReadBytes
void functionalAccess(PacketPtr pkt)
Perform an untimed memory read or write without changing anything but the memory itself.
statistics::Scalar numWrRetry
virtual std::string name() const
RequestorID requestorId() const
void print(std::ostream &o, int verbosity=0, const std::string &prefix="") const
BusState selectNextBusState()
Returns next bus direction (READ or WRITE) based on configured policy.
A queued port is a port that has an infinite queue for outgoing packets and thus decouples the module...
#define ADD_STAT(n,...)
Convenience macro to add a stat to a statistics group.
A Packet is used to encapsulate a transfer between two objects in the memory system (e....
bool readQueueFull(unsigned int pkt_count) const
Check if the read queue has room for more entries.
uint64_t Tick
Tick count type.
bool isDram() const
Return true if its a DRAM access.
void recvFunctional(PacketPtr pkt)
Tick commandOffset() const override
bool recvTimingReq(PacketPtr pkt)
A memory packet stores packets along with the timestamp of when the packet entered the queue,...
DrainState drain() override
Draining is the process of clearing out the states of SimObjects.These are the SimObjects that are pa...
System * system() const
read the system pointer
virtual void init() override
init() is called after all C++ SimObjects have been created and all ports are connected.
void qosValue(const uint8_t qv)
Set the packet QoS value (interface compatibility with Packet)
void pruneBurstTick()
Remove commands that have already issued from burstTicks.
std::unique_ptr< Packet > pendingDelete
Upstream caches need this packet until true is returned, so hold it for deletion until a subsequent c...
uint64_t totalWriteQueueSize
Total write request packets queue length in #packets.
void checkRefreshState(uint8_t rank)
Check the refresh state to determine if refresh needs to be kicked back into action after a read resp...
bool allIntfDrained() const
Ensure that all interfaced have drained commands.
statistics::Vector rdQLenPdf
MemPacketQueue::iterator chooseNextFRFCFS(MemPacketQueue &queue, Tick extra_col_delay)
For FR-FCFS policy reorder the read/write queue depending on row buffer hits and earliest bursts avai...
bool isConnected() const
Is this port currently connected to a peer?
statistics::Scalar bytesWrittenSys
void recvFunctional(PacketPtr pkt) override
Receive a functional request packet from the peer.
statistics::Scalar writeReqs
unsigned int burstsServiced
Number of bursts serviced so far for a system packet.
void regStats() override
Callback to set stat parameters.
RequestorID requestorId() const
Get the packet RequestorID (interface compatibility with Packet)
@ Drained
Buffers drained, ready for serialization/handover.
bool isTimingMode() const
Is the system in timing mode?
Tick minReadToWriteDataGap() const
bool isTimingMode
Remember if the memory system is in timing mode.
void pushLabel(const std::string &lbl)
Push label for PrintReq (safe to call unconditionally).
void popLabel()
Pop label for PrintReq (safe to call unconditionally).
bool needsResponse() const
statistics::Scalar numRdRetry
uint64_t Addr
Address type This will probably be moved somewhere else in the near future.
std::deque< MemPacket * > respQueue
Response queue where read packets wait after we're done working with them, but it's not time to send ...
const Tick frontendLatency
Pipeline latency of the controller frontend.
statistics::Vector wrQLenPdf
Tick recvAtomicBackdoor(PacketPtr pkt, MemBackdoorPtr &backdoor) override
Receive an atomic request packet from the peer, and optionally provide a backdoor to the data being a...
const std::string & name()
statistics::Scalar servicedByWrQ
statistics::Scalar totGap
void sendRangeChange() const
Called by the owner to send a range change.
bool burstReady(MemPacket *pkt) const override
Check if a burst operation can be issued to the NVM.
void getBackdoor(MemBackdoorPtr &bd_ptr)
void chooseRead(MemPacketQueue &queue)
Select read command to issue asynchronously.
std::vector< MemPacketQueue > readQueue
The controller's main read and write queues, with support for QoS reordering.
statistics::Scalar readBursts
static constexpr T divCeil(const T &a, const U &b)
std::unordered_multiset< Tick > burstTicks
Holds count of commands issued in burst window starting at defined Tick.
void signalDrainDone() const
Signal that an object is drained.
Tick commandOffset() const override
#define panic_if(cond,...)
Conditional panic macro that checks the supplied condition and only panics if the condition is true a...
statistics::Scalar readReqs
void doBurstAccess(MemPacket *mem_pkt)
Actually do the burst based on media specific access function.
std::pair< Tick, Tick > doBurstAccess(MemPacket *pkt, Tick next_burst_at)
Actually do the burst and update stats.
Tick minReadToWriteDataGap()
Calculate the minimum delay used when scheduling a read-to-write transision.
std::pair< Tick, Tick > doBurstAccess(MemPacket *mem_pkt, Tick next_burst_at, const std::vector< MemPacketQueue > &queue)
Actually do the burst - figure out the latency it will take to service the req based on bank state,...
static constexpr int ceilLog2(const T &n)
const uint32_t writeLowThreshold
AddrRangeList getAddrRanges() const override
Get a list of the non-overlapping address ranges the owner is responsible for.
Tick verifyMultiCmd(Tick cmd_tick, Tick max_cmds_per_burst, Tick max_multi_cmd_split=0)
Check for command bus contention for multi-cycle (2 currently) command.
Port & getPort(const std::string &if_name, PortID idx=InvalidPortID) override
Get a port with a given name and index.
Tick recvAtomic(PacketPtr pkt) override
Receive an atomic request packet from the peer.
Ports are used to interface objects to each other.
const uint8_t rank
Will be populated by address decoder.
Addr burstAlign(Addr addr, bool is_dram) const
Burst-align an address.
bool writeQueueFull(unsigned int pkt_count) const
Check if the write queue has room for more entries.
uint8_t numPriorities() const
Gets the total number of priority levels in the QoS memory controller.
statistics::Formula & simSeconds
statistics::Scalar writeBursts
statistics::Histogram rdPerTurnAround
A burst helper helps organize and manage a packet that is larger than the memory burst size.
void recordTurnaroundStats()
Record statistics on turnarounds based on busStateNext and busState values.
Tick getBurstWindow(Tick cmd_tick)
Calculate burst window aligned tick.
MemPacketQueue::iterator chooseNext(MemPacketQueue &queue, Tick extra_col_delay)
The memory schduler/arbiter - picks which request needs to go next, based on the specified policy suc...
MemPacket * decodePacket(const PacketPtr pkt, Addr pkt_addr, unsigned int size, bool is_read, bool is_dram)
Address decoder to figure out physical mapping onto ranks, banks, and rows.
const uint32_t minWritesPerSwitch
uint32_t bytesPerBurst() const
virtual void drainResume() override
Resume execution after a successful drain.
const Tick commandWindow
Length of a command window, used to check command bandwidth.
statistics::Scalar mergedWrBursts
void schedTimingResp(PacketPtr pkt, Tick when)
Schedule the sending of a timing response.
void printQs() const
Used for debugging to observe the contents of the queues.
#define fatal_if(cond,...)
Conditional fatal macro that checks the supplied condition and only causes a fatal error if the condi...
void access(PacketPtr pkt)
Perform an untimed memory access and update all the state (e.g.
Tick nextReqTime
The soonest you have to start thinking about the next request is the longest access time that can occ...
const unsigned int burstCount
Number of bursts requred for a system packet.
Reference material can be found at the JEDEC website: UFS standard http://www.jedec....
bool isRead() const
Return true if its a read packet (interface compatibility with Packet)
statistics::Average avgRdQLen
void logRequest(BusState dir, RequestorID id, uint8_t _qos, Addr addr, uint64_t entries)
Called upon receiving a request or updates statistics and updates queues status.
void addToReadQueue(PacketPtr pkt, unsigned int pkt_count, bool is_dram)
When a new read comes in, first check if the write q has a pending request to the same address....
statistics::Histogram wrPerTurnAround
statistics::Vector requestorWriteBytes
@ Draining
Draining buffers pending serialization/handover.
bool scheduled() const
Determine if the current event is scheduled.
void startup() override
Iterate through dram ranks and instantiate per rank startup routine.
#define panic(...)
This implements a cprintf based panic() function.
Generated on Wed May 4 2022 12:14:00 for gem5 by doxygen 1.8.17