48 #include "debug/LLSC.hh" 49 #include "debug/MemoryAccess.hh" 50 #include "debug/ProtocolTrace.hh" 51 #include "debug/RubySequencer.hh" 52 #include "debug/RubyStats.hh" 55 #include "mem/ruby/protocol/PrefetchBit.hh" 56 #include "mem/ruby/protocol/RubyAccessMode.hh" 64 RubySequencerParams::create()
70 :
RubyPort(p), m_IncompleteTimes(MachineType_NUM),
71 deadlockCheckEvent([this]{
wakeup(); },
"Sequencer deadlock check")
99 DPRINTF(LLSC,
"LLSC Monitor - inserting load linked - " 100 "addr=0x%lx - cpu=%u\n", claddr,
m_version);
110 DPRINTF(LLSC,
"LLSC Monitor - clearing due to store - " 111 "addr=0x%lx - cpu=%u\n", claddr,
m_version);
122 DPRINTF(LLSC,
"LLSC Monitor - clearing due to " 123 "store conditional - " 124 "addr=0x%lx - cpu=%u\n",
160 int total_outstanding = 0;
163 for (
const auto seq_req : table_entry.second) {
167 panic(
"Possible Deadlock detected. Aborting!\n version: %d " 168 "request.paddr: 0x%x m_readRequestTable: %d current time: " 169 "%u issue_time: %d difference: %d\n",
m_version,
170 seq_req.pkt->getAddr(), table_entry.second.size(),
175 total_outstanding += table_entry.second.size();
192 for (
const auto& seq_req : table_entry.second) {
193 if (seq_req.functionalWrite(func_pkt))
207 for (
int i = 0;
i < RubyRequestType_NUM;
i++) {
211 for (
int j = 0;
j < MachineType_NUM;
j++) {
217 for (
int i = 0;
i < MachineType_NUM;
i++) {
234 RubyRequestType secondary_type)
246 seq_req_list.emplace_back(pkt, primary_type, secondary_type,
curCycle());
249 if (seq_req_list.size() > 1) {
250 return RequestStatus_Aliased;
255 return RequestStatus_Ready;
266 const MachineType respondingMach,
267 bool isExternalHit,
Cycles initialRequestTime,
268 Cycles forwardRequestTime,
276 Cycles total_lat = completion_time - issued_time;
278 if (initialRequestTime < issued_time) {
287 DPRINTFR(ProtocolTrace,
"%15s %3s %10s%20s %6s>%-6s %s %d cycles\n",
298 if (respondingMach != MachineType_NUM) {
302 if ((issued_time <= initialRequestTime) &&
303 (initialRequestTime <= forwardRequestTime) &&
304 (forwardRequestTime <= firstResponseTime) &&
305 (firstResponseTime <= completion_time)) {
308 initialRequestTime - issued_time);
310 forwardRequestTime - initialRequestTime);
312 firstResponseTime - forwardRequestTime);
314 completion_time - firstResponseTime);
323 if (respondingMach != MachineType_NUM) {
339 const bool externalHit,
const MachineType mach,
340 const Cycles initialRequestTime,
341 const Cycles forwardRequestTime,
342 const Cycles firstResponseTime)
355 bool ruby_request =
true;
356 int aliased_stores = 0;
357 int aliased_loads = 0;
358 while (!seq_req_list.empty()) {
361 assert(seq_req.
m_type != RubyRequestType_LD);
362 assert(seq_req.
m_type != RubyRequestType_Load_Linked);
363 assert(seq_req.
m_type != RubyRequestType_IFETCH);
367 if ((seq_req.
m_type != RubyRequestType_LD) &&
368 (seq_req.
m_type != RubyRequestType_Load_Linked) &&
369 (seq_req.
m_type != RubyRequestType_IFETCH)) {
371 bool success =
false;
373 if (seq_req.
m_type != RubyRequestType_Store_Conditional) {
381 seq_req.
pkt->
req->setExtraData(success ? 1 : 0);
387 if (seq_req.
m_type == RubyRequestType_Locked_RMW_Read) {
395 }
else if (seq_req.
m_type == RubyRequestType_Locked_RMW_Write) {
401 initialRequestTime, forwardRequestTime,
407 ruby_request =
false;
408 hitCallback(&seq_req, data, success, mach, externalHit,
409 initialRequestTime, forwardRequestTime,
413 assert(!ruby_request);
415 ruby_request =
false;
417 hitCallback(&seq_req, data,
true, mach, externalHit,
418 initialRequestTime, forwardRequestTime,
421 seq_req_list.pop_front();
425 if (seq_req_list.empty()) {
432 bool externalHit,
const MachineType mach,
433 Cycles initialRequestTime,
434 Cycles forwardRequestTime,
448 bool ruby_request =
true;
449 int aliased_loads = 0;
450 while (!seq_req_list.empty()) {
453 assert((seq_req.
m_type == RubyRequestType_LD) ||
454 (seq_req.
m_type == RubyRequestType_Load_Linked) ||
455 (seq_req.
m_type == RubyRequestType_IFETCH));
459 if ((seq_req.
m_type != RubyRequestType_LD) &&
460 (seq_req.
m_type != RubyRequestType_Load_Linked) &&
461 (seq_req.
m_type != RubyRequestType_IFETCH)) {
468 initialRequestTime, forwardRequestTime,
472 ruby_request =
false;
473 hitCallback(&seq_req, data,
true, mach, externalHit,
474 initialRequestTime, forwardRequestTime,
476 seq_req_list.pop_front();
480 if (seq_req_list.empty()) {
488 const MachineType mach,
const bool externalHit,
489 const Cycles initialRequestTime,
490 const Cycles forwardRequestTime,
491 const Cycles firstResponseTime)
493 warn_once(
"Replacement policy updates recently became the responsibility " 494 "of SLICC state machines. Make sure to setMRU() near callbacks " 502 if (type == RubyRequestType_Load_Linked) {
512 if ((type == RubyRequestType_LD) ||
513 (type == RubyRequestType_IFETCH) ||
514 (type == RubyRequestType_RMW_Read) ||
515 (type == RubyRequestType_Locked_RMW_Read) ||
516 (type == RubyRequestType_Load_Linked)) {
519 DPRINTF(RubySequencer,
"read data %s\n", data);
520 }
else if (pkt->
req->isSwap()) {
525 data.
setData(&overwrite_val[0],
527 DPRINTF(RubySequencer,
"swap data %s\n", data);
528 }
else if (type != RubyRequestType_Store_Conditional || llscSuccess) {
533 DPRINTF(RubySequencer,
"set data %s\n", data);
541 DPRINTF(RubySequencer,
"hitCallback %s 0x%x using RubyTester\n",
545 assert(testerSenderState);
546 testerSenderState->subBlock.mergeFrom(data);
573 return RequestStatus_BufferFull;
576 RubyRequestType primary_type = RubyRequestType_NULL;
577 RubyRequestType secondary_type = RubyRequestType_NULL;
591 DPRINTF(RubySequencer,
"Issuing SC\n");
592 primary_type = RubyRequestType_Store_Conditional;
593 #ifdef PROTOCOL_MESI_Three_Level 594 secondary_type = RubyRequestType_Store_Conditional;
596 secondary_type = RubyRequestType_ST;
599 DPRINTF(RubySequencer,
"Issuing LL\n");
601 primary_type = RubyRequestType_Load_Linked;
602 secondary_type = RubyRequestType_LD;
604 }
else if (pkt->
req->isLockedRMW()) {
612 DPRINTF(RubySequencer,
"Issuing Locked RMW Write\n");
613 primary_type = RubyRequestType_Locked_RMW_Write;
615 DPRINTF(RubySequencer,
"Issuing Locked RMW Read\n");
617 primary_type = RubyRequestType_Locked_RMW_Read;
619 secondary_type = RubyRequestType_ST;
630 primary_type = secondary_type = RubyRequestType_ST;
631 }
else if (pkt->
isRead()) {
632 if (pkt->
req->isInstFetch()) {
633 primary_type = secondary_type = RubyRequestType_IFETCH;
635 bool storeCheck =
false;
638 uint32_t flags = pkt->
req->getFlags();
643 primary_type = RubyRequestType_RMW_Read;
644 secondary_type = RubyRequestType_ST;
646 primary_type = secondary_type = RubyRequestType_LD;
650 primary_type = secondary_type = RubyRequestType_FLUSH;
652 panic(
"Unsupported ruby packet type\n");
658 (primary_type != RubyRequestType_Locked_RMW_Write)) {
662 return RequestStatus_Aliased;
668 if (status != RequestStatus_Ready && status != RequestStatus_Aliased)
672 if (status != RequestStatus_Aliased)
676 return RequestStatus_Issued;
690 if (pkt->
req->hasPC()) {
691 pc = pkt->
req->getPC();
696 std::shared_ptr<RubyRequest> msg =
699 nullptr : pkt->
getPtr<uint8_t>(),
701 RubyAccessMode_Supervisor, pkt,
702 PrefetchBit_No, proc_id, core_id);
704 DPRINTFR(ProtocolTrace,
"%15s %3s %10s%20s %6s>%-6s %#x %s\n",
707 RubyRequestType_to_string(secondary_type));
717 template <
class KEY,
class VALUE>
719 operator<<(ostream &out, const std::unordered_map<KEY, VALUE> &map)
721 for (
const auto &table_entry : map) {
722 out <<
"[ " << table_entry.first <<
" =";
723 for (
const auto &seq_req : table_entry.second) {
724 out <<
" " << RubyRequestType_to_string(seq_req.m_second_type);
751 DPRINTF(RubyStats,
"Recorded statistic: %s\n",
752 SequencerRequestType_to_string(requestType));
775 for (
int i = 0;
i < RubyRequestType_NUM;
i++) {
786 for (
int i = 0;
i < MachineType_NUM;
i++) {
806 for (
int i = 0;
i < RubyRequestType_NUM;
i++) {
810 for (
int j = 0;
j < MachineType_NUM;
j++) {
void enqueueNextFetchRequest()
Function for fetching warming up the memory and the caches.
#define panic(...)
This implements a cprintf based panic() function.
int m_max_outstanding_requests
const uint8_t * getData(int offset, int len) const
std::unordered_map< Addr, std::list< SequencerRequest > > m_RequestTable
Cycles is a wrapper class for representing cycle counts, i.e.
T * findNextSenderState() const
Go through the sender state stack and return the first instance that is of type T (as determined by a...
void recordMissLatency(SequencerRequest *srequest, bool llscSuccess, const MachineType respondingMach, bool isExternalHit, Cycles initialRequestTime, Cycles forwardRequestTime, Cycles firstResponseTime)
std::vector< Stats::Histogram * > m_missTypeLatencyHist
void print(std::ostream &out) const
RequestStatus makeRequest(PacketPtr pkt) override
bool isLocked(int context) const
void llscClearMonitor(const Addr)
Removes the cache line address from the global monitor.
CacheRecorder * m_cache_recorder
EventFunctionWrapper deadlockCheckEvent
AbstractController * m_controller
void llscLoadLinked(const Addr)
Places the cache line address into the global monitor tagged with this Sequencer object's version id...
void regStats() override
Callback to set stat parameters.
std::vector< Stats::Histogram * > m_FirstResponseToCompletionDelayHist
void recordRequestType(SequencerRequestType requestType)
Histogram & init(size_type size)
Set the parameters of this histogram.
Overload hash function for BasicBlockRange type.
Stats::Histogram m_missLatencyHist
Histogram for holding latency profile of all requests that miss in the controller connected to this s...
Stats::Histogram m_latencyHist
Histogram for holding latency profile of all requests.
CacheMemory * m_dataCache_ptr
std::vector< Stats::Histogram * > m_typeLatencyHist
T * getPtr()
get a pointer to the data ptr.
RequestStatus insertRequest(PacketPtr pkt, RubyRequestType primary_type, RubyRequestType secondary_type)
RubySystem * m_ruby_system
Stats::Histogram m_outstandReqHist
Histogram for number of outstanding requests per cycle.
std::vector< Stats::Histogram * > m_ForwardToFirstResponseDelayHist
CacheMemory * m_instCache_ptr
RequestPtr req
A pointer to the original request.
DrainState drainState() const
Return the current drain state of an object.
std::vector< Stats::Histogram * > m_InitialToForwardDelayHist
static bool getWarmupEnabled()
Tick cyclesToTicks(Cycles c) const
Tick curTick()
The current simulated tick.
void enqueueNextFlushRequest()
Function for flushing the memory contents of the caches to the main memory.
void ruby_eviction_callback(Addr address)
void setData(const uint8_t *p)
Copy data into the packet from the provided pointer.
uint64_t Tick
Tick count type.
void evictionCallback(Addr address)
std::vector< Stats::Histogram * > m_hitTypeLatencyHist
std::vector< Stats::Histogram * > m_IssueToInitialDelayHist
Histograms for recording the breakdown of miss latency.
Cycles curCycle() const
Determine the current cycle, corresponding to a tick aligned to a clock edge.
void writeData(uint8_t *p) const
Copy data from the packet to the memory at the provided pointer.
void schedule(Event &event, Tick when)
Addr getOffset(Addr addr)
void writeCallbackScFail(Addr address, DataBlock &data)
Proxy function to writeCallback that first invalidates the line address in the local monitor...
virtual int functionalWrite(Packet *func_pkt) override
static bool getCooldownEnabled()
RubyRequestType m_second_type
std::vector< Stats::Histogram * > m_missMachLatencyHist
Histograms for profiling the latencies for requests that required external messages.
void ruby_hit_callback(PacketPtr pkt)
void setLocked(int context)
uint64_t Addr
Address type This will probably be moved somewhere else in the near future.
void issueRequest(PacketPtr pkt, RubyRequestType type)
A Packet is used to encapsulate a transfer between two objects in the memory system (e...
Addr makeLineAddress(Addr addr)
Tick clockEdge(Cycles cycles=Cycles(0)) const
Determine the tick when a cycle begins, by default the current one, but the argument also enables the...
std::string printAddress(Addr addr)
void reset()
Reset stat value to default.
void blockOnQueue(Addr, MessageBuffer *)
bool scheduled() const
Determine if the current event is scheduled.
MessageBuffer * m_mandatory_q_ptr
void hitCallback(SequencerRequest *srequest, DataBlock &data, bool llscSuccess, const MachineType mach, const bool externalHit, const Cycles initialRequestTime, const Cycles forwardRequestTime, const Cycles firstResponseTime)
std::vector< std::vector< Stats::Histogram * > > m_missTypeMachLatencyHist
std::vector< Stats::Counter > m_IncompleteTimes
Declaration of the Packet class.
virtual Cycles mandatoryQueueLatency(const RubyRequestType ¶m_type)
bool m_runningGarnetStandalone
void resetStats() override
Callback to reset stats.
const ContextID InvalidContextID
const T * getConstPtr() const
void setData(const uint8_t *data, int offset, int len)
std::vector< Stats::Histogram * > m_hitMachLatencyHist
Histograms for profiling the latencies for requests that did not required external messages...
virtual int functionalWrite(Packet *func_pkt)
Cycles m_deadlock_threshold
virtual void regStats()
Callback to set stat parameters.
const std::string & cmdString() const
Return the string name of the cmd field (for debugging and tracing).
void readCallback(Addr address, DataBlock &data, const bool externalHit=false, const MachineType mach=MachineType_NUM, const Cycles initialRequestTime=Cycles(0), const Cycles forwardRequestTime=Cycles(0), const Cycles firstResponseTime=Cycles(0))
bool llscCheckMonitor(const Addr)
Searches for cache line address in the global monitor tagged with this Sequencer object's version id...
AbstractCacheEntry * lookup(Addr address)
bool isBlocked(Addr) const
void writeCallback(Addr address, DataBlock &data, const bool externalHit=false, const MachineType mach=MachineType_NUM, const Cycles initialRequestTime=Cycles(0), const Cycles forwardRequestTime=Cycles(0), const Cycles firstResponseTime=Cycles(0))
bool llscStoreConditional(const Addr)
Searches for cache line address in the global monitor tagged with this Sequencer object's version id...
void enqueue(MsgPtr message, Tick curTime, Tick delta)
Stats::Histogram m_hitLatencyHist
Histogram for holding latency profile of all requests that hit in the controller connected to this se...
int ContextID
Globally unique thread context ID.
void sample(const U &v, int n=1)
Add a value to the distribtion n times.
void checkCoherence(Addr address)
std::vector< std::vector< Stats::Histogram * > > m_hitTypeMachLatencyHist
Arch getArch() const
Get the architecture.
Sequencer(const Params *)