49#include "debug/LLSC.hh"
50#include "debug/MemoryAccess.hh"
51#include "debug/ProtocolTrace.hh"
52#include "debug/RubyHitMiss.hh"
53#include "debug/RubySequencer.hh"
54#include "debug/RubyStats.hh"
57#include "mem/ruby/protocol/PrefetchBit.hh"
58#include "mem/ruby/protocol/RubyAccessMode.hh"
71 :
RubyPort(
p), m_IncompleteTimes(MachineType_NUM),
72 deadlockCheckEvent([this]{
wakeup(); },
"Sequencer deadlock check")
74 m_outstanding_count = 0;
76 m_dataCache_ptr =
p.dcache;
77 m_max_outstanding_requests =
p.max_outstanding_requests;
78 m_deadlock_threshold =
p.deadlock_threshold;
81 assert(m_max_outstanding_requests > 0);
82 assert(m_deadlock_threshold > 0);
84 m_unaddressedTransactionCnt = 0;
86 m_runningGarnetStandalone =
p.garnet_standalone;
92 m_outstandReqHist.init(10);
93 m_latencyHist.init(10);
94 m_hitLatencyHist.init(10);
95 m_missLatencyHist.init(10);
97 for (
int i = 0;
i < RubyRequestType_NUM;
i++) {
99 m_typeLatencyHist[
i]->init(10);
102 m_hitTypeLatencyHist[
i]->init(10);
105 m_missTypeLatencyHist[
i]->init(10);
108 for (
int i = 0;
i < MachineType_NUM;
i++) {
110 m_hitMachLatencyHist[
i]->init(10);
113 m_missMachLatencyHist[
i]->init(10);
116 m_IssueToInitialDelayHist[
i]->init(10);
119 m_InitialToForwardDelayHist[
i]->init(10);
121 m_ForwardToFirstResponseDelayHist.push_back(
123 m_ForwardToFirstResponseDelayHist[
i]->init(10);
125 m_FirstResponseToCompletionDelayHist.push_back(
127 m_FirstResponseToCompletionDelayHist[
i]->init(10);
130 for (
int i = 0;
i < RubyRequestType_NUM;
i++) {
131 m_hitTypeMachLatencyHist.push_back(
133 m_missTypeMachLatencyHist.push_back(
136 for (
int j = 0;
j < MachineType_NUM;
j++) {
137 m_hitTypeMachLatencyHist[
i].push_back(
new statistics::Histogram());
138 m_hitTypeMachLatencyHist[
i][
j]->init(10);
140 m_missTypeMachLatencyHist[
i].push_back(
141 new statistics::Histogram());
142 m_missTypeMachLatencyHist[
i][
j]->init(10);
156 "%s must have a dcache object to support LLSC requests.",
name());
160 DPRINTF(LLSC,
"LLSC Monitor - inserting load linked - "
161 "addr=0x%lx - cpu=%u\n", claddr,
m_version);
174 DPRINTF(LLSC,
"LLSC Monitor - clearing due to store - "
175 "addr=0x%lx - cpu=%u\n", claddr,
m_version);
183 "%s must have a dcache object to support LLSC requests.",
name());
188 DPRINTF(LLSC,
"LLSC Monitor - clearing due to "
189 "store conditional - "
190 "addr=0x%lx - cpu=%u\n",
233 [[maybe_unused]]
int total_outstanding = 0;
236 for (
const auto &seq_req : table_entry.second) {
240 panic(
"Possible Deadlock detected. Aborting!\n version: %d "
241 "request.paddr: 0x%x m_readRequestTable: %d current time: "
242 "%u issue_time: %d difference: %d\n",
m_version,
243 seq_req.pkt->getAddr(), table_entry.second.size(),
248 total_outstanding += table_entry.second.size();
265 for (
const auto& seq_req : table_entry.second) {
266 if (seq_req.functionalWrite(func_pkt))
280 for (
int i = 0;
i < RubyRequestType_NUM;
i++) {
284 for (
int j = 0;
j < MachineType_NUM;
j++) {
290 for (
int i = 0;
i < MachineType_NUM;
i++) {
307 RubyRequestType secondary_type)
316 assert(primary_type == secondary_type);
318 switch (primary_type) {
319 case RubyRequestType_TLBI_EXT_SYNC_COMP:
322 case RubyRequestType_TLBI:
323 case RubyRequestType_TLBI_SYNC:
328 [[maybe_unused]]
auto insert_data = \
332 pkt, primary_type, secondary_type,
curCycle()));
335 assert(insert_data.second &&
336 "Another TLBI request with the same ID exists");
338 DPRINTF(RubySequencer,
"Inserting TLBI request %016x\n",
345 panic(
"Unexpected TLBI RubyRequestType");
348 return RequestStatus_Ready;
355 seq_req_list.emplace_back(pkt, primary_type,
359 if (seq_req_list.size() > 1) {
360 return RequestStatus_Aliased;
365 return RequestStatus_Ready;
376 const MachineType respondingMach,
377 bool isExternalHit,
Cycles initialRequestTime,
378 Cycles forwardRequestTime,
386 Cycles total_lat = completion_time - issued_time;
388 if ((initialRequestTime != 0) && (initialRequestTime < issued_time)) {
397 DPRINTFR(ProtocolTrace,
"%15s %3s %10s%20s %6s>%-6s %s %d cycles\n",
408 if (respondingMach != MachineType_NUM) {
412 if ((issued_time <= initialRequestTime) &&
413 (initialRequestTime <= forwardRequestTime) &&
414 (forwardRequestTime <= firstResponseTime) &&
415 (firstResponseTime <= completion_time)) {
418 initialRequestTime - issued_time);
420 forwardRequestTime - initialRequestTime);
422 firstResponseTime - forwardRequestTime);
424 completion_time - firstResponseTime);
433 if (respondingMach != MachineType_NUM) {
449 const bool externalHit,
const MachineType mach,
450 const Cycles initialRequestTime,
451 const Cycles forwardRequestTime,
452 const Cycles firstResponseTime,
466 bool ruby_request =
true;
467 while (!seq_req_list.empty()) {
470 if (noCoales && !ruby_request) {
479 assert(seq_req.
m_type != RubyRequestType_LD);
480 assert(seq_req.
m_type != RubyRequestType_Load_Linked);
481 assert(seq_req.
m_type != RubyRequestType_IFETCH);
485 if ((seq_req.
m_type != RubyRequestType_LD) &&
486 (seq_req.
m_type != RubyRequestType_Load_Linked) &&
487 (seq_req.
m_type != RubyRequestType_IFETCH)) {
491 if (seq_req.
m_type != RubyRequestType_Store_Conditional) {
499 seq_req.
pkt->
req->setExtraData(success ? 1 : 0);
505 if (seq_req.
m_type == RubyRequestType_Locked_RMW_Read) {
513 }
else if (seq_req.
m_type == RubyRequestType_Locked_RMW_Write) {
519 initialRequestTime, forwardRequestTime,
525 initialRequestTime, forwardRequestTime,
526 firstResponseTime, !ruby_request);
527 ruby_request =
false;
530 assert(!ruby_request);
533 initialRequestTime, forwardRequestTime,
534 firstResponseTime, !ruby_request);
536 seq_req_list.pop_front();
540 if (seq_req_list.empty()) {
547 bool externalHit,
const MachineType mach,
548 Cycles initialRequestTime,
549 Cycles forwardRequestTime,
563 bool ruby_request =
true;
564 while (!seq_req_list.empty()) {
567 assert((seq_req.
m_type == RubyRequestType_LD) ||
568 (seq_req.
m_type == RubyRequestType_Load_Linked) ||
569 (seq_req.
m_type == RubyRequestType_IFETCH));
571 if ((seq_req.
m_type != RubyRequestType_LD) &&
572 (seq_req.
m_type != RubyRequestType_Load_Linked) &&
573 (seq_req.
m_type != RubyRequestType_IFETCH)) {
580 initialRequestTime, forwardRequestTime,
585 initialRequestTime, forwardRequestTime,
586 firstResponseTime, !ruby_request);
587 ruby_request =
false;
588 seq_req_list.pop_front();
592 if (seq_req_list.empty()) {
600 const MachineType mach,
const bool externalHit,
601 const Cycles initialRequestTime,
602 const Cycles forwardRequestTime,
603 const Cycles firstResponseTime,
604 const bool was_coalesced)
606 warn_once(
"Replacement policy updates recently became the responsibility "
607 "of SLICC state machines. Make sure to setMRU() near callbacks "
623 if (
type == RubyRequestType_Load_Linked) {
628 DPRINTF(RubyHitMiss,
"Cache %s at %#x\n",
629 externalHit ?
"miss" :
"hit",
636 if ((
type == RubyRequestType_LD) ||
637 (
type == RubyRequestType_IFETCH) ||
638 (
type == RubyRequestType_RMW_Read) ||
639 (
type == RubyRequestType_Locked_RMW_Read) ||
640 (
type == RubyRequestType_Load_Linked)) {
644 }
else if (pkt->
req->isSwap()) {
650 data.setData(&overwrite_val[0],
657 DPRINTF(RubySequencer,
"AMO original data %s\n",
data);
662 }
else if (
type != RubyRequestType_Store_Conditional || llscSuccess) {
674 DPRINTF(RubySequencer,
"hitCallback %s 0x%x using RubyTester\n",
678 assert(testerSenderState);
686 rs->m_cache_recorder->enqueueNextFetchRequest();
689 rs->m_cache_recorder->enqueueNextFlushRequest();
698 RubyRequestType reqType,
699 const MachineType mach,
700 const Cycles initialRequestTime,
701 const Cycles forwardRequestTime,
702 const Cycles firstResponseTime)
704 DPRINTF(RubySequencer,
"unaddressedCallback ID:%08x type:%d\n",
705 unaddressedReqId, reqType);
708 case RubyRequestType_TLBI_EXT_SYNC:
718 case RubyRequestType_TLBI:
719 case RubyRequestType_TLBI_SYNC:
730 assert(seq_req.
m_type == reqType);
742 panic(
"Unexpected TLBI RubyRequestType");
759 !pkt->
req->isHTMAbort()) {
760 return RequestStatus_BufferFull;
763 RubyRequestType primary_type = RubyRequestType_NULL;
764 RubyRequestType secondary_type = RubyRequestType_NULL;
778 DPRINTF(RubySequencer,
"Issuing SC\n");
779 primary_type = RubyRequestType_Store_Conditional;
780#if defined (PROTOCOL_MESI_Three_Level) || defined (PROTOCOL_MESI_Three_Level_HTM)
781 secondary_type = RubyRequestType_Store_Conditional;
783 secondary_type = RubyRequestType_ST;
786 DPRINTF(RubySequencer,
"Issuing LL\n");
788 primary_type = RubyRequestType_Load_Linked;
789 secondary_type = RubyRequestType_LD;
791 }
else if (pkt->
req->isLockedRMW()) {
799 DPRINTF(RubySequencer,
"Issuing Locked RMW Write\n");
800 primary_type = RubyRequestType_Locked_RMW_Write;
802 DPRINTF(RubySequencer,
"Issuing Locked RMW Read\n");
804 primary_type = RubyRequestType_Locked_RMW_Read;
806 secondary_type = RubyRequestType_ST;
807 }
else if (pkt->
req->isTlbiCmd()) {
809 DPRINTF(RubySequencer,
"Issuing TLBI\n");
820 primary_type = secondary_type = RubyRequestType_ST;
821 }
else if (pkt->
isRead()) {
823 if (pkt->
req->isHTMCmd()) {
825 }
else if (pkt->
req->isInstFetch()) {
826 primary_type = secondary_type = RubyRequestType_IFETCH;
828 if (pkt->
req->isReadModifyWrite()) {
829 primary_type = RubyRequestType_RMW_Read;
830 secondary_type = RubyRequestType_ST;
832 primary_type = secondary_type = RubyRequestType_LD;
836 primary_type = secondary_type = RubyRequestType_FLUSH;
838 panic(
"Unsupported ruby packet type\n");
843 if (!pkt->
req->isMemMgmt() &&
845 (primary_type != RubyRequestType_Locked_RMW_Write)) {
849 return RequestStatus_Aliased;
855 if (
status != RequestStatus_Ready &&
status != RequestStatus_Aliased)
859 if (
status != RequestStatus_Aliased)
863 return RequestStatus_Issued;
877 if (pkt->
req->hasPC()) {
878 pc = pkt->
req->getPC();
883 std::shared_ptr<RubyRequest> msg;
884 if (pkt->
req->isMemMgmt()) {
885 msg = std::make_shared<RubyRequest>(
clockEdge(),
887 RubyAccessMode_Supervisor, pkt,
890 DPRINTFR(ProtocolTrace,
"%15s %3s %10s%20s %6s>%-6s %s\n",
892 RubyRequestType_to_string(secondary_type));
894 if (pkt->
req->isTlbiCmd()) {
895 msg->m_isTlbi =
true;
896 switch (secondary_type) {
897 case RubyRequestType_TLBI_EXT_SYNC_COMP:
898 msg->m_tlbiTransactionUid = pkt->
req->getExtraData();
900 case RubyRequestType_TLBI:
901 case RubyRequestType_TLBI_SYNC:
902 msg->m_tlbiTransactionUid = \
903 getCurrentUnaddressedTransactionID();
906 panic(
"Unexpected TLBI RubyRequestType");
908 DPRINTF(RubySequencer,
"Issuing TLBI %016x\n",
909 msg->m_tlbiTransactionUid);
914 RubyAccessMode_Supervisor, pkt,
915 PrefetchBit_No, proc_id, core_id);
917 DPRINTFR(ProtocolTrace,
"%15s %3s %10s%20s %6s>%-6s %#x %s\n",
920 RubyRequestType_to_string(secondary_type));
927 msg->m_htmFromTransaction =
true;
939template <
class KEY,
class VALUE>
941operator<<(std::ostream &out,
const std::unordered_map<KEY, VALUE> &map)
943 for (
const auto &table_entry : map) {
944 out <<
"[ " << table_entry.first <<
" =";
945 for (
const auto &seq_req : table_entry.second) {
946 out <<
" " << RubyRequestType_to_string(seq_req.m_second_type);
965 DPRINTF(RubyStats,
"Recorded statistic: %s\n",
966 SequencerRequestType_to_string(requestType));
982 uint64_t aligned_txid = \
985 if (aligned_txid > 0xFFFFFFFFull) {
994 uint64_t(
m_version & 0xFFFFFFFF) << 32) |
Cycles curCycle() const
Determine the current cycle, corresponding to a tick aligned to a clock edge.
Tick clockEdge(Cycles cycles=Cycles(0)) const
Determine the tick when a cycle begins, by default the current one, but the argument also enables the...
Tick cyclesToTicks(Cycles c) const
Cycles is a wrapper class for representing cycle counts, i.e.
virtual std::string name() const
A Packet is used to encapsulate a transfer between two objects in the memory system (e....
const std::string & cmdString() const
Return the string name of the cmd field (for debugging and tracing).
T * findNextSenderState() const
Go through the sender state stack and return the first instance that is of type T (as determined by a...
bool isHtmTransactional() const
Returns whether or not this packet/request originates in the CPU executing in transactional mode,...
void setData(const uint8_t *p)
Copy data into the packet from the provided pointer.
uint64_t getHtmTransactionUid() const
If a packet/request originates in a CPU executing in transactional mode, i.e.
RequestPtr req
A pointer to the original request.
AtomicOpFunctor * getAtomicOp() const
Accessor function to atomic op.
bool isMaskedWrite() const
void writeData(uint8_t *p) const
Copy data from the packet to the memory at the provided pointer.
bool isLocked(int context) const
void setLocked(int context)
virtual void notifyCoalesced(const Addr &addr, const RubyRequestType &type, const RequestPtr &req, const DataBlock &data_blk, const bool &was_miss)
Notifies controller of a request coalesced at the sequencer.
virtual Cycles mandatoryQueueLatency(const RubyRequestType ¶m_type)
bool isBlocked(Addr) const
void blockOnQueue(Addr, MessageBuffer *)
void clearLockedAll(int context)
AbstractCacheEntry * lookup(Addr address)
void enqueue(MsgPtr message, Tick curTime, Tick delta)
void ruby_hit_callback(PacketPtr pkt)
void ruby_unaddressed_callback(PacketPtr pkt)
void ruby_stale_translation_callback(Addr txnId)
RubySystem * m_ruby_system
virtual int functionalWrite(Packet *func_pkt)
AbstractController * m_controller
void ruby_eviction_callback(Addr address)
MessageBuffer * m_mandatory_q_ptr
static bool getWarmupEnabled()
static uint32_t getBlockSizeBits()
static bool getCooldownEnabled()
statistics::Histogram m_outstandReqHist
Histogram for number of outstanding requests per cycle.
void resetStats() override
Callback to reset stats.
Cycles m_deadlock_threshold
std::unordered_map< Addr, std::list< SequencerRequest > > m_RequestTable
virtual bool empty() const
void writeCallback(Addr address, DataBlock &data, const bool externalHit=false, const MachineType mach=MachineType_NUM, const Cycles initialRequestTime=Cycles(0), const Cycles forwardRequestTime=Cycles(0), const Cycles firstResponseTime=Cycles(0), const bool noCoales=false)
std::vector< statistics::Counter > m_IncompleteTimes
bool llscCheckMonitor(const Addr)
Searches for cache line address in the global monitor tagged with this Sequencer object's version id.
virtual int functionalWrite(Packet *func_pkt) override
std::vector< statistics::Histogram * > m_InitialToForwardDelayHist
std::vector< statistics::Histogram * > m_hitMachLatencyHist
Histograms for profiling the latencies for requests that did not required external messages.
void recordMissLatency(SequencerRequest *srequest, bool llscSuccess, const MachineType respondingMach, bool isExternalHit, Cycles initialRequestTime, Cycles forwardRequestTime, Cycles firstResponseTime)
std::vector< statistics::Histogram * > m_typeLatencyHist
void writeCallbackScFail(Addr address, DataBlock &data)
Proxy function to writeCallback that first invalidates the line address in the local monitor.
CacheMemory * m_dataCache_ptr
std::vector< statistics::Histogram * > m_FirstResponseToCompletionDelayHist
void incrementUnaddressedTransactionCnt()
Increment the unaddressed transaction counter.
void hitCallback(SequencerRequest *srequest, DataBlock &data, bool llscSuccess, const MachineType mach, const bool externalHit, const Cycles initialRequestTime, const Cycles forwardRequestTime, const Cycles firstResponseTime, const bool was_coalesced)
void llscLoadLinked(const Addr)
Places the cache line address into the global monitor tagged with this Sequencer object's version id.
uint64_t getCurrentUnaddressedTransactionID() const
Generate the current unaddressed transaction ID based on the counter and the Sequencer object's versi...
Sequencer(const Params &)
statistics::Histogram m_latencyHist
Histogram for holding latency profile of all requests.
void issueRequest(PacketPtr pkt, RubyRequestType type)
bool llscStoreConditional(const Addr)
Searches for cache line address in the global monitor tagged with this Sequencer object's version id.
void unaddressedCallback(Addr unaddressedReqId, RubyRequestType requestType, const MachineType mach=MachineType_NUM, const Cycles initialRequestTime=Cycles(0), const Cycles forwardRequestTime=Cycles(0), const Cycles firstResponseTime=Cycles(0))
virtual RequestStatus insertRequest(PacketPtr pkt, RubyRequestType primary_type, RubyRequestType secondary_type)
EventFunctionWrapper deadlockCheckEvent
std::unordered_map< uint64_t, SequencerRequest > m_UnaddressedRequestTable
uint64_t m_unaddressedTransactionCnt
std::vector< statistics::Histogram * > m_hitTypeLatencyHist
statistics::Histogram m_hitLatencyHist
Histogram for holding latency profile of all requests that hit in the controller connected to this se...
RequestStatus makeRequest(PacketPtr pkt) override
int m_max_outstanding_requests
void llscClearLocalMonitor()
Removes all addresses from the local monitor.
void recordRequestType(SequencerRequestType requestType)
std::vector< statistics::Histogram * > m_IssueToInitialDelayHist
Histograms for recording the breakdown of miss latency.
virtual void print(std::ostream &out) const
RubySequencerParams Params
std::vector< std::vector< statistics::Histogram * > > m_hitTypeMachLatencyHist
void llscClearMonitor(const Addr)
Removes the cache line address from the global monitor.
void evictionCallback(Addr address)
std::vector< statistics::Histogram * > m_ForwardToFirstResponseDelayHist
void readCallback(Addr address, DataBlock &data, const bool externalHit=false, const MachineType mach=MachineType_NUM, const Cycles initialRequestTime=Cycles(0), const Cycles forwardRequestTime=Cycles(0), const Cycles firstResponseTime=Cycles(0))
std::vector< statistics::Histogram * > m_missMachLatencyHist
Histograms for profiling the latencies for requests that required external messages.
std::vector< statistics::Histogram * > m_missTypeLatencyHist
statistics::Histogram m_missLatencyHist
Histogram for holding latency profile of all requests that miss in the controller connected to this s...
std::vector< std::vector< statistics::Histogram * > > m_missTypeMachLatencyHist
void mergeFrom(const DataBlock &data)
void sample(const U &v, int n=1)
Add a value to the distribtion n times.
void reset()
Reset stat value to default.
DrainState drainState() const
Return the current drain state of an object.
@ Draining
Draining buffers pending serialization/handover.
bool scheduled() const
Determine if the current event is scheduled.
void schedule(Event &event, Tick when)
#define panic(...)
This implements a cprintf based panic() function.
#define fatal_if(cond,...)
Conditional fatal macro that checks the supplied condition and only causes a fatal error if the condi...
RubyRequestType tlbiCmdToRubyRequestType(const Packet *pkt)
bool isTlbiCmdRequest(RubyRequestType type)
Addr makeLineAddress(Addr addr)
Addr getOffset(Addr addr)
RubyRequestType htmCmdToRubyRequestType(const Packet *pkt)
std::string printAddress(Addr addr)
Reference material can be found at the JEDEC website: UFS standard http://www.jedec....
Tick curTick()
The universal simulation clock.
uint64_t Addr
Address type This will probably be moved somewhere else in the near future.
std::ostream & operator<<(std::ostream &os, const ArmSemihosting::InPlaceArg &ipa)
uint64_t Tick
Tick count type.
int ContextID
Globally unique thread context ID.
const ContextID InvalidContextID
Declaration of the Packet class.
RubyRequestType m_second_type