gem5  v20.0.0.0
All Classes Namespaces Files Functions Variables Typedefs Enumerations Enumerator Friends Macros Modules Pages
Sequencer.cc
Go to the documentation of this file.
1 /*
2  * Copyright (c) 2019-2020 ARM Limited
3  * All rights reserved.
4  *
5  * The license below extends only to copyright in the software and shall
6  * not be construed as granting a license to any other intellectual
7  * property including but not limited to intellectual property relating
8  * to a hardware implementation of the functionality of the software
9  * licensed hereunder. You may use the software subject to the license
10  * terms below provided that you ensure that this notice is replicated
11  * unmodified and in its entirety in all distributions of the software,
12  * modified or unmodified, in source code or in binary form.
13  *
14  * Copyright (c) 1999-2008 Mark D. Hill and David A. Wood
15  * Copyright (c) 2013 Advanced Micro Devices, Inc.
16  * All rights reserved.
17  *
18  * Redistribution and use in source and binary forms, with or without
19  * modification, are permitted provided that the following conditions are
20  * met: redistributions of source code must retain the above copyright
21  * notice, this list of conditions and the following disclaimer;
22  * redistributions in binary form must reproduce the above copyright
23  * notice, this list of conditions and the following disclaimer in the
24  * documentation and/or other materials provided with the distribution;
25  * neither the name of the copyright holders nor the names of its
26  * contributors may be used to endorse or promote products derived from
27  * this software without specific prior written permission.
28  *
29  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
30  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
31  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
32  * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
33  * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
34  * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
35  * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
36  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
37  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
38  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
39  * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
40  */
41 
43 
44 #include "arch/x86/ldstflags.hh"
45 #include "base/logging.hh"
46 #include "base/str.hh"
48 #include "debug/LLSC.hh"
49 #include "debug/MemoryAccess.hh"
50 #include "debug/ProtocolTrace.hh"
51 #include "debug/RubySequencer.hh"
52 #include "debug/RubyStats.hh"
53 #include "mem/packet.hh"
55 #include "mem/ruby/protocol/PrefetchBit.hh"
56 #include "mem/ruby/protocol/RubyAccessMode.hh"
59 #include "sim/system.hh"
60 
61 using namespace std;
62 
63 Sequencer *
64 RubySequencerParams::create()
65 {
66  return new Sequencer(this);
67 }
68 
70  : RubyPort(p), m_IncompleteTimes(MachineType_NUM),
71  deadlockCheckEvent([this]{ wakeup(); }, "Sequencer deadlock check")
72 {
74 
75  m_instCache_ptr = p->icache;
76  m_dataCache_ptr = p->dcache;
77  m_max_outstanding_requests = p->max_outstanding_requests;
78  m_deadlock_threshold = p->deadlock_threshold;
79 
80  m_coreId = p->coreid; // for tracking the two CorePair sequencers
81  assert(m_max_outstanding_requests > 0);
82  assert(m_deadlock_threshold > 0);
83  assert(m_instCache_ptr != NULL);
84  assert(m_dataCache_ptr != NULL);
85 
86  m_runningGarnetStandalone = p->garnet_standalone;
87 }
88 
90 {
91 }
92 
93 void
95 {
96  AbstractCacheEntry *line = m_dataCache_ptr->lookup(claddr);
97  if (line) {
98  line->setLocked(m_version);
99  DPRINTF(LLSC, "LLSC Monitor - inserting load linked - "
100  "addr=0x%lx - cpu=%u\n", claddr, m_version);
101  }
102 }
103 
104 void
106 {
107  AbstractCacheEntry *line = m_dataCache_ptr->lookup(claddr);
108  if (line && line->isLocked(m_version)) {
109  line->clearLocked();
110  DPRINTF(LLSC, "LLSC Monitor - clearing due to store - "
111  "addr=0x%lx - cpu=%u\n", claddr, m_version);
112  }
113 }
114 
115 bool
117 {
118  AbstractCacheEntry *line = m_dataCache_ptr->lookup(claddr);
119  if (!line)
120  return false;
121 
122  DPRINTF(LLSC, "LLSC Monitor - clearing due to "
123  "store conditional - "
124  "addr=0x%lx - cpu=%u\n",
125  claddr, m_version);
126 
127  if (line->isLocked(m_version)) {
128  line->clearLocked();
129  return true;
130  } else {
131  line->clearLocked();
132  return false;
133  }
134 }
135 
136 bool
138 {
139  const Addr claddr = makeLineAddress(address);
140  AbstractCacheEntry *line = m_dataCache_ptr->lookup(claddr);
141  if (!line)
142  return false;
143 
144  if (line->isLocked(m_version)) {
145  return true;
146  } else {
147  return false;
148  }
149 }
150 
151 void
153 {
154  assert(drainState() != DrainState::Draining);
155 
156  // Check for deadlock of any of the requests
157  Cycles current_time = curCycle();
158 
159  // Check across all outstanding requests
160  int total_outstanding = 0;
161 
162  for (const auto &table_entry : m_RequestTable) {
163  for (const auto seq_req : table_entry.second) {
164  if (current_time - seq_req.issue_time < m_deadlock_threshold)
165  continue;
166 
167  panic("Possible Deadlock detected. Aborting!\n version: %d "
168  "request.paddr: 0x%x m_readRequestTable: %d current time: "
169  "%u issue_time: %d difference: %d\n", m_version,
170  seq_req.pkt->getAddr(), table_entry.second.size(),
171  current_time * clockPeriod(), seq_req.issue_time
172  * clockPeriod(), (current_time * clockPeriod())
173  - (seq_req.issue_time * clockPeriod()));
174  }
175  total_outstanding += table_entry.second.size();
176  }
177 
178  assert(m_outstanding_count == total_outstanding);
179 
180  if (m_outstanding_count > 0) {
181  // If there are still outstanding requests, keep checking
183  }
184 }
185 
186 int
188 {
189  int num_written = RubyPort::functionalWrite(func_pkt);
190 
191  for (const auto &table_entry : m_RequestTable) {
192  for (const auto& seq_req : table_entry.second) {
193  if (seq_req.functionalWrite(func_pkt))
194  ++num_written;
195  }
196  }
197 
198  return num_written;
199 }
200 
202 {
207  for (int i = 0; i < RubyRequestType_NUM; i++) {
208  m_typeLatencyHist[i]->reset();
209  m_hitTypeLatencyHist[i]->reset();
210  m_missTypeLatencyHist[i]->reset();
211  for (int j = 0; j < MachineType_NUM; j++) {
212  m_hitTypeMachLatencyHist[i][j]->reset();
213  m_missTypeMachLatencyHist[i][j]->reset();
214  }
215  }
216 
217  for (int i = 0; i < MachineType_NUM; i++) {
218  m_missMachLatencyHist[i]->reset();
219  m_hitMachLatencyHist[i]->reset();
220 
221  m_IssueToInitialDelayHist[i]->reset();
222  m_InitialToForwardDelayHist[i]->reset();
225 
226  m_IncompleteTimes[i] = 0;
227  }
228 }
229 
230 // Insert the request in the request table. Return RequestStatus_Aliased
231 // if the entry was already present.
232 RequestStatus
233 Sequencer::insertRequest(PacketPtr pkt, RubyRequestType primary_type,
234  RubyRequestType secondary_type)
235 {
236  // See if we should schedule a deadlock check
237  if (!deadlockCheckEvent.scheduled() &&
240  }
241 
242  Addr line_addr = makeLineAddress(pkt->getAddr());
243  // Check if there is any outstanding request for the same cache line.
244  auto &seq_req_list = m_RequestTable[line_addr];
245  // Create a default entry
246  seq_req_list.emplace_back(pkt, primary_type, secondary_type, curCycle());
248 
249  if (seq_req_list.size() > 1) {
250  return RequestStatus_Aliased;
251  }
252 
254 
255  return RequestStatus_Ready;
256 }
257 
258 void
260 {
262 }
263 
264 void
265 Sequencer::recordMissLatency(SequencerRequest* srequest, bool llscSuccess,
266  const MachineType respondingMach,
267  bool isExternalHit, Cycles initialRequestTime,
268  Cycles forwardRequestTime,
269  Cycles firstResponseTime)
270 {
271  RubyRequestType type = srequest->m_type;
272  Cycles issued_time = srequest->issue_time;
273  Cycles completion_time = curCycle();
274 
275  assert(curCycle() >= issued_time);
276  Cycles total_lat = completion_time - issued_time;
277 
278  if (initialRequestTime < issued_time) {
279  // if the request was combined in the protocol with an earlier request
280  // for the same address, it is possible that it will return an
281  // initialRequestTime corresponding the earlier request. Since Cycles
282  // is unsigned, we can't let this request get profiled below.
283 
284  total_lat = Cycles(0);
285  }
286 
287  DPRINTFR(ProtocolTrace, "%15s %3s %10s%20s %6s>%-6s %s %d cycles\n",
288  curTick(), m_version, "Seq", llscSuccess ? "Done" : "SC_Failed",
289  "", "", printAddress(srequest->pkt->getAddr()), total_lat);
290 
291  m_latencyHist.sample(total_lat);
292  m_typeLatencyHist[type]->sample(total_lat);
293 
294  if (isExternalHit) {
295  m_missLatencyHist.sample(total_lat);
296  m_missTypeLatencyHist[type]->sample(total_lat);
297 
298  if (respondingMach != MachineType_NUM) {
299  m_missMachLatencyHist[respondingMach]->sample(total_lat);
300  m_missTypeMachLatencyHist[type][respondingMach]->sample(total_lat);
301 
302  if ((issued_time <= initialRequestTime) &&
303  (initialRequestTime <= forwardRequestTime) &&
304  (forwardRequestTime <= firstResponseTime) &&
305  (firstResponseTime <= completion_time)) {
306 
307  m_IssueToInitialDelayHist[respondingMach]->sample(
308  initialRequestTime - issued_time);
309  m_InitialToForwardDelayHist[respondingMach]->sample(
310  forwardRequestTime - initialRequestTime);
311  m_ForwardToFirstResponseDelayHist[respondingMach]->sample(
312  firstResponseTime - forwardRequestTime);
313  m_FirstResponseToCompletionDelayHist[respondingMach]->sample(
314  completion_time - firstResponseTime);
315  } else {
316  m_IncompleteTimes[respondingMach]++;
317  }
318  }
319  } else {
320  m_hitLatencyHist.sample(total_lat);
321  m_hitTypeLatencyHist[type]->sample(total_lat);
322 
323  if (respondingMach != MachineType_NUM) {
324  m_hitMachLatencyHist[respondingMach]->sample(total_lat);
325  m_hitTypeMachLatencyHist[type][respondingMach]->sample(total_lat);
326  }
327  }
328 }
329 
330 void
332 {
333  llscClearMonitor(address);
334  writeCallback(address, data);
335 }
336 
337 void
339  const bool externalHit, const MachineType mach,
340  const Cycles initialRequestTime,
341  const Cycles forwardRequestTime,
342  const Cycles firstResponseTime)
343 {
344  //
345  // Free the whole list as we assume we have had the exclusive access
346  // to this cache line when response for the write comes back
347  //
348  assert(address == makeLineAddress(address));
349  assert(m_RequestTable.find(address) != m_RequestTable.end());
350  auto &seq_req_list = m_RequestTable[address];
351 
352  // Perform hitCallback on every cpu request made to this cache block while
353  // ruby request was outstanding. Since only 1 ruby request was made,
354  // profile the ruby latency once.
355  bool ruby_request = true;
356  int aliased_stores = 0;
357  int aliased_loads = 0;
358  while (!seq_req_list.empty()) {
359  SequencerRequest &seq_req = seq_req_list.front();
360  if (ruby_request) {
361  assert(seq_req.m_type != RubyRequestType_LD);
362  assert(seq_req.m_type != RubyRequestType_Load_Linked);
363  assert(seq_req.m_type != RubyRequestType_IFETCH);
364  }
365 
366  // handle write request
367  if ((seq_req.m_type != RubyRequestType_LD) &&
368  (seq_req.m_type != RubyRequestType_Load_Linked) &&
369  (seq_req.m_type != RubyRequestType_IFETCH)) {
370  // LL/SC support (tested with ARMv8)
371  bool success = false;
372 
373  if (seq_req.m_type != RubyRequestType_Store_Conditional) {
374  // Regular stores to addresses being monitored
375  // will fail (remove) the monitor entry.
376  llscClearMonitor(address);
377  } else {
378  // Store conditionals must first check the monitor
379  // if they will succeed or not
380  success = llscStoreConditional(address);
381  seq_req.pkt->req->setExtraData(success ? 1 : 0);
382  }
383 
384  // Handle SLICC block_on behavior for Locked_RMW accesses. NOTE: the
385  // address variable here is assumed to be a line address, so when
386  // blocking buffers, must check line addresses.
387  if (seq_req.m_type == RubyRequestType_Locked_RMW_Read) {
388  // blockOnQueue blocks all first-level cache controller queues
389  // waiting on memory accesses for the specified address that go
390  // to the specified queue. In this case, a Locked_RMW_Write must
391  // go to the mandatory_q before unblocking the first-level
392  // controller. This will block standard loads, stores, ifetches,
393  // etc.
395  } else if (seq_req.m_type == RubyRequestType_Locked_RMW_Write) {
396  m_controller->unblock(address);
397  }
398 
399  if (ruby_request) {
400  recordMissLatency(&seq_req, success, mach, externalHit,
401  initialRequestTime, forwardRequestTime,
402  firstResponseTime);
403  } else {
404  aliased_stores++;
405  }
406  markRemoved();
407  ruby_request = false;
408  hitCallback(&seq_req, data, success, mach, externalHit,
409  initialRequestTime, forwardRequestTime,
410  firstResponseTime);
411  } else {
412  // handle read request
413  assert(!ruby_request);
414  markRemoved();
415  ruby_request = false;
416  aliased_loads++;
417  hitCallback(&seq_req, data, true, mach, externalHit,
418  initialRequestTime, forwardRequestTime,
419  firstResponseTime);
420  }
421  seq_req_list.pop_front();
422  }
423 
424  // free all outstanding requests corresponding to this address
425  if (seq_req_list.empty()) {
426  m_RequestTable.erase(address);
427  }
428 }
429 
430 void
432  bool externalHit, const MachineType mach,
433  Cycles initialRequestTime,
434  Cycles forwardRequestTime,
435  Cycles firstResponseTime)
436 {
437  //
438  // Free up read requests until we hit the first Write request
439  // or end of the corresponding list.
440  //
441  assert(address == makeLineAddress(address));
442  assert(m_RequestTable.find(address) != m_RequestTable.end());
443  auto &seq_req_list = m_RequestTable[address];
444 
445  // Perform hitCallback on every cpu request made to this cache block while
446  // ruby request was outstanding. Since only 1 ruby request was made,
447  // profile the ruby latency once.
448  bool ruby_request = true;
449  int aliased_loads = 0;
450  while (!seq_req_list.empty()) {
451  SequencerRequest &seq_req = seq_req_list.front();
452  if (ruby_request) {
453  assert((seq_req.m_type == RubyRequestType_LD) ||
454  (seq_req.m_type == RubyRequestType_Load_Linked) ||
455  (seq_req.m_type == RubyRequestType_IFETCH));
456  } else {
457  aliased_loads++;
458  }
459  if ((seq_req.m_type != RubyRequestType_LD) &&
460  (seq_req.m_type != RubyRequestType_Load_Linked) &&
461  (seq_req.m_type != RubyRequestType_IFETCH)) {
462  // Write request: reissue request to the cache hierarchy
463  issueRequest(seq_req.pkt, seq_req.m_second_type);
464  break;
465  }
466  if (ruby_request) {
467  recordMissLatency(&seq_req, true, mach, externalHit,
468  initialRequestTime, forwardRequestTime,
469  firstResponseTime);
470  }
471  markRemoved();
472  ruby_request = false;
473  hitCallback(&seq_req, data, true, mach, externalHit,
474  initialRequestTime, forwardRequestTime,
475  firstResponseTime);
476  seq_req_list.pop_front();
477  }
478 
479  // free all outstanding requests corresponding to this address
480  if (seq_req_list.empty()) {
481  m_RequestTable.erase(address);
482  }
483 }
484 
485 void
487  bool llscSuccess,
488  const MachineType mach, const bool externalHit,
489  const Cycles initialRequestTime,
490  const Cycles forwardRequestTime,
491  const Cycles firstResponseTime)
492 {
493  warn_once("Replacement policy updates recently became the responsibility "
494  "of SLICC state machines. Make sure to setMRU() near callbacks "
495  "in .sm files!");
496 
497  PacketPtr pkt = srequest->pkt;
498  Addr request_address(pkt->getAddr());
499  RubyRequestType type = srequest->m_type;
500 
501  // Load-linked handling
502  if (type == RubyRequestType_Load_Linked) {
503  Addr line_addr = makeLineAddress(request_address);
504  llscLoadLinked(line_addr);
505  }
506 
507  // update the data unless it is a non-data-carrying flush
509  data.setData(pkt->getConstPtr<uint8_t>(),
510  getOffset(request_address), pkt->getSize());
511  } else if (!pkt->isFlush()) {
512  if ((type == RubyRequestType_LD) ||
513  (type == RubyRequestType_IFETCH) ||
514  (type == RubyRequestType_RMW_Read) ||
515  (type == RubyRequestType_Locked_RMW_Read) ||
516  (type == RubyRequestType_Load_Linked)) {
517  pkt->setData(
518  data.getData(getOffset(request_address), pkt->getSize()));
519  DPRINTF(RubySequencer, "read data %s\n", data);
520  } else if (pkt->req->isSwap()) {
521  std::vector<uint8_t> overwrite_val(pkt->getSize());
522  pkt->writeData(&overwrite_val[0]);
523  pkt->setData(
524  data.getData(getOffset(request_address), pkt->getSize()));
525  data.setData(&overwrite_val[0],
526  getOffset(request_address), pkt->getSize());
527  DPRINTF(RubySequencer, "swap data %s\n", data);
528  } else if (type != RubyRequestType_Store_Conditional || llscSuccess) {
529  // Types of stores set the actual data here, apart from
530  // failed Store Conditional requests
531  data.setData(pkt->getConstPtr<uint8_t>(),
532  getOffset(request_address), pkt->getSize());
533  DPRINTF(RubySequencer, "set data %s\n", data);
534  }
535  }
536 
537  // If using the RubyTester, update the RubyTester sender state's
538  // subBlock with the recieved data. The tester will later access
539  // this state.
540  if (m_usingRubyTester) {
541  DPRINTF(RubySequencer, "hitCallback %s 0x%x using RubyTester\n",
542  pkt->cmdString(), pkt->getAddr());
543  RubyTester::SenderState* testerSenderState =
545  assert(testerSenderState);
546  testerSenderState->subBlock.mergeFrom(data);
547  }
548 
551  assert(pkt->req);
552  delete pkt;
554  } else if (RubySystem::getCooldownEnabled()) {
555  delete pkt;
557  } else {
558  ruby_hit_callback(pkt);
560  }
561 }
562 
563 bool
565 {
566  return m_RequestTable.empty();
567 }
568 
569 RequestStatus
571 {
573  return RequestStatus_BufferFull;
574  }
575 
576  RubyRequestType primary_type = RubyRequestType_NULL;
577  RubyRequestType secondary_type = RubyRequestType_NULL;
578 
579  if (pkt->isLLSC()) {
580  // LL/SC instructions need to be handled carefully by the cache
581  // coherence protocol to ensure they follow the proper semantics. In
582  // particular, by identifying the operations as atomic, the protocol
583  // should understand that migratory sharing optimizations should not
584  // be performed (i.e. a load between the LL and SC should not steal
585  // away exclusive permission).
586  //
587  // The following logic works correctly with the semantics
588  // of armV8 LDEX/STEX instructions.
589 
590  if (pkt->isWrite()) {
591  DPRINTF(RubySequencer, "Issuing SC\n");
592  primary_type = RubyRequestType_Store_Conditional;
593 #ifdef PROTOCOL_MESI_Three_Level
594  secondary_type = RubyRequestType_Store_Conditional;
595 #else
596  secondary_type = RubyRequestType_ST;
597 #endif
598  } else {
599  DPRINTF(RubySequencer, "Issuing LL\n");
600  assert(pkt->isRead());
601  primary_type = RubyRequestType_Load_Linked;
602  secondary_type = RubyRequestType_LD;
603  }
604  } else if (pkt->req->isLockedRMW()) {
605  //
606  // x86 locked instructions are translated to store cache coherence
607  // requests because these requests should always be treated as read
608  // exclusive operations and should leverage any migratory sharing
609  // optimization built into the protocol.
610  //
611  if (pkt->isWrite()) {
612  DPRINTF(RubySequencer, "Issuing Locked RMW Write\n");
613  primary_type = RubyRequestType_Locked_RMW_Write;
614  } else {
615  DPRINTF(RubySequencer, "Issuing Locked RMW Read\n");
616  assert(pkt->isRead());
617  primary_type = RubyRequestType_Locked_RMW_Read;
618  }
619  secondary_type = RubyRequestType_ST;
620  } else {
621  //
622  // To support SwapReq, we need to check isWrite() first: a SwapReq
623  // should always be treated like a write, but since a SwapReq implies
624  // both isWrite() and isRead() are true, check isWrite() first here.
625  //
626  if (pkt->isWrite()) {
627  //
628  // Note: M5 packets do not differentiate ST from RMW_Write
629  //
630  primary_type = secondary_type = RubyRequestType_ST;
631  } else if (pkt->isRead()) {
632  if (pkt->req->isInstFetch()) {
633  primary_type = secondary_type = RubyRequestType_IFETCH;
634  } else {
635  bool storeCheck = false;
636  // only X86 need the store check
637  if (system->getArch() == Arch::X86ISA) {
638  uint32_t flags = pkt->req->getFlags();
639  storeCheck = flags &
641  }
642  if (storeCheck) {
643  primary_type = RubyRequestType_RMW_Read;
644  secondary_type = RubyRequestType_ST;
645  } else {
646  primary_type = secondary_type = RubyRequestType_LD;
647  }
648  }
649  } else if (pkt->isFlush()) {
650  primary_type = secondary_type = RubyRequestType_FLUSH;
651  } else {
652  panic("Unsupported ruby packet type\n");
653  }
654  }
655 
656  // Check if the line is blocked for a Locked_RMW
658  (primary_type != RubyRequestType_Locked_RMW_Write)) {
659  // Return that this request's cache line address aliases with
660  // a prior request that locked the cache line. The request cannot
661  // proceed until the cache line is unlocked by a Locked_RMW_Write
662  return RequestStatus_Aliased;
663  }
664 
665  RequestStatus status = insertRequest(pkt, primary_type, secondary_type);
666 
667  // It is OK to receive RequestStatus_Aliased, it can be considered Issued
668  if (status != RequestStatus_Ready && status != RequestStatus_Aliased)
669  return status;
670  // non-aliased with any existing request in the request table, just issue
671  // to the cache
672  if (status != RequestStatus_Aliased)
673  issueRequest(pkt, secondary_type);
674 
675  // TODO: issue hardware prefetches here
676  return RequestStatus_Issued;
677 }
678 
679 void
680 Sequencer::issueRequest(PacketPtr pkt, RubyRequestType secondary_type)
681 {
682  assert(pkt != NULL);
683  ContextID proc_id = pkt->req->hasContextId() ?
684  pkt->req->contextId() : InvalidContextID;
685 
686  ContextID core_id = coreId();
687 
688  // If valid, copy the pc to the ruby request
689  Addr pc = 0;
690  if (pkt->req->hasPC()) {
691  pc = pkt->req->getPC();
692  }
693 
694  // check if the packet has data as for example prefetch and flush
695  // requests do not
696  std::shared_ptr<RubyRequest> msg =
697  std::make_shared<RubyRequest>(clockEdge(), pkt->getAddr(),
698  pkt->isFlush() ?
699  nullptr : pkt->getPtr<uint8_t>(),
700  pkt->getSize(), pc, secondary_type,
701  RubyAccessMode_Supervisor, pkt,
702  PrefetchBit_No, proc_id, core_id);
703 
704  DPRINTFR(ProtocolTrace, "%15s %3s %10s%20s %6s>%-6s %#x %s\n",
705  curTick(), m_version, "Seq", "Begin", "", "",
706  printAddress(msg->getPhysicalAddress()),
707  RubyRequestType_to_string(secondary_type));
708 
709  Tick latency = cyclesToTicks(
710  m_controller->mandatoryQueueLatency(secondary_type));
711  assert(latency > 0);
712 
713  assert(m_mandatory_q_ptr != NULL);
714  m_mandatory_q_ptr->enqueue(msg, clockEdge(), latency);
715 }
716 
717 template <class KEY, class VALUE>
718 std::ostream &
719 operator<<(ostream &out, const std::unordered_map<KEY, VALUE> &map)
720 {
721  for (const auto &table_entry : map) {
722  out << "[ " << table_entry.first << " =";
723  for (const auto &seq_req : table_entry.second) {
724  out << " " << RubyRequestType_to_string(seq_req.m_second_type);
725  }
726  }
727  out << " ]";
728 
729  return out;
730 }
731 
732 void
733 Sequencer::print(ostream& out) const
734 {
735  out << "[Sequencer: " << m_version
736  << ", outstanding requests: " << m_outstanding_count
737  << ", request table: " << m_RequestTable
738  << "]";
739 }
740 
741 // this can be called from setState whenever coherence permissions are
742 // upgraded when invoked, coherence violations will be checked for the
743 // given block
744 void
746 {
747 }
748 
749 void
750 Sequencer::recordRequestType(SequencerRequestType requestType) {
751  DPRINTF(RubyStats, "Recorded statistic: %s\n",
752  SequencerRequestType_to_string(requestType));
753 }
754 
755 void
757 {
758  llscClearMonitor(address);
759  ruby_eviction_callback(address);
760 }
761 
762 void
764 {
766 
767  // These statistical variables are not for display.
768  // The profiler will collate these across different
769  // sequencers and display those collated statistics.
771  m_latencyHist.init(10);
774 
775  for (int i = 0; i < RubyRequestType_NUM; i++) {
776  m_typeLatencyHist.push_back(new Stats::Histogram());
777  m_typeLatencyHist[i]->init(10);
778 
779  m_hitTypeLatencyHist.push_back(new Stats::Histogram());
780  m_hitTypeLatencyHist[i]->init(10);
781 
782  m_missTypeLatencyHist.push_back(new Stats::Histogram());
783  m_missTypeLatencyHist[i]->init(10);
784  }
785 
786  for (int i = 0; i < MachineType_NUM; i++) {
787  m_hitMachLatencyHist.push_back(new Stats::Histogram());
788  m_hitMachLatencyHist[i]->init(10);
789 
790  m_missMachLatencyHist.push_back(new Stats::Histogram());
791  m_missMachLatencyHist[i]->init(10);
792 
794  m_IssueToInitialDelayHist[i]->init(10);
795 
797  m_InitialToForwardDelayHist[i]->init(10);
798 
801 
804  }
805 
806  for (int i = 0; i < RubyRequestType_NUM; i++) {
809 
810  for (int j = 0; j < MachineType_NUM; j++) {
811  m_hitTypeMachLatencyHist[i].push_back(new Stats::Histogram());
812  m_hitTypeMachLatencyHist[i][j]->init(10);
813 
814  m_missTypeMachLatencyHist[i].push_back(new Stats::Histogram());
815  m_missTypeMachLatencyHist[i][j]->init(10);
816  }
817  }
818 }
void enqueueNextFetchRequest()
Function for fetching warming up the memory and the caches.
#define panic(...)
This implements a cprintf based panic() function.
Definition: logging.hh:163
#define DPRINTF(x,...)
Definition: trace.hh:225
int m_max_outstanding_requests
Definition: Sequencer.hh:206
const uint8_t * getData(int offset, int len) const
Definition: DataBlock.cc:95
std::unordered_map< Addr, std::list< SequencerRequest > > m_RequestTable
Definition: Sequencer.hh:220
const int FlagShift
Definition: ldstflags.hh:50
Cycles is a wrapper class for representing cycle counts, i.e.
Definition: types.hh:81
T * findNextSenderState() const
Go through the sender state stack and return the first instance that is of type T (as determined by a...
Definition: packet.hh:504
Cycles issue_time
Definition: Sequencer.hh:61
void recordMissLatency(SequencerRequest *srequest, bool llscSuccess, const MachineType respondingMach, bool isExternalHit, Cycles initialRequestTime, Cycles forwardRequestTime, Cycles firstResponseTime)
Definition: Sequencer.cc:265
Bitfield< 7 > i
std::vector< Stats::Histogram * > m_missTypeLatencyHist
Definition: Sequencer.hh:250
~Sequencer()
Definition: Sequencer.cc:89
void print(std::ostream &out) const
Definition: Sequencer.cc:733
RequestStatus makeRequest(PacketPtr pkt) override
Definition: Sequencer.cc:570
bool isLocked(int context) const
void llscClearMonitor(const Addr)
Removes the cache line address from the global monitor.
Definition: Sequencer.cc:105
CacheRecorder * m_cache_recorder
Definition: RubySystem.hh:139
EventFunctionWrapper deadlockCheckEvent
Definition: Sequencer.hh:264
int coreId() const
Definition: Sequencer.hh:131
AbstractController * m_controller
Definition: RubyPort.hh:189
ip6_addr_t addr
Definition: inet.hh:330
void llscLoadLinked(const Addr)
Places the cache line address into the global monitor tagged with this Sequencer object&#39;s version id...
Definition: Sequencer.cc:94
void regStats() override
Callback to set stat parameters.
Definition: Sequencer.cc:763
PacketPtr pkt
Definition: Sequencer.hh:58
std::vector< Stats::Histogram * > m_FirstResponseToCompletionDelayHist
Definition: Sequencer.hh:261
void recordRequestType(SequencerRequestType requestType)
Definition: Sequencer.cc:750
Histogram & init(size_type size)
Set the parameters of this histogram.
Definition: statistics.hh:2641
Overload hash function for BasicBlockRange type.
Definition: vec_reg.hh:587
Stats::Histogram m_missLatencyHist
Histogram for holding latency profile of all requests that miss in the controller connected to this s...
Definition: Sequencer.hh:249
Stats::Histogram m_latencyHist
Histogram for holding latency profile of all requests.
Definition: Sequencer.hh:234
CacheMemory * m_dataCache_ptr
Definition: Sequencer.hh:209
Tick clockPeriod() const
std::vector< Stats::Histogram * > m_typeLatencyHist
Definition: Sequencer.hh:235
bool isWrite() const
Definition: packet.hh:523
T * getPtr()
get a pointer to the data ptr.
Definition: packet.hh:1084
RequestStatus insertRequest(PacketPtr pkt, RubyRequestType primary_type, RubyRequestType secondary_type)
Definition: Sequencer.cc:233
bool isRead() const
Definition: packet.hh:522
RubySystem * m_ruby_system
Definition: RubyPort.hh:187
RubyPortParams Params
Definition: RubyPort.hh:145
Stats::Histogram m_outstandReqHist
Histogram for number of outstanding requests per cycle.
Definition: Sequencer.hh:231
std::vector< Stats::Histogram * > m_ForwardToFirstResponseDelayHist
Definition: Sequencer.hh:260
CacheMemory * m_instCache_ptr
Definition: Sequencer.hh:210
RequestPtr req
A pointer to the original request.
Definition: packet.hh:321
RubyRequestType m_type
Definition: Sequencer.hh:59
DrainState drainState() const
Return the current drain state of an object.
Definition: drain.hh:308
Bitfield< 5, 0 > status
std::vector< Stats::Histogram * > m_InitialToForwardDelayHist
Definition: Sequencer.hh:259
unsigned getSize() const
Definition: packet.hh:730
static bool getWarmupEnabled()
Definition: RubySystem.hh:62
Tick cyclesToTicks(Cycles c) const
uint8_t type
Definition: inet.hh:328
Tick curTick()
The current simulated tick.
Definition: core.hh:44
Bitfield< 4 > pc
void enqueueNextFlushRequest()
Function for flushing the memory contents of the caches to the main memory.
void ruby_eviction_callback(Addr address)
Definition: RubyPort.cc:598
void setData(const uint8_t *p)
Copy data into the packet from the provided pointer.
Definition: packet.hh:1152
uint64_t Tick
Tick count type.
Definition: types.hh:61
System * system
Definition: RubyPort.hh:192
void evictionCallback(Addr address)
Definition: Sequencer.cc:756
std::vector< Stats::Histogram * > m_hitTypeLatencyHist
Definition: Sequencer.hh:240
uint32_t m_version
Definition: RubyPort.hh:188
std::vector< Stats::Histogram * > m_IssueToInitialDelayHist
Histograms for recording the breakdown of miss latency.
Definition: Sequencer.hh:258
A simple histogram stat.
Definition: statistics.hh:2626
Addr getAddr() const
Definition: packet.hh:720
Cycles curCycle() const
Determine the current cycle, corresponding to a tick aligned to a clock edge.
void writeData(uint8_t *p) const
Copy data from the packet to the memory at the provided pointer.
Definition: packet.hh:1181
bool m_usingRubyTester
Definition: RubyPort.hh:191
void schedule(Event &event, Tick when)
Definition: eventq.hh:934
Addr getOffset(Addr addr)
Definition: Address.cc:48
void writeCallbackScFail(Addr address, DataBlock &data)
Proxy function to writeCallback that first invalidates the line address in the local monitor...
Definition: Sequencer.cc:331
virtual int functionalWrite(Packet *func_pkt) override
Definition: Sequencer.cc:187
static bool getCooldownEnabled()
Definition: RubySystem.hh:63
RubyRequestType m_second_type
Definition: Sequencer.hh:60
std::vector< Stats::Histogram * > m_missMachLatencyHist
Histograms for profiling the latencies for requests that required external messages.
Definition: Sequencer.hh:254
void ruby_hit_callback(PacketPtr pkt)
Definition: RubyPort.cc:426
void setLocked(int context)
uint64_t Addr
Address type This will probably be moved somewhere else in the near future.
Definition: types.hh:140
void issueRequest(PacketPtr pkt, RubyRequestType type)
Definition: Sequencer.cc:680
A Packet is used to encapsulate a transfer between two objects in the memory system (e...
Definition: packet.hh:249
Addr makeLineAddress(Addr addr)
Definition: Address.cc:54
Tick clockEdge(Cycles cycles=Cycles(0)) const
Determine the tick when a cycle begins, by default the current one, but the argument also enables the...
#define warn_once(...)
Definition: logging.hh:212
std::string printAddress(Addr addr)
Definition: Address.cc:67
void reset()
Reset stat value to default.
Definition: statistics.hh:1920
int m_coreId
Definition: Sequencer.hh:226
void blockOnQueue(Addr, MessageBuffer *)
Bitfield< 24 > j
bool scheduled() const
Determine if the current event is scheduled.
Definition: eventq.hh:459
bool isLLSC() const
Definition: packet.hh:548
MessageBuffer * m_mandatory_q_ptr
Definition: RubyPort.hh:190
void hitCallback(SequencerRequest *srequest, DataBlock &data, bool llscSuccess, const MachineType mach, const bool externalHit, const Cycles initialRequestTime, const Cycles forwardRequestTime, const Cycles firstResponseTime)
Definition: Sequencer.cc:486
std::vector< std::vector< Stats::Histogram * > > m_missTypeMachLatencyHist
Definition: Sequencer.hh:255
std::vector< Stats::Counter > m_IncompleteTimes
Definition: Sequencer.hh:262
void testDrainComplete()
Definition: RubyPort.cc:475
Declaration of the Packet class.
virtual Cycles mandatoryQueueLatency(const RubyRequestType &param_type)
bool empty() const
Definition: Sequencer.cc:564
bool m_runningGarnetStandalone
Definition: Sequencer.hh:228
void resetStats() override
Callback to reset stats.
Definition: Sequencer.cc:201
const ContextID InvalidContextID
Definition: types.hh:230
void markRemoved()
Definition: Sequencer.cc:259
const T * getConstPtr() const
Definition: packet.hh:1093
void setData(const uint8_t *data, int offset, int len)
Definition: DataBlock.cc:108
std::vector< Stats::Histogram * > m_hitMachLatencyHist
Histograms for profiling the latencies for requests that did not required external messages...
Definition: Sequencer.hh:244
virtual int functionalWrite(Packet *func_pkt)
Definition: RubyPort.cc:632
Cycles m_deadlock_threshold
Definition: Sequencer.hh:207
virtual void regStats()
Callback to set stat parameters.
Definition: group.cc:64
void wakeup()
Definition: Sequencer.cc:152
const std::string & cmdString() const
Return the string name of the cmd field (for debugging and tracing).
Definition: packet.hh:517
void readCallback(Addr address, DataBlock &data, const bool externalHit=false, const MachineType mach=MachineType_NUM, const Cycles initialRequestTime=Cycles(0), const Cycles forwardRequestTime=Cycles(0), const Cycles firstResponseTime=Cycles(0))
Definition: Sequencer.cc:431
bool llscCheckMonitor(const Addr)
Searches for cache line address in the global monitor tagged with this Sequencer object&#39;s version id...
Definition: Sequencer.cc:137
Bitfield< 0 > p
Running normally.
Bitfield< 9, 8 > rs
const char data[]
bool isFlush() const
Definition: packet.hh:551
AbstractCacheEntry * lookup(Addr address)
Definition: CacheMemory.cc:342
bool isBlocked(Addr) const
void writeCallback(Addr address, DataBlock &data, const bool externalHit=false, const MachineType mach=MachineType_NUM, const Cycles initialRequestTime=Cycles(0), const Cycles forwardRequestTime=Cycles(0), const Cycles firstResponseTime=Cycles(0))
Definition: Sequencer.cc:338
bool llscStoreConditional(const Addr)
Searches for cache line address in the global monitor tagged with this Sequencer object&#39;s version id...
Definition: Sequencer.cc:116
void enqueue(MsgPtr message, Tick curTime, Tick delta)
int m_outstanding_count
Definition: Sequencer.hh:223
Stats::Histogram m_hitLatencyHist
Histogram for holding latency profile of all requests that hit in the controller connected to this se...
Definition: Sequencer.hh:239
int ContextID
Globally unique thread context ID.
Definition: types.hh:229
void sample(const U &v, int n=1)
Add a value to the distribtion n times.
Definition: statistics.hh:1896
void checkCoherence(Addr address)
Definition: Sequencer.cc:745
std::vector< std::vector< Stats::Histogram * > > m_hitTypeMachLatencyHist
Definition: Sequencer.hh:245
Arch getArch() const
Get the architecture.
Definition: system.hh:248
#define DPRINTFR(...)
Definition: trace.hh:227
Sequencer(const Params *)
Definition: Sequencer.cc:69

Generated on Thu May 28 2020 16:21:35 for gem5 by doxygen 1.8.13