52 #include "debug/RubyCacheTrace.hh"
53 #include "debug/RubySystem.hh"
81 m_cache_recorder(NULL)
119 for (
int idx = 0; idx <
m_networks.size(); ++idx) {
125 fatal_if(network_id < 0,
"Could not add MachineID %s. Network not found",
148 MachineID mach_id = cntrl->getMachineID();
152 "No machineID %s. Does not belong to a Ruby network?",
163 for (
auto id = 0;
id <
params().system->maxRequestors(); ++
id) {
177 uint64_t cache_trace_size,
178 uint64_t block_size_bytes)
185 if (sequencer_ptr == NULL) {
186 sequencer_ptr = sequencer_map[cntrl];
190 assert(sequencer_ptr != NULL);
193 if (sequencer_map[cntrl] == NULL) {
194 sequencer_map[cntrl] = sequencer_ptr;
205 sequencer_map, block_size_bytes);
214 DPRINTF(RubyCacheTrace,
"Recording Cache Trace\n");
219 DPRINTF(RubyCacheTrace,
"Cache Trace Complete\n");
223 DPRINTF(RubyCacheTrace,
"Recording current tick %ld\n", curtick_original);
231 DPRINTF(RubyCacheTrace,
"Event %s auto-deletes when descheduled,"
232 " not recording\n", curr_head->
name());
234 original_events.push_back(
235 std::make_pair(curr_head, curr_head->
when()));
241 DPRINTF(RubyCacheTrace,
"Starting cache flush\n");
244 DPRINTF(RubyCacheTrace,
"Cache flush complete\n");
257 while (!original_events.empty()) {
260 original_events.pop_back();
274 warn_once(
"Ruby memory writeback is experimental. Continuing simulation "
275 "afterwards may not always work as intended.");
283 uint64_t uncompressed_trace_size)
288 int fd = creat(thefile.c_str(), 0664);
291 fatal(
"Can't open memory trace file '%s'\n", filename);
294 gzFile compressedMemory = gzdopen(
fd,
"wb");
295 if (compressedMemory == NULL)
296 fatal(
"Insufficient memory to allocate compression state for %s\n",
299 if (gzwrite(compressedMemory, raw_data, uncompressed_trace_size) !=
300 uncompressed_trace_size) {
301 fatal(
"Write failed on memory trace file '%s'\n", filename);
304 if (gzclose(compressedMemory)) {
305 fatal(
"Close failed on memory trace file '%s'\n", filename);
323 fatal(
"Call memWriteback() before serialize() to create ruby trace");
327 uint8_t *raw_data =
new uint8_t[4096];
330 std::string cache_trace_file =
name() +
".cache.gz";
350 uint64_t &uncompressed_trace_size)
353 gzFile compressedTrace;
356 int fd = open(filename.c_str(), O_RDONLY);
359 fatal(
"Unable to open trace file %s", filename);
362 compressedTrace = gzdopen(
fd,
"rb");
363 if (compressedTrace == NULL) {
364 fatal(
"Insufficient memory to allocate compression state for %s\n",
368 raw_data =
new uint8_t[uncompressed_trace_size];
369 if (gzread(compressedTrace, raw_data, uncompressed_trace_size) <
370 uncompressed_trace_size) {
371 fatal(
"Unable to read complete trace from file %s\n", filename);
374 if (gzclose(compressedTrace)) {
375 fatal(
"Failed to close cache trace file '%s'\n", filename);
382 uint8_t *uncompressed_trace = NULL;
390 std::string cache_trace_file;
391 uint64_t cache_trace_size = 0;
395 cache_trace_file = cp.
getCptDir() +
"/" + cache_trace_file;
433 DPRINTF(RubyCacheTrace,
"Starting ruby cache warmup\n");
478 network->resetStats();
482 #ifndef PARTIAL_FUNC_READS
489 AccessPermission access_perm = AccessPermission_NotPresent;
493 unsigned int num_ro = 0;
494 unsigned int num_rw = 0;
495 unsigned int num_busy = 0;
496 unsigned int num_maybe_stale = 0;
497 unsigned int num_backing_store = 0;
498 unsigned int num_invalid = 0;
511 for (
auto& cntrl :
netCntrls[request_net_id]) {
512 access_perm = cntrl-> getAccessPermission(line_address);
513 if (access_perm == AccessPermission_Read_Only){
515 if (ctrl_ro ==
nullptr) ctrl_ro = cntrl;
517 else if (access_perm == AccessPermission_Read_Write){
519 if (ctrl_rw ==
nullptr) ctrl_rw = cntrl;
521 else if (access_perm == AccessPermission_Busy)
523 else if (access_perm == AccessPermission_Maybe_Stale)
525 else if (access_perm == AccessPermission_Backing_Store) {
531 if (ctrl_backing_store ==
nullptr)
532 ctrl_backing_store = cntrl;
534 else if (access_perm == AccessPermission_Invalid ||
535 access_perm == AccessPermission_NotPresent)
547 int num_controllers =
netCntrls[request_net_id].size();
548 if (num_invalid == (num_controllers - 1) && num_backing_store == 1) {
552 }
else if (num_ro > 0 || num_rw >= 1) {
557 warn(
"More than one Abstract Controller with RW permission for "
558 "addr: %#x on cacheline: %#x.", address, line_address);
566 num_maybe_stale, num_busy, num_ro, num_rw);
576 }
else if ((num_busy + num_maybe_stale) > 0) {
581 "(num_maybe_stale=%d, num_busy = %d)\n",
582 num_maybe_stale, num_busy);
583 for (
auto& cntrl :
netCntrls[request_net_id]) {
584 if (cntrl->functionalReadBuffers(pkt))
588 "(num_maybe_stale=%d, num_busy = %d)\n",
589 num_maybe_stale, num_busy);
591 if (network->functionalRead(pkt))
615 switch(ctrl->getAccessPermission(line_address)) {
616 case AccessPermission_Read_Only:
617 ctrl_ro.push_back(ctrl);
619 case AccessPermission_Busy:
620 ctrl_busy.push_back(ctrl);
622 case AccessPermission_Read_Write:
623 assert(ctrl_rw ==
nullptr);
626 case AccessPermission_Backing_Store:
627 assert(ctrl_bs ==
nullptr);
630 case AccessPermission_Backing_Store_Busy:
631 assert(ctrl_bs ==
nullptr);
633 ctrl_busy.push_back(ctrl);
636 ctrl_others.push_back(ctrl);
642 "backing_store=%d\n",
643 ctrl_ro.size(), ctrl_busy.size(),
644 ctrl_rw !=
nullptr, ctrl_bs !=
nullptr);
649 if (ctrl_rw !=
nullptr) {
657 for (
auto ctrl : ctrl_ro)
658 ctrl->functionalRead(line_address, pkt, bytes);
666 if (!ctrl_busy.empty() || !bytes.isFull()) {
668 "buffers and networks\n");
669 if (ctrl_rw !=
nullptr)
671 for (
auto ctrl : ctrl_ro)
672 ctrl->functionalReadBuffers(pkt, bytes);
673 if (ctrl_bs !=
nullptr)
675 for (
auto ctrl : ctrl_busy) {
676 ctrl->functionalRead(line_address, pkt, bytes);
677 ctrl->functionalReadBuffers(pkt, bytes);
680 network->functionalRead(pkt, bytes);
682 for (
auto ctrl : ctrl_others) {
683 ctrl->functionalRead(line_address, pkt, bytes);
684 ctrl->functionalReadBuffers(pkt, bytes);
688 panic_if(!(bytes.isFull() || bytes.isEmpty()),
689 "Inconsistent state on functional read for %#x %s\n",
692 return bytes.isFull();
705 AccessPermission access_perm = AccessPermission_NotPresent;
709 [[maybe_unused]] uint32_t num_functional_writes = 0;
716 for (
auto& cntrl :
netCntrls[request_net_id]) {
717 num_functional_writes += cntrl->functionalWriteBuffers(pkt);
719 access_perm = cntrl->getAccessPermission(line_addr);
720 if (access_perm != AccessPermission_Invalid &&
721 access_perm != AccessPermission_NotPresent) {
722 num_functional_writes +=
723 cntrl->functionalWrite(line_addr, pkt);
728 if (cntrl->getCPUSequencer()) {
729 num_functional_writes +=
730 cntrl->getCPUSequencer()->functionalWrite(pkt);
732 if (cntrl->getDMASequencer()) {
733 num_functional_writes +=
734 cntrl->getDMASequencer()->functionalWrite(pkt);
739 num_functional_writes += network->functionalWrite(pkt);