58 #include "debug/Mwait.hh"
59 #include "debug/SyscallVerbose.hh"
60 #include "debug/Thread.hh"
62 #include "params/BaseCPU.hh"
77 std::unique_ptr<BaseCPU::GlobalStats> BaseCPU::globalStats;
86 CPUProgressEvent::CPUProgressEvent(BaseCPU *_cpu,
Tick ival)
87 :
Event(
Event::Progress_Event_Pri), _interval(ival), lastNumInst(0),
88 cpu(_cpu), _repeatEvent(true)
91 cpu->schedule(
this,
curTick() + _interval);
95 CPUProgressEvent::process()
100 cpu->schedule(
this,
curTick() + _interval);
102 if (cpu->switchedOut()) {
107 double ipc = double(temp - lastNumInst) / (_interval / cpu->clockPeriod());
109 DPRINTFN(
"%s progress event, total committed:%i, progress insts committed: "
110 "%lli, IPC: %0.8d\n", cpu->name(), temp, temp - lastNumInst,
114 cprintf(
"%lli: %s progress event, total committed:%i, progress insts "
115 "committed: %lli\n",
curTick(), cpu->name(), temp,
122 CPUProgressEvent::description()
const
124 return "CPU Progress";
127 BaseCPU::BaseCPU(
const Params &
p,
bool is_checker)
128 : ClockedObject(
p), instCnt(0), _cpuId(
p.cpu_id), _socketId(
p.socket_id),
129 _instRequestorId(
p.
system->getRequestorId(this,
"inst")),
130 _dataRequestorId(
p.
system->getRequestorId(this,
"data")),
131 _taskId(context_switch_task_id::
Unknown), _pid(invldPid),
132 _switchedOut(
p.switched_out), _cacheLineSize(
p.
system->cacheLineSize()),
134 previousCycle(0), previousState(CPU_STATE_SLEEP),
135 functionTraceStream(nullptr), currentFunctionStart(0),
136 currentFunctionEnd(0), functionEntryTick(0),
138 addressMonitor(
p.numThreads),
139 syscallRetryLatency(
p.syscallRetryLatency),
140 pwrGatingLatency(
p.pwr_gating_latency),
141 powerGatingOnIdle(
p.power_gating_on_idle),
142 enterPwrGatingEvent([this]{ enterPwrGating(); },
name())
146 _cpuId = cpuList.size();
150 cpuList.push_back(
this);
152 DPRINTF(SyscallVerbose,
"Constructing CPU with id %d, socket id %d\n",
158 functionTracingEnabled =
false;
159 if (
p.function_trace) {
160 const std::string fname =
csprintf(
"ftrace.%s",
name());
163 currentFunctionStart = currentFunctionEnd = 0;
164 functionEntryTick =
p.function_trace_start;
166 if (
p.function_trace_start == 0) {
167 functionTracingEnabled =
true;
169 Event *
event =
new EventFunctionWrapper(
170 [
this]{ enableFunctionTrace(); },
name(),
true);
171 schedule(
event,
p.function_trace_start);
175 tracer = params().tracer;
177 if (params().isa.size() != numThreads) {
178 fatal(
"Number of ISAs (%i) assigned to the CPU does not equal number "
179 "of threads (%i).\n", params().isa.size(), numThreads);
184 BaseCPU::enableFunctionTrace()
186 functionTracingEnabled =
true;
196 interrupts[tid]->post(int_num,
index);
207 assert(tid < numThreads);
208 AddressMonitor &monitor = addressMonitor[tid];
210 monitor.armed =
true;
211 monitor.vAddr = address;
213 DPRINTF(Mwait,
"[tid:%d] Armed monitor (vAddr=0x%lx)\n", tid, address);
219 assert(tid < numThreads);
220 AddressMonitor &monitor = addressMonitor[tid];
222 if (!monitor.gotWakeup) {
223 int block_size = cacheLineSize();
224 uint64_t
mask = ~((uint64_t)(block_size - 1));
226 assert(pkt->req->hasPaddr());
227 monitor.pAddr = pkt->getAddr() &
mask;
228 monitor.waiting =
true;
230 DPRINTF(Mwait,
"[tid:%d] mwait called (vAddr=0x%lx, "
231 "line's paddr=0x%lx)\n", tid, monitor.vAddr, monitor.pAddr);
234 monitor.gotWakeup =
false;
240 BaseCPU::mwaitAtomic(
ThreadID tid, ThreadContext *tc, BaseMMU *mmu)
242 assert(tid < numThreads);
243 AddressMonitor &monitor = addressMonitor[tid];
248 int block_size = cacheLineSize();
249 uint64_t
mask = ~((uint64_t)(block_size - 1));
250 int size = block_size;
255 if (secondAddr >
addr)
256 size = secondAddr -
addr;
258 req->setVirt(
addr, size, 0x0, dataRequestorId(),
259 tc->pcState().instAddr());
262 Fault fault = mmu->translateAtomic(req, tc, BaseMMU::Read);
265 monitor.pAddr = req->getPaddr() &
mask;
266 monitor.waiting =
true;
268 DPRINTF(Mwait,
"[tid:%d] mwait called (vAddr=0x%lx, line's paddr=0x%lx)\n",
269 tid, monitor.vAddr, monitor.pAddr);
277 if (params().max_insts_any_thread != 0) {
278 const char *cause =
"a thread reached the max instruction count";
279 for (
ThreadID tid = 0; tid < numThreads; ++tid)
280 scheduleInstStop(tid, params().max_insts_any_thread, cause);
287 if (!params().simpoint_start_insts.empty()) {
288 const char *cause =
"simpoint starting point found";
289 for (
size_t i = 0;
i < params().simpoint_start_insts.size(); ++
i)
290 scheduleInstStop(0, params().simpoint_start_insts[
i], cause);
293 if (params().max_insts_all_threads != 0) {
294 const char *cause =
"all threads reached the max instruction count";
299 int *counter =
new int;
300 *counter = numThreads;
301 for (
ThreadID tid = 0; tid < numThreads; ++tid) {
302 Event *
event =
new CountedExitEvent(cause, *counter);
303 threadContexts[tid]->scheduleInstCountEvent(
304 event, params().max_insts_all_threads);
308 if (!params().switched_out) {
309 registerThreadContexts();
318 if (params().progress_interval) {
319 new CPUProgressEvent(
this, params().progress_interval);
326 if (powerState->get() == enums::PwrState::UNDEFINED)
327 powerState->set(enums::PwrState::ON);
332 BaseCPU::pmuProbePoint(
const char *
name)
341 BaseCPU::regProbePoints()
343 ppAllCycles = pmuProbePoint(
"Cycles");
344 ppActiveCycles = pmuProbePoint(
"ActiveCycles");
346 ppRetiredInsts = pmuProbePoint(
"RetiredInsts");
347 ppRetiredInstsPC = pmuProbePoint(
"RetiredInstsPC");
348 ppRetiredLoads = pmuProbePoint(
"RetiredLoads");
349 ppRetiredStores = pmuProbePoint(
"RetiredStores");
350 ppRetiredBranches = pmuProbePoint(
"RetiredBranches");
352 ppSleeping =
new ProbePointArg<bool>(this->getProbeManager(),
359 if (!inst->isMicroop() || inst->isLastMicroop()) {
360 ppRetiredInsts->notify(1);
361 ppRetiredInstsPC->notify(
pc);
365 ppRetiredLoads->notify(1);
367 if (inst->isStore() || inst->isAtomic())
368 ppRetiredStores->notify(1);
370 if (inst->isControl())
371 ppRetiredBranches->notify(1);
375 BaseCPUStats::BaseCPUStats(statistics::Group *parent)
376 : statistics::Group(parent),
377 ADD_STAT(numCycles, statistics::units::Cycle::get(),
378 "Number of cpu cycles simulated"),
379 ADD_STAT(numWorkItemsStarted, statistics::units::Count::get(),
380 "Number of work items this cpu started"),
381 ADD_STAT(numWorkItemsCompleted, statistics::units::Count::get(),
382 "Number of work items this cpu completed")
389 ClockedObject::regStats();
394 globalStats.reset(
new GlobalStats(Root::root()));
397 using namespace statistics;
399 int size = threadContexts.size();
401 for (
int i = 0;
i < size; ++
i) {
402 std::stringstream namestr;
404 threadContexts[
i]->regStats(namestr.str());
406 }
else if (size == 1)
407 threadContexts[0]->regStats(
name());
411 BaseCPU::getPort(
const std::string &if_name,
PortID idx)
416 if (if_name ==
"dcache_port")
417 return getDataPort();
418 else if (if_name ==
"icache_port")
419 return getInstPort();
421 return ClockedObject::getPort(if_name, idx);
425 BaseCPU::registerThreadContexts()
427 assert(
system->multiThread || numThreads == 1);
429 fatal_if(interrupts.size() != numThreads,
430 "CPU %s has %i interrupt controllers, but is expecting one "
432 name(), interrupts.size(), numThreads);
434 ThreadID size = threadContexts.size();
435 for (
ThreadID tid = 0; tid < size; ++tid) {
436 ThreadContext *tc = threadContexts[tid];
438 if (
system->multiThread) {
439 system->registerThreadContext(tc);
441 system->registerThreadContext(tc, _cpuId);
445 tc->getProcessPtr()->assignThreadContext(tc->contextId());
447 interrupts[tid]->setThreadContext(tc);
448 tc->getIsaPtr()->setThreadContext(tc);
453 BaseCPU::deschedulePowerGatingEvent()
455 if (enterPwrGatingEvent.scheduled()){
456 deschedule(enterPwrGatingEvent);
461 BaseCPU::schedulePowerGatingEvent()
463 for (
auto tc : threadContexts) {
464 if (tc->status() == ThreadContext::Active)
468 if (powerState->get() == enums::PwrState::CLK_GATED &&
470 assert(!enterPwrGatingEvent.scheduled());
473 schedule(enterPwrGatingEvent, clockEdge(pwrGatingLatency));
478 BaseCPU::findContext(ThreadContext *tc)
480 ThreadID size = threadContexts.size();
481 for (
ThreadID tid = 0; tid < size; ++tid) {
482 if (tc == threadContexts[tid])
489 BaseCPU::activateContext(
ThreadID thread_num)
491 DPRINTF(Thread,
"activate contextId %d\n",
492 threadContexts[thread_num]->contextId());
494 if (enterPwrGatingEvent.scheduled())
495 deschedule(enterPwrGatingEvent);
497 powerState->set(enums::PwrState::ON);
499 updateCycleCounters(CPU_STATE_WAKEUP);
503 BaseCPU::suspendContext(
ThreadID thread_num)
505 DPRINTF(Thread,
"suspend contextId %d\n",
506 threadContexts[thread_num]->contextId());
508 for (
auto t : threadContexts) {
509 if (
t->status() != ThreadContext::Suspended) {
515 updateCycleCounters(CPU_STATE_SLEEP);
518 powerState->set(enums::PwrState::CLK_GATED);
521 if (powerGatingOnIdle) {
524 schedule(enterPwrGatingEvent, clockEdge(pwrGatingLatency));
529 BaseCPU::haltContext(
ThreadID thread_num)
531 updateCycleCounters(BaseCPU::CPU_STATE_SLEEP);
535 BaseCPU::enterPwrGating(
void)
543 assert(!_switchedOut);
557 assert(threadContexts.size() == oldCPU->threadContexts.size());
558 assert(_cpuId == oldCPU->cpuId());
559 assert(_switchedOut);
560 assert(oldCPU !=
this);
561 _pid = oldCPU->getPid();
562 _taskId = oldCPU->taskId();
564 powerState->set(oldCPU->powerState->get());
566 previousState = oldCPU->previousState;
567 previousCycle = oldCPU->previousCycle;
569 _switchedOut =
false;
571 ThreadID size = threadContexts.size();
573 ThreadContext *newTC = threadContexts[
i];
574 ThreadContext *oldTC = oldCPU->threadContexts[
i];
576 newTC->getIsaPtr()->setThreadContext(newTC);
578 newTC->takeOverFrom(oldTC);
580 assert(newTC->contextId() == oldTC->contextId());
581 assert(newTC->threadId() == oldTC->threadId());
582 system->replaceThreadContext(newTC, newTC->contextId());
591 newTC->getMMUPtr()->takeOverFrom(oldTC->getMMUPtr());
595 CheckerCPU *old_checker = oldTC->getCheckerCpuPtr();
596 CheckerCPU *new_checker = newTC->getCheckerCpuPtr();
597 if (old_checker && new_checker) {
598 new_checker->getMMUPtr()->takeOverFrom(old_checker->getMMUPtr());
602 interrupts = oldCPU->interrupts;
603 for (
ThreadID tid = 0; tid < numThreads; tid++) {
604 interrupts[tid]->setThreadContext(threadContexts[tid]);
606 oldCPU->interrupts.clear();
612 getInstPort().takeOverFrom(&oldCPU->getInstPort());
613 getDataPort().takeOverFrom(&oldCPU->getDataPort());
619 for (
ThreadID i = 0;
i < threadContexts.size(); ++
i) {
620 ThreadContext &tc(*threadContexts[
i]);
621 CheckerCPU *checker(tc.getCheckerCpuPtr());
623 tc.getMMUPtr()->flushAll();
625 checker->getMMUPtr()->flushAll();
644 ScopedCheckpointSection sec(cp,
csprintf(
"xc.%i",
i));
645 interrupts[
i]->serialize(cp);
646 serializeThread(cp,
i);
661 ScopedCheckpointSection sec(cp,
csprintf(
"xc.%i",
i));
662 interrupts[
i]->unserialize(cp);
663 unserializeThread(cp,
i);
669 BaseCPU::scheduleInstStop(
ThreadID tid,
Counter insts,
const char *cause)
671 const Tick now(getCurrentInstCount(tid));
672 Event *
event(
new LocalSimLoopExitEvent(cause, 0));
674 threadContexts[tid]->scheduleInstCountEvent(
event, now + insts);
678 BaseCPU::getCurrentInstCount(
ThreadID tid)
680 return threadContexts[tid]->getCurrentInstCount();
683 AddressMonitor::AddressMonitor()
693 assert(pkt->req->hasPaddr());
694 if (armed && waiting) {
695 if (pAddr == pkt->getAddr()) {
696 DPRINTF(Mwait,
"pAddr=0x%lx invalidated: waking up core\n",
707 BaseCPU::traceFunctionsInternal(
Addr pc)
714 if (pc < currentFunctionStart || pc >= currentFunctionEnd) {
716 pc, currentFunctionEnd);
722 currentFunctionStart =
pc;
723 currentFunctionEnd =
pc + 1;
726 currentFunctionStart = it->address;
729 ccprintf(*functionTraceStream,
" (%d)\n%d: %s",
736 BaseCPU::GlobalStats::GlobalStats(statistics::Group *parent)
737 : statistics::Group(parent),
738 ADD_STAT(simInsts, statistics::units::Count::get(),
739 "Number of instructions simulated"),
740 ADD_STAT(simOps, statistics::units::Count::get(),
741 "Number of ops (including micro ops) simulated"),
742 ADD_STAT(hostInstRate, statistics::units::Rate<
743 statistics::units::Count, statistics::units::Second>::get(),
744 "Simulator instruction rate (inst/s)"),
745 ADD_STAT(hostOpRate, statistics::units::Rate<
746 statistics::units::Count, statistics::units::Second>::get(),
747 "Simulator op (including micro ops) rate (op/s)")
750 .functor(BaseCPU::numSimulatedInsts)
756 .functor(BaseCPU::numSimulatedOps)