gem5  v21.0.0.0
All Classes Namespaces Files Functions Variables Typedefs Enumerations Enumerator Friends Macros Modules Pages
fetch2.cc
Go to the documentation of this file.
1 /*
2  * Copyright (c) 2013-2014,2016 ARM Limited
3  * All rights reserved
4  *
5  * The license below extends only to copyright in the software and shall
6  * not be construed as granting a license to any other intellectual
7  * property including but not limited to intellectual property relating
8  * to a hardware implementation of the functionality of the software
9  * licensed hereunder. You may use the software subject to the license
10  * terms below provided that you ensure that this notice is replicated
11  * unmodified and in its entirety in all distributions of the software,
12  * modified or unmodified, in source code or in binary form.
13  *
14  * Redistribution and use in source and binary forms, with or without
15  * modification, are permitted provided that the following conditions are
16  * met: redistributions of source code must retain the above copyright
17  * notice, this list of conditions and the following disclaimer;
18  * redistributions in binary form must reproduce the above copyright
19  * notice, this list of conditions and the following disclaimer in the
20  * documentation and/or other materials provided with the distribution;
21  * neither the name of the copyright holders nor the names of its
22  * contributors may be used to endorse or promote products derived from
23  * this software without specific prior written permission.
24  *
25  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
26  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
27  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
28  * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
29  * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
30  * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
31  * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
32  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
33  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
34  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
35  * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
36  */
37 
38 #include "cpu/minor/fetch2.hh"
39 
40 #include <string>
41 
42 #include "arch/decoder.hh"
43 #include "arch/utility.hh"
44 #include "base/logging.hh"
45 #include "base/trace.hh"
46 #include "cpu/minor/pipeline.hh"
47 #include "cpu/pred/bpred_unit.hh"
48 #include "debug/Branch.hh"
49 #include "debug/Fetch.hh"
50 #include "debug/MinorTrace.hh"
51 
52 namespace Minor
53 {
54 
55 Fetch2::Fetch2(const std::string &name,
56  MinorCPU &cpu_,
57  const MinorCPUParams &params,
59  Latch<BranchData>::Output branchInp_,
60  Latch<BranchData>::Input predictionOut_,
62  std::vector<InputBuffer<ForwardInstData>> &next_stage_input_buffer) :
63  Named(name),
64  cpu(cpu_),
65  inp(inp_),
66  branchInp(branchInp_),
67  predictionOut(predictionOut_),
68  out(out_),
69  nextStageReserve(next_stage_input_buffer),
70  outputWidth(params.decodeInputWidth),
71  processMoreThanOneInput(params.fetch2CycleInput),
72  branchPredictor(*params.branchPred),
73  fetchInfo(params.numThreads),
74  threadPriority(0), stats(&cpu_)
75 {
76  if (outputWidth < 1)
77  fatal("%s: decodeInputWidth must be >= 1 (%d)\n", name, outputWidth);
78 
79  if (params.fetch2InputBufferSize < 1) {
80  fatal("%s: fetch2InputBufferSize must be >= 1 (%d)\n", name,
81  params.fetch2InputBufferSize);
82  }
83 
84  /* Per-thread input buffers */
85  for (ThreadID tid = 0; tid < params.numThreads; tid++) {
86  inputBuffer.push_back(
88  name + ".inputBuffer" + std::to_string(tid), "lines",
89  params.fetch2InputBufferSize));
90  }
91 }
92 
93 const ForwardLineData *
95 {
96  /* Get a line from the inputBuffer to work with */
97  if (!inputBuffer[tid].empty()) {
98  return &(inputBuffer[tid].front());
99  } else {
100  return NULL;
101  }
102 }
103 
104 void
106 {
107  if (!inputBuffer[tid].empty()) {
108  inputBuffer[tid].front().freeLine();
109  inputBuffer[tid].pop();
110  }
111 
112  fetchInfo[tid].inputIndex = 0;
113 }
114 
115 void
117 {
118  DPRINTF(Fetch, "Dumping whole input buffer\n");
119  while (!inputBuffer[tid].empty())
120  popInput(tid);
121 
122  fetchInfo[tid].inputIndex = 0;
123 }
124 
125 void
127 {
128  MinorDynInstPtr inst = branch.inst;
129 
130  /* Don't even consider instructions we didn't try to predict or faults */
131  if (inst->isFault() || !inst->triedToPredict)
132  return;
133 
134  switch (branch.reason) {
136  /* No data to update */
137  break;
139  /* Never try to predict interrupts */
140  break;
142  /* Don't need to act on suspends */
143  break;
145  /* Don't need to act on fetch wakeup */
146  break;
148  /* Shouldn't happen. Fetch2 is the only source of
149  * BranchPredictions */
150  break;
152  /* Unpredicted branch or barrier */
153  DPRINTF(Branch, "Unpredicted branch seen inst: %s\n", *inst);
154  branchPredictor.squash(inst->id.fetchSeqNum,
155  branch.target, true, inst->id.threadId);
156  // Update after squashing to accomodate O3CPU
157  // using the branch prediction code.
158  branchPredictor.update(inst->id.fetchSeqNum,
159  inst->id.threadId);
160  break;
162  /* Predicted taken, was taken */
163  DPRINTF(Branch, "Branch predicted correctly inst: %s\n", *inst);
164  branchPredictor.update(inst->id.fetchSeqNum,
165  inst->id.threadId);
166  break;
168  /* Predicted taken, not taken */
169  DPRINTF(Branch, "Branch mis-predicted inst: %s\n", *inst);
170  branchPredictor.squash(inst->id.fetchSeqNum,
171  branch.target /* Not used */, false, inst->id.threadId);
172  // Update after squashing to accomodate O3CPU
173  // using the branch prediction code.
174  branchPredictor.update(inst->id.fetchSeqNum,
175  inst->id.threadId);
176  break;
178  /* Predicted taken, was taken but to a different target */
179  DPRINTF(Branch, "Branch mis-predicted target inst: %s target: %s\n",
180  *inst, branch.target);
181  branchPredictor.squash(inst->id.fetchSeqNum,
182  branch.target, true, inst->id.threadId);
183  break;
184  }
185 }
186 
187 void
189 {
190  Fetch2ThreadInfo &thread = fetchInfo[inst->id.threadId];
191  TheISA::PCState inst_pc = inst->pc;
192 
193  assert(!inst->predictedTaken);
194 
195  /* Skip non-control/sys call instructions */
196  if (inst->staticInst->isControl() ||
197  inst->staticInst->isSyscall())
198  {
199  /* Tried to predict */
200  inst->triedToPredict = true;
201 
202  DPRINTF(Branch, "Trying to predict for inst: %s\n", *inst);
203 
204  if (branchPredictor.predict(inst->staticInst,
205  inst->id.fetchSeqNum, inst_pc,
206  inst->id.threadId))
207  {
208  inst->predictedTaken = true;
209  inst->predictedTarget = inst_pc;
210  branch.target = inst_pc;
211  }
212  } else {
213  DPRINTF(Branch, "Not attempting prediction for inst: %s\n", *inst);
214  }
215 
216  /* If we predict taken, set branch and update sequence numbers */
217  if (inst->predictedTaken) {
218  /* Update the predictionSeqNum and remember the streamSeqNum that it
219  * was associated with */
220  thread.expectedStreamSeqNum = inst->id.streamSeqNum;
221 
223  inst->id.threadId,
224  inst->id.streamSeqNum, thread.predictionSeqNum + 1,
225  inst->predictedTarget, inst);
226 
227  /* Mark with a new prediction number by the stream number of the
228  * instruction causing the prediction */
229  thread.predictionSeqNum++;
230  branch = new_branch;
231 
232  DPRINTF(Branch, "Branch predicted taken inst: %s target: %s"
233  " new predictionSeqNum: %d\n",
234  *inst, inst->predictedTarget, thread.predictionSeqNum);
235  }
236 }
237 
238 void
240 {
241  /* Push input onto appropriate input buffer */
242  if (!inp.outputWire->isBubble())
243  inputBuffer[inp.outputWire->id.threadId].setTail(*inp.outputWire);
244 
245  ForwardInstData &insts_out = *out.inputWire;
246  BranchData prediction;
247  BranchData &branch_inp = *branchInp.outputWire;
248 
249  assert(insts_out.isBubble());
250 
251  /* React to branches from Execute to update local branch prediction
252  * structures */
253  updateBranchPrediction(branch_inp);
254 
255  /* If a branch arrives, don't try and do anything about it. Only
256  * react to your own predictions */
257  if (branch_inp.isStreamChange()) {
258  DPRINTF(Fetch, "Dumping all input as a stream changing branch"
259  " has arrived\n");
260  dumpAllInput(branch_inp.threadId);
261  fetchInfo[branch_inp.threadId].havePC = false;
262  }
263 
264  assert(insts_out.isBubble());
265  /* Even when blocked, clear out input lines with the wrong
266  * prediction sequence number */
267  for (ThreadID tid = 0; tid < cpu.numThreads; tid++) {
268  Fetch2ThreadInfo &thread = fetchInfo[tid];
269 
270  thread.blocked = !nextStageReserve[tid].canReserve();
271 
272  const ForwardLineData *line_in = getInput(tid);
273 
274  while (line_in &&
275  thread.expectedStreamSeqNum == line_in->id.streamSeqNum &&
276  thread.predictionSeqNum != line_in->id.predictionSeqNum)
277  {
278  DPRINTF(Fetch, "Discarding line %s"
279  " due to predictionSeqNum mismatch (expected: %d)\n",
280  line_in->id, thread.predictionSeqNum);
281 
282  popInput(tid);
283  fetchInfo[tid].havePC = false;
284 
286  DPRINTF(Fetch, "Wrapping\n");
287  line_in = getInput(tid);
288  } else {
289  line_in = NULL;
290  }
291  }
292  }
293 
295  DPRINTF(Fetch, "Scheduled Thread: %d\n", tid);
296 
297  assert(insts_out.isBubble());
298  if (tid != InvalidThreadID) {
299  Fetch2ThreadInfo &fetch_info = fetchInfo[tid];
300 
301  const ForwardLineData *line_in = getInput(tid);
302 
303  unsigned int output_index = 0;
304 
305  /* Pack instructions into the output while we can. This may involve
306  * using more than one input line. Note that lineWidth will be 0
307  * for faulting lines */
308  while (line_in &&
309  (line_in->isFault() ||
310  fetch_info.inputIndex < line_in->lineWidth) && /* More input */
311  output_index < outputWidth && /* More output to fill */
312  prediction.isBubble() /* No predicted branch */)
313  {
314  ThreadContext *thread = cpu.getContext(line_in->id.threadId);
315  TheISA::Decoder *decoder = thread->getDecoderPtr();
316 
317  /* Discard line due to prediction sequence number being wrong but
318  * without the streamSeqNum number having changed */
319  bool discard_line =
320  fetch_info.expectedStreamSeqNum == line_in->id.streamSeqNum &&
321  fetch_info.predictionSeqNum != line_in->id.predictionSeqNum;
322 
323  /* Set the PC if the stream changes. Setting havePC to false in
324  * a previous cycle handles all other change of flow of control
325  * issues */
326  bool set_pc = fetch_info.lastStreamSeqNum != line_in->id.streamSeqNum;
327 
328  if (!discard_line && (!fetch_info.havePC || set_pc)) {
329  /* Set the inputIndex to be the MachInst-aligned offset
330  * from lineBaseAddr of the new PC value */
331  fetch_info.inputIndex =
332  (line_in->pc.instAddr() & BaseCPU::PCMask) -
333  line_in->lineBaseAddr;
334  DPRINTF(Fetch, "Setting new PC value: %s inputIndex: 0x%x"
335  " lineBaseAddr: 0x%x lineWidth: 0x%x\n",
336  line_in->pc, fetch_info.inputIndex, line_in->lineBaseAddr,
337  line_in->lineWidth);
338  fetch_info.pc = line_in->pc;
339  fetch_info.havePC = true;
340  decoder->reset();
341  }
342 
343  /* The generated instruction. Leave as NULL if no instruction
344  * is to be packed into the output */
345  MinorDynInstPtr dyn_inst = NULL;
346 
347  if (discard_line) {
348  /* Rest of line was from an older prediction in the same
349  * stream */
350  DPRINTF(Fetch, "Discarding line %s (from inputIndex: %d)"
351  " due to predictionSeqNum mismatch (expected: %d)\n",
352  line_in->id, fetch_info.inputIndex,
353  fetch_info.predictionSeqNum);
354  } else if (line_in->isFault()) {
355  /* Pack a fault as a MinorDynInst with ->fault set */
356 
357  /* Make a new instruction and pick up the line, stream,
358  * prediction, thread ids from the incoming line */
359  dyn_inst = new MinorDynInst(
361 
362  /* Fetch and prediction sequence numbers originate here */
363  dyn_inst->id.fetchSeqNum = fetch_info.fetchSeqNum;
364  dyn_inst->id.predictionSeqNum = fetch_info.predictionSeqNum;
365  /* To complete the set, test that exec sequence number has
366  * not been set */
367  assert(dyn_inst->id.execSeqNum == 0);
368 
369  dyn_inst->pc = fetch_info.pc;
370 
371  /* Pack a faulting instruction but allow other
372  * instructions to be generated. (Fetch2 makes no
373  * immediate judgement about streamSeqNum) */
374  dyn_inst->fault = line_in->fault;
375  DPRINTF(Fetch, "Fault being passed output_index: "
376  "%d: %s\n", output_index, dyn_inst->fault->name());
377  } else {
378  uint8_t *line = line_in->line;
379 
380  /* The instruction is wholly in the line, can just
381  * assign */
382  auto inst_word = *reinterpret_cast<TheISA::MachInst *>
383  (line + fetch_info.inputIndex);
384 
385  if (!decoder->instReady()) {
386  decoder->moreBytes(fetch_info.pc,
387  line_in->lineBaseAddr + fetch_info.inputIndex,
388  inst_word);
389  DPRINTF(Fetch, "Offering MachInst to decoder addr: 0x%x\n",
390  line_in->lineBaseAddr + fetch_info.inputIndex);
391  }
392 
393  /* Maybe make the above a loop to accomodate ISAs with
394  * instructions longer than sizeof(MachInst) */
395 
396  if (decoder->instReady()) {
397  /* Note that the decoder can update the given PC.
398  * Remember not to assign it until *after* calling
399  * decode */
400  StaticInstPtr decoded_inst =
401  decoder->decode(fetch_info.pc);
402 
403  /* Make a new instruction and pick up the line, stream,
404  * prediction, thread ids from the incoming line */
405  dyn_inst = new MinorDynInst(decoded_inst, line_in->id);
406 
407  /* Fetch and prediction sequence numbers originate here */
408  dyn_inst->id.fetchSeqNum = fetch_info.fetchSeqNum;
409  dyn_inst->id.predictionSeqNum = fetch_info.predictionSeqNum;
410  /* To complete the set, test that exec sequence number
411  * has not been set */
412  assert(dyn_inst->id.execSeqNum == 0);
413 
414  dyn_inst->pc = fetch_info.pc;
415  DPRINTF(Fetch, "decoder inst %s\n", *dyn_inst);
416 
417  // Collect some basic inst class stats
418  if (decoded_inst->isLoad())
420  else if (decoded_inst->isStore())
422  else if (decoded_inst->isAtomic())
424  else if (decoded_inst->isVector())
426  else if (decoded_inst->isFloating())
428  else if (decoded_inst->isInteger())
430 
431  DPRINTF(Fetch, "Instruction extracted from line %s"
432  " lineWidth: %d output_index: %d inputIndex: %d"
433  " pc: %s inst: %s\n",
434  line_in->id,
435  line_in->lineWidth, output_index, fetch_info.inputIndex,
436  fetch_info.pc, *dyn_inst);
437 
438 #if THE_ISA == X86_ISA || THE_ISA == ARM_ISA
439  /* In SE mode, it's possible to branch to a microop when
440  * replaying faults such as page faults (or simply
441  * intra-microcode branches in X86). Unfortunately,
442  * as Minor has micro-op decomposition in a separate
443  * pipeline stage from instruction decomposition, the
444  * following advancePC (which may follow a branch with
445  * microPC() != 0) *must* see a fresh macroop. This
446  * kludge should be improved with an addition to PCState
447  * but I offer it in this form for the moment
448  *
449  * X86 can branch within microops so we need to deal with
450  * the case that, after a branch, the first un-advanced PC
451  * may be pointing to a microop other than 0. Once
452  * advanced, however, the microop number *must* be 0 */
453  fetch_info.pc.upc(0);
454  fetch_info.pc.nupc(1);
455 #endif
456 
457  /* Advance PC for the next instruction */
458  TheISA::advancePC(fetch_info.pc, decoded_inst);
459 
460  /* Predict any branches and issue a branch if
461  * necessary */
462  predictBranch(dyn_inst, prediction);
463  } else {
464  DPRINTF(Fetch, "Inst not ready yet\n");
465  }
466 
467  /* Step on the pointer into the line if there's no
468  * complete instruction waiting */
469  if (decoder->needMoreBytes()) {
470  fetch_info.inputIndex += sizeof(TheISA::MachInst);
471 
472  DPRINTF(Fetch, "Updated inputIndex value PC: %s"
473  " inputIndex: 0x%x lineBaseAddr: 0x%x lineWidth: 0x%x\n",
474  line_in->pc, fetch_info.inputIndex, line_in->lineBaseAddr,
475  line_in->lineWidth);
476  }
477  }
478 
479  if (dyn_inst) {
480  /* Step to next sequence number */
481  fetch_info.fetchSeqNum++;
482 
483  /* Correctly size the output before writing */
484  if (output_index == 0) {
485  insts_out.resize(outputWidth);
486  }
487  /* Pack the generated dynamic instruction into the output */
488  insts_out.insts[output_index] = dyn_inst;
489  output_index++;
490 
491  /* Output MinorTrace instruction info for
492  * pre-microop decomposition macroops */
493  if (DTRACE(MinorTrace) && !dyn_inst->isFault() &&
494  dyn_inst->staticInst->isMacroop())
495  {
496  dyn_inst->minorTraceInst(*this);
497  }
498  }
499 
500  /* Remember the streamSeqNum of this line so we can tell when
501  * we change stream */
502  fetch_info.lastStreamSeqNum = line_in->id.streamSeqNum;
503 
504  /* Asked to discard line or there was a branch or fault */
505  if (!prediction.isBubble() || /* The remains of a
506  line with a prediction in it */
507  line_in->isFault() /* A line which is just a fault */)
508  {
509  DPRINTF(Fetch, "Discarding all input on branch/fault\n");
510  dumpAllInput(tid);
511  fetch_info.havePC = false;
512  line_in = NULL;
513  } else if (discard_line) {
514  /* Just discard one line, one's behind it may have new
515  * stream sequence numbers. There's a DPRINTF above
516  * for this event */
517  popInput(tid);
518  fetch_info.havePC = false;
519  line_in = NULL;
520  } else if (fetch_info.inputIndex == line_in->lineWidth) {
521  /* Got to end of a line, pop the line but keep PC
522  * in case this is a line-wrapping inst. */
523  popInput(tid);
524  line_in = NULL;
525  }
526 
527  if (!line_in && processMoreThanOneInput) {
528  DPRINTF(Fetch, "Wrapping\n");
529  line_in = getInput(tid);
530  }
531  }
532 
533  /* The rest of the output (if any) should already have been packed
534  * with bubble instructions by insts_out's initialisation */
535  }
536  if (tid == InvalidThreadID) {
537  assert(insts_out.isBubble());
538  }
540  *predictionOut.inputWire = prediction;
541 
542  /* If we generated output, reserve space for the result in the next stage
543  * and mark the stage as being active this cycle */
544  if (!insts_out.isBubble()) {
545  /* Note activity of following buffer */
547  insts_out.threadId = tid;
548  nextStageReserve[tid].reserve();
549  }
550 
551  /* If we still have input to process and somewhere to put it,
552  * mark stage as active */
553  for (ThreadID i = 0; i < cpu.numThreads; i++)
554  {
555  if (getInput(i) && nextStageReserve[i].canReserve()) {
557  break;
558  }
559  }
560 
561  /* Make sure the input (if any left) is pushed */
562  if (!inp.outputWire->isBubble())
563  inputBuffer[inp.outputWire->id.threadId].pushTail();
564 }
565 
566 inline ThreadID
568 {
569  /* Select thread via policy. */
570  std::vector<ThreadID> priority_list;
571 
572  switch (cpu.threadPolicy) {
573  case Enums::SingleThreaded:
574  priority_list.push_back(0);
575  break;
576  case Enums::RoundRobin:
577  priority_list = cpu.roundRobinPriority(threadPriority);
578  break;
579  case Enums::Random:
580  priority_list = cpu.randomPriority();
581  break;
582  default:
583  panic("Unknown fetch policy");
584  }
585 
586  for (auto tid : priority_list) {
587  if (getInput(tid) && !fetchInfo[tid].blocked) {
588  threadPriority = tid;
589  return tid;
590  }
591  }
592 
593  return InvalidThreadID;
594 }
595 
596 bool
598 {
599  for (const auto &buffer : inputBuffer) {
600  if (!buffer.empty())
601  return false;
602  }
603 
604  return (*inp.outputWire).isBubble() &&
605  (*predictionOut.inputWire).isBubble();
606 }
607 
609  : Stats::Group(cpu, "fetch2"),
610  ADD_STAT(intInstructions, UNIT_COUNT,
611  "Number of integer instructions successfully decoded"),
612  ADD_STAT(fpInstructions, UNIT_COUNT,
613  "Number of floating point instructions successfully decoded"),
614  ADD_STAT(vecInstructions, UNIT_COUNT,
615  "Number of SIMD instructions successfully decoded"),
616  ADD_STAT(loadInstructions, UNIT_COUNT,
617  "Number of memory load instructions successfully decoded"),
618  ADD_STAT(storeInstructions, UNIT_COUNT,
619  "Number of memory store instructions successfully decoded"),
620  ADD_STAT(amoInstructions, UNIT_COUNT,
621  "Number of memory atomic instructions successfully decoded")
622 {
635 }
636 
637 void
639 {
640  std::ostringstream data;
641 
642  if (fetchInfo[0].blocked)
643  data << 'B';
644  else
645  (*out.inputWire).reportData(data);
646 
647  MINORTRACE("inputIndex=%d havePC=%d predictionSeqNum=%d insts=%s\n",
648  fetchInfo[0].inputIndex, fetchInfo[0].havePC, fetchInfo[0].predictionSeqNum, data.str());
649  inputBuffer[0].minorTrace();
650 }
651 
652 }
InvalidThreadID
const ThreadID InvalidThreadID
Definition: types.hh:234
fatal
#define fatal(...)
This implements a cprintf based fatal() function.
Definition: logging.hh:183
Minor::BranchData::threadId
ThreadID threadId
ThreadID associated with branch.
Definition: pipe_data.hh:112
Minor::Fetch2::threadPriority
ThreadID threadPriority
Definition: fetch2.hh:166
Minor::BranchData::target
TheISA::PCState target
Starting PC of that stream.
Definition: pipe_data.hh:119
Minor::ForwardInstData::threadId
ThreadID threadId
Thread associated with these instructions.
Definition: pipe_data.hh:263
Minor::Fetch2::Fetch2ThreadInfo::predictionSeqNum
InstSeqNum predictionSeqNum
Fetch2 is the source of prediction sequence numbers.
Definition: fetch2.hh:159
Minor::Fetch2::fetchInfo
std::vector< Fetch2ThreadInfo > fetchInfo
Definition: fetch2.hh:165
Minor::BranchData::inst
MinorDynInstPtr inst
Instruction which caused this branch.
Definition: pipe_data.hh:122
Minor::ForwardInstData
Forward flowing data between Fetch2,Decode,Execute carrying a packet of instructions of a width appro...
Definition: pipe_data.hh:253
Minor::BranchData::BranchPrediction
@ BranchPrediction
Definition: pipe_data.hh:80
Minor::Fetch2::cpu
MinorCPU & cpu
Pointer back to the containing CPU.
Definition: fetch2.hh:65
data
const char data[]
Definition: circlebuf.test.cc:47
ArmISA::i
Bitfield< 7 > i
Definition: miscregs_types.hh:63
ThreadID
int16_t ThreadID
Thread index/ID type.
Definition: types.hh:233
ArmISA::MachInst
uint32_t MachInst
Definition: types.hh:52
Minor::Fetch2::getScheduledThread
ThreadID getScheduledThread()
Use the current threading policy to determine the next thread to fetch from.
Definition: fetch2.cc:567
Minor::Latch::Input
Encapsulate wires on either input or output of the latch.
Definition: buffers.hh:247
MinorCPU::activityRecorder
Minor::MinorActivityRecorder * activityRecorder
Activity recording for pipeline.
Definition: cpu.hh:88
Minor::Fetch2::Fetch2ThreadInfo::lastStreamSeqNum
InstSeqNum lastStreamSeqNum
Stream sequence number of the last seen line used to identify changes of instruction stream.
Definition: fetch2.hh:144
Minor::Fetch2::Fetch2ThreadInfo::fetchSeqNum
InstSeqNum fetchSeqNum
Fetch2 is the source of fetch sequence numbers.
Definition: fetch2.hh:148
Minor::Fetch2::branchPredictor
BPredUnit & branchPredictor
Branch predictor passed from Python configuration.
Definition: fetch2.hh:91
Minor::ForwardLineData
Line fetch data in the forward direction.
Definition: pipe_data.hh:171
Minor::Pipeline::Fetch2StageId
@ Fetch2StageId
Definition: pipeline.hh:100
sc_dt::to_string
const std::string to_string(sc_enc enc)
Definition: sc_fxdefs.cc:91
Minor::BranchData::HaltFetch
@ HaltFetch
Definition: pipe_data.hh:95
Minor::Fetch2::out
Latch< ForwardInstData >::Input out
Output port carrying instructions into Decode.
Definition: fetch2.hh:78
MinorCPU::threadPolicy
Enums::ThreadPolicy threadPolicy
Thread Scheduling Policy (RoundRobin, Random, etc)
Definition: cpu.hh:112
Minor::BranchData::reason
Reason reason
Explanation for this branch.
Definition: pipe_data.hh:109
MINORTRACE
#define MINORTRACE(...)
DPRINTFN for MinorTrace reporting.
Definition: trace.hh:60
DTRACE
#define DTRACE(x)
Definition: debug.hh:156
Minor::Fetch2::Fetch2ThreadInfo
Data members after this line are cycle-to-cycle state.
Definition: fetch2.hh:100
std::vector
STL vector class.
Definition: stl.hh:37
Minor::ForwardLineData::id
InstId id
Thread, stream, prediction ...
Definition: pipe_data.hh:195
Minor::BranchData::UnpredictedBranch
@ UnpredictedBranch
Definition: pipe_data.hh:78
BaseCPU::getContext
virtual ThreadContext * getContext(int tn)
Given a thread num get tho thread context for it.
Definition: base.hh:300
Minor::BranchData::isStreamChange
static bool isStreamChange(const BranchData::Reason reason)
Is a request with this reason actually a request to change the PC rather than a bubble or branch pred...
Definition: pipe_data.cc:81
Minor::Fetch2::predictBranch
void predictBranch(MinorDynInstPtr inst, BranchData &branch)
Predicts branches for the given instruction.
Definition: fetch2.cc:188
Minor::BranchData::BadlyPredictedBranchTarget
@ BadlyPredictedBranchTarget
Definition: pipe_data.hh:82
Minor::Fetch2::Fetch2Stats::loadInstructions
Stats::Scalar loadInstructions
Definition: fetch2.hh:175
BPredUnit::update
void update(const InstSeqNum &done_sn, ThreadID tid)
Tells the branch predictor to commit any updates until the given sequence number.
Definition: bpred_unit.cc:295
Minor
Definition: activity.cc:44
ArmISA::advancePC
void advancePC(PCState &pc, const StaticInstPtr &inst)
Definition: utility.hh:392
Minor::BranchData::CorrectlyPredictedBranch
@ CorrectlyPredictedBranch
Definition: pipe_data.hh:73
Minor::ForwardLineData::lineWidth
unsigned int lineWidth
Explicit line width, don't rely on data.size.
Definition: pipe_data.hh:187
Stats::DataWrap::flags
Derived & flags(Flags _flags)
Set the flags and marks this stat to print at the end of simulation.
Definition: statistics.hh:339
StaticInst::isInteger
bool isInteger() const
Definition: static_inst.hh:175
Minor::Fetch2::evaluate
void evaluate()
Pass on input/buffer data to the output if you can.
Definition: fetch2.cc:239
fetch2.hh
Minor::Fetch2::inp
Latch< ForwardLineData >::Output inp
Input port carrying lines from Fetch1.
Definition: fetch2.hh:68
Minor::BranchData::isBubble
bool isBubble() const
Definition: pipe_data.hh:148
Minor::Fetch2::branchInp
Latch< BranchData >::Output branchInp
Input port carrying branches from Execute.
Definition: fetch2.hh:72
StaticInst::isStore
bool isStore() const
Definition: static_inst.hh:167
Minor::Fetch2::popInput
void popInput(ThreadID tid)
Pop an element off the input buffer, if there are any.
Definition: fetch2.cc:105
StaticInst::isLoad
bool isLoad() const
Definition: static_inst.hh:166
Minor::ForwardLineData::fault
Fault fault
This line has a fault.
Definition: pipe_data.hh:192
Minor::BranchData::NoBranch
@ NoBranch
Definition: pipe_data.hh:70
Minor::BranchData::Interrupt
@ Interrupt
Definition: pipe_data.hh:93
Minor::Fetch2::Fetch2ThreadInfo::havePC
bool havePC
PC is currently valid.
Definition: fetch2.hh:140
Minor::Fetch2::getInput
const ForwardLineData * getInput(ThreadID tid)
Get a piece of data to work on from the inputBuffer, or 0 if there is no data.
Definition: fetch2.cc:94
ThreadContext
ThreadContext is the external interface to all thread state for anything outside of the CPU.
Definition: thread_context.hh:88
Minor::InstId::streamSeqNum
InstSeqNum streamSeqNum
The 'stream' this instruction belongs to.
Definition: dyn_inst.hh:88
Minor::MinorDynInst
Dynamic instruction for Minor.
Definition: dyn_inst.hh:157
MinorCPU
MinorCPU is an in-order CPU model with four fixed pipeline stages:
Definition: cpu.hh:77
Minor::Latch::Output
Definition: buffers.hh:258
Minor::Fetch2::Fetch2ThreadInfo::blocked
bool blocked
Blocked indication for report.
Definition: fetch2.hh:162
Minor::Fetch2::processMoreThanOneInput
bool processMoreThanOneInput
If true, more than one input word can be processed each cycle if there is room in the output to conta...
Definition: fetch2.hh:88
Minor::ForwardLineData::lineBaseAddr
Addr lineBaseAddr
First byte address in the line.
Definition: pipe_data.hh:181
DPRINTF
#define DPRINTF(x,...)
Definition: trace.hh:237
ADD_STAT
#define ADD_STAT(n,...)
Convenience macro to add a stat to a statistics group.
Definition: group.hh:71
StaticInst::isAtomic
bool isAtomic() const
Definition: static_inst.hh:168
Minor::Fetch2::outputWidth
unsigned int outputWidth
Width of output of this stage/input of next in instructions.
Definition: fetch2.hh:84
Minor::Fetch2::Fetch2ThreadInfo::expectedStreamSeqNum
InstSeqNum expectedStreamSeqNum
Stream sequence number remembered from last time the predictionSeqNum changed.
Definition: fetch2.hh:154
Minor::Fetch2::Fetch2Stats::storeInstructions
Stats::Scalar storeInstructions
Definition: fetch2.hh:176
Minor::Fetch2::Fetch2Stats::vecInstructions
Stats::Scalar vecInstructions
Definition: fetch2.hh:174
Minor::ForwardInstData::isBubble
bool isBubble() const
BubbleIF interface.
Definition: pipe_data.cc:247
pipeline.hh
Minor::Fetch2::updateBranchPrediction
void updateBranchPrediction(const BranchData &branch)
Update local branch prediction structures from feedback from Execute.
Definition: fetch2.cc:126
Minor::ForwardInstData::insts
MinorDynInstPtr insts[MAX_FORWARD_INSTS]
Array of carried insts, ref counted.
Definition: pipe_data.hh:257
Minor::Fetch2::stats
Minor::Fetch2::Fetch2Stats stats
Minor::Fetch2::dumpAllInput
void dumpAllInput(ThreadID tid)
Dump the whole contents of the input buffer.
Definition: fetch2.cc:116
Minor::InputBuffer
Like a Queue but with a restricted interface and a setTail function which, when the queue is empty,...
Definition: buffers.hh:567
Minor::InstId::threadId
ThreadID threadId
The thread to which this line/instruction belongs.
Definition: dyn_inst.hh:83
UNIT_COUNT
#define UNIT_COUNT
Definition: units.hh:49
StaticInst::nullStaticInstPtr
static StaticInstPtr nullStaticInstPtr
Pointer to a statically allocated "null" instruction object.
Definition: static_inst.hh:253
name
const std::string & name()
Definition: trace.cc:48
Minor::ForwardLineData::isFault
bool isFault() const
This is a fault, not a line.
Definition: pipe_data.hh:220
Minor::Fetch2::Fetch2Stats::Fetch2Stats
Fetch2Stats(MinorCPU *cpu)
Definition: fetch2.cc:608
Minor::Fetch2::predictionOut
Latch< BranchData >::Input predictionOut
Output port carrying predictions back to Fetch1.
Definition: fetch2.hh:75
Named
Definition: trace.hh:150
Minor::BranchData
Forward data betwen Execute and Fetch1 carrying change-of-address/stream information.
Definition: pipe_data.hh:62
Minor::BranchData::BadlyPredictedBranch
@ BadlyPredictedBranch
Definition: pipe_data.hh:86
Minor::Fetch2::Fetch2Stats::amoInstructions
Stats::Scalar amoInstructions
Definition: fetch2.hh:177
MipsISA::PCState
GenericISA::DelaySlotPCState< MachInst > PCState
Definition: types.hh:41
bpred_unit.hh
ActivityRecorder::activity
void activity()
Records that there is activity this cycle.
Definition: activity.cc:52
MinorCPU::roundRobinPriority
std::vector< ThreadID > roundRobinPriority(ThreadID priority)
Thread scheduling utility functions.
Definition: cpu.hh:165
Minor::Fetch2::inputBuffer
std::vector< InputBuffer< ForwardLineData > > inputBuffer
Definition: fetch2.hh:95
Minor::ForwardLineData::pc
TheISA::PCState pc
PC of the first requested inst within this line.
Definition: pipe_data.hh:184
Minor::Fetch2::Fetch2ThreadInfo::pc
TheISA::PCState pc
Remembered program counter value.
Definition: fetch2.hh:135
BaseCPU::PCMask
static const Addr PCMask
Definition: base.hh:281
Minor::Fetch2::Fetch2ThreadInfo::inputIndex
unsigned int inputIndex
Index into an incompletely processed input line that instructions are to be extracted from.
Definition: fetch2.hh:126
BPredUnit::predict
bool predict(const StaticInstPtr &inst, const InstSeqNum &seqNum, TheISA::PCState &pc, ThreadID tid)
Predicts whether or not the instruction is a taken branch, and the target of the branch if it is take...
Definition: bpred_unit.cc:120
BaseCPU::numThreads
ThreadID numThreads
Number of threads we're actually simulating (<= SMT_MAX_THREADS).
Definition: base.hh:378
logging.hh
Minor::Fetch2::isDrained
bool isDrained()
Is this stage drained? For Fetch2, draining is initiated by Execute halting Fetch1 causing Fetch2 to ...
Definition: fetch2.cc:597
Minor::Fetch2::Fetch2Stats::intInstructions
Stats::Scalar intInstructions
Stats.
Definition: fetch2.hh:172
ThreadContext::getDecoderPtr
virtual TheISA::Decoder * getDecoderPtr()=0
Stats
Definition: statistics.cc:53
ActivityRecorder::activateStage
void activateStage(const int idx)
Marks a stage as active.
Definition: activity.cc:88
StaticInst::isFloating
bool isFloating() const
Definition: static_inst.hh:176
MinorCPU::randomPriority
std::vector< ThreadID > randomPriority()
Definition: cpu.hh:174
RefCountingPtr< MinorDynInst >
trace.hh
Minor::ForwardInstData::resize
void resize(unsigned int width)
Resize a bubble/empty ForwardInstData and fill with bubbles.
Definition: pipe_data.cc:260
decoder
output decoder
Definition: nop.cc:61
Named::name
const std::string & name() const
Definition: trace.hh:159
Minor::InstId::predictionSeqNum
InstSeqNum predictionSeqNum
The predicted qualifier to stream, attached by Fetch2 as a consequence of branch prediction.
Definition: dyn_inst.hh:92
Stats::total
const FlagsType total
Print the total.
Definition: info.hh:50
Minor::ForwardLineData::line
uint8_t * line
Line data.
Definition: pipe_data.hh:199
BPredUnit::squash
void squash(const InstSeqNum &squashed_sn, ThreadID tid)
Squashes all outstanding updates until a given sequence number.
Definition: bpred_unit.cc:318
Minor::Fetch2::Fetch2
Fetch2(const std::string &name, MinorCPU &cpu_, const MinorCPUParams &params, Latch< ForwardLineData >::Output inp_, Latch< BranchData >::Output branchInp_, Latch< BranchData >::Input predictionOut_, Latch< ForwardInstData >::Input out_, std::vector< InputBuffer< ForwardInstData >> &next_stage_input_buffer)
Definition: fetch2.cc:55
StaticInst::isVector
bool isVector() const
Definition: static_inst.hh:177
Minor::Fetch2::minorTrace
void minorTrace() const
Definition: fetch2.cc:638
Minor::Fetch2::Fetch2Stats::fpInstructions
Stats::Scalar fpInstructions
Definition: fetch2.hh:173
Minor::BranchData::SuspendThread
@ SuspendThread
Definition: pipe_data.hh:91
Minor::Fetch2::nextStageReserve
std::vector< InputBuffer< ForwardInstData > > & nextStageReserve
Interface to reserve space in the next stage.
Definition: fetch2.hh:81
panic
#define panic(...)
This implements a cprintf based panic() function.
Definition: logging.hh:171

Generated on Tue Mar 23 2021 19:41:24 for gem5 by doxygen 1.8.17