gem5  v22.1.0.0
fetch2.cc
Go to the documentation of this file.
1 /*
2  * Copyright (c) 2013-2014,2016 ARM Limited
3  * All rights reserved
4  *
5  * The license below extends only to copyright in the software and shall
6  * not be construed as granting a license to any other intellectual
7  * property including but not limited to intellectual property relating
8  * to a hardware implementation of the functionality of the software
9  * licensed hereunder. You may use the software subject to the license
10  * terms below provided that you ensure that this notice is replicated
11  * unmodified and in its entirety in all distributions of the software,
12  * modified or unmodified, in source code or in binary form.
13  *
14  * Redistribution and use in source and binary forms, with or without
15  * modification, are permitted provided that the following conditions are
16  * met: redistributions of source code must retain the above copyright
17  * notice, this list of conditions and the following disclaimer;
18  * redistributions in binary form must reproduce the above copyright
19  * notice, this list of conditions and the following disclaimer in the
20  * documentation and/or other materials provided with the distribution;
21  * neither the name of the copyright holders nor the names of its
22  * contributors may be used to endorse or promote products derived from
23  * this software without specific prior written permission.
24  *
25  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
26  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
27  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
28  * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
29  * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
30  * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
31  * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
32  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
33  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
34  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
35  * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
36  */
37 
38 #include "cpu/minor/fetch2.hh"
39 
40 #include <string>
41 
42 #include "arch/generic/decoder.hh"
43 #include "base/logging.hh"
44 #include "base/trace.hh"
45 #include "cpu/minor/pipeline.hh"
46 #include "cpu/null_static_inst.hh"
47 #include "cpu/pred/bpred_unit.hh"
48 #include "debug/Branch.hh"
49 #include "debug/Fetch.hh"
50 #include "debug/MinorTrace.hh"
51 
52 namespace gem5
53 {
54 
56 namespace minor
57 {
58 
59 Fetch2::Fetch2(const std::string &name,
60  MinorCPU &cpu_,
61  const BaseMinorCPUParams &params,
63  Latch<BranchData>::Output branchInp_,
64  Latch<BranchData>::Input predictionOut_,
66  std::vector<InputBuffer<ForwardInstData>> &next_stage_input_buffer) :
67  Named(name),
68  cpu(cpu_),
69  inp(inp_),
70  branchInp(branchInp_),
71  predictionOut(predictionOut_),
72  out(out_),
73  nextStageReserve(next_stage_input_buffer),
74  outputWidth(params.decodeInputWidth),
75  processMoreThanOneInput(params.fetch2CycleInput),
76  branchPredictor(*params.branchPred),
77  fetchInfo(params.numThreads),
78  threadPriority(0), stats(&cpu_)
79 {
80  if (outputWidth < 1)
81  fatal("%s: decodeInputWidth must be >= 1 (%d)\n", name, outputWidth);
82 
83  if (params.fetch2InputBufferSize < 1) {
84  fatal("%s: fetch2InputBufferSize must be >= 1 (%d)\n", name,
85  params.fetch2InputBufferSize);
86  }
87 
88  /* Per-thread input buffers */
89  for (ThreadID tid = 0; tid < params.numThreads; tid++) {
90  inputBuffer.push_back(
92  name + ".inputBuffer" + std::to_string(tid), "lines",
93  params.fetch2InputBufferSize));
94  }
95 }
96 
97 const ForwardLineData *
99 {
100  /* Get a line from the inputBuffer to work with */
101  if (!inputBuffer[tid].empty()) {
102  return &(inputBuffer[tid].front());
103  } else {
104  return NULL;
105  }
106 }
107 
108 void
110 {
111  if (!inputBuffer[tid].empty()) {
112  inputBuffer[tid].front().freeLine();
113  inputBuffer[tid].pop();
114  }
115 
116  fetchInfo[tid].inputIndex = 0;
117 }
118 
119 void
121 {
122  DPRINTF(Fetch, "Dumping whole input buffer\n");
123  while (!inputBuffer[tid].empty())
124  popInput(tid);
125 
126  fetchInfo[tid].inputIndex = 0;
127 }
128 
129 void
131 {
132  MinorDynInstPtr inst = branch.inst;
133 
134  /* Don't even consider instructions we didn't try to predict or faults */
135  if (inst->isFault() || !inst->triedToPredict)
136  return;
137 
138  switch (branch.reason) {
140  /* No data to update */
141  break;
143  /* Never try to predict interrupts */
144  break;
146  /* Don't need to act on suspends */
147  break;
149  /* Don't need to act on fetch wakeup */
150  break;
152  /* Shouldn't happen. Fetch2 is the only source of
153  * BranchPredictions */
154  break;
156  /* Unpredicted branch or barrier */
157  DPRINTF(Branch, "Unpredicted branch seen inst: %s\n", *inst);
158  branchPredictor.squash(inst->id.fetchSeqNum,
159  *branch.target, true, inst->id.threadId);
160  // Update after squashing to accomodate O3CPU
161  // using the branch prediction code.
162  branchPredictor.update(inst->id.fetchSeqNum,
163  inst->id.threadId);
164  break;
166  /* Predicted taken, was taken */
167  DPRINTF(Branch, "Branch predicted correctly inst: %s\n", *inst);
168  branchPredictor.update(inst->id.fetchSeqNum,
169  inst->id.threadId);
170  break;
172  /* Predicted taken, not taken */
173  DPRINTF(Branch, "Branch mis-predicted inst: %s\n", *inst);
174  branchPredictor.squash(inst->id.fetchSeqNum,
175  *branch.target /* Not used */, false, inst->id.threadId);
176  // Update after squashing to accomodate O3CPU
177  // using the branch prediction code.
178  branchPredictor.update(inst->id.fetchSeqNum,
179  inst->id.threadId);
180  break;
182  /* Predicted taken, was taken but to a different target */
183  DPRINTF(Branch, "Branch mis-predicted target inst: %s target: %s\n",
184  *inst, *branch.target);
185  branchPredictor.squash(inst->id.fetchSeqNum,
186  *branch.target, true, inst->id.threadId);
187  break;
188  }
189 }
190 
191 void
193 {
194  Fetch2ThreadInfo &thread = fetchInfo[inst->id.threadId];
195 
196  assert(!inst->predictedTaken);
197 
198  /* Skip non-control/sys call instructions */
199  if (inst->staticInst->isControl() || inst->staticInst->isSyscall()){
200  std::unique_ptr<PCStateBase> inst_pc(inst->pc->clone());
201 
202  /* Tried to predict */
203  inst->triedToPredict = true;
204 
205  DPRINTF(Branch, "Trying to predict for inst: %s\n", *inst);
206 
207  if (branchPredictor.predict(inst->staticInst,
208  inst->id.fetchSeqNum, *inst_pc, inst->id.threadId)) {
209  set(branch.target, *inst_pc);
210  inst->predictedTaken = true;
211  set(inst->predictedTarget, inst_pc);
212  }
213  } else {
214  DPRINTF(Branch, "Not attempting prediction for inst: %s\n", *inst);
215  }
216 
217  /* If we predict taken, set branch and update sequence numbers */
218  if (inst->predictedTaken) {
219  /* Update the predictionSeqNum and remember the streamSeqNum that it
220  * was associated with */
221  thread.expectedStreamSeqNum = inst->id.streamSeqNum;
222 
224  inst->id.threadId,
225  inst->id.streamSeqNum, thread.predictionSeqNum + 1,
226  *inst->predictedTarget, inst);
227 
228  /* Mark with a new prediction number by the stream number of the
229  * instruction causing the prediction */
230  thread.predictionSeqNum++;
231  branch = new_branch;
232 
233  DPRINTF(Branch, "Branch predicted taken inst: %s target: %s"
234  " new predictionSeqNum: %d\n",
235  *inst, *inst->predictedTarget, thread.predictionSeqNum);
236  }
237 }
238 
239 void
241 {
242  /* Push input onto appropriate input buffer */
243  if (!inp.outputWire->isBubble())
244  inputBuffer[inp.outputWire->id.threadId].setTail(*inp.outputWire);
245 
246  ForwardInstData &insts_out = *out.inputWire;
247  BranchData prediction;
248  BranchData &branch_inp = *branchInp.outputWire;
249 
250  assert(insts_out.isBubble());
251 
252  /* React to branches from Execute to update local branch prediction
253  * structures */
254  updateBranchPrediction(branch_inp);
255 
256  /* If a branch arrives, don't try and do anything about it. Only
257  * react to your own predictions */
258  if (branch_inp.isStreamChange()) {
259  DPRINTF(Fetch, "Dumping all input as a stream changing branch"
260  " has arrived\n");
261  dumpAllInput(branch_inp.threadId);
262  fetchInfo[branch_inp.threadId].havePC = false;
263  }
264 
265  assert(insts_out.isBubble());
266  /* Even when blocked, clear out input lines with the wrong
267  * prediction sequence number */
268  for (ThreadID tid = 0; tid < cpu.numThreads; tid++) {
269  Fetch2ThreadInfo &thread = fetchInfo[tid];
270 
271  thread.blocked = !nextStageReserve[tid].canReserve();
272 
273  const ForwardLineData *line_in = getInput(tid);
274 
275  while (line_in &&
276  thread.expectedStreamSeqNum == line_in->id.streamSeqNum &&
277  thread.predictionSeqNum != line_in->id.predictionSeqNum)
278  {
279  DPRINTF(Fetch, "Discarding line %s"
280  " due to predictionSeqNum mismatch (expected: %d)\n",
281  line_in->id, thread.predictionSeqNum);
282 
283  popInput(tid);
284  fetchInfo[tid].havePC = false;
285 
287  DPRINTF(Fetch, "Wrapping\n");
288  line_in = getInput(tid);
289  } else {
290  line_in = NULL;
291  }
292  }
293  }
294 
296  DPRINTF(Fetch, "Scheduled Thread: %d\n", tid);
297 
298  assert(insts_out.isBubble());
299  if (tid != InvalidThreadID) {
300  Fetch2ThreadInfo &fetch_info = fetchInfo[tid];
301 
302  const ForwardLineData *line_in = getInput(tid);
303 
304  unsigned int output_index = 0;
305 
306  /* Pack instructions into the output while we can. This may involve
307  * using more than one input line. Note that lineWidth will be 0
308  * for faulting lines */
309  while (line_in &&
310  (line_in->isFault() ||
311  fetch_info.inputIndex < line_in->lineWidth) && /* More input */
312  output_index < outputWidth && /* More output to fill */
313  prediction.isBubble() /* No predicted branch */)
314  {
315  ThreadContext *thread = cpu.getContext(line_in->id.threadId);
316  InstDecoder *decoder = thread->getDecoderPtr();
317 
318  /* Discard line due to prediction sequence number being wrong but
319  * without the streamSeqNum number having changed */
320  bool discard_line =
321  fetch_info.expectedStreamSeqNum == line_in->id.streamSeqNum &&
322  fetch_info.predictionSeqNum != line_in->id.predictionSeqNum;
323 
324  /* Set the PC if the stream changes. Setting havePC to false in
325  * a previous cycle handles all other change of flow of control
326  * issues */
327  bool set_pc = fetch_info.lastStreamSeqNum != line_in->id.streamSeqNum;
328 
329  if (!discard_line && (!fetch_info.havePC || set_pc)) {
330  /* Set the inputIndex to be the MachInst-aligned offset
331  * from lineBaseAddr of the new PC value */
332  fetch_info.inputIndex =
333  (line_in->pc->instAddr() & decoder->pcMask()) -
334  line_in->lineBaseAddr;
335  DPRINTF(Fetch, "Setting new PC value: %s inputIndex: 0x%x"
336  " lineBaseAddr: 0x%x lineWidth: 0x%x\n",
337  *line_in->pc, fetch_info.inputIndex, line_in->lineBaseAddr,
338  line_in->lineWidth);
339  set(fetch_info.pc, line_in->pc);
340  fetch_info.havePC = true;
341  decoder->reset();
342  }
343 
344  /* The generated instruction. Leave as NULL if no instruction
345  * is to be packed into the output */
346  MinorDynInstPtr dyn_inst = NULL;
347 
348  if (discard_line) {
349  /* Rest of line was from an older prediction in the same
350  * stream */
351  DPRINTF(Fetch, "Discarding line %s (from inputIndex: %d)"
352  " due to predictionSeqNum mismatch (expected: %d)\n",
353  line_in->id, fetch_info.inputIndex,
354  fetch_info.predictionSeqNum);
355  } else if (line_in->isFault()) {
356  /* Pack a fault as a MinorDynInst with ->fault set */
357 
358  /* Make a new instruction and pick up the line, stream,
359  * prediction, thread ids from the incoming line */
360  dyn_inst = new MinorDynInst(nullStaticInstPtr, line_in->id);
361 
362  /* Fetch and prediction sequence numbers originate here */
363  dyn_inst->id.fetchSeqNum = fetch_info.fetchSeqNum;
364  dyn_inst->id.predictionSeqNum = fetch_info.predictionSeqNum;
365  /* To complete the set, test that exec sequence number has
366  * not been set */
367  assert(dyn_inst->id.execSeqNum == 0);
368 
369  set(dyn_inst->pc, fetch_info.pc);
370 
371  /* Pack a faulting instruction but allow other
372  * instructions to be generated. (Fetch2 makes no
373  * immediate judgement about streamSeqNum) */
374  dyn_inst->fault = line_in->fault;
375  DPRINTF(Fetch, "Fault being passed output_index: "
376  "%d: %s\n", output_index, dyn_inst->fault->name());
377  } else {
378  uint8_t *line = line_in->line;
379 
380  /* The instruction is wholly in the line, can just copy. */
381  memcpy(decoder->moreBytesPtr(), line + fetch_info.inputIndex,
382  decoder->moreBytesSize());
383 
384  if (!decoder->instReady()) {
385  decoder->moreBytes(*fetch_info.pc,
386  line_in->lineBaseAddr + fetch_info.inputIndex);
387  DPRINTF(Fetch, "Offering MachInst to decoder addr: 0x%x\n",
388  line_in->lineBaseAddr + fetch_info.inputIndex);
389  }
390 
391  /* Maybe make the above a loop to accomodate ISAs with
392  * instructions longer than sizeof(MachInst) */
393 
394  if (decoder->instReady()) {
395  /* Note that the decoder can update the given PC.
396  * Remember not to assign it until *after* calling
397  * decode */
398  StaticInstPtr decoded_inst =
399  decoder->decode(*fetch_info.pc);
400 
401  /* Make a new instruction and pick up the line, stream,
402  * prediction, thread ids from the incoming line */
403  dyn_inst = new MinorDynInst(decoded_inst, line_in->id);
404 
405  /* Fetch and prediction sequence numbers originate here */
406  dyn_inst->id.fetchSeqNum = fetch_info.fetchSeqNum;
407  dyn_inst->id.predictionSeqNum = fetch_info.predictionSeqNum;
408  /* To complete the set, test that exec sequence number
409  * has not been set */
410  assert(dyn_inst->id.execSeqNum == 0);
411 
412  set(dyn_inst->pc, fetch_info.pc);
413  DPRINTF(Fetch, "decoder inst %s\n", *dyn_inst);
414 
415  // Collect some basic inst class stats
416  if (decoded_inst->isLoad())
418  else if (decoded_inst->isStore())
420  else if (decoded_inst->isAtomic())
422  else if (decoded_inst->isVector())
424  else if (decoded_inst->isFloating())
426  else if (decoded_inst->isInteger())
428 
429  DPRINTF(Fetch, "Instruction extracted from line %s"
430  " lineWidth: %d output_index: %d inputIndex: %d"
431  " pc: %s inst: %s\n",
432  line_in->id,
433  line_in->lineWidth, output_index, fetch_info.inputIndex,
434  *fetch_info.pc, *dyn_inst);
435 
436  /*
437  * In SE mode, it's possible to branch to a microop when
438  * replaying faults such as page faults (or simply
439  * intra-microcode branches in X86). Unfortunately,
440  * as Minor has micro-op decomposition in a separate
441  * pipeline stage from instruction decomposition, the
442  * following advancePC (which may follow a branch with
443  * microPC() != 0) *must* see a fresh macroop.
444  *
445  * X86 can branch within microops so we need to deal with
446  * the case that, after a branch, the first un-advanced PC
447  * may be pointing to a microop other than 0. Once
448  * advanced, however, the microop number *must* be 0
449  */
450  fetch_info.pc->uReset();
451 
452  /* Advance PC for the next instruction */
453  decoded_inst->advancePC(*fetch_info.pc);
454 
455  /* Predict any branches and issue a branch if
456  * necessary */
457  predictBranch(dyn_inst, prediction);
458  } else {
459  DPRINTF(Fetch, "Inst not ready yet\n");
460  }
461 
462  /* Step on the pointer into the line if there's no
463  * complete instruction waiting */
464  if (decoder->needMoreBytes()) {
465  fetch_info.inputIndex += decoder->moreBytesSize();
466 
467  DPRINTF(Fetch, "Updated inputIndex value PC: %s"
468  " inputIndex: 0x%x lineBaseAddr: 0x%x lineWidth: 0x%x\n",
469  *line_in->pc, fetch_info.inputIndex, line_in->lineBaseAddr,
470  line_in->lineWidth);
471  }
472  }
473 
474  if (dyn_inst) {
475  /* Step to next sequence number */
476  fetch_info.fetchSeqNum++;
477 
478  /* Correctly size the output before writing */
479  if (output_index == 0) {
480  insts_out.resize(outputWidth);
481  }
482  /* Pack the generated dynamic instruction into the output */
483  insts_out.insts[output_index] = dyn_inst;
484  output_index++;
485 
486  /* Output MinorTrace instruction info for
487  * pre-microop decomposition macroops */
488  if (debug::MinorTrace && !dyn_inst->isFault() &&
489  dyn_inst->staticInst->isMacroop()) {
490  dyn_inst->minorTraceInst(*this);
491  }
492  }
493 
494  /* Remember the streamSeqNum of this line so we can tell when
495  * we change stream */
496  fetch_info.lastStreamSeqNum = line_in->id.streamSeqNum;
497 
498  /* Asked to discard line or there was a branch or fault */
499  if (!prediction.isBubble() || /* The remains of a
500  line with a prediction in it */
501  line_in->isFault() /* A line which is just a fault */)
502  {
503  DPRINTF(Fetch, "Discarding all input on branch/fault\n");
504  dumpAllInput(tid);
505  fetch_info.havePC = false;
506  line_in = NULL;
507  } else if (discard_line) {
508  /* Just discard one line, one's behind it may have new
509  * stream sequence numbers. There's a DPRINTF above
510  * for this event */
511  popInput(tid);
512  fetch_info.havePC = false;
513  line_in = NULL;
514  } else if (fetch_info.inputIndex == line_in->lineWidth) {
515  /* Got to end of a line, pop the line but keep PC
516  * in case this is a line-wrapping inst. */
517  popInput(tid);
518  line_in = NULL;
519  }
520 
521  if (!line_in && processMoreThanOneInput) {
522  DPRINTF(Fetch, "Wrapping\n");
523  line_in = getInput(tid);
524  }
525  }
526 
527  /* The rest of the output (if any) should already have been packed
528  * with bubble instructions by insts_out's initialisation */
529  }
530  if (tid == InvalidThreadID) {
531  assert(insts_out.isBubble());
532  }
534  *predictionOut.inputWire = prediction;
535 
536  /* If we generated output, reserve space for the result in the next stage
537  * and mark the stage as being active this cycle */
538  if (!insts_out.isBubble()) {
539  /* Note activity of following buffer */
541  insts_out.threadId = tid;
542  nextStageReserve[tid].reserve();
543  }
544 
545  /* If we still have input to process and somewhere to put it,
546  * mark stage as active */
547  for (ThreadID i = 0; i < cpu.numThreads; i++)
548  {
549  if (getInput(i) && nextStageReserve[i].canReserve()) {
551  break;
552  }
553  }
554 
555  /* Make sure the input (if any left) is pushed */
556  if (!inp.outputWire->isBubble())
557  inputBuffer[inp.outputWire->id.threadId].pushTail();
558 }
559 
560 inline ThreadID
562 {
563  /* Select thread via policy. */
564  std::vector<ThreadID> priority_list;
565 
566  switch (cpu.threadPolicy) {
567  case enums::SingleThreaded:
568  priority_list.push_back(0);
569  break;
570  case enums::RoundRobin:
571  priority_list = cpu.roundRobinPriority(threadPriority);
572  break;
573  case enums::Random:
574  priority_list = cpu.randomPriority();
575  break;
576  default:
577  panic("Unknown fetch policy");
578  }
579 
580  for (auto tid : priority_list) {
581  if (getInput(tid) && !fetchInfo[tid].blocked) {
582  threadPriority = tid;
583  return tid;
584  }
585  }
586 
587  return InvalidThreadID;
588 }
589 
590 bool
592 {
593  for (const auto &buffer : inputBuffer) {
594  if (!buffer.empty())
595  return false;
596  }
597 
598  return (*inp.outputWire).isBubble() &&
599  (*predictionOut.inputWire).isBubble();
600 }
601 
603  : statistics::Group(cpu, "fetch2"),
604  ADD_STAT(intInstructions, statistics::units::Count::get(),
605  "Number of integer instructions successfully decoded"),
606  ADD_STAT(fpInstructions, statistics::units::Count::get(),
607  "Number of floating point instructions successfully decoded"),
608  ADD_STAT(vecInstructions, statistics::units::Count::get(),
609  "Number of SIMD instructions successfully decoded"),
610  ADD_STAT(loadInstructions, statistics::units::Count::get(),
611  "Number of memory load instructions successfully decoded"),
612  ADD_STAT(storeInstructions, statistics::units::Count::get(),
613  "Number of memory store instructions successfully decoded"),
614  ADD_STAT(amoInstructions, statistics::units::Count::get(),
615  "Number of memory atomic instructions successfully decoded")
616 {
629 }
630 
631 void
633 {
634  std::ostringstream data;
635 
636  if (fetchInfo[0].blocked)
637  data << 'B';
638  else
639  (*out.inputWire).reportData(data);
640 
641  minor::minorTrace("inputIndex=%d havePC=%d predictionSeqNum=%d insts=%s\n",
642  fetchInfo[0].inputIndex, fetchInfo[0].havePC,
643  fetchInfo[0].predictionSeqNum, data.str());
644  inputBuffer[0].minorTrace();
645 }
646 
647 } // namespace minor
648 } // namespace gem5
#define DPRINTF(x,...)
Definition: trace.hh:186
const char data[]
void activity()
Records that there is activity this cycle.
Definition: activity.cc:55
void activateStage(const int idx)
Marks a stage as active.
Definition: activity.cc:91
virtual ThreadContext * getContext(int tn)
Given a thread num get tho thread context for it.
Definition: base.hh:284
ThreadID numThreads
Number of threads we're actually simulating (<= SMT_MAX_THREADS).
Definition: base.hh:367
MinorCPU is an in-order CPU model with four fixed pipeline stages:
Definition: cpu.hh:86
minor::MinorActivityRecorder * activityRecorder
Activity recording for pipeline.
Definition: cpu.hh:96
std::vector< ThreadID > roundRobinPriority(ThreadID priority)
Thread scheduling utility functions.
Definition: cpu.hh:173
std::vector< ThreadID > randomPriority()
Definition: cpu.hh:182
enums::ThreadPolicy threadPolicy
Thread Scheduling Policy (RoundRobin, Random, etc)
Definition: cpu.hh:120
Interface for things with names.
Definition: named.hh:39
virtual std::string name() const
Definition: named.hh:47
Base class for branch operations.
Definition: branch.hh:49
bool isInteger() const
Definition: static_inst.hh:156
bool isLoad() const
Definition: static_inst.hh:147
virtual void advancePC(PCStateBase &pc_state) const =0
bool isFloating() const
Definition: static_inst.hh:157
bool isVector() const
Definition: static_inst.hh:158
bool isStore() const
Definition: static_inst.hh:148
bool isAtomic() const
Definition: static_inst.hh:149
ThreadContext is the external interface to all thread state for anything outside of the CPU.
virtual InstDecoder * getDecoderPtr()=0
void update(const InstSeqNum &done_sn, ThreadID tid)
Tells the branch predictor to commit any updates until the given sequence number.
Definition: bpred_unit.cc:310
bool predict(const StaticInstPtr &inst, const InstSeqNum &seqNum, PCStateBase &pc, ThreadID tid)
Predicts whether or not the instruction is a taken branch, and the target of the branch if it is take...
Definition: bpred_unit.cc:132
void squash(const InstSeqNum &squashed_sn, ThreadID tid)
Squashes all outstanding updates until a given sequence number.
Definition: bpred_unit.cc:333
Forward data betwen Execute and Fetch1 carrying change-of-address/stream information.
Definition: pipe_data.hh:67
MinorDynInstPtr inst
Instruction which caused this branch.
Definition: pipe_data.hh:126
static bool isStreamChange(const BranchData::Reason reason)
Is a request with this reason actually a request to change the PC rather than a bubble or branch pred...
Definition: pipe_data.cc:85
Reason reason
Explanation for this branch.
Definition: pipe_data.hh:113
ThreadID threadId
ThreadID associated with branch.
Definition: pipe_data.hh:116
bool isBubble() const
Definition: pipe_data.hh:164
std::unique_ptr< PCStateBase > target
Starting PC of that stream.
Definition: pipe_data.hh:123
gem5::minor::Fetch2::Fetch2Stats stats
void evaluate()
Pass on input/buffer data to the output if you can.
Definition: fetch2.cc:240
void updateBranchPrediction(const BranchData &branch)
Update local branch prediction structures from feedback from Execute.
Definition: fetch2.cc:130
const ForwardLineData * getInput(ThreadID tid)
Get a piece of data to work on from the inputBuffer, or 0 if there is no data.
Definition: fetch2.cc:98
std::vector< InputBuffer< ForwardLineData > > inputBuffer
Definition: fetch2.hh:100
void popInput(ThreadID tid)
Pop an element off the input buffer, if there are any.
Definition: fetch2.cc:109
MinorCPU & cpu
Pointer back to the containing CPU.
Definition: fetch2.hh:70
Fetch2(const std::string &name, MinorCPU &cpu_, const BaseMinorCPUParams &params, Latch< ForwardLineData >::Output inp_, Latch< BranchData >::Output branchInp_, Latch< BranchData >::Input predictionOut_, Latch< ForwardInstData >::Input out_, std::vector< InputBuffer< ForwardInstData >> &next_stage_input_buffer)
Definition: fetch2.cc:59
branch_prediction::BPredUnit & branchPredictor
Branch predictor passed from Python configuration.
Definition: fetch2.hh:96
Latch< ForwardLineData >::Output inp
Input port carrying lines from Fetch1.
Definition: fetch2.hh:73
bool isDrained()
Is this stage drained? For Fetch2, draining is initiated by Execute halting Fetch1 causing Fetch2 to ...
Definition: fetch2.cc:591
Latch< BranchData >::Input predictionOut
Output port carrying predictions back to Fetch1.
Definition: fetch2.hh:80
ThreadID threadPriority
Definition: fetch2.hh:162
ThreadID getScheduledThread()
Use the current threading policy to determine the next thread to fetch from.
Definition: fetch2.cc:561
Latch< ForwardInstData >::Input out
Output port carrying instructions into Decode.
Definition: fetch2.hh:83
void dumpAllInput(ThreadID tid)
Dump the whole contents of the input buffer.
Definition: fetch2.cc:120
void predictBranch(MinorDynInstPtr inst, BranchData &branch)
Predicts branches for the given instruction.
Definition: fetch2.cc:192
std::vector< InputBuffer< ForwardInstData > > & nextStageReserve
Interface to reserve space in the next stage.
Definition: fetch2.hh:86
bool processMoreThanOneInput
If true, more than one input word can be processed each cycle if there is room in the output to conta...
Definition: fetch2.hh:93
std::vector< Fetch2ThreadInfo > fetchInfo
Definition: fetch2.hh:161
unsigned int outputWidth
Width of output of this stage/input of next in instructions.
Definition: fetch2.hh:89
Latch< BranchData >::Output branchInp
Input port carrying branches from Execute.
Definition: fetch2.hh:77
void minorTrace() const
Definition: fetch2.cc:632
Forward flowing data between Fetch2,Decode,Execute carrying a packet of instructions of a width appro...
Definition: pipe_data.hh:285
ThreadID threadId
Thread associated with these instructions.
Definition: pipe_data.hh:294
void resize(unsigned int width)
Resize a bubble/empty ForwardInstData and fill with bubbles.
Definition: pipe_data.cc:264
bool isBubble() const
BubbleIF interface.
Definition: pipe_data.cc:251
MinorDynInstPtr insts[MAX_FORWARD_INSTS]
Array of carried insts, ref counted.
Definition: pipe_data.hh:288
Line fetch data in the forward direction.
Definition: pipe_data.hh:188
unsigned int lineWidth
Explicit line width, don't rely on data.size.
Definition: pipe_data.hh:207
uint8_t * line
Line data.
Definition: pipe_data.hh:219
InstId id
Thread, stream, prediction ...
Definition: pipe_data.hh:215
std::unique_ptr< PCStateBase > pc
PC of the first inst within this sequence.
Definition: pipe_data.hh:201
bool isFault() const
This is a fault, not a line.
Definition: pipe_data.hh:251
Fault fault
This line has a fault.
Definition: pipe_data.hh:212
Addr lineBaseAddr
First byte address in the line.
Definition: pipe_data.hh:198
Like a Queue but with a restricted interface and a setTail function which, when the queue is empty,...
Definition: buffers.hh:573
ThreadID threadId
The thread to which this line/instruction belongs.
Definition: dyn_inst.hh:89
InstSeqNum streamSeqNum
The 'stream' this instruction belongs to.
Definition: dyn_inst.hh:94
InstSeqNum predictionSeqNum
The predicted qualifier to stream, attached by Fetch2 as a consequence of branch prediction.
Definition: dyn_inst.hh:98
Encapsulate wires on either input or output of the latch.
Definition: buffers.hh:253
Dynamic instruction for Minor.
Definition: dyn_inst.hh:164
Derived & flags(Flags _flags)
Set the flags and marks this stat to print at the end of simulation.
Definition: statistics.hh:358
Statistics container.
Definition: group.hh:94
STL vector class.
Definition: stl.hh:37
Fetch2 receives lines of data from Fetch1, separates them into instructions and passes them to Decode...
#define ADD_STAT(n,...)
Convenience macro to add a stat to a statistics group.
Definition: group.hh:75
#define panic(...)
This implements a cprintf based panic() function.
Definition: logging.hh:178
#define fatal(...)
This implements a cprintf based fatal() function.
Definition: logging.hh:190
Bitfield< 7 > i
Definition: misc_types.hh:67
Bitfield< 12, 11 > set
Definition: misc_types.hh:709
void minorTrace(const char *fmt, Args ...args)
DPRINTFN for MinorTrace reporting.
Definition: trace.hh:67
const FlagsType total
Print the total.
Definition: info.hh:60
Reference material can be found at the JEDEC website: UFS standard http://www.jedec....
int16_t ThreadID
Thread index/ID type.
Definition: types.hh:235
const ThreadID InvalidThreadID
Definition: types.hh:236
const StaticInstPtr nullStaticInstPtr
Statically allocated null StaticInstPtr.
GEM5_DEPRECATED_NAMESPACE(GuestABI, guest_abi)
Minor contains all the definitions within the MinorCPU apart from the CPU class itself.
const std::string to_string(sc_enc enc)
Definition: sc_fxdefs.cc:60
output decoder
Definition: nop.cc:61
The constructed pipeline.
statistics::Scalar loadInstructions
Definition: fetch2.hh:171
statistics::Scalar fpInstructions
Definition: fetch2.hh:169
statistics::Scalar intInstructions
Stats.
Definition: fetch2.hh:168
statistics::Scalar storeInstructions
Definition: fetch2.hh:172
statistics::Scalar amoInstructions
Definition: fetch2.hh:173
statistics::Scalar vecInstructions
Definition: fetch2.hh:170
Data members after this line are cycle-to-cycle state.
Definition: fetch2.hh:106
InstSeqNum expectedStreamSeqNum
Stream sequence number remembered from last time the predictionSeqNum changed.
Definition: fetch2.hh:150
InstSeqNum fetchSeqNum
Fetch2 is the source of fetch sequence numbers.
Definition: fetch2.hh:144
bool havePC
PC is currently valid.
Definition: fetch2.hh:136
InstSeqNum lastStreamSeqNum
Stream sequence number of the last seen line used to identify changes of instruction stream.
Definition: fetch2.hh:140
std::unique_ptr< PCStateBase > pc
Remembered program counter value.
Definition: fetch2.hh:131
unsigned int inputIndex
Index into an incompletely processed input line that instructions are to be extracted from.
Definition: fetch2.hh:122
InstSeqNum predictionSeqNum
Fetch2 is the source of prediction sequence numbers.
Definition: fetch2.hh:155
bool blocked
Blocked indication for report.
Definition: fetch2.hh:158
const std::string & name()
Definition: trace.cc:49

Generated on Wed Dec 21 2022 10:22:30 for gem5 by doxygen 1.9.1