gem5  v22.1.0.0
scalar_memory_pipeline.cc
Go to the documentation of this file.
1 /*
2  * Copyright (c) 2016-2017 Advanced Micro Devices, Inc.
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions are met:
7  *
8  * 1. Redistributions of source code must retain the above copyright notice,
9  * this list of conditions and the following disclaimer.
10  *
11  * 2. Redistributions in binary form must reproduce the above copyright notice,
12  * this list of conditions and the following disclaimer in the documentation
13  * and/or other materials provided with the distribution.
14  *
15  * 3. Neither the name of the copyright holder nor the names of its
16  * contributors may be used to endorse or promote products derived from this
17  * software without specific prior written permission.
18  *
19  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
20  * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
21  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
22  * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
23  * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29  * POSSIBILITY OF SUCH DAMAGE.
30  */
31 
33 
34 #include "debug/GPUMem.hh"
35 #include "debug/GPUReg.hh"
39 #include "gpu-compute/shader.hh"
40 #include "gpu-compute/wavefront.hh"
41 
42 namespace gem5
43 {
44 
45 ScalarMemPipeline::ScalarMemPipeline(const ComputeUnitParams &p,
46  ComputeUnit &cu)
47  : computeUnit(cu), _name(cu.name() + ".ScalarMemPipeline"),
48  queueSize(p.scalar_mem_queue_size),
49  inflightStores(0), inflightLoads(0)
50 {
51 }
52 
53 void
55 {
56  // afind oldest scalar request whose data has arrived
57  GPUDynInstPtr m = !returnedLoads.empty() ? returnedLoads.front() :
58  !returnedStores.empty() ? returnedStores.front() : nullptr;
59 
60  Wavefront *w = nullptr;
61 
62  bool accessSrf = true;
63  // check the SRF to see if the operands of a load (or load component
64  // of an atomic) are accessible
65  if ((m) && (m->isLoad() || m->isAtomicRet())) {
66  w = m->wavefront();
67 
68  accessSrf =
69  w->computeUnit->srf[w->simdId]->
70  canScheduleWriteOperandsFromLoad(w, m);
71  }
72 
73  if ((!returnedStores.empty() || !returnedLoads.empty()) &&
74  m->latency.rdy() && computeUnit.scalarMemToSrfBus.rdy() &&
75  accessSrf &&
78 
79  w = m->wavefront();
80 
81  if (m->isLoad() || m->isAtomicRet()) {
82  w->computeUnit->srf[w->simdId]->
83  scheduleWriteOperandsFromLoad(w, m);
84  }
85 
86  m->completeAcc(m);
87  w->decLGKMInstsIssued();
88 
89  if (m->isLoad() || m->isAtomic()) {
90  returnedLoads.pop();
91  assert(inflightLoads > 0);
92  --inflightLoads;
93  } else {
94  returnedStores.pop();
95  assert(inflightStores > 0);
97  }
98 
99  // Decrement outstanding register count
100  computeUnit.shader->ScheduleAdd(&w->outstandingReqs, m->time, -1);
101 
102  if (m->isStore() || m->isAtomic()) {
103  computeUnit.shader->ScheduleAdd(&w->scalarOutstandingReqsWrGm,
104  m->time, -1);
105  }
106 
107  if (m->isLoad() || m->isAtomic()) {
108  computeUnit.shader->ScheduleAdd(&w->scalarOutstandingReqsRdGm,
109  m->time, -1);
110  }
111 
112  // Mark write bus busy for appropriate amount of time
115  w->computeUnit->scalarMemUnit.set(m->time);
116  }
117 
118  // If pipeline has executed a global memory instruction
119  // execute global memory packets and issue global
120  // memory packets to DTLB
121  if (!issuedRequests.empty()) {
122  GPUDynInstPtr mp = issuedRequests.front();
123  if (mp->isLoad() || mp->isAtomic()) {
124 
125  if (inflightLoads >= queueSize) {
126  return;
127  } else {
128  ++inflightLoads;
129  }
130  } else {
131  if (inflightStores >= queueSize) {
132  return;
133  } else {
134  ++inflightStores;
135  }
136  }
137  mp->initiateAcc(mp);
138  issuedRequests.pop();
139 
140  DPRINTF(GPUMem, "CU%d: WF[%d][%d] Popping scalar mem_op\n",
141  computeUnit.cu_id, mp->simdId, mp->wfSlotId);
142  }
143 }
144 
145 void
147 {
148  Wavefront *wf = gpuDynInst->wavefront();
149  if (gpuDynInst->isLoad()) {
150  wf->scalarRdGmReqsInPipe--;
152  } else if (gpuDynInst->isStore()) {
153  wf->scalarWrGmReqsInPipe--;
155  }
156 
157  wf->outstandingReqs++;
159 
160  issuedRequests.push(gpuDynInst);
161 }
162 
163 } // namespace gem5
#define DPRINTF(x,...)
Definition: trace.hh:186
WaitClass scalarMemUnit
WaitClass scalarMemToSrfBus
std::queue< GPUDynInstPtr > returnedLoads
ScalarMemPipeline(const ComputeUnitParams &p, ComputeUnit &cu)
std::queue< GPUDynInstPtr > issuedRequests
std::queue< GPUDynInstPtr > returnedStores
void issueRequest(GPUDynInstPtr gpuDynInst)
void ScheduleAdd(int *val, Tick when, int x)
Definition: shader.cc:357
int coissue_return
Definition: shader.hh:229
void set(uint64_t i)
Definition: misc.hh:82
bool rdy(Cycles cycles=Cycles(0)) const
Definition: misc.hh:93
void validateRequestCounters()
Definition: wavefront.cc:770
int scalarWrGmReqsInPipe
Definition: wavefront.hh:189
int scalarOutstandingReqsWrGm
Definition: wavefront.hh:183
int scalarOutstandingReqsRdGm
Definition: wavefront.hh:181
int scalarRdGmReqsInPipe
Definition: wavefront.hh:188
Bitfield< 11 > mp
Definition: misc_types.hh:827
Bitfield< 6 > w
Definition: pagetable.hh:59
Bitfield< 54 > p
Definition: pagetable.hh:70
Reference material can be found at the JEDEC website: UFS standard http://www.jedec....
std::shared_ptr< GPUDynInst > GPUDynInstPtr
Definition: misc.hh:49
const std::string & name()
Definition: trace.cc:49

Generated on Wed Dec 21 2022 10:22:35 for gem5 by doxygen 1.9.1