gem5 v24.1.0.1
All Classes Namespaces Files Functions Variables Typedefs Enumerations Enumerator Friends Macros Modules Pages
gpu_command_processor.hh
Go to the documentation of this file.
1/*
2 * Copyright (c) 2018 Advanced Micro Devices, Inc.
3 * All rights reserved.
4 *
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions are met:
7 *
8 * 1. Redistributions of source code must retain the above copyright notice,
9 * this list of conditions and the following disclaimer.
10 *
11 * 2. Redistributions in binary form must reproduce the above copyright notice,
12 * this list of conditions and the following disclaimer in the documentation
13 * and/or other materials provided with the distribution.
14 *
15 * 3. Neither the name of the copyright holder nor the names of its
16 * contributors may be used to endorse or promote products derived from this
17 * software without specific prior written permission.
18 *
19 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
20 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
21 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
22 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
23 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
24 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
25 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
26 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
27 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
28 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
29 * POSSIBILITY OF SUCH DAMAGE.
30 */
31
43#ifndef __DEV_HSA_GPU_COMMAND_PROCESSOR_HH__
44#define __DEV_HSA_GPU_COMMAND_PROCESSOR_HH__
45
46#include <cstdint>
47#include <functional>
48
50#include "base/logging.hh"
51#include "base/trace.hh"
52#include "base/types.hh"
53#include "debug/GPUCommandProc.hh"
56#include "dev/hsa/hsa_signal.hh"
60#include "params/GPUCommandProcessor.hh"
61#include "sim/full_system.hh"
62
63namespace gem5
64{
65
66struct GPUCommandProcessorParams;
67class GPUComputeDriver;
68class GPUDispatcher;
69class Shader;
70
72{
73 public:
74 typedef GPUCommandProcessorParams Params;
75 typedef std::function<void(const uint64_t &)> HsaSignalCallbackFunction;
76
79
82
83 void setGPUDevice(AMDGPUDevice *gpu_device);
84 void setShader(Shader *shader);
85 Shader* shader();
87
96
98
100 {
101 Nop = 0,
102 Steal = 1
103 };
104
106
107 void completeTimingRead();
108
109 void submitAgentDispatchPkt(void *raw_pkt, uint32_t queue_id,
110 Addr host_pkt_addr);
111 void submitDispatchPkt(void *raw_pkt, uint32_t queue_id,
112 Addr host_pkt_addr);
113 void submitVendorPkt(void *raw_pkt, uint32_t queue_id,
114 Addr host_pkt_addr);
116
117 void dispatchKernelObject(AMDKernelCode *akc, void *raw_pkt,
118 uint32_t queue_id, Addr host_pkt_addr);
119 void dispatchPkt(HSAQueueEntry *task);
120 void signalWakeupEvent(uint32_t event_id);
121
122 Tick write(PacketPtr pkt) override { return 0; }
123 Tick read(PacketPtr pkt) override { return 0; }
124 AddrRangeList getAddrRanges() const override;
125 System *system();
126
127 void sendCompletionSignal(Addr signal_handle);
128 void updateHsaSignal(Addr signal_handle, uint64_t signal_value,
130 [] (const uint64_t &) { });
131 void updateHsaSignalAsync(Addr signal_handle, int64_t diff);
132 void updateHsaSignalData(Addr value_addr, int64_t diff,
133 uint64_t *prev_value);
134 void updateHsaSignalDone(uint64_t *signal_value);
135 void updateHsaMailboxData(Addr signal_handle, uint64_t *mailbox_value);
136 void updateHsaEventData(Addr signal_handle, uint64_t *event_value);
137 void updateHsaEventTs(Addr signal_handle, amd_event_t *event_value);
138
139 uint64_t functionalReadHsaSignal(Addr signal_handle);
140
142 {
143 return signal_handle + offsetof(amd_signal_t, value);
144 }
145
147 {
148 return signal_handle + offsetof(amd_signal_t, event_mailbox_ptr);
149 }
150
152 {
153 return signal_handle + offsetof(amd_signal_t, event_id);
154 }
155
156 private:
162
163 // Typedefing dmaRead and dmaWrite function pointer
164 typedef void (DmaDevice::*DmaFnPtr)(Addr, int, Event*, uint8_t*, Tick);
165 void initABI(HSAQueueEntry *task);
166 void sanityCheckAKC(AMDKernelCode *akc);
168 TranslationGenPtr translate(Addr vaddr, Addr size) override;
169
170 // Running counter of dispatched tasks
172
173 // Running counter of dispatched user (non-blit) kernels
175
176 // Skip all user (non-blit) kernels until reaching this kernel
178
179 // Keep track of start times for task dispatches.
180 std::unordered_map<Addr, Tick> dispatchStartTime;
181
194 void
196 const uint32_t &readDispIdOffset)
197 {
206 task->queueId())->hostReadIndexPtr - readDispIdOffset;
207
212 auto *mqdDmaEvent = new DmaVirtCallback<int>(
213 [ = ] (const int &) { MQDDmaEvent(task); });
214
216 sizeof(_amd_queue_t), mqdDmaEvent, &task->amdQueue);
217 }
218
226 void
228 {
243 // TODO: Raising this signal will potentially nuke scratch
244 // space for in-flight kernels that were launched from this
245 // queue. We need to drain all kernels and deschedule the
246 // queue before raising this signal. For now, just assert if
247 // there are any in-flight kernels and tell the user that this
248 // feature still needs to be implemented.
249 fatal_if(hsaPP->inFlightPkts(task->queueId()) > 1,
250 "Needed more scratch, but kernels are in flight for "
251 "this queue and it is unsafe to reallocate scratch. "
252 "We need to implement additional intelligence in the "
253 "hardware scheduling logic to support CP-driven "
254 "queue draining and scheduling.");
255 DPRINTF(GPUCommandProc, "Not enough scratch space to launch "
256 "kernel (%x available, %x requested bytes per "
257 "workitem). Asking host runtime to allocate more "
258 "space.\n",
260 task->privMemPerItem());
261
263 [ = ] (const uint64_t &dma_buffer)
264 { WaitScratchDmaEvent(task, dma_buffer); });
265
266 } else {
267 DPRINTF(GPUCommandProc, "Sufficient scratch space, launching "
268 "kernel (%x available, %x requested bytes per "
269 "workitem).\n",
271 task->privMemPerItem());
272 dispatchPkt(task);
273 }
274 }
275
280 void
281 WaitScratchDmaEvent(HSAQueueEntry *task, const uint64_t &dmaBuffer)
282 {
283 if (dmaBuffer == 0) {
284 DPRINTF(GPUCommandProc, "Host scratch allocation complete. "
285 "Attempting to re-read MQD\n");
294 auto cb = new DmaVirtCallback<int>(
295 [ = ] (const int &) { MQDDmaEvent(task); });
296
297 dmaReadVirt(task->hostAMDQueueAddr, sizeof(_amd_queue_t), cb,
298 &task->amdQueue);
299 } else {
304 Addr value_addr = getHsaSignalValueAddr(
306 DPRINTF(GPUCommandProc, "Polling queue inactive signal at "
307 "%p.\n", value_addr);
308 auto cb = new DmaVirtCallback<uint64_t>(
309 [ = ] (const uint64_t &dma_buffer)
310 { WaitScratchDmaEvent(task, dma_buffer); } );
311
319 dmaReadVirt(value_addr, sizeof(Addr), cb, &cb->dmaBuffer, 1e9);
320 }
321 }
322};
323
324} // namespace gem5
325
326#endif // __DEV_HSA_GPU_COMMAND_PROCESSOR_HH__
#define DPRINTF(x,...)
Definition trace.hh:209
Defines global host-dependent types: Counter, Tick, and (indirectly) {int,uint}{8,...
Device model for an AMD GPU.
Wraps a std::function object in a DmaCallback.
void dmaReadVirt(Addr host_addr, unsigned size, DmaCallback *cb, void *data, Tick delay=0)
Initiate a DMA read from virtual address host_addr.
void sendCompletionSignal(Addr signal_handle)
void submitDispatchPkt(void *raw_pkt, uint32_t queue_id, Addr host_pkt_addr)
submitDispatchPkt() is the entry point into the CP from the HSAPP and is only meant to be used with A...
void ReadDispIdOffsetDmaEvent(HSAQueueEntry *task, const uint32_t &readDispIdOffset)
Perform a DMA read of the read_dispatch_id_field_base_byte_offset field, which follows directly after...
RequestorID vramRequestorId()
Forward the VRAM requestor ID needed for device memory from GPU device.
Addr getHsaSignalMailboxAddr(Addr signal_handle)
void(DmaDevice::* DmaFnPtr)(Addr, int, Event *, uint8_t *, Tick)
void setGPUDevice(AMDGPUDevice *gpu_device)
TranslationGenPtr translate(Addr vaddr, Addr size) override
Function used to translate a range of addresses from virtual to physical addresses.
void signalWakeupEvent(uint32_t event_id)
void updateHsaSignal(Addr signal_handle, uint64_t signal_value, HsaSignalCallbackFunction function=[](const uint64_t &) { })
void updateHsaSignalDone(uint64_t *signal_value)
HSAPacketProcessor & hsaPacketProc()
void submitAgentDispatchPkt(void *raw_pkt, uint32_t queue_id, Addr host_pkt_addr)
submitAgentDispatchPkt() is for accepting agent dispatch packets.
std::list< struct KernelDispatchData > kernelDispatchList
Addr getHsaSignalValueAddr(Addr signal_handle)
void updateHsaEventTs(Addr signal_handle, amd_event_t *event_value)
void dispatchKernelObject(AMDKernelCode *akc, void *raw_pkt, uint32_t queue_id, Addr host_pkt_addr)
void MQDDmaEvent(HSAQueueEntry *task)
Perform a DMA read of the MQD that corresponds to a hardware queue descriptor (HQD).
void attachDriver(GPUComputeDriver *driver)
void initABI(HSAQueueEntry *task)
The CP is responsible for traversing all HSA-ABI-related data structures from memory and initializing...
void updateHsaSignalAsync(Addr signal_handle, int64_t diff)
std::unordered_map< Addr, Tick > dispatchStartTime
Addr getHsaSignalEventAddr(Addr signal_handle)
AddrRangeList getAddrRanges() const override
Every PIO device is obliged to provide an implementation that returns the address ranges the device r...
void submitVendorPkt(void *raw_pkt, uint32_t queue_id, Addr host_pkt_addr)
submitVendorPkt() is for accepting vendor-specific packets from the HSAPP.
void sanityCheckAKC(AMDKernelCode *akc)
Tick write(PacketPtr pkt) override
Pure virtual function that the device must implement.
GPUCommandProcessorParams Params
void dispatchPkt(HSAQueueEntry *task)
Once the CP has finished extracting all relevant information about a task and has initialized the ABI...
Tick read(PacketPtr pkt) override
Pure virtual function that the device must implement.
void updateHsaMailboxData(Addr signal_handle, uint64_t *mailbox_value)
void performTimingRead(PacketPtr pkt)
void updateHsaEventData(Addr signal_handle, uint64_t *event_value)
std::function< void(const uint64_t &)> HsaSignalCallbackFunction
uint64_t functionalReadHsaSignal(Addr signal_handle)
void WaitScratchDmaEvent(HSAQueueEntry *task, const uint64_t &dmaBuffer)
Poll on queue_inactive signal until the runtime can get around to taking care of our lack of scratch ...
void updateHsaSignalData(Addr value_addr, int64_t diff, uint64_t *prev_value)
uint64_t inFlightPkts(uint32_t queId)
HSAQueueDescriptor * getQueueDesc(uint32_t queId)
_amd_queue_t amdQueue
Keep a copy of the AMD HSA queue because we need info from some of its fields to initialize register ...
uint32_t queueId() const
Addr hostAMDQueueAddr
Host-side addr of the amd_queue_t on which this task was queued.
A Packet is used to encapsulate a transfer between two objects in the memory system (e....
Definition packet.hh:295
Nop class.
Definition nop.hh:49
STL list class.
Definition stl.hh:51
The GPUDispatcher is the component of the shader that is responsible for creating and dispatching WGs...
The GPUComputeDriver implements an HSADriver for an HSA AMD GPU agent.
#define fatal_if(cond,...)
Conditional fatal macro that checks the supplied condition and only causes a fatal error if the condi...
Definition logging.hh:236
HSAQueuEntry is the simulator's internal representation of an AQL queue entry (task).
Bitfield< 0 > p
const int NumVecElemPerVecReg(64)
Copyright (c) 2024 Arm Limited All rights reserved.
Definition binary32.hh:36
uint64_t Addr
Address type This will probably be moved somewhere else in the near future.
Definition types.hh:147
uint64_t Tick
Tick count type.
Definition types.hh:58
uint16_t RequestorID
Definition request.hh:95
std::unique_ptr< TranslationGen > TranslationGenPtr
PM4 packets.
_hsa_signal_t queue_inactive_signal
Definition hsa_queue.hh:87
uint32_t compute_tmpring_size_wavesize
Definition hsa_queue.hh:79

Generated on Mon Jan 13 2025 04:28:36 for gem5 by doxygen 1.9.8