59 if (gpuDynInst->exec_mask.none()) {
68 gpuDynInst->latency.init(gpuDynInst->computeUnit());
69 gpuDynInst->latency.set(gpuDynInst->computeUnit()->clockPeriod());
79 initMemRead<VecElemU8>(gpuDynInst);
88 if (gpuDynInst->exec_mask[lane]) {
90 gpuDynInst->d_data))[lane]);
115 if (gpuDynInst->exec_mask.none()) {
124 gpuDynInst->latency.init(gpuDynInst->computeUnit());
125 gpuDynInst->latency.set(gpuDynInst->computeUnit()->clockPeriod());
135 initMemRead<VecElemI8>(gpuDynInst);
144 if (gpuDynInst->exec_mask[lane]) {
146 gpuDynInst->d_data))[lane]);
171 if (gpuDynInst->exec_mask.none()) {
180 gpuDynInst->latency.init(gpuDynInst->computeUnit());
181 gpuDynInst->latency.set(gpuDynInst->computeUnit()->clockPeriod());
191 initMemRead<VecElemU16>(gpuDynInst);
200 if (gpuDynInst->exec_mask[lane]) {
202 gpuDynInst->d_data))[lane]);
258 if (gpuDynInst->exec_mask.none()) {
267 gpuDynInst->latency.init(gpuDynInst->computeUnit());
268 gpuDynInst->latency.set(gpuDynInst->computeUnit()->clockPeriod());
278 initMemRead<VecElemU32>(gpuDynInst);
287 if (gpuDynInst->exec_mask[lane]) {
289 gpuDynInst->d_data))[lane];
315 if (gpuDynInst->exec_mask.none()) {
324 gpuDynInst->latency.init(gpuDynInst->computeUnit());
325 gpuDynInst->latency.set(gpuDynInst->computeUnit()->clockPeriod());
335 initMemRead<2>(gpuDynInst);
347 gpuDynInst->d_data))[lane * 2];
349 gpuDynInst->d_data))[lane * 2 + 1];
354 gpuDynInst->d_data))[lane];
384 if (gpuDynInst->exec_mask.none()) {
393 gpuDynInst->latency.init(gpuDynInst->computeUnit());
394 gpuDynInst->latency.set(gpuDynInst->computeUnit()->clockPeriod());
404 initMemRead<3>(gpuDynInst);
417 gpuDynInst->d_data))[lane * 3];
419 gpuDynInst->d_data))[lane * 3 + 1];
421 gpuDynInst->d_data))[lane * 3 + 2];
426 gpuDynInst->d_data))[lane];
459 if (gpuDynInst->exec_mask.none()) {
468 gpuDynInst->latency.init(gpuDynInst->computeUnit());
469 gpuDynInst->latency.set(gpuDynInst->computeUnit()->clockPeriod());
479 initMemRead<4>(gpuDynInst);
493 gpuDynInst->d_data))[lane * 4];
495 gpuDynInst->d_data))[lane * 4 + 1];
497 gpuDynInst->d_data))[lane * 4 + 2];
499 gpuDynInst->d_data))[lane * 4 + 3];
504 gpuDynInst->d_data))[lane];
539 if (gpuDynInst->exec_mask.none()) {
549 gpuDynInst->latency.init(gpuDynInst->computeUnit());
550 gpuDynInst->latency.set(gpuDynInst->computeUnit()->clockPeriod());
559 if (gpuDynInst->exec_mask[lane]) {
560 (
reinterpret_cast<VecElemU8*
>(gpuDynInst->d_data))[lane]
571 initMemWrite<VecElemU8>(gpuDynInst);
598 if (gpuDynInst->exec_mask.none()) {
608 gpuDynInst->latency.init(gpuDynInst->computeUnit());
609 gpuDynInst->latency.set(gpuDynInst->computeUnit()->clockPeriod());
618 if (gpuDynInst->exec_mask[lane]) {
619 (
reinterpret_cast<VecElemU16*
>(gpuDynInst->d_data))[lane]
630 initMemWrite<VecElemU16>(gpuDynInst);
641 :
Inst_FLAT(iFmt,
"flat_store_short_d16_hi")
658 if (gpuDynInst->exec_mask.none()) {
668 gpuDynInst->latency.init(gpuDynInst->computeUnit());
669 gpuDynInst->latency.set(gpuDynInst->computeUnit()->clockPeriod());
678 if (gpuDynInst->exec_mask[lane]) {
679 (
reinterpret_cast<VecElemU16*
>(gpuDynInst->d_data))[lane]
680 = (
data[lane] >> 16);
690 initMemWrite<VecElemU16>(gpuDynInst);
717 if (gpuDynInst->exec_mask.none()) {
727 gpuDynInst->latency.init(gpuDynInst->computeUnit());
728 gpuDynInst->latency.set(gpuDynInst->computeUnit()->clockPeriod());
737 if (gpuDynInst->exec_mask[lane]) {
738 (
reinterpret_cast<VecElemU32*
>(gpuDynInst->d_data))[lane]
749 initMemWrite<VecElemU32>(gpuDynInst);
777 if (gpuDynInst->exec_mask.none()) {
787 gpuDynInst->latency.init(gpuDynInst->computeUnit());
788 gpuDynInst->latency.set(gpuDynInst->computeUnit()->clockPeriod());
797 if (gpuDynInst->exec_mask[lane]) {
798 (
reinterpret_cast<VecElemU64*
>(gpuDynInst->d_data))[lane]
809 initMemWrite<2>(gpuDynInst);
837 if (gpuDynInst->exec_mask.none()) {
847 gpuDynInst->latency.init(gpuDynInst->computeUnit());
848 gpuDynInst->latency.set(gpuDynInst->computeUnit()->clockPeriod());
861 if (gpuDynInst->exec_mask[lane]) {
863 gpuDynInst->d_data))[lane * 3] = data0[lane];
865 gpuDynInst->d_data))[lane * 3 + 1] = data1[lane];
867 gpuDynInst->d_data))[lane * 3 + 2] = data2[lane];
877 initMemWrite<3>(gpuDynInst);
905 if (gpuDynInst->exec_mask.none()) {
915 gpuDynInst->latency.init(gpuDynInst->computeUnit());
916 gpuDynInst->latency.set(gpuDynInst->computeUnit()->clockPeriod());
931 if (gpuDynInst->exec_mask[lane]) {
933 gpuDynInst->d_data))[lane * 4] = data0[lane];
935 gpuDynInst->d_data))[lane * 4 + 1] = data1[lane];
937 gpuDynInst->d_data))[lane * 4 + 2] = data2[lane];
939 gpuDynInst->d_data))[lane * 4 + 3] = data3[lane];
949 initMemWrite<4>(gpuDynInst);
982 atomicExecute<ConstVecOperandU32, VecElemU32>(gpuDynInst);
988 initAtomicAccess<VecElemU32>(gpuDynInst);
994 atomicComplete<VecOperandU32, VecElemU32>(gpuDynInst);
999 Inst_FLAT__FLAT_ATOMIC_CMPSWAP
1000 ::Inst_FLAT__FLAT_ATOMIC_CMPSWAP(
InFmt_FLAT *iFmt)
1001 :
Inst_FLAT(iFmt,
"flat_atomic_cmpswap")
999 Inst_FLAT__FLAT_ATOMIC_CMPSWAP {
…}
1026 atomicExecute<ConstVecOperandU32, VecElemU32, 1>(gpuDynInst);
1032 initAtomicAccess<VecElemU32>(gpuDynInst);
1038 atomicComplete<VecOperandU32, VecElemU32>(gpuDynInst);
1066 atomicExecute<ConstVecOperandU32, VecElemU32>(gpuDynInst);
1072 initAtomicAccess<VecElemU32>(gpuDynInst);
1078 atomicComplete<VecOperandU32, VecElemU32>(gpuDynInst);
1106 atomicExecute<ConstVecOperandU32, VecElemU32>(gpuDynInst);
1112 initAtomicAccess<VecElemU32>(gpuDynInst);
1118 atomicComplete<VecOperandU32, VecElemU32>(gpuDynInst);
1146 atomicExecute<ConstVecOperandI32, VecElemI32>(gpuDynInst);
1152 initAtomicAccess<VecElemI32>(gpuDynInst);
1158 atomicComplete<VecOperandI32, VecElemI32>(gpuDynInst);
1186 atomicExecute<ConstVecOperandU32, VecElemU32>(gpuDynInst);
1192 initAtomicAccess<VecElemU32>(gpuDynInst);
1198 atomicComplete<VecOperandU32, VecElemU32>(gpuDynInst);
1226 atomicExecute<ConstVecOperandI32, VecElemI32>(gpuDynInst);
1232 initAtomicAccess<VecElemI32>(gpuDynInst);
1238 atomicComplete<VecOperandI32, VecElemI32>(gpuDynInst);
1266 atomicExecute<ConstVecOperandU32, VecElemU32>(gpuDynInst);
1272 initAtomicAccess<VecElemU32>(gpuDynInst);
1278 atomicComplete<VecOperandU32, VecElemU32>(gpuDynInst);
1306 atomicExecute<ConstVecOperandU32, VecElemU32>(gpuDynInst);
1312 initAtomicAccess<VecElemU32>(gpuDynInst);
1318 atomicComplete<VecOperandU32, VecElemU32>(gpuDynInst);
1346 atomicExecute<ConstVecOperandU32, VecElemU32>(gpuDynInst);
1352 initAtomicAccess<VecElemU32>(gpuDynInst);
1358 atomicComplete<VecOperandU32, VecElemU32>(gpuDynInst);
1387 atomicExecute<ConstVecOperandU32, VecElemU32>(gpuDynInst);
1393 initAtomicAccess<VecElemU32>(gpuDynInst);
1399 atomicComplete<VecOperandU32, VecElemU32>(gpuDynInst);
1427 atomicExecute<ConstVecOperandU32, VecElemU32>(gpuDynInst);
1433 initAtomicAccess<VecElemU32>(gpuDynInst);
1439 atomicComplete<VecOperandU32, VecElemU32>(gpuDynInst);
1467 atomicExecute<ConstVecOperandU32, VecElemU32>(gpuDynInst);
1473 initAtomicAccess<VecElemU32>(gpuDynInst);
1479 atomicComplete<VecOperandU32, VecElemU32>(gpuDynInst);
1485 :
Inst_FLAT(iFmt,
"flat_atomic_swap_x2")
1508 atomicExecute<ConstVecOperandU64, VecElemU64>(gpuDynInst);
1514 initAtomicAccess<VecElemU64>(gpuDynInst);
1520 atomicComplete<VecOperandU64, VecElemU64>(gpuDynInst);
1526 :
Inst_FLAT(iFmt,
"flat_atomic_cmpswap_x2")
1551 atomicExecute<ConstVecOperandU64, VecElemU64, 2>(gpuDynInst);
1557 initAtomicAccess<VecElemU64>(gpuDynInst);
1563 atomicComplete<VecOperandU64, VecElemU64>(gpuDynInst);
1592 atomicExecute<ConstVecOperandU64, VecElemU64>(gpuDynInst);
1598 initAtomicAccess<VecElemU64>(gpuDynInst);
1604 atomicComplete<VecOperandU64, VecElemU64>(gpuDynInst);
1633 atomicExecute<ConstVecOperandU64, VecElemU64>(gpuDynInst);
1639 initAtomicAccess<VecElemU64>(gpuDynInst);
1645 atomicComplete<VecOperandU64, VecElemU64>(gpuDynInst);
1651 :
Inst_FLAT(iFmt,
"flat_atomic_smin_x2")
1674 atomicExecute<ConstVecOperandI64, VecElemI64>(gpuDynInst);
1680 initAtomicAccess<VecElemI64>(gpuDynInst);
1686 atomicComplete<VecOperandI64, VecElemI64>(gpuDynInst);
1692 :
Inst_FLAT(iFmt,
"flat_atomic_umin_x2")
1715 atomicExecute<ConstVecOperandU64, VecElemU64>(gpuDynInst);
1721 initAtomicAccess<VecElemU64>(gpuDynInst);
1727 atomicComplete<VecOperandU64, VecElemU64>(gpuDynInst);
1733 :
Inst_FLAT(iFmt,
"flat_atomic_smax_x2")
1756 atomicExecute<ConstVecOperandI64, VecElemI64>(gpuDynInst);
1762 initAtomicAccess<VecElemI64>(gpuDynInst);
1768 atomicComplete<VecOperandI64, VecElemI64>(gpuDynInst);
1774 :
Inst_FLAT(iFmt,
"flat_atomic_umax_x2")
1797 atomicExecute<ConstVecOperandU64, VecElemU64>(gpuDynInst);
1803 initAtomicAccess<VecElemU64>(gpuDynInst);
1809 atomicComplete<VecOperandU64, VecElemU64>(gpuDynInst);
1838 atomicExecute<ConstVecOperandU64, VecElemU64>(gpuDynInst);
1844 initAtomicAccess<VecElemU64>(gpuDynInst);
1850 atomicComplete<VecOperandU64, VecElemU64>(gpuDynInst);
1879 atomicExecute<ConstVecOperandU64, VecElemU64>(gpuDynInst);
1885 initAtomicAccess<VecElemU64>(gpuDynInst);
1891 atomicComplete<VecOperandU64, VecElemU64>(gpuDynInst);
1920 atomicExecute<ConstVecOperandU64, VecElemU64>(gpuDynInst);
1926 initAtomicAccess<VecElemU64>(gpuDynInst);
1932 atomicComplete<VecOperandU64, VecElemU64>(gpuDynInst);
1961 atomicExecute<ConstVecOperandU64, VecElemU64>(gpuDynInst);
1967 initAtomicAccess<VecElemU64>(gpuDynInst);
1973 atomicComplete<VecOperandU64, VecElemU64>(gpuDynInst);
2003 atomicExecute<ConstVecOperandU64, VecElemU64>(gpuDynInst);
2009 initAtomicAccess<VecElemU64>(gpuDynInst);
2015 atomicComplete<VecOperandU64, VecElemU64>(gpuDynInst);
2021 :
Inst_FLAT(iFmt,
"flat_atomic_add_f32")
2039 atomicExecute<ConstVecOperandF32, VecElemF32>(gpuDynInst);
2045 initAtomicAccess<VecElemF32>(gpuDynInst);
2051 atomicComplete<VecOperandF32, VecElemF32>(gpuDynInst);
2057 :
Inst_FLAT(iFmt,
"flat_atomic_pk_add_f16")
2091 :
Inst_FLAT(iFmt,
"flat_atomic_add_f64")
2109 atomicExecute<ConstVecOperandF64, VecElemF64>(gpuDynInst);
2115 initAtomicAccess<VecElemF64>(gpuDynInst);
2121 atomicComplete<VecOperandF64, VecElemF64>(gpuDynInst);
2127 :
Inst_FLAT(iFmt,
"flat_atomic_min_f64")
2145 atomicExecute<ConstVecOperandF64, VecElemF64>(gpuDynInst);
2151 initAtomicAccess<VecElemF64>(gpuDynInst);
2157 atomicComplete<VecOperandF64, VecElemF64>(gpuDynInst);
2163 :
Inst_FLAT(iFmt,
"flat_atomic_max_f64")
2181 atomicExecute<ConstVecOperandF64, VecElemF64>(gpuDynInst);
2187 initAtomicAccess<VecElemF64>(gpuDynInst);
2193 atomicComplete<VecOperandF64, VecElemF64>(gpuDynInst);
bool isFlatScratch() const
void completeAcc(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_ADD_F32()
Inst_FLAT__FLAT_ATOMIC_ADD_F32(InFmt_FLAT *)
void execute(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_ADD_F64()
Inst_FLAT__FLAT_ATOMIC_ADD_F64(InFmt_FLAT *)
void completeAcc(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_ADD_X2()
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_ADD_X2(InFmt_FLAT *)
void initiateAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_ADD(InFmt_FLAT *)
~Inst_FLAT__FLAT_ATOMIC_ADD()
void execute(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_AND_X2()
void initiateAcc(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_AND_X2(InFmt_FLAT *)
void execute(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_AND()
Inst_FLAT__FLAT_ATOMIC_AND(InFmt_FLAT *)
void execute(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_CMPSWAP_X2()
void initiateAcc(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_CMPSWAP_X2(InFmt_FLAT *)
void completeAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_CMPSWAP()
void initiateAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_DEC_X2()
void completeAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_DEC_X2(InFmt_FLAT *)
void initiateAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_DEC(InFmt_FLAT *)
~Inst_FLAT__FLAT_ATOMIC_DEC()
void completeAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_INC_X2()
void completeAcc(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_INC_X2(InFmt_FLAT *)
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_INC(InFmt_FLAT *)
~Inst_FLAT__FLAT_ATOMIC_INC()
void execute(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_MAX_F64(InFmt_FLAT *)
~Inst_FLAT__FLAT_ATOMIC_MAX_F64()
void execute(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_MIN_F64()
Inst_FLAT__FLAT_ATOMIC_MIN_F64(InFmt_FLAT *)
void initiateAcc(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_OR_X2(InFmt_FLAT *)
~Inst_FLAT__FLAT_ATOMIC_OR_X2()
void completeAcc(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_OR()
void completeAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_OR(InFmt_FLAT *)
void initiateAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_PK_ADD_F16()
void initiateAcc(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_PK_ADD_F16(InFmt_FLAT *)
void initiateAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_SMAX_X2()
Inst_FLAT__FLAT_ATOMIC_SMAX_X2(InFmt_FLAT *)
void completeAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_SMAX(InFmt_FLAT *)
void completeAcc(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_SMAX()
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_SMIN_X2(InFmt_FLAT *)
void completeAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_SMIN_X2()
void initiateAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_SMIN(InFmt_FLAT *)
void completeAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_SMIN()
void initiateAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_SUB_X2(InFmt_FLAT *)
void initiateAcc(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_SUB_X2()
void completeAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_SUB()
Inst_FLAT__FLAT_ATOMIC_SUB(InFmt_FLAT *)
void initiateAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_SWAP_X2(InFmt_FLAT *)
~Inst_FLAT__FLAT_ATOMIC_SWAP_X2()
void execute(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_SWAP()
void initiateAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_SWAP(InFmt_FLAT *)
void completeAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_UMAX_X2()
void initiateAcc(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_UMAX_X2(InFmt_FLAT *)
~Inst_FLAT__FLAT_ATOMIC_UMAX()
void initiateAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_UMAX(InFmt_FLAT *)
void completeAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_UMIN_X2(InFmt_FLAT *)
void completeAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_UMIN_X2()
~Inst_FLAT__FLAT_ATOMIC_UMIN()
Inst_FLAT__FLAT_ATOMIC_UMIN(InFmt_FLAT *)
void initiateAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_XOR_X2(InFmt_FLAT *)
void completeAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_XOR_X2()
void initiateAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
~Inst_FLAT__FLAT_ATOMIC_XOR()
void completeAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_ATOMIC_XOR(InFmt_FLAT *)
void initiateAcc(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_LOAD_DWORDX2()
Inst_FLAT__FLAT_LOAD_DWORDX2(InFmt_FLAT *)
Inst_FLAT__FLAT_LOAD_DWORDX3(InFmt_FLAT *)
void initiateAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_LOAD_DWORDX3()
void execute(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_LOAD_DWORDX4()
Inst_FLAT__FLAT_LOAD_DWORDX4(InFmt_FLAT *)
void completeAcc(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_LOAD_DWORD(InFmt_FLAT *)
~Inst_FLAT__FLAT_LOAD_DWORD()
void execute(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_LOAD_SBYTE(InFmt_FLAT *)
~Inst_FLAT__FLAT_LOAD_SBYTE()
void completeAcc(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
Inst_FLAT__FLAT_LOAD_SSHORT(InFmt_FLAT *)
~Inst_FLAT__FLAT_LOAD_SSHORT()
Inst_FLAT__FLAT_LOAD_UBYTE(InFmt_FLAT *)
void initiateAcc(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
~Inst_FLAT__FLAT_LOAD_UBYTE()
void initiateAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_LOAD_USHORT()
void execute(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_LOAD_USHORT(InFmt_FLAT *)
void completeAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_STORE_BYTE(InFmt_FLAT *)
void execute(GPUDynInstPtr) override
~Inst_FLAT__FLAT_STORE_BYTE()
void initiateAcc(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
~Inst_FLAT__FLAT_STORE_DWORDX2()
Inst_FLAT__FLAT_STORE_DWORDX2(InFmt_FLAT *)
void completeAcc(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_STORE_DWORDX3(InFmt_FLAT *)
void execute(GPUDynInstPtr) override
~Inst_FLAT__FLAT_STORE_DWORDX3()
void completeAcc(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_STORE_DWORDX4()
Inst_FLAT__FLAT_STORE_DWORDX4(InFmt_FLAT *)
void execute(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_STORE_DWORD(InFmt_FLAT *)
void execute(GPUDynInstPtr) override
void completeAcc(GPUDynInstPtr) override
~Inst_FLAT__FLAT_STORE_DWORD()
~Inst_FLAT__FLAT_STORE_SHORT_D16_HI()
void completeAcc(GPUDynInstPtr) override
Inst_FLAT__FLAT_STORE_SHORT_D16_HI(InFmt_FLAT *)
void execute(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
void execute(GPUDynInstPtr) override
~Inst_FLAT__FLAT_STORE_SHORT()
Inst_FLAT__FLAT_STORE_SHORT(InFmt_FLAT *)
void completeAcc(GPUDynInstPtr) override
void initiateAcc(GPUDynInstPtr) override
void calcAddr(GPUDynInstPtr gpuDynInst, ScalarRegU32 vaddr, ScalarRegU32 saddr, ScalarRegI32 offset)
void issueRequestHelper(GPUDynInstPtr gpuDynInst)
void panicUnimplemented() const
void read() override
read from the vrf.
void write() override
write to the vrf.
void decVMemInstsIssued()
void decLGKMInstsIssued()
constexpr unsigned NumVecElemPerVecReg
Copyright (c) 2024 Arm Limited All rights reserved.
std::shared_ptr< GPUDynInst > GPUDynInstPtr