Go to the documentation of this file.
46 #include "debug/Config.hh"
47 #include "debug/Drain.hh"
48 #include "debug/Ruby.hh"
49 #include "mem/ruby/protocol/AccessPermission.hh"
63 m_controller(NULL), m_mandatory_q_ptr(NULL),
65 pioRequestPort(
csprintf(
"%s.pio-request-port",
name()), this),
66 pioResponsePort(
csprintf(
"%s.pio-response-port",
name()), this),
67 memRequestPort(
csprintf(
"%s.mem-request-port",
name()), this),
68 memResponsePort(
csprintf(
"%s-mem-response-port",
name()), this,
69 p.ruby_system->getAccessBackingStore(), -1,
71 gotAddrRanges(
p.port_interrupt_out_port_connection_count),
72 m_isCPUSequencer(
p.is_cpu_sequencer)
77 for (
size_t i = 0;
i <
p.port_in_ports_connection_count; ++
i) {
79 (
"%s.response_ports%d",
name(),
i),
this,
80 p.ruby_system->getAccessBackingStore(),
81 i,
p.no_retry_on_stall));
85 for (
size_t i = 0;
i <
p.port_interrupt_out_port_connection_count; ++
i) {
87 "%s.request_ports%d",
name(),
i),
this));
97 response_port->sendRangeChange();
103 if (if_name ==
"mem_request_port") {
105 }
else if (if_name ==
"pio_request_port") {
107 }
else if (if_name ==
"mem_response_port") {
109 }
else if (if_name ==
"pio_response_port") {
111 }
else if (if_name ==
"interrupt_out_port") {
115 panic(
"%s: unknown %s index (%d)\n", __func__, if_name, idx);
119 }
else if (if_name ==
"in_ports") {
123 panic(
"%s: unknown %s index (%d)\n", __func__, if_name, idx);
136 reqQueue(*_port, *this), snoopRespQueue(*_port, *this)
151 reqQueue(*_port, *this), snoopRespQueue(*_port, *this)
158 bool _access_backing_store,
PortID id,
159 bool _no_retry_on_stall)
161 access_backing_store(_access_backing_store),
162 no_retry_on_stall(_no_retry_on_stall)
190 assert(port != NULL);
212 for (
auto it =
l.begin(); it !=
l.end(); ++it) {
213 if (it->contains(pkt->
getAddr())) {
216 GEM5_VAR_USED
bool success =
223 panic(
"Should never reach here!\n");
232 panic(
"Ruby supports atomic accesses only in noncaching mode\n");
237 for (
auto it =
l.begin(); it !=
l.end(); ++it) {
238 if (it->contains(pkt->
getAddr())) {
243 panic(
"Could not find address in Ruby PIO address ranges!\n");
254 panic(
"RubyPort should never see request with the "
255 "cacheResponding flag set\n");
259 if (pkt->
req->isCacheMaintenance()) {
260 warn_once(
"Cache maintenance operations are not supported in Ruby.\n");
262 schedTimingResp(pkt,
curTick());
268 if (!isPhysMemAddress(pkt)) {
269 assert(!pkt->
req->isHTMCmd());
272 "pio address\n", pkt->
getAddr());
291 RequestStatus requestStatus = ruby_port->
makeRequest(pkt);
296 if (requestStatus == RequestStatus_Issued) {
308 "Request %s for address %#x did not issue because %s\n",
310 RequestStatus_to_string(requestStatus));
324 panic(
"Ruby supports atomic accesses only in noncaching mode\n");
330 if (!isPhysMemAddress(pkt)) {
333 "pio address\n", pkt->
getAddr());
353 pkt->
getAddr(), MachineType_Directory);
356 rs->m_abstract_controls[
id.getType()][
id.getNum()];
358 if (access_backing_store)
359 rs->getPhysMem()->access(pkt);
373 if (!no_retry_on_stall && !ruby_port->
onRetryList(
this)) {
388 if (!isPhysMemAddress(pkt)) {
390 assert(rp->pioRequestPort.isConnected());
391 rp->pioRequestPort.sendFunctional(pkt);
398 if (access_backing_store) {
403 rs->getPhysMem()->functionalAccess(pkt);
405 bool accessSucceeded =
false;
410 accessSucceeded =
rs->functionalRead(pkt);
412 accessSucceeded =
rs->functionalWrite(pkt);
420 fatal(
"Ruby functional %s failed for address %#x\n",
436 accessSucceeded ?
"successful":
"failed");
455 assert(port != NULL);
480 for (
auto i = curRetryList.begin();
i != curRetryList.end(); ++
i) {
482 "Sequencer may now be free. SendRetry to port %s\n",
484 (*i)->sendRetryReq();
495 DPRINTF(Drain,
"Drain count: %u\n", drainCount);
496 if (drainCount == 0) {
497 DPRINTF(Drain,
"RubyPort done draining, signaling drain done\n");
516 DPRINTF(Drain,
"RubyPort not drained\n");
530 bool accessPhysMem = access_backing_store;
534 if (pkt->
req->getExtraData() != 0) {
544 accessPhysMem =
false;
558 accessPhysMem =
false;
561 if (pkt->
req->isKernel()) {
562 accessPhysMem =
false;
563 needsResponse =
true;
577 rs->getPhysMem()->access(pkt);
579 panic(
"Packet is in neither device nor system memory!");
581 }
else if (needsResponse) {
591 schedTimingResp(pkt,
curTick());
607 ranges.splice(ranges.begin(),
610 for (GEM5_VAR_USED
const auto &
r : ranges)
621 for (
auto it = ranges.begin(); it != ranges.end(); ++it) {
622 if (it->contains(
addr)) {
646 auto request = std::make_shared<Request>(
656 if ((*p)->isSnooping()) {
658 (*p)->sendTimingSnoopReq(&pkt);
669 r.pioResponsePort.sendRangeChange();
678 if (port->trySatisfyFunctional(func_pkt)) {
Tick curTick()
The universal simulation clock.
#define fatal(...)
This implements a cprintf based fatal() function.
int16_t PortID
Port index/ID type, and a symbolic name for an invalid port id.
const std::string & cmdString() const
Return the string name of the cmd field (for debugging and tracing).
virtual Port & getPort(const std::string &if_name, PortID idx=InvalidPortID)
Get a port with a given name and index.
Port & getPort(const std::string &if_name, PortID idx=InvalidPortID) override
Get a port with a given name and index.
void setFunctionalResponseStatus(bool success)
const std::string name() const
Return port name (for DPRINTF).
virtual void descheduleDeadlockEvent()=0
DrainState drainState() const
Return the current drain state of an object.
void recvRangeChange()
Called to receive an address range change from the peer response port.
void recvFunctional(PacketPtr pkt)
Receive a functional request packet from the peer.
The QueuedRequestPort combines two queues, a request queue and a snoop response queue,...
void pushSenderState(SenderState *sender_state)
Push a new sender state to the packet and make the current sender state the predecessor of the new on...
RequestPtr req
A pointer to the original request.
void convertScToWrite()
It has been determined that the SC packet should successfully update memory.
std::vector< MemResponsePort * >::iterator CpuPortIter
Vector of M5 Ports attached to this Ruby port.
MemResponsePort memResponsePort
std::vector< MemResponsePort * > response_ports
Tick sendAtomic(PacketPtr pkt)
Send an atomic request packet, where the data is moved and the state is updated in zero time,...
PioResponsePort pioResponsePort
bool recvTimingResp(PacketPtr pkt)
Receive a timing response from the peer.
Tick recvAtomic(PacketPtr pkt)
Receive an atomic request packet from the peer.
bool bypassCaches() const
Should caches be bypassed?
bool cacheResponding() const
void ruby_hit_callback(PacketPtr pkt)
AddrRangeList getShadowRomRanges() const
static uint32_t getBlockSizeBytes()
std::vector< MemResponsePort * > retryList
std::string csprintf(const char *format, const Args &...args)
memory::AbstractMemory * getDeviceMemory(const PacketPtr &pkt) const
Return a pointer to the device memory.
bool isShadowRomAddress(Addr addr) const
Addr makeLineAddress(Addr addr)
void hitCallback(PacketPtr pkt)
MessageBuffer * m_mandatory_q_ptr
void convertLlToRead()
When ruby is in use, Ruby will monitor the cache line and the phys memory should treat LL ops as norm...
MemResponsePort(const std::string &_name, RubyPort *_port, bool _access_backing_store, PortID id, bool _no_retry_on_stall)
DrainState
Object drain/handover states.
bool isPhysMemAddress(PacketPtr pkt) const
void init() override
init() is called after all C++ SimObjects have been created and all ports are connected.
virtual RequestStatus makeRequest(PacketPtr pkt)=0
void ruby_eviction_callback(Addr address)
bool isMemAddr(Addr addr) const
Check if a physical address is within a range of a memory that is part of the global address map.
virtual std::string name() const
PioResponsePort(const std::string &_name, RubyPort *_port)
A queued port is a port that has an infinite queue for outgoing packets and thus decouples the module...
A Packet is used to encapsulate a transfer between two objects in the memory system (e....
bool htmTransactionFailedInCache() const
Returns whether or not this packet/request has returned from the cache hierarchy in a failed transact...
MachineID mapAddressToMachine(Addr addr, MachineType mtype) const
Map an address to the correct MachineID.
uint64_t Tick
Tick count type.
std::vector< PioRequestPort * > request_ports
virtual bool isDeadlockEventScheduled() const =0
Tick recvAtomic(PacketPtr pkt)
bool suppressFuncError() const
PioRequestPort(const std::string &_name, RubyPort *_port)
RubyPort(const Params &p)
Addr getOffset(Addr addr)
virtual int outstandingCount() const =0
MemRequestPort memRequestPort
bool isConnected() const
Is this port currently connected to a peer?
@ Drained
Buffers drained, ready for serialization/handover.
MemCmd cmd
The command field of the packet.
bool needsResponse() const
uint64_t Addr
Address type This will probably be moved somewhere else in the near future.
const std::string & name()
RubySystem * m_ruby_system
The ClockedObject class extends the SimObject with a clock and accessor functions to relate ticks to ...
SenderState * popSenderState()
Pop the top of the state stack and return a pointer to it.
void signalDrainDone() const
Signal that an object is drained.
bool FullSystem
The FullSystem variable can be used to determine the current mode of simulation.
bool isDeviceMemAddr(const PacketPtr &pkt) const
Similar to isMemAddr but for devices.
virtual int functionalWrite(Packet *func_pkt)
AddrRangeList getAddrRanges() const
Get a list of the non-overlapping address ranges the owner is responsible for.
@ funcRequestorId
This requestor id is used for functional requests that don't come from a particular device.
Tick recvAtomic(PacketPtr pkt)
Receive an atomic request packet from the peer.
Ports are used to interface objects to each other.
AbstractController * m_controller
void makeResponse()
Take a request packet and modify it in place to be suitable for returning as a response to that reque...
void addToRetryList(MemResponsePort *port)
void schedTimingReq(PacketPtr pkt, Tick when)
Schedule the sending of a timing request.
bool onRetryList(MemResponsePort *port)
Cycles ticksToCycles(Tick t) const
virtual MessageBuffer * getMandatoryQueue() const =0
void schedTimingResp(PacketPtr pkt, Tick when)
Schedule the sending of a timing response.
void access(PacketPtr pkt)
Perform an untimed memory access and update all the state (e.g.
bool recvTimingReq(PacketPtr pkt)
Receive a timing request from the peer.
Reference material can be found at the JEDEC website: UFS standard http://www.jedec....
DrainState drain() override
Draining is the process of clearing out the states of SimObjects.These are the SimObjects that are pa...
MemRequestPort(const std::string &_name, RubyPort *_port)
@ Draining
Draining buffers pending serialization/handover.
bool recvTimingReq(PacketPtr pkt)
Receive a timing request from the peer.
#define panic(...)
This implements a cprintf based panic() function.
RubyTester::SenderState SenderState
bool recvTimingResp(PacketPtr pkt)
Receive a timing response from the peer.
PioRequestPort pioRequestPort
Generated on Wed Jul 28 2021 12:10:29 for gem5 by doxygen 1.8.17