| /* |
| * Copyright (c) 2010-2015 Advanced Micro Devices, Inc. |
| * All rights reserved. |
| * |
| * For use for simulation and test purposes only |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions are met: |
| * |
| * 1. Redistributions of source code must retain the above copyright notice, |
| * this list of conditions and the following disclaimer. |
| * |
| * 2. Redistributions in binary form must reproduce the above copyright notice, |
| * this list of conditions and the following disclaimer in the documentation |
| * and/or other materials provided with the distribution. |
| * |
| * 3. Neither the name of the copyright holder nor the names of its |
| * contributors may be used to endorse or promote products derived from this |
| * software without specific prior written permission. |
| * |
| * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
| * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
| * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE |
| * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
| * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
| * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
| * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
| * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
| * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
| * POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| machine(MachineType:RegionBuffer, "Region Buffer for AMD_Base-like protocol") |
| : CacheMemory *cacheMemory; // stores only region addresses. Must set block size same as below |
| bool isOnCPU; |
| int blocksPerRegion := 64; // 4k regions |
| Cycles toDirLatency := 5; // Latency to fwd requests to directory |
| Cycles toRegionDirLatency := 5; // Latency for requests and acks to directory |
| Cycles nextEvictLatency := 1; // latency added between each block while evicting region |
| bool noTCCdir := "False"; |
| int TCC_select_num_bits := 1; |
| |
| // From the Cores |
| MessageBuffer * requestFromCore, network="From", virtual_network="0", vnet_type="request"; |
| MessageBuffer * responseFromCore, network="From", virtual_network="2", vnet_type="response"; |
| |
| // Requests to the cores or directory |
| MessageBuffer * requestToNetwork, network="To", virtual_network="0", vnet_type="request"; |
| |
| // From Region-Dir |
| MessageBuffer * notifyFromRegionDir, network="From", virtual_network="7", vnet_type="request"; |
| MessageBuffer * probeFromRegionDir, network="From", virtual_network="8", vnet_type="request"; |
| |
| // From the directory |
| MessageBuffer * unblockFromDir, network="From", virtual_network="4", vnet_type="unblock"; |
| |
| // To the region-Dir |
| MessageBuffer * responseToRegDir, network="To", virtual_network="2", vnet_type="response"; |
| |
| MessageBuffer * triggerQueue; |
| { |
| |
| // States |
| state_declaration(State, desc="Region states", default="RegionBuffer_State_NP") { |
| NP, AccessPermission:Invalid, desc="Not present in region directory"; |
| P, AccessPermission:Invalid, desc="Region is private to the cache"; |
| S, AccessPermission:Invalid, desc="Region is possibly shared with others"; |
| |
| NP_PS, AccessPermission:Invalid, desc="Intermediate state waiting for notify from r-dir"; |
| S_P, AccessPermission:Invalid, desc="Intermediate state while upgrading region"; |
| |
| P_NP, AccessPermission:Invalid, desc="Intermediate state while evicting all lines in region"; |
| P_S, AccessPermission:Invalid, desc="Intermediate state while downgrading all lines in region"; |
| |
| S_NP_PS, AccessPermission:Invalid, desc="Got an inv in S_P, waiting for all inv acks, then going to since the write is already out there NP_PS"; |
| P_NP_NP, AccessPermission:Invalid, desc="Evicting region on repl, then got an inv. Need to re-evict"; |
| |
| P_NP_O, AccessPermission:Invalid, desc="Waiting for all outstanding requests"; |
| P_S_O, AccessPermission:Invalid, desc="Waiting for all outstanding requests"; |
| S_O, AccessPermission:Invalid, desc="Waiting for all outstanding requests"; |
| S_NP_PS_O, AccessPermission:Invalid, desc="Waiting for all outstanding requests"; |
| |
| SS_P, AccessPermission:Invalid, desc="Waiting for CPU write that we know is there"; |
| |
| P_NP_W, AccessPermission:Invalid, desc="Waiting for writeback ack"; |
| |
| NP_W, AccessPermission:Invalid, desc="Got a done ack before request, waiting for that victim"; |
| } |
| |
| enumeration(Event, desc="Region directory events") { |
| CPURead, desc="Access from CPU core"; |
| CPUWrite, desc="Access from CPU core"; |
| CPUWriteback, desc="Writeback request from CPU core"; |
| |
| ReplRegion, desc="Start a replace on a region"; |
| |
| PrivateNotify, desc="Update entry to private state"; |
| SharedNotify, desc="Update entry to shared state"; |
| WbNotify, desc="Writeback notification received"; |
| InvRegion, desc="Start invalidating a region"; |
| DowngradeRegion,desc="Start invalidating a region"; |
| |
| InvAck, desc="Ack from core"; |
| |
| DoneAck, desc="Ack from core that request has finished"; |
| AllOutstanding, desc="All outstanding requests have now finished"; |
| |
| Evict, desc="Loopback to evict each block"; |
| LastAck_PrbResp, desc="Done eviciting all the blocks, got the last ack from core, now respond to region dir"; |
| LastAck_CleanWb, desc="Done eviciting all the blocks, got the last ack from core, now start clean writeback (note the dir has already been updated)"; |
| |
| StallAccess, desc="Wait for the done ack on the address before proceeding"; |
| StallDoneAck, desc="Wait for the access on the address before proceeding"; |
| |
| StaleRequest, desc="Got a stale victim from the cache, fwd it without incrementing outstanding"; |
| } |
| |
| enumeration(RequestType, desc="To communicate stats from transitions to recordStats") { |
| TagArrayRead, desc="Read the data array"; |
| TagArrayWrite, desc="Write the data array"; |
| } |
| |
| structure(BoolVec, external="yes") { |
| bool at(int); |
| void resize(int); |
| void clear(); |
| int size(); |
| } |
| |
| structure(Entry, desc="Region entry", interface="AbstractCacheEntry") { |
| Addr addr, desc="Base address of this region"; |
| State RegionState, desc="Region state"; |
| DataBlock DataBlk, desc="Data for the block (always empty in region buffer)"; |
| BoolVec ValidBlocks, desc="A vector to keep track of valid blocks"; |
| int NumValidBlocks, desc="Number of trues in ValidBlocks to avoid iterating"; |
| BoolVec UsedBlocks, desc="A vector to keep track of blocks ever valid"; |
| bool dirty, desc="Dirty as best known by the region buffer"; |
| // This is needed so we don't ack an invalidate until all requests are ordered |
| int NumOutstandingReqs, desc="Total outstanding private/shared requests"; |
| BoolVec OutstandingReqs, desc="Blocks that have outstanding private/shared requests"; |
| bool MustDowngrade, desc="Set when we got a downgrade before the shd or pvt permissions"; |
| Cycles ProbeRequestTime, default="Cycles(0)", desc="Time region dir started the probe"; |
| Cycles InitialRequestTime, default="Cycles(0)", desc="Time message was sent to region dir"; |
| bool MsgSentToDir, desc="True if the current request required a message to the dir"; |
| bool clearOnDone, default="false", desc="clear valid bit when request completes"; |
| Addr clearOnDoneAddr, desc="clear valid bit when request completes"; |
| } |
| |
| structure(TBE, desc="...") { |
| State TBEState, desc="Transient state"; |
| //int NumValidBlocks, desc="Number of blocks valid so we don't have to count a BoolVec"; |
| BoolVec ValidBlocks, desc="A vector to keep track of valid blocks"; |
| bool AllAcksReceived, desc="Got all necessary acks from dir"; |
| bool DoneEvicting, desc="Done iterating through blocks checking for valids"; |
| BoolVec AcksReceived, desc="Received acks for theses blocks\n"; |
| bool SendAck, desc="If true, send an ack to the r-dir at end of inv"; |
| ProbeRequestType MsgType, desc="Type of message to send while 'evicting' "; |
| int NumOutstandingReqs, desc="Total outstanding private/shared requests"; |
| BoolVec OutstandingReqs, desc="Blocks that have outstanding private/shared requests"; |
| MachineID Requestor, desc="Requestor for three hop transactions"; |
| bool DemandRequest, default="false", desc="Associated with a demand request"; |
| Addr DemandAddress, desc="Address for the demand request"; |
| bool DoneAckReceived, default="false", desc="True if the done ack arrived before the message"; |
| Addr DoneAckAddr, desc="Address of the done ack received early"; |
| int OutstandingThreshold, desc="Number of outstanding requests to trigger AllOutstanding on"; |
| |
| ProbeRequestType NewMsgType, desc="Type of message to send while 'evicting' "; |
| MachineID NewRequestor, desc="Requestor for three hop transactions"; |
| bool NewDemandRequest, default="false", desc="Associated with a demand request"; |
| Addr NewDemandAddress, desc="Address for the demand request"; |
| bool dirty, desc="dirty"; |
| bool AllOutstandingTriggered, default="false", desc="bit for only one all outstanding"; |
| int OutstandingAcks, default="0", desc="number of acks to wait for"; |
| } |
| |
| structure(TBETable, external="yes") { |
| TBE lookup(Addr); |
| void allocate(Addr); |
| void deallocate(Addr); |
| bool isPresent(Addr); |
| } |
| |
| // Stores only region addresses |
| TBETable TBEs, template="<RegionBuffer_TBE>", constructor="m_number_of_TBEs"; |
| int TCC_select_low_bit, default="RubySystem::getBlockSizeBits()"; |
| |
| Tick clockEdge(); |
| Tick cyclesToTicks(Cycles c); |
| |
| void set_cache_entry(AbstractCacheEntry b); |
| void unset_cache_entry(); |
| void set_tbe(TBE b); |
| void unset_tbe(); |
| void wakeUpAllBuffers(); |
| void wakeUpBuffers(Addr a); |
| Cycles curCycle(); |
| MachineID mapAddressToMachine(Addr addr, MachineType mtype); |
| |
| int blockBits, default="RubySystem::getBlockSizeBits()"; |
| int blockBytes, default="RubySystem::getBlockSizeBytes()"; |
| int regionBits, default="log2(m_blocksPerRegion)"; |
| |
| // Functions |
| |
| int getRegionOffset(Addr addr) { |
| if (blocksPerRegion > 1) { |
| Addr offset := bitSelect(addr, blockBits, regionBits+blockBits-1); |
| int ret := addressToInt(offset); |
| assert(ret < blocksPerRegion); |
| return ret; |
| } else { |
| return 0; |
| } |
| } |
| |
| Addr getRegionBase(Addr addr) { |
| return maskLowOrderBits(addr, blockBits+regionBits); |
| } |
| |
| Addr getNextBlock(Addr addr) { |
| Addr a := addr; |
| return makeNextStrideAddress(a, 1); |
| } |
| |
| MachineID getPeer(MachineID mach, Addr address) { |
| if (isOnCPU) { |
| return createMachineID(MachineType:CorePair, intToID(0)); |
| } else if (noTCCdir) { |
| return mapAddressToRange(address,MachineType:TCC, |
| TCC_select_low_bit, TCC_select_num_bits); |
| } else { |
| return createMachineID(MachineType:TCCdir, intToID(0)); |
| } |
| } |
| |
| bool isOutstanding(TBE tbe, Entry cache_entry, Addr addr) { |
| if (is_valid(tbe) && tbe.OutstandingReqs.size() > 0) { |
| DPRINTF(RubySlicc, " outstanding tbe reqs %s %s %d %d\n", |
| tbe.OutstandingReqs, addr, getRegionOffset(addr), |
| tbe.OutstandingReqs.at(getRegionOffset(addr))); |
| return tbe.OutstandingReqs.at(getRegionOffset(addr)); |
| } else if (is_valid(cache_entry)) { |
| DPRINTF(RubySlicc, " outstanding cache reqs %s %s %d %d\n", |
| cache_entry.OutstandingReqs, addr, getRegionOffset(addr), |
| cache_entry.OutstandingReqs.at(getRegionOffset(addr))); |
| return cache_entry.OutstandingReqs.at(getRegionOffset(addr)); |
| } else { |
| return false; |
| } |
| } |
| |
| bool isOnGPU() { |
| if (isOnCPU) { |
| return false; |
| } |
| return true; |
| } |
| |
| bool isRead(CoherenceRequestType type) { |
| return (type == CoherenceRequestType:RdBlk || type == CoherenceRequestType:RdBlkS || |
| type == CoherenceRequestType:VicClean); |
| } |
| |
| bool presentOrAvail(Addr addr) { |
| return cacheMemory.isTagPresent(getRegionBase(addr)) || cacheMemory.cacheAvail(getRegionBase(addr)); |
| } |
| |
| // Returns a region entry! |
| Entry getCacheEntry(Addr addr), return_by_pointer="yes" { |
| return static_cast(Entry, "pointer", cacheMemory.lookup(getRegionBase(addr))); |
| } |
| |
| TBE getTBE(Addr addr), return_by_pointer="yes" { |
| return TBEs.lookup(getRegionBase(addr)); |
| } |
| |
| DataBlock getDataBlock(Addr addr), return_by_ref="yes" { |
| return getCacheEntry(getRegionBase(addr)).DataBlk; |
| } |
| |
| State getState(TBE tbe, Entry cache_entry, Addr addr) { |
| if (is_valid(tbe)) { |
| return tbe.TBEState; |
| } else if (is_valid(cache_entry)) { |
| return cache_entry.RegionState; |
| } |
| return State:NP; |
| } |
| |
| void setState(TBE tbe, Entry cache_entry, Addr addr, State state) { |
| if (is_valid(tbe)) { |
| tbe.TBEState := state; |
| } |
| if (is_valid(cache_entry)) { |
| cache_entry.RegionState := state; |
| } |
| } |
| |
| AccessPermission getAccessPermission(Addr addr) { |
| TBE tbe := getTBE(addr); |
| if(is_valid(tbe)) { |
| return RegionBuffer_State_to_permission(tbe.TBEState); |
| } |
| Entry cache_entry := getCacheEntry(addr); |
| if(is_valid(cache_entry)) { |
| return RegionBuffer_State_to_permission(cache_entry.RegionState); |
| } |
| return AccessPermission:NotPresent; |
| } |
| |
| void functionalRead(Addr addr, Packet *pkt) { |
| functionalMemoryRead(pkt); |
| } |
| |
| int functionalWrite(Addr addr, Packet *pkt) { |
| if (functionalMemoryWrite(pkt)) { |
| return 1; |
| } else { |
| return 0; |
| } |
| } |
| |
| void setAccessPermission(Entry cache_entry, Addr addr, State state) { |
| if (is_valid(cache_entry)) { |
| cache_entry.changePermission(RegionBuffer_State_to_permission(state)); |
| } |
| } |
| |
| void recordRequestType(RequestType stat, Addr addr) { |
| if (stat == RequestType:TagArrayRead) { |
| cacheMemory.recordRequestType(CacheRequestType:TagArrayRead, addr); |
| } else if (stat == RequestType:TagArrayWrite) { |
| cacheMemory.recordRequestType(CacheRequestType:TagArrayWrite, addr); |
| } |
| } |
| |
| bool checkResourceAvailable(RequestType request_type, Addr addr) { |
| if (request_type == RequestType:TagArrayRead) { |
| return cacheMemory.checkResourceAvailable(CacheResourceType:TagArray, addr); |
| } else if (request_type == RequestType:TagArrayWrite) { |
| return cacheMemory.checkResourceAvailable(CacheResourceType:TagArray, addr); |
| } else { |
| error("Invalid RequestType type in checkResourceAvailable"); |
| return true; |
| } |
| } |
| |
| out_port(triggerQueue_out, TriggerMsg, triggerQueue); |
| |
| // Overloaded outgoing request nework for both probes to cores and reqeusts |
| // to the directory. |
| // Fix Me: These forwarded requests need to be on a separate virtual channel |
| // to avoid deadlock! |
| out_port(requestNetwork_out, CPURequestMsg, requestToNetwork); |
| out_port(probeNetwork_out, NBProbeRequestMsg, requestToNetwork); |
| |
| out_port(responseNetwork_out, ResponseMsg, responseToRegDir); |
| |
| in_port(triggerQueue_in, TriggerMsg, triggerQueue, rank=4) { |
| if (triggerQueue_in.isReady(clockEdge())) { |
| peek(triggerQueue_in, TriggerMsg) { |
| Entry cache_entry := getCacheEntry(in_msg.addr); |
| TBE tbe := getTBE(in_msg.addr); |
| DPRINTF(RubySlicc, "trigger msg: %s (%s)\n", in_msg, getRegionBase(in_msg.addr)); |
| assert(is_valid(tbe)); |
| if (in_msg.Type == TriggerType:AcksComplete) { |
| if (tbe.SendAck) { |
| trigger(Event:LastAck_PrbResp, in_msg.addr, cache_entry, tbe); |
| } else { |
| trigger(Event:LastAck_CleanWb, in_msg.addr, cache_entry, tbe); |
| } |
| } else if (in_msg.Type == TriggerType:AllOutstanding) { |
| trigger(Event:AllOutstanding, in_msg.addr, cache_entry, tbe); |
| } else { |
| assert(in_msg.Type == TriggerType:InvNext); |
| trigger(Event:Evict, in_msg.addr, cache_entry, tbe); |
| } |
| } |
| } |
| } |
| |
| in_port(unblockNetwork_in, UnblockMsg, unblockFromDir, rank=3) { |
| if (unblockNetwork_in.isReady(clockEdge())) { |
| peek(unblockNetwork_in, UnblockMsg) { |
| TBE tbe := getTBE(in_msg.addr); |
| Entry cache_entry := getCacheEntry(in_msg.addr); |
| if (in_msg.DoneAck) { |
| if (isOutstanding(tbe, cache_entry, in_msg.addr)) { |
| trigger(Event:DoneAck, in_msg.addr, cache_entry, tbe); |
| } else { |
| trigger(Event:StallDoneAck, in_msg.addr, cache_entry, tbe); |
| } |
| } else { |
| assert(is_valid(tbe)); |
| trigger(Event:InvAck, in_msg.addr, cache_entry, tbe); |
| } |
| } |
| } |
| } |
| |
| in_port(probeNetwork_in, NBProbeRequestMsg, probeFromRegionDir, rank=2) { |
| if (probeNetwork_in.isReady(clockEdge())) { |
| peek(probeNetwork_in, NBProbeRequestMsg) { |
| TBE tbe := getTBE(in_msg.addr); |
| Entry cache_entry := getCacheEntry(in_msg.addr); |
| assert(getRegionBase(in_msg.addr) == in_msg.addr); |
| if (in_msg.Type == ProbeRequestType:PrbInv) { |
| trigger(Event:InvRegion, in_msg.addr, cache_entry, tbe); |
| } else if (in_msg.Type == ProbeRequestType:PrbDowngrade) { |
| trigger(Event:DowngradeRegion, in_msg.addr, cache_entry, tbe); |
| } else { |
| error("Unknown probe message\n"); |
| } |
| } |
| } |
| } |
| |
| in_port(notifyNetwork_in, CPURequestMsg, notifyFromRegionDir, rank=1) { |
| if (notifyNetwork_in.isReady(clockEdge())) { |
| peek(notifyNetwork_in, CPURequestMsg) { |
| TBE tbe := getTBE(in_msg.addr); |
| Entry cache_entry := getCacheEntry(in_msg.addr); |
| //Fix Me...add back in: assert(is_valid(cache_entry)); |
| if (in_msg.Type == CoherenceRequestType:WbNotify) { |
| trigger(Event:WbNotify, in_msg.addr, cache_entry, tbe); |
| } else if (in_msg.Type == CoherenceRequestType:SharedNotify) { |
| trigger(Event:SharedNotify, in_msg.addr, cache_entry, tbe); |
| } else if (in_msg.Type == CoherenceRequestType:PrivateNotify) { |
| trigger(Event:PrivateNotify, in_msg.addr, cache_entry, tbe); |
| } else { |
| error("Unknown notify message\n"); |
| } |
| } |
| } |
| } |
| |
| // In from cores |
| // NOTE: We get the cache / TBE entry based on the region address, |
| // but pass the block address to the actions |
| in_port(requestNetwork_in, CPURequestMsg, requestFromCore, rank=0) { |
| if (requestNetwork_in.isReady(clockEdge())) { |
| peek(requestNetwork_in, CPURequestMsg) { |
| TBE tbe := getTBE(in_msg.addr); |
| Entry cache_entry := getCacheEntry(in_msg.addr); |
| if (is_valid(tbe) && tbe.DoneAckReceived && tbe.DoneAckAddr == in_msg.addr) { |
| DPRINTF(RubySlicc, "Stale/Stall request %s\n", in_msg.Type); |
| if (in_msg.Type == CoherenceRequestType:VicDirty || in_msg.Type == CoherenceRequestType:VicClean ) |
| { |
| trigger(Event:StaleRequest, in_msg.addr, cache_entry, tbe); |
| } else { |
| trigger(Event:StallAccess, in_msg.addr, cache_entry, tbe); |
| } |
| } else if (isOutstanding(tbe, cache_entry, in_msg.addr)) { |
| DPRINTF(RubySlicc, "Stall outstanding request %s\n", in_msg.Type); |
| trigger(Event:StallAccess, in_msg.addr, cache_entry, tbe); |
| } else { |
| if (presentOrAvail(in_msg.addr)) { |
| if (in_msg.Type == CoherenceRequestType:RdBlkM ) { |
| trigger(Event:CPUWrite, in_msg.addr, cache_entry, tbe); |
| } else if (in_msg.Type == CoherenceRequestType:WriteThrough ) { |
| trigger(Event:CPUWrite, in_msg.addr, cache_entry, tbe); |
| } else if (in_msg.Type == CoherenceRequestType:Atomic ) { |
| trigger(Event:CPUWrite, in_msg.addr, cache_entry, tbe); |
| } else { |
| if (in_msg.Type == CoherenceRequestType:VicDirty || |
| in_msg.Type == CoherenceRequestType:VicClean) { |
| trigger(Event:CPUWriteback, in_msg.addr, cache_entry, tbe); |
| } else { |
| trigger(Event:CPURead, in_msg.addr, cache_entry, tbe); |
| } |
| } |
| } else { |
| Addr victim := cacheMemory.cacheProbe(getRegionBase(in_msg.addr)); |
| TBE victim_tbe := getTBE(victim); |
| Entry victim_entry := getCacheEntry(victim); |
| DPRINTF(RubySlicc, "Replacing region %s for %s(%s)\n", victim, in_msg.addr, getRegionBase(in_msg.addr)); |
| trigger(Event:ReplRegion, victim, victim_entry, victim_tbe); |
| } |
| } |
| } |
| } |
| } |
| |
| // Actions |
| action(f_fwdReqToDir, "f", desc="Forward CPU request to directory") { |
| peek(requestNetwork_in, CPURequestMsg) { |
| enqueue(requestNetwork_out, CPURequestMsg, toDirLatency) { |
| out_msg.addr := in_msg.addr; |
| out_msg.Type := in_msg.Type; |
| out_msg.DataBlk := in_msg.DataBlk; |
| out_msg.Dirty := in_msg.Dirty; |
| out_msg.Requestor := in_msg.Requestor; |
| out_msg.WTRequestor := in_msg.WTRequestor; |
| out_msg.Destination.add(mapAddressToMachine(in_msg.addr, MachineType:Directory)); |
| out_msg.Shared := in_msg.Shared; |
| out_msg.MessageSize := in_msg.MessageSize; |
| out_msg.Private := true; |
| out_msg.InitialRequestTime := curCycle(); |
| out_msg.ProbeRequestStartTime := curCycle(); |
| if (getState(tbe, cache_entry, address) == State:S) { |
| out_msg.ForceShared := true; |
| } |
| DPRINTF(RubySlicc, "Fwd: %s\n", out_msg); |
| //assert(getState(tbe, cache_entry, address) == State:P || getState(tbe, cache_entry, address) == State:S); |
| if (getState(tbe, cache_entry, address) == State:NP_W) { |
| APPEND_TRANSITION_COMMENT(" fwding stale request: "); |
| APPEND_TRANSITION_COMMENT(out_msg.Type); |
| } |
| } |
| } |
| } |
| |
| action(u_updateRegionEntry, "u", desc="Update the entry for profiling") { |
| peek(requestNetwork_in, CPURequestMsg) { |
| if (is_valid(cache_entry)) { |
| if (in_msg.CtoDSinked == false) { |
| APPEND_TRANSITION_COMMENT(" incr outstanding "); |
| cache_entry.NumOutstandingReqs := 1 + cache_entry.NumOutstandingReqs; |
| assert(cache_entry.OutstandingReqs.at(getRegionOffset(address)) == false); |
| cache_entry.OutstandingReqs.at(getRegionOffset(address)) := true; |
| assert(cache_entry.NumOutstandingReqs == countBoolVec(cache_entry.OutstandingReqs)); |
| } else { |
| APPEND_TRANSITION_COMMENT(" NOT incr outstanding "); |
| assert(in_msg.Type == CoherenceRequestType:RdBlkM || in_msg.Type == CoherenceRequestType:RdBlkS); |
| } |
| APPEND_TRANSITION_COMMENT(cache_entry.NumOutstandingReqs); |
| if (in_msg.Type == CoherenceRequestType:RdBlkM || in_msg.Type == CoherenceRequestType:Atomic || |
| in_msg.Type == CoherenceRequestType:WriteThrough ) |
| { |
| cache_entry.dirty := true; |
| } |
| if (in_msg.Type == CoherenceRequestType:VicDirty || |
| in_msg.Type == CoherenceRequestType:VicClean) { |
| DPRINTF(RubySlicc, "Got %s for addr %s\n", in_msg.Type, address); |
| //assert(cache_entry.ValidBlocks.at(getRegionOffset(address))); |
| // can in fact be inv if core got an inv after a vicclean before it got here |
| if (cache_entry.ValidBlocks.at(getRegionOffset(address))) { |
| cache_entry.clearOnDone := true; |
| cache_entry.clearOnDoneAddr := address; |
| //cache_entry.ValidBlocks.at(getRegionOffset(address)) := false; |
| //cache_entry.NumValidBlocks := cache_entry.NumValidBlocks - 1; |
| } |
| } else { |
| if (cache_entry.ValidBlocks.at(getRegionOffset(address)) == false) { |
| cache_entry.NumValidBlocks := cache_entry.NumValidBlocks + 1; |
| } |
| DPRINTF(RubySlicc, "before valid addr %s bits %s\n", |
| in_msg.Type, address, cache_entry.ValidBlocks); |
| cache_entry.ValidBlocks.at(getRegionOffset(address)) := true; |
| DPRINTF(RubySlicc, "after valid addr %s bits %s\n", |
| in_msg.Type, address, cache_entry.ValidBlocks); |
| cache_entry.UsedBlocks.at(getRegionOffset(address)) := true; |
| } |
| assert(cache_entry.NumValidBlocks <= blocksPerRegion); |
| assert(cache_entry.NumValidBlocks >= 0); |
| APPEND_TRANSITION_COMMENT(" valid blocks "); |
| APPEND_TRANSITION_COMMENT(cache_entry.ValidBlocks); |
| } else { |
| error("This shouldn't happen anymore I think"); |
| //tbe.ValidBlocks.at(getRegionOffest(address)) := true; |
| assert(getState(tbe, cache_entry, address) == State:P_NP); |
| } |
| } |
| } |
| |
| action(uw_updatePossibleWriteback, "uw", desc="writeback request complete") { |
| peek(unblockNetwork_in, UnblockMsg) { |
| if (is_valid(cache_entry) && in_msg.validToInvalid && |
| cache_entry.clearOnDone && cache_entry.clearOnDoneAddr == address) { |
| DPRINTF(RubySlicc, "I have no idea what is going on here\n"); |
| cache_entry.ValidBlocks.at(getRegionOffset(address)) := false; |
| cache_entry.NumValidBlocks := cache_entry.NumValidBlocks - 1; |
| cache_entry.clearOnDone := false; |
| } |
| } |
| } |
| |
| |
| action(rp_requestPrivate, "rp", desc="Send private request r-dir") { |
| peek(requestNetwork_in, CPURequestMsg) { |
| // No need to send acks on replacements |
| assert(is_invalid(tbe)); |
| enqueue(requestNetwork_out, CPURequestMsg, toRegionDirLatency) { |
| out_msg.addr := address; // use the actual address so the demand request can be fulfilled |
| out_msg.DemandAddress := address; |
| out_msg.Type := CoherenceRequestType:PrivateRequest; |
| out_msg.OriginalType := in_msg.Type; |
| out_msg.Requestor := machineID; |
| out_msg.WTRequestor := in_msg.WTRequestor; |
| out_msg.InitialRequestTime := curCycle(); |
| // will this always be ok? probably not for multisocket |
| out_msg.Destination.add(mapAddressToMachine(address, MachineType:RegionDir)); |
| out_msg.MessageSize := MessageSizeType:Request_Control; |
| DPRINTF(RubySlicc, "Private request %s\n", out_msg); |
| } |
| cache_entry.ProbeRequestTime := curCycle(); |
| cache_entry.MsgSentToDir := true; |
| APPEND_TRANSITION_COMMENT(getRegionBase(address)); |
| } |
| } |
| |
| action(ru_requestUpgrade, "ru", desc="Send upgrade request r-dir") { |
| peek(requestNetwork_in, CPURequestMsg) { |
| // No need to send acks on replacements |
| assert(is_invalid(tbe)); |
| enqueue(requestNetwork_out, CPURequestMsg, toRegionDirLatency) { |
| out_msg.addr := address; // use the actual address so the demand request can be fulfilled |
| out_msg.Type := CoherenceRequestType:UpgradeRequest; |
| out_msg.OriginalType := in_msg.Type; |
| out_msg.Requestor := machineID; |
| out_msg.WTRequestor := in_msg.WTRequestor; |
| out_msg.InitialRequestTime := curCycle(); |
| // will this always be ok? probably not for multisocket |
| out_msg.Destination.add(mapAddressToMachine(address, MachineType:RegionDir)); |
| out_msg.MessageSize := MessageSizeType:Request_Control; |
| } |
| cache_entry.ProbeRequestTime := curCycle(); |
| cache_entry.MsgSentToDir := true; |
| APPEND_TRANSITION_COMMENT(getRegionBase(address)); |
| } |
| } |
| |
| action(rw_requestWriteback, "rq", desc="Send writeback request") { |
| // No need to send acks on replacements |
| enqueue(requestNetwork_out, CPURequestMsg, toRegionDirLatency) { |
| out_msg.addr := getRegionBase(address); // use the actual address so the demand request can be fulfilled |
| out_msg.Type := CoherenceRequestType:CleanWbRequest; |
| out_msg.Requestor := machineID; |
| // will this always be ok? probably not for multisocket |
| out_msg.Destination.add(mapAddressToMachine(address, MachineType:RegionDir)); |
| out_msg.MessageSize := MessageSizeType:Request_Control; |
| out_msg.Dirty := tbe.dirty; |
| APPEND_TRANSITION_COMMENT(getRegionBase(address)); |
| } |
| } |
| |
| action(rs_requestShared, "rs", desc="Send shared request r-dir") { |
| peek(requestNetwork_in, CPURequestMsg) { |
| // No need to send acks on replacements |
| assert(is_invalid(tbe)); |
| enqueue(requestNetwork_out, CPURequestMsg, toRegionDirLatency) { |
| out_msg.addr := address; // use the actual address so the demand request can be fulfilled |
| out_msg.Type := CoherenceRequestType:SharedRequest; |
| out_msg.OriginalType := in_msg.Type; |
| out_msg.Requestor := machineID; |
| out_msg.WTRequestor := in_msg.WTRequestor; |
| out_msg.InitialRequestTime := curCycle(); |
| // will this always be ok? probably not for multisocket |
| out_msg.Destination.add(mapAddressToMachine(address, MachineType:RegionDir)); |
| out_msg.MessageSize := MessageSizeType:Request_Control; |
| } |
| cache_entry.ProbeRequestTime := curCycle(); |
| cache_entry.MsgSentToDir := true; |
| APPEND_TRANSITION_COMMENT(getRegionBase(address)); |
| } |
| } |
| |
| action(ai_ackRegionInv, "ai", desc="Send ack to r-dir on region inv if tbe says so") { |
| // No need to send acks on replacements |
| assert(is_valid(tbe)); |
| enqueue(responseNetwork_out, ResponseMsg, toRegionDirLatency) { |
| out_msg.addr := getRegionBase(address); |
| out_msg.Type := CoherenceResponseType:CPUPrbResp; |
| out_msg.Sender := machineID; |
| // will this always be ok? probably not for multisocket |
| out_msg.Destination.add(mapAddressToMachine(address, MachineType:RegionDir)); |
| out_msg.MessageSize := MessageSizeType:Response_Control; |
| } |
| } |
| |
| action(ad_ackDircetory, "ad", desc="send probe response to directory") { |
| if (noTCCdir && tbe.MsgType == ProbeRequestType:PrbDowngrade && isOnGPU()) { //VIPER tcc doesnt understand PrbShrData |
| assert(tbe.DemandRequest); //So, let RegionBuffer take care of sending back ack |
| enqueue(responseNetwork_out, ResponseMsg, toDirLatency) { |
| out_msg.addr := tbe.DemandAddress; |
| out_msg.Type := CoherenceResponseType:CPUPrbResp; // L3 and CPUs respond in same way to probes |
| out_msg.Sender := getPeer(machineID,address); |
| // will this always be ok? probably not for multisocket |
| out_msg.Destination.add(mapAddressToMachine(address, MachineType:Directory)); |
| out_msg.Dirty := false; // only true if sending back data i think |
| out_msg.Hit := false; |
| out_msg.Ntsl := false; |
| out_msg.State := CoherenceState:NA; |
| out_msg.NoAckNeeded := true; |
| out_msg.MessageSize := MessageSizeType:Response_Control; |
| DPRINTF(RubySlicc, "%s\n", out_msg); |
| } |
| } |
| } |
| |
| action(aie_ackRegionExclusiveInv, "aie", desc="Send ack to r-dir on region inv if tbe says so") { |
| // No need to send acks on replacements |
| assert(is_valid(tbe)); |
| enqueue(responseNetwork_out, ResponseMsg, toRegionDirLatency) { |
| out_msg.addr := getRegionBase(address); |
| out_msg.Type := CoherenceResponseType:CPUPrbResp; |
| out_msg.Sender := machineID; |
| out_msg.NotCached := true; |
| // will this always be ok? probably not for multisocket |
| out_msg.Destination.add(mapAddressToMachine(address, MachineType:RegionDir)); |
| out_msg.MessageSize := MessageSizeType:Response_Control; |
| out_msg.Dirty := tbe.dirty; |
| } |
| } |
| |
| action(ain_ackRegionInvNow, "ain", desc="Send ack to r-dir on region inv") { |
| enqueue(responseNetwork_out, ResponseMsg, toRegionDirLatency) { |
| out_msg.addr := getRegionBase(address); |
| out_msg.Type := CoherenceResponseType:CPUPrbResp; |
| out_msg.Sender := machineID; |
| // will this always be ok? probably not for multisocket |
| out_msg.Destination.add(mapAddressToMachine(address, MachineType:RegionDir)); |
| out_msg.MessageSize := MessageSizeType:Response_Control; |
| } |
| } |
| |
| action(aine_ackRegionInvExlusiveNow, "aine", desc="Send ack to r-dir on region inv with exlusive permission") { |
| enqueue(responseNetwork_out, ResponseMsg, toRegionDirLatency) { |
| out_msg.addr := getRegionBase(address); |
| out_msg.Type := CoherenceResponseType:CPUPrbResp; |
| out_msg.Sender := machineID; |
| out_msg.NotCached := true; |
| // will this always be ok? probably not for multisocket |
| out_msg.Destination.add(mapAddressToMachine(address, MachineType:RegionDir)); |
| out_msg.MessageSize := MessageSizeType:Response_Control; |
| } |
| } |
| |
| action(ap_ackPrivateNotify, "ap", desc="Send ack to r-dir on private notify") { |
| enqueue(responseNetwork_out, ResponseMsg, toRegionDirLatency) { |
| out_msg.addr := getRegionBase(address); |
| out_msg.Type := CoherenceResponseType:PrivateAck; |
| out_msg.Sender := machineID; |
| // will this always be ok? probably not for multisocket |
| out_msg.Destination.add(mapAddressToMachine(address, MachineType:RegionDir)); |
| out_msg.MessageSize := MessageSizeType:Response_Control; |
| } |
| } |
| |
| action(aw_ackWbNotify, "aw", desc="Send ack to r-dir on writeback notify") { |
| peek(notifyNetwork_in, CPURequestMsg) { |
| if (in_msg.NoAckNeeded == false) { |
| enqueue(responseNetwork_out, ResponseMsg, toRegionDirLatency) { |
| out_msg.addr := getRegionBase(address); |
| out_msg.Type := CoherenceResponseType:RegionWbAck; |
| out_msg.Sender := machineID; |
| // will this always be ok? probably not for multisocket |
| out_msg.Destination.add(mapAddressToMachine(address, MachineType:RegionDir)); |
| out_msg.MessageSize := MessageSizeType:Response_Control; |
| } |
| } |
| } |
| } |
| |
| action(e_evictCurrent, "e", desc="Evict this block in the region") { |
| // send force invalidate message to directory to invalidate this block |
| // must invalidate all blocks since region buffer could have privitized it |
| if (tbe.ValidBlocks.at(getRegionOffset(address)) && |
| (tbe.DemandRequest == false || tbe.DemandAddress != address)) { |
| DPRINTF(RubySlicc, "trying to evict address %s (base: %s, offset: %d)\n", address, getRegionBase(address), getRegionOffset(address)); |
| DPRINTF(RubySlicc, "tbe valid blocks %s\n", tbe.ValidBlocks); |
| |
| enqueue(probeNetwork_out, NBProbeRequestMsg, 1) { |
| out_msg.addr := address; |
| out_msg.Type := tbe.MsgType; |
| out_msg.ReturnData := true; |
| if (address == tbe.DemandAddress) { |
| out_msg.DemandRequest := true; |
| } |
| out_msg.MessageSize := MessageSizeType:Control; |
| out_msg.Destination.add(getPeer(machineID,address)); |
| DPRINTF(RubySlicc, "%s\n", out_msg); |
| } |
| APPEND_TRANSITION_COMMENT(" current "); |
| APPEND_TRANSITION_COMMENT(tbe.ValidBlocks.at(getRegionOffset(address))); |
| tbe.AllAcksReceived := false; |
| } else { |
| DPRINTF(RubySlicc, "Not evicting demand %s\n", address); |
| } |
| } |
| |
| action(ed_evictDemand, "ed", desc="Evict the demand request if it's valid") { |
| if (noTCCdir && tbe.MsgType == ProbeRequestType:PrbDowngrade && isOnGPU()) { |
| tbe.OutstandingAcks := 0; |
| tbe.AllAcksReceived := true; |
| tbe.DoneEvicting := true; |
| enqueue(triggerQueue_out, TriggerMsg, 1) { |
| out_msg.Type := TriggerType:AcksComplete; |
| out_msg.addr := getRegionBase(address); |
| } |
| } else if (tbe.DemandRequest) { |
| enqueue(probeNetwork_out, NBProbeRequestMsg, 1) { |
| out_msg.addr := tbe.DemandAddress; |
| out_msg.Type := tbe.MsgType; |
| out_msg.ReturnData := true; |
| out_msg.DemandRequest := true; |
| out_msg.MessageSize := MessageSizeType:Control; |
| out_msg.Destination.add(getPeer(machineID,address)); |
| DPRINTF(RubySlicc, "%s\n", out_msg); |
| tbe.AllAcksReceived := false; |
| } |
| if (tbe.ValidBlocks.at(getRegionOffset(tbe.DemandAddress)) == false) { |
| tbe.OutstandingAcks := tbe.OutstandingAcks + 1; |
| } |
| APPEND_TRANSITION_COMMENT("Evicting demand "); |
| APPEND_TRANSITION_COMMENT(tbe.DemandAddress); |
| } |
| APPEND_TRANSITION_COMMENT("waiting acks "); |
| APPEND_TRANSITION_COMMENT(tbe.OutstandingAcks); |
| } |
| |
| action(adp_AckDemandProbe, "fp", desc="forward demand probe even if we know that the core is invalid") { |
| peek(probeNetwork_in, NBProbeRequestMsg) { |
| if (in_msg.DemandRequest) { |
| enqueue(responseNetwork_out, ResponseMsg, toDirLatency) { |
| out_msg.addr := in_msg.DemandAddress; |
| out_msg.Type := CoherenceResponseType:CPUPrbResp; // L3 and CPUs respond in same way to probes |
| out_msg.Sender := getPeer(machineID,address); |
| // will this always be ok? probably not for multisocket |
| out_msg.Destination.add(mapAddressToMachine(address, MachineType:Directory)); |
| out_msg.Dirty := false; // only true if sending back data i think |
| out_msg.Hit := false; |
| out_msg.Ntsl := false; |
| out_msg.State := CoherenceState:NA; |
| out_msg.NoAckNeeded := true; |
| out_msg.MessageSize := MessageSizeType:Response_Control; |
| DPRINTF(RubySlicc, "%s\n", out_msg); |
| } |
| } |
| } |
| } |
| |
| action(en_enqueueNextEvict, "en", desc="Queue evict the next block in the region") { |
| // increment in_msg.addr by blockSize bytes and enqueue on triggerPort |
| // Only enqueue if the next address doesn't overrun the region bound |
| if (getRegionBase(getNextBlock(address)) == getRegionBase(address)) { |
| enqueue(triggerQueue_out, TriggerMsg, nextEvictLatency) { |
| out_msg.Type := TriggerType:InvNext; |
| out_msg.addr := getNextBlock(address); |
| } |
| } else { |
| tbe.DoneEvicting := true; |
| DPRINTF(RubySlicc, "Done evicing region %s\n", getRegionBase(address)); |
| DPRINTF(RubySlicc, "Waiting for %s acks\n", tbe.OutstandingAcks); |
| if (tbe.AllAcksReceived == true) { |
| enqueue(triggerQueue_out, TriggerMsg, 1) { |
| out_msg.Type := TriggerType:AcksComplete; |
| out_msg.addr := getRegionBase(address); |
| } |
| } |
| } |
| } |
| |
| action(ef_enqueueFirstEvict, "ef", desc="Queue the first block in the region to be evicted") { |
| if (tbe.DoneEvicting == false) { |
| enqueue(triggerQueue_out, TriggerMsg, nextEvictLatency) { |
| out_msg.Type := TriggerType:InvNext; |
| out_msg.addr := getRegionBase(address); |
| } |
| } |
| } |
| |
| action(ra_receiveAck, "ra", desc="Mark TBE entry as received this ack") { |
| DPRINTF(RubySlicc, "received ack for %s reg: %s vec: %s pos: %d\n", |
| address, getRegionBase(address), tbe.ValidBlocks, getRegionOffset(address)); |
| peek(unblockNetwork_in, UnblockMsg) { |
| // |
| // Note the tbe ValidBlock vec will be a conservative list of the |
| // valid blocks since the cache entry ValidBlock vec is set on the |
| // request |
| // |
| if (in_msg.wasValid) { |
| assert(tbe.ValidBlocks.at(getRegionOffset(address))); |
| } |
| } |
| tbe.OutstandingAcks := tbe.OutstandingAcks - 1; |
| tbe.AcksReceived.at(getRegionOffset(address)) := true; |
| assert(tbe.OutstandingAcks >= 0); |
| if (tbe.OutstandingAcks == 0) { |
| tbe.AllAcksReceived := true; |
| if (tbe.DoneEvicting) { |
| enqueue(triggerQueue_out, TriggerMsg, 1) { |
| out_msg.Type := TriggerType:AcksComplete; |
| out_msg.addr := getRegionBase(address); |
| } |
| } |
| } |
| |
| APPEND_TRANSITION_COMMENT(getRegionBase(address)); |
| APPEND_TRANSITION_COMMENT(" Acks left receive "); |
| APPEND_TRANSITION_COMMENT(tbe.OutstandingAcks); |
| } |
| |
| action(do_decrementOutstanding, "do", desc="Decrement outstanding requests") { |
| APPEND_TRANSITION_COMMENT(" decr outstanding "); |
| if (is_valid(cache_entry)) { |
| cache_entry.NumOutstandingReqs := cache_entry.NumOutstandingReqs - 1; |
| assert(cache_entry.OutstandingReqs.at(getRegionOffset(address))); |
| cache_entry.OutstandingReqs.at(getRegionOffset(address)) := false; |
| assert(cache_entry.NumOutstandingReqs >= 0); |
| assert(cache_entry.NumOutstandingReqs == countBoolVec(cache_entry.OutstandingReqs)); |
| APPEND_TRANSITION_COMMENT(cache_entry.NumOutstandingReqs); |
| } |
| if (is_valid(tbe)) { |
| tbe.NumOutstandingReqs := tbe.NumOutstandingReqs - 1; |
| assert(tbe.OutstandingReqs.at(getRegionOffset(address))); |
| tbe.OutstandingReqs.at(getRegionOffset(address)) := false; |
| assert(tbe.NumOutstandingReqs >= 0); |
| assert(tbe.NumOutstandingReqs == countBoolVec(tbe.OutstandingReqs)); |
| APPEND_TRANSITION_COMMENT(tbe.NumOutstandingReqs); |
| } |
| } |
| |
| action(co_checkOutstanding, "co", desc="check if there are no more outstanding requests") { |
| assert(is_valid(tbe)); |
| if ((tbe.NumOutstandingReqs <= tbe.OutstandingThreshold) && |
| (tbe.AllOutstandingTriggered == false)) { |
| APPEND_TRANSITION_COMMENT(" no more outstanding: "); |
| APPEND_TRANSITION_COMMENT(tbe.NumOutstandingReqs); |
| APPEND_TRANSITION_COMMENT(tbe.OutstandingThreshold); |
| enqueue(triggerQueue_out, TriggerMsg, 1) { |
| out_msg.Type := TriggerType:AllOutstanding; |
| if (tbe.DemandRequest) { |
| out_msg.addr := tbe.DemandAddress; |
| } else { |
| out_msg.addr := getRegionBase(address); |
| } |
| DPRINTF(RubySlicc, "co enqueuing %s\n", out_msg); |
| tbe.AllOutstandingTriggered := true; |
| } |
| } else { |
| APPEND_TRANSITION_COMMENT(" still more outstanding "); |
| } |
| } |
| |
| action(ro_resetAllOutstanding, "ro", desc="Reset all outstanding") { |
| tbe.AllOutstandingTriggered := false; |
| } |
| |
| action(so_setOutstandingCheckOne, "so", desc="Check outstanding is waiting for 1, not 0") { |
| // Need this for S_P because one request is outstanding between here and r-dir |
| tbe.OutstandingThreshold := 1; |
| } |
| |
| action(a_allocateRegionEntry, "a", desc="Allocate a new entry") { |
| set_cache_entry(cacheMemory.allocate(getRegionBase(address), new Entry)); |
| cache_entry.ValidBlocks.clear(); |
| cache_entry.ValidBlocks.resize(blocksPerRegion); |
| cache_entry.UsedBlocks.clear(); |
| cache_entry.UsedBlocks.resize(blocksPerRegion); |
| cache_entry.dirty := false; |
| cache_entry.NumOutstandingReqs := 0; |
| cache_entry.OutstandingReqs.clear(); |
| cache_entry.OutstandingReqs.resize(blocksPerRegion); |
| } |
| |
| action(d_deallocateRegionEntry, "d", desc="Deallocate region entry") { |
| cacheMemory.deallocate(getRegionBase(address)); |
| unset_cache_entry(); |
| } |
| |
| action(t_allocateTBE, "t", desc="allocate TBE Entry") { |
| check_allocate(TBEs); |
| TBEs.allocate(getRegionBase(address)); |
| set_tbe(getTBE(address)); |
| tbe.OutstandingAcks := 0; |
| tbe.AllAcksReceived := true; // starts true since the region could be empty |
| tbe.DoneEvicting := false; |
| tbe.AcksReceived.clear(); |
| tbe.AcksReceived.resize(blocksPerRegion); |
| tbe.SendAck := false; |
| tbe.OutstandingThreshold := 0; |
| if (is_valid(cache_entry)) { |
| tbe.NumOutstandingReqs := cache_entry.NumOutstandingReqs; |
| tbe.OutstandingReqs := cache_entry.OutstandingReqs; |
| assert(tbe.NumOutstandingReqs == countBoolVec(tbe.OutstandingReqs)); |
| tbe.dirty := cache_entry.dirty; |
| tbe.ValidBlocks := cache_entry.ValidBlocks; |
| tbe.OutstandingAcks := countBoolVec(tbe.ValidBlocks); |
| APPEND_TRANSITION_COMMENT(" tbe valid blocks "); |
| APPEND_TRANSITION_COMMENT(tbe.ValidBlocks); |
| APPEND_TRANSITION_COMMENT(" cache valid blocks "); |
| APPEND_TRANSITION_COMMENT(cache_entry.ValidBlocks); |
| } else { |
| tbe.dirty := false; |
| } |
| } |
| |
| action(m_markSendAck, "m", desc="Mark TBE that we need to ack at end") { |
| assert(is_valid(tbe)); |
| tbe.SendAck := true; |
| } |
| |
| action(db_markDirtyBit, "db", desc="Mark TBE dirty bit") { |
| peek(unblockNetwork_in, UnblockMsg) { |
| if (is_valid(tbe)) { |
| tbe.dirty := tbe.dirty || in_msg.Dirty; |
| } |
| } |
| } |
| |
| action(dr_markDoneAckReceived, "dr", desc="Mark TBE that a done ack has been received") { |
| assert(is_valid(tbe)); |
| tbe.DoneAckReceived := true; |
| tbe.DoneAckAddr := address; |
| APPEND_TRANSITION_COMMENT(" marking done ack on TBE "); |
| } |
| |
| action(se_setTBE, "se", desc="Set msg type to evict") { |
| peek(probeNetwork_in, NBProbeRequestMsg) { |
| tbe.MsgType := in_msg.Type; |
| tbe.Requestor := in_msg.Requestor; |
| tbe.DemandAddress := in_msg.DemandAddress; |
| tbe.DemandRequest := in_msg.DemandRequest; |
| } |
| } |
| |
| action(sne_setNewTBE, "sne", desc="Set msg type to evict") { |
| peek(probeNetwork_in, NBProbeRequestMsg) { |
| tbe.NewMsgType := in_msg.Type; |
| tbe.NewRequestor := in_msg.Requestor; |
| tbe.NewDemandAddress := in_msg.DemandAddress; |
| tbe.NewDemandRequest := in_msg.DemandRequest; |
| } |
| } |
| |
| action(soe_setOldTBE, "soe", desc="Set msg type to evict") { |
| tbe.MsgType := tbe.NewMsgType; |
| tbe.Requestor := tbe.NewRequestor; |
| tbe.DemandAddress := tbe.NewDemandAddress; |
| tbe.DemandRequest := tbe.NewDemandRequest; |
| tbe.OutstandingAcks := countBoolVec(tbe.ValidBlocks); |
| tbe.AllAcksReceived := true; // starts true since the region could be empty |
| tbe.DoneEvicting := false; |
| tbe.AcksReceived.clear(); |
| tbe.AcksReceived.resize(blocksPerRegion); |
| tbe.SendAck := false; |
| } |
| |
| action(ser_setTBE, "ser", desc="Set msg type to evict repl") { |
| tbe.MsgType := ProbeRequestType:PrbInv; |
| } |
| |
| action(md_setMustDowngrade, "md", desc="When permissions finally get here, must be shared") { |
| assert(is_valid(cache_entry)); |
| cache_entry.MustDowngrade := true; |
| } |
| |
| action(dt_deallocateTBE, "dt", desc="deallocate TBE Entry") { |
| TBEs.deallocate(getRegionBase(address)); |
| unset_tbe(); |
| } |
| |
| action(p_popRequestQueue, "p", desc="Pop the request queue") { |
| requestNetwork_in.dequeue(clockEdge()); |
| } |
| |
| action(pl_popUnblockQueue, "pl", desc="Pop the unblock queue") { |
| unblockNetwork_in.dequeue(clockEdge()); |
| } |
| |
| action(pn_popNotifyQueue, "pn", desc="Pop the notify queue") { |
| notifyNetwork_in.dequeue(clockEdge()); |
| } |
| |
| action(pp_popProbeQueue, "pp", desc="Pop the probe queue") { |
| probeNetwork_in.dequeue(clockEdge()); |
| } |
| |
| action(pt_popTriggerQueue, "pt", desc="Pop the trigger queue") { |
| DPRINTF(RubySlicc, "Trigger Before Contents: %s\n", triggerQueue_in); |
| triggerQueue_in.dequeue(clockEdge()); |
| DPRINTF(RubySlicc, "Trigger After Contents: %s\n", triggerQueue_in); |
| } |
| |
| // Must always use wake all, since non-region address wait on region addresses |
| action(wa_wakeUpAllDependents, "wa", desc="Wake up any requests waiting for this region") { |
| wakeUpAllBuffers(); |
| } |
| |
| action(zz_stallAndWaitRequestQueue, "\z", desc="recycle request queue") { |
| Addr regAddr := getRegionBase(address); |
| DPRINTF(RubySlicc, "Stalling address %s\n", regAddr); |
| stall_and_wait(requestNetwork_in, regAddr); |
| } |
| |
| action(yy_stallAndWaitProbeQueue, "\y", desc="stall probe queue") { |
| Addr regAddr := getRegionBase(address); |
| stall_and_wait(probeNetwork_in, regAddr); |
| } |
| |
| action(yyy_recycleProbeQueue, "\yy", desc="recycle probe queue") { |
| probeNetwork_in.recycle(clockEdge(), cyclesToTicks(recycle_latency)); |
| } |
| |
| action(zzz_recycleRequestQueue, "\zz", desc="recycle request queue") { |
| requestNetwork_in.recycle(clockEdge(), cyclesToTicks(recycle_latency)); |
| } |
| |
| action(www_recycleUnblockNetwork, "\ww", desc="recycle unblock queue") { |
| unblockNetwork_in.recycle(clockEdge(), cyclesToTicks(recycle_latency)); |
| } |
| |
| action(z_stall, "z", desc="stall request queue") { |
| // fake state |
| } |
| |
| action(mru_setMRU, "mru", desc="set MRU") { |
| cacheMemory.setMRU(address, cache_entry.NumValidBlocks); |
| } |
| |
| // Transitions |
| |
| transition({NP_PS, S_P, S_NP_PS, P_NP, P_S, P_NP_O, S_NP_PS_O, P_S_O, S_O, P_NP_W, P_NP_NP, NP_W}, {CPURead, CPUWriteback, CPUWrite}) {} { |
| zz_stallAndWaitRequestQueue; |
| } |
| |
| transition(SS_P, {CPURead, CPUWriteback}) { |
| zz_stallAndWaitRequestQueue; |
| } |
| |
| transition({NP, S, P, NP_PS, S_P, S_NP_PS, P_NP, P_S, P_NP_O, S_NP_PS_O, P_S_O, S_O, SS_P, NP_W, P_NP_NP}, StallAccess) {} { |
| zz_stallAndWaitRequestQueue; |
| } |
| |
| transition({S, P, NP_PS, S_P, S_NP_PS, P_NP, P_S, P_NP_O, S_NP_PS_O, P_S_O, S_O, SS_P, P_NP_W, P_NP_NP, NP_W}, StallDoneAck) { |
| www_recycleUnblockNetwork; |
| } |
| |
| transition(NP, StallDoneAck, NP_W) { |
| t_allocateTBE; |
| db_markDirtyBit; |
| dr_markDoneAckReceived; |
| pl_popUnblockQueue; |
| } |
| |
| transition(NP_W, StaleRequest, NP) { |
| f_fwdReqToDir; |
| dt_deallocateTBE; |
| wa_wakeUpAllDependents; |
| p_popRequestQueue; |
| } |
| |
| transition(P_NP_O, DowngradeRegion) {} { |
| z_stall; // should stall and wait |
| } |
| |
| transition({NP_PS, S_NP_PS, S_P, P_S, P_NP_O, S_NP_PS_O, P_S_O, S_O, SS_P}, ReplRegion) {} { |
| zz_stallAndWaitRequestQueue; // can't let things get out of order! |
| } |
| |
| transition({P_NP_O, S_O, SS_P}, InvRegion) {} { |
| yyy_recycleProbeQueue; // can't be z_stall because there could be a RdBlkM in the requestQueue which has the sinked flag which is blocking the inv |
| } |
| |
| transition(P_NP, {InvRegion, DowngradeRegion}, P_NP_NP) {} { |
| sne_setNewTBE; |
| pp_popProbeQueue; |
| } |
| |
| transition(S_P, DowngradeRegion) {} { |
| adp_AckDemandProbe; |
| ain_ackRegionInvNow; |
| pp_popProbeQueue; |
| } |
| |
| transition(P_NP_W, InvRegion) { |
| adp_AckDemandProbe; |
| ain_ackRegionInvNow; |
| pp_popProbeQueue; |
| } |
| |
| transition(P_NP_W, DowngradeRegion) { |
| adp_AckDemandProbe; |
| aine_ackRegionInvExlusiveNow; |
| pp_popProbeQueue; |
| } |
| |
| transition({P, S}, {CPURead, CPUWriteback}) {TagArrayRead, TagArrayWrite} { |
| mru_setMRU; |
| f_fwdReqToDir; |
| u_updateRegionEntry; |
| p_popRequestQueue; |
| } |
| |
| transition(P, CPUWrite) {TagArrayRead, TagArrayWrite} { |
| mru_setMRU; |
| f_fwdReqToDir; |
| u_updateRegionEntry; |
| p_popRequestQueue; |
| } |
| |
| transition(S, CPUWrite, S_O) {TagArrayRead} { |
| mru_setMRU; |
| t_allocateTBE; |
| co_checkOutstanding; |
| zz_stallAndWaitRequestQueue; |
| } |
| |
| transition(S_O, AllOutstanding, SS_P) { |
| wa_wakeUpAllDependents; |
| ro_resetAllOutstanding; |
| pt_popTriggerQueue; |
| } |
| |
| transition(SS_P, CPUWrite, S_P) { |
| mru_setMRU; |
| dt_deallocateTBE; |
| ru_requestUpgrade; |
| u_updateRegionEntry; |
| p_popRequestQueue; |
| } |
| |
| transition(NP, {CPURead, CPUWriteback}, NP_PS) {TagArrayRead, TagArrayWrite} { |
| a_allocateRegionEntry; |
| rs_requestShared; |
| u_updateRegionEntry; |
| p_popRequestQueue;//zz_stallAndWaitRequestQueue; |
| } |
| |
| transition(NP, CPUWrite, NP_PS) {TagArrayRead, TagArrayWrite} { |
| a_allocateRegionEntry; |
| rp_requestPrivate; |
| u_updateRegionEntry; |
| p_popRequestQueue;//zz_stallAndWaitRequestQueue; |
| } |
| |
| transition(NP_PS, PrivateNotify, P) {} { |
| ap_ackPrivateNotify; |
| wa_wakeUpAllDependents; |
| pn_popNotifyQueue; |
| } |
| |
| transition(S_P, PrivateNotify, P) {} { |
| ap_ackPrivateNotify; |
| wa_wakeUpAllDependents; |
| pn_popNotifyQueue; |
| } |
| |
| transition(NP_PS, SharedNotify, S) {} { |
| ap_ackPrivateNotify; |
| wa_wakeUpAllDependents; |
| pn_popNotifyQueue; |
| } |
| |
| transition(P_NP_W, WbNotify, NP) {} { |
| aw_ackWbNotify; |
| wa_wakeUpAllDependents; |
| dt_deallocateTBE; |
| pn_popNotifyQueue; |
| } |
| |
| transition({P, S}, ReplRegion, P_NP_O) {TagArrayRead, TagArrayWrite} { |
| t_allocateTBE; |
| ser_setTBE; |
| d_deallocateRegionEntry; |
| co_checkOutstanding; |
| } |
| |
| transition({P, S}, InvRegion, P_NP_O) {TagArrayRead, TagArrayWrite} { |
| t_allocateTBE; |
| se_setTBE; |
| m_markSendAck; |
| d_deallocateRegionEntry; |
| co_checkOutstanding; |
| pp_popProbeQueue; |
| } |
| |
| transition(P_NP_O, AllOutstanding, P_NP) {} { |
| ed_evictDemand; |
| ef_enqueueFirstEvict; |
| ro_resetAllOutstanding; |
| pt_popTriggerQueue; |
| } |
| |
| transition(S_P, InvRegion, S_NP_PS_O) {TagArrayRead} { |
| t_allocateTBE; |
| se_setTBE; |
| m_markSendAck; |
| so_setOutstandingCheckOne; |
| co_checkOutstanding; |
| pp_popProbeQueue; |
| } |
| |
| transition(S_NP_PS_O, AllOutstanding, S_NP_PS) { |
| ed_evictDemand; |
| ef_enqueueFirstEvict; |
| ro_resetAllOutstanding; |
| pt_popTriggerQueue; |
| } |
| |
| transition(P, DowngradeRegion, P_S_O) {TagArrayRead, TagArrayWrite} { |
| t_allocateTBE; |
| se_setTBE; |
| m_markSendAck; |
| co_checkOutstanding; |
| pp_popProbeQueue; |
| } |
| |
| transition(P_S_O, AllOutstanding, P_S) {} { |
| ed_evictDemand; |
| ef_enqueueFirstEvict; |
| ro_resetAllOutstanding; |
| pt_popTriggerQueue; |
| } |
| |
| transition({P, S}, DoneAck) {TagArrayWrite} { |
| do_decrementOutstanding; |
| wa_wakeUpAllDependents; |
| db_markDirtyBit; |
| uw_updatePossibleWriteback; |
| pl_popUnblockQueue; |
| } |
| |
| transition({S_P, NP_PS, S_NP_PS}, DoneAck) {TagArrayWrite} { |
| www_recycleUnblockNetwork; |
| } |
| |
| transition({P_NP_O, S_NP_PS_O, P_S_O, S_O}, DoneAck) {} { |
| do_decrementOutstanding; |
| co_checkOutstanding; |
| db_markDirtyBit; |
| uw_updatePossibleWriteback; |
| pl_popUnblockQueue; |
| } |
| |
| transition({P_NP, P_S, S_NP_PS, P_NP_NP}, Evict) {} { |
| e_evictCurrent; |
| en_enqueueNextEvict; |
| pt_popTriggerQueue; |
| } |
| |
| transition({P_NP, P_S, S_NP_PS, P_NP_NP}, InvAck) {} { |
| ra_receiveAck; |
| db_markDirtyBit; |
| pl_popUnblockQueue; |
| } |
| |
| transition(P_NP, LastAck_CleanWb, P_NP_W) {} { |
| rw_requestWriteback; |
| pt_popTriggerQueue; |
| } |
| |
| transition(P_NP_NP, LastAck_CleanWb, P_NP) {} { |
| soe_setOldTBE; |
| m_markSendAck; |
| ed_evictDemand; |
| ef_enqueueFirstEvict; |
| pt_popTriggerQueue; |
| } |
| |
| transition(P_NP, LastAck_PrbResp, NP) {} { |
| aie_ackRegionExclusiveInv; |
| dt_deallocateTBE; |
| wa_wakeUpAllDependents; |
| pt_popTriggerQueue; |
| } |
| |
| transition(S_NP_PS, LastAck_PrbResp, NP_PS) {} { |
| aie_ackRegionExclusiveInv; |
| dt_deallocateTBE; |
| wa_wakeUpAllDependents; |
| pt_popTriggerQueue; |
| } |
| |
| transition(P_S, LastAck_PrbResp, S) {} { |
| ai_ackRegionInv; |
| ad_ackDircetory; |
| dt_deallocateTBE; |
| wa_wakeUpAllDependents; |
| pt_popTriggerQueue; |
| } |
| |
| } |
| |