| # Copyright (c) 2011-2015 Advanced Micro Devices, Inc. |
| # All rights reserved. |
| # |
| # Redistribution and use in source and binary forms, with or without |
| # modification, are permitted provided that the following conditions are met: |
| # |
| # 1. Redistributions of source code must retain the above copyright notice, |
| # this list of conditions and the following disclaimer. |
| # |
| # 2. Redistributions in binary form must reproduce the above copyright notice, |
| # this list of conditions and the following disclaimer in the documentation |
| # and/or other materials provided with the distribution. |
| # |
| # 3. Neither the name of the copyright holder nor the names of its |
| # contributors may be used to endorse or promote products derived from this |
| # software without specific prior written permission. |
| # |
| # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
| # AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
| # ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE |
| # LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
| # CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
| # SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
| # INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
| # CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
| # ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
| # POSSIBILITY OF SUCH DAMAGE. |
| |
| import math |
| import m5 |
| from m5.objects import * |
| from m5.defines import buildEnv |
| from m5.util import addToPath |
| from .Ruby import create_topology |
| from .Ruby import send_evicts |
| from common import ObjectList |
| from common import MemConfig |
| from common import FileSystemConfig |
| |
| addToPath("../") |
| |
| from topologies.Cluster import Cluster |
| from topologies.Crossbar import Crossbar |
| |
| |
| class CntrlBase: |
| _seqs = 0 |
| |
| @classmethod |
| def seqCount(cls): |
| # Use SeqCount not class since we need global count |
| CntrlBase._seqs += 1 |
| return CntrlBase._seqs - 1 |
| |
| _cntrls = 0 |
| |
| @classmethod |
| def cntrlCount(cls): |
| # Use CntlCount not class since we need global count |
| CntrlBase._cntrls += 1 |
| return CntrlBase._cntrls - 1 |
| |
| _version = 0 |
| |
| @classmethod |
| def versionCount(cls): |
| cls._version += 1 # Use count for this particular type |
| return cls._version - 1 |
| |
| |
| class L1Cache(RubyCache): |
| resourceStalls = False |
| dataArrayBanks = 2 |
| tagArrayBanks = 2 |
| dataAccessLatency = 1 |
| tagAccessLatency = 1 |
| |
| def create(self, size, assoc, options): |
| self.size = MemorySize(size) |
| self.assoc = assoc |
| self.replacement_policy = TreePLRURP() |
| |
| |
| class L2Cache(RubyCache): |
| resourceStalls = False |
| assoc = 16 |
| dataArrayBanks = 16 |
| tagArrayBanks = 16 |
| |
| def create(self, size, assoc, options): |
| self.size = MemorySize(size) |
| self.assoc = assoc |
| self.replacement_policy = TreePLRURP() |
| |
| |
| class CPCntrl(CorePair_Controller, CntrlBase): |
| def create(self, options, ruby_system, system): |
| self.version = self.versionCount() |
| |
| self.L1Icache = L1Cache() |
| self.L1Icache.create(options.l1i_size, options.l1i_assoc, options) |
| self.L1D0cache = L1Cache() |
| self.L1D0cache.create(options.l1d_size, options.l1d_assoc, options) |
| self.L1D1cache = L1Cache() |
| self.L1D1cache.create(options.l1d_size, options.l1d_assoc, options) |
| self.L2cache = L2Cache() |
| self.L2cache.create(options.l2_size, options.l2_assoc, options) |
| |
| self.sequencer = RubySequencer() |
| self.sequencer.version = self.seqCount() |
| self.sequencer.dcache = self.L1D0cache |
| self.sequencer.ruby_system = ruby_system |
| self.sequencer.coreid = 0 |
| self.sequencer.is_cpu_sequencer = True |
| |
| self.sequencer1 = RubySequencer() |
| self.sequencer1.version = self.seqCount() |
| self.sequencer1.dcache = self.L1D1cache |
| self.sequencer1.ruby_system = ruby_system |
| self.sequencer1.coreid = 1 |
| self.sequencer1.is_cpu_sequencer = True |
| |
| self.issue_latency = options.cpu_to_dir_latency |
| self.send_evictions = send_evicts(options) |
| |
| self.ruby_system = ruby_system |
| |
| if options.recycle_latency: |
| self.recycle_latency = options.recycle_latency |
| |
| |
| class TCPCache(RubyCache): |
| size = "16kB" |
| assoc = 16 |
| dataArrayBanks = 16 # number of data banks |
| tagArrayBanks = 16 # number of tag banks |
| dataAccessLatency = 4 |
| tagAccessLatency = 1 |
| |
| def create(self, options): |
| self.size = MemorySize(options.tcp_size) |
| self.assoc = options.tcp_assoc |
| self.resourceStalls = options.no_tcc_resource_stalls |
| self.replacement_policy = TreePLRURP() |
| |
| |
| class TCPCntrl(TCP_Controller, CntrlBase): |
| def create(self, options, ruby_system, system): |
| self.version = self.versionCount() |
| |
| self.L1cache = TCPCache( |
| tagAccessLatency=options.TCP_latency, |
| dataAccessLatency=options.TCP_latency, |
| ) |
| self.L1cache.resourceStalls = options.no_resource_stalls |
| self.L1cache.create(options) |
| self.issue_latency = 1 |
| # TCP_Controller inherits this from RubyController |
| self.mandatory_queue_latency = options.mandatory_queue_latency |
| |
| self.coalescer = VIPERCoalescer() |
| self.coalescer.version = self.seqCount() |
| self.coalescer.icache = self.L1cache |
| self.coalescer.dcache = self.L1cache |
| self.coalescer.ruby_system = ruby_system |
| self.coalescer.support_inst_reqs = False |
| self.coalescer.is_cpu_sequencer = False |
| if options.tcp_deadlock_threshold: |
| self.coalescer.deadlock_threshold = options.tcp_deadlock_threshold |
| self.coalescer.max_coalesces_per_cycle = ( |
| options.max_coalesces_per_cycle |
| ) |
| |
| self.sequencer = RubySequencer() |
| self.sequencer.version = self.seqCount() |
| self.sequencer.dcache = self.L1cache |
| self.sequencer.ruby_system = ruby_system |
| self.sequencer.is_cpu_sequencer = True |
| |
| self.use_seq_not_coal = False |
| |
| self.ruby_system = ruby_system |
| if hasattr(options, "gpu_clock") and hasattr(options, "gpu_voltage"): |
| self.clk_domain = SrcClockDomain( |
| clock=options.gpu_clock, |
| voltage_domain=VoltageDomain(voltage=options.gpu_voltage), |
| ) |
| |
| if options.recycle_latency: |
| self.recycle_latency = options.recycle_latency |
| |
| def createCP(self, options, ruby_system, system): |
| self.version = self.versionCount() |
| |
| self.L1cache = TCPCache( |
| tagAccessLatency=options.TCP_latency, |
| dataAccessLatency=options.TCP_latency, |
| ) |
| self.L1cache.resourceStalls = options.no_resource_stalls |
| self.L1cache.create(options) |
| self.issue_latency = 1 |
| |
| self.coalescer = VIPERCoalescer() |
| self.coalescer.version = self.seqCount() |
| self.coalescer.icache = self.L1cache |
| self.coalescer.dcache = self.L1cache |
| self.coalescer.ruby_system = ruby_system |
| self.coalescer.support_inst_reqs = False |
| self.coalescer.is_cpu_sequencer = False |
| |
| self.sequencer = RubySequencer() |
| self.sequencer.version = self.seqCount() |
| self.sequencer.dcache = self.L1cache |
| self.sequencer.ruby_system = ruby_system |
| self.sequencer.is_cpu_sequencer = True |
| |
| self.use_seq_not_coal = True |
| |
| self.ruby_system = ruby_system |
| |
| if options.recycle_latency: |
| self.recycle_latency = options.recycle_latency |
| |
| |
| class SQCCache(RubyCache): |
| dataArrayBanks = 8 |
| tagArrayBanks = 8 |
| dataAccessLatency = 1 |
| tagAccessLatency = 1 |
| |
| def create(self, options): |
| self.size = MemorySize(options.sqc_size) |
| self.assoc = options.sqc_assoc |
| self.replacement_policy = TreePLRURP() |
| |
| |
| class SQCCntrl(SQC_Controller, CntrlBase): |
| def create(self, options, ruby_system, system): |
| self.version = self.versionCount() |
| |
| self.L1cache = SQCCache() |
| self.L1cache.create(options) |
| self.L1cache.resourceStalls = options.no_resource_stalls |
| |
| self.sequencer = RubySequencer() |
| |
| self.sequencer.version = self.seqCount() |
| self.sequencer.dcache = self.L1cache |
| self.sequencer.ruby_system = ruby_system |
| self.sequencer.support_data_reqs = False |
| self.sequencer.is_cpu_sequencer = False |
| if options.sqc_deadlock_threshold: |
| self.sequencer.deadlock_threshold = options.sqc_deadlock_threshold |
| |
| self.ruby_system = ruby_system |
| if hasattr(options, "gpu_clock") and hasattr(options, "gpu_voltage"): |
| self.clk_domain = SrcClockDomain( |
| clock=options.gpu_clock, |
| voltage_domain=VoltageDomain(voltage=options.gpu_voltage), |
| ) |
| |
| if options.recycle_latency: |
| self.recycle_latency = options.recycle_latency |
| |
| |
| class TCC(RubyCache): |
| size = MemorySize("256kB") |
| assoc = 16 |
| dataAccessLatency = 8 |
| tagAccessLatency = 2 |
| resourceStalls = True |
| |
| def create(self, options): |
| self.assoc = options.tcc_assoc |
| if hasattr(options, "bw_scalor") and options.bw_scalor > 0: |
| s = options.num_compute_units |
| tcc_size = s * 128 |
| tcc_size = str(tcc_size) + "kB" |
| self.size = MemorySize(tcc_size) |
| self.dataArrayBanks = 64 |
| self.tagArrayBanks = 64 |
| else: |
| self.size = MemorySize(options.tcc_size) |
| self.dataArrayBanks = ( |
| 256 / options.num_tccs |
| ) # number of data banks |
| self.tagArrayBanks = 256 / options.num_tccs # number of tag banks |
| self.size.value = self.size.value / options.num_tccs |
| if (self.size.value / int(self.assoc)) < 128: |
| self.size.value = int(128 * self.assoc) |
| self.start_index_bit = math.log(options.cacheline_size, 2) + math.log( |
| options.num_tccs, 2 |
| ) |
| self.replacement_policy = TreePLRURP() |
| |
| |
| class TCCCntrl(TCC_Controller, CntrlBase): |
| def create(self, options, ruby_system, system): |
| self.version = self.versionCount() |
| self.L2cache = TCC() |
| self.L2cache.create(options) |
| self.L2cache.resourceStalls = options.no_tcc_resource_stalls |
| |
| self.ruby_system = ruby_system |
| if hasattr(options, "gpu_clock") and hasattr(options, "gpu_voltage"): |
| self.clk_domain = SrcClockDomain( |
| clock=options.gpu_clock, |
| voltage_domain=VoltageDomain(voltage=options.gpu_voltage), |
| ) |
| |
| if options.recycle_latency: |
| self.recycle_latency = options.recycle_latency |
| |
| |
| class L3Cache(RubyCache): |
| dataArrayBanks = 16 |
| tagArrayBanks = 16 |
| |
| def create(self, options, ruby_system, system): |
| self.size = MemorySize(options.l3_size) |
| self.size.value /= options.num_dirs |
| self.assoc = options.l3_assoc |
| self.dataArrayBanks /= options.num_dirs |
| self.tagArrayBanks /= options.num_dirs |
| self.dataArrayBanks /= options.num_dirs |
| self.tagArrayBanks /= options.num_dirs |
| self.dataAccessLatency = options.l3_data_latency |
| self.tagAccessLatency = options.l3_tag_latency |
| self.resourceStalls = False |
| self.replacement_policy = TreePLRURP() |
| |
| |
| class L3Cntrl(L3Cache_Controller, CntrlBase): |
| def create(self, options, ruby_system, system): |
| self.version = self.versionCount() |
| self.L3cache = L3Cache() |
| self.L3cache.create(options, ruby_system, system) |
| |
| self.l3_response_latency = max( |
| self.L3cache.dataAccessLatency, self.L3cache.tagAccessLatency |
| ) |
| self.ruby_system = ruby_system |
| |
| if options.recycle_latency: |
| self.recycle_latency = options.recycle_latency |
| |
| def connectWireBuffers( |
| self, |
| req_to_dir, |
| resp_to_dir, |
| l3_unblock_to_dir, |
| req_to_l3, |
| probe_to_l3, |
| resp_to_l3, |
| ): |
| self.reqToDir = req_to_dir |
| self.respToDir = resp_to_dir |
| self.l3UnblockToDir = l3_unblock_to_dir |
| self.reqToL3 = req_to_l3 |
| self.probeToL3 = probe_to_l3 |
| self.respToL3 = resp_to_l3 |
| |
| |
| class DirCntrl(Directory_Controller, CntrlBase): |
| def create(self, options, dir_ranges, ruby_system, system): |
| self.version = self.versionCount() |
| |
| self.response_latency = 30 |
| |
| self.addr_ranges = dir_ranges |
| self.directory = RubyDirectoryMemory() |
| |
| self.L3CacheMemory = L3Cache() |
| self.L3CacheMemory.create(options, ruby_system, system) |
| |
| self.l3_hit_latency = max( |
| self.L3CacheMemory.dataAccessLatency, |
| self.L3CacheMemory.tagAccessLatency, |
| ) |
| |
| self.number_of_TBEs = options.num_tbes |
| |
| self.ruby_system = ruby_system |
| |
| if options.recycle_latency: |
| self.recycle_latency = options.recycle_latency |
| |
| def connectWireBuffers( |
| self, |
| req_to_dir, |
| resp_to_dir, |
| l3_unblock_to_dir, |
| req_to_l3, |
| probe_to_l3, |
| resp_to_l3, |
| ): |
| self.reqToDir = req_to_dir |
| self.respToDir = resp_to_dir |
| self.l3UnblockToDir = l3_unblock_to_dir |
| self.reqToL3 = req_to_l3 |
| self.probeToL3 = probe_to_l3 |
| self.respToL3 = resp_to_l3 |
| |
| |
| def define_options(parser): |
| parser.add_argument("--num-subcaches", type=int, default=4) |
| parser.add_argument("--l3-data-latency", type=int, default=20) |
| parser.add_argument("--l3-tag-latency", type=int, default=15) |
| parser.add_argument("--cpu-to-dir-latency", type=int, default=120) |
| parser.add_argument("--gpu-to-dir-latency", type=int, default=120) |
| parser.add_argument( |
| "--no-resource-stalls", action="store_false", default=True |
| ) |
| parser.add_argument( |
| "--no-tcc-resource-stalls", action="store_false", default=True |
| ) |
| parser.add_argument("--use-L3-on-WT", action="store_true", default=False) |
| parser.add_argument("--num-tbes", type=int, default=256) |
| parser.add_argument("--l2-latency", type=int, default=50) # load to use |
| parser.add_argument( |
| "--num-tccs", |
| type=int, |
| default=1, |
| help="number of TCC banks in the GPU", |
| ) |
| parser.add_argument( |
| "--sqc-size", type=str, default="32kB", help="SQC cache size" |
| ) |
| parser.add_argument( |
| "--sqc-assoc", type=int, default=8, help="SQC cache assoc" |
| ) |
| parser.add_argument( |
| "--sqc-deadlock-threshold", |
| type=int, |
| help="Set the SQC deadlock threshold to some value", |
| ) |
| |
| parser.add_argument( |
| "--WB_L1", action="store_true", default=False, help="writeback L1" |
| ) |
| parser.add_argument( |
| "--WB_L2", action="store_true", default=False, help="writeback L2" |
| ) |
| parser.add_argument( |
| "--TCP_latency", |
| type=int, |
| default=4, |
| help="In combination with the number of banks for the " |
| "TCP, this determines how many requests can happen " |
| "per cycle (i.e., the bandwidth)", |
| ) |
| parser.add_argument( |
| "--mandatory_queue_latency", |
| type=int, |
| default=1, |
| help="Hit latency for TCP", |
| ) |
| parser.add_argument( |
| "--TCC_latency", type=int, default=16, help="TCC latency" |
| ) |
| parser.add_argument( |
| "--tcc-size", type=str, default="256kB", help="agregate tcc size" |
| ) |
| parser.add_argument("--tcc-assoc", type=int, default=16, help="tcc assoc") |
| parser.add_argument( |
| "--tcp-size", type=str, default="16kB", help="tcp size" |
| ) |
| parser.add_argument("--tcp-assoc", type=int, default=16, help="tcp assoc") |
| parser.add_argument( |
| "--tcp-deadlock-threshold", |
| type=int, |
| help="Set the TCP deadlock threshold to some value", |
| ) |
| parser.add_argument( |
| "--max-coalesces-per-cycle", |
| type=int, |
| default=1, |
| help="Maximum insts that may coalesce in a cycle", |
| ) |
| |
| parser.add_argument( |
| "--noL1", action="store_true", default=False, help="bypassL1" |
| ) |
| parser.add_argument( |
| "--scalar-buffer-size", |
| type=int, |
| default=128, |
| help="Size of the mandatory queue in the GPU scalar " |
| "cache controller", |
| ) |
| |
| |
| def construct_dirs(options, system, ruby_system, network): |
| |
| dir_cntrl_nodes = [] |
| |
| # For an odd number of CPUs, still create the right number of controllers |
| TCC_bits = int(math.log(options.num_tccs, 2)) |
| |
| if options.numa_high_bit: |
| numa_bit = options.numa_high_bit |
| else: |
| # if the numa_bit is not specified, set the directory bits as the |
| # lowest bits above the block offset bits, and the numa_bit as the |
| # highest of those directory bits |
| dir_bits = int(math.log(options.num_dirs, 2)) |
| block_size_bits = int(math.log(options.cacheline_size, 2)) |
| numa_bit = block_size_bits + dir_bits - 1 |
| |
| for i in range(options.num_dirs): |
| dir_ranges = [] |
| for r in system.mem_ranges: |
| addr_range = m5.objects.AddrRange( |
| r.start, |
| size=r.size(), |
| intlvHighBit=numa_bit, |
| intlvBits=dir_bits, |
| intlvMatch=i, |
| ) |
| dir_ranges.append(addr_range) |
| |
| dir_cntrl = DirCntrl(noTCCdir=True, TCC_select_num_bits=TCC_bits) |
| dir_cntrl.create(options, dir_ranges, ruby_system, system) |
| dir_cntrl.number_of_TBEs = options.num_tbes |
| dir_cntrl.useL3OnWT = options.use_L3_on_WT |
| # the number_of_TBEs is inclusive of TBEs below |
| |
| # Connect the Directory controller to the ruby network |
| dir_cntrl.requestFromCores = MessageBuffer(ordered=True) |
| dir_cntrl.requestFromCores.in_port = network.out_port |
| |
| dir_cntrl.responseFromCores = MessageBuffer() |
| dir_cntrl.responseFromCores.in_port = network.out_port |
| |
| dir_cntrl.unblockFromCores = MessageBuffer() |
| dir_cntrl.unblockFromCores.in_port = network.out_port |
| |
| dir_cntrl.probeToCore = MessageBuffer() |
| dir_cntrl.probeToCore.out_port = network.in_port |
| |
| dir_cntrl.responseToCore = MessageBuffer() |
| dir_cntrl.responseToCore.out_port = network.in_port |
| |
| dir_cntrl.triggerQueue = MessageBuffer(ordered=True) |
| dir_cntrl.L3triggerQueue = MessageBuffer(ordered=True) |
| dir_cntrl.requestToMemory = MessageBuffer(ordered=True) |
| dir_cntrl.responseFromMemory = MessageBuffer(ordered=True) |
| |
| dir_cntrl.requestFromDMA = MessageBuffer(ordered=True) |
| dir_cntrl.requestFromDMA.in_port = network.out_port |
| |
| dir_cntrl.responseToDMA = MessageBuffer() |
| dir_cntrl.responseToDMA.out_port = network.in_port |
| |
| exec("ruby_system.dir_cntrl%d = dir_cntrl" % i) |
| dir_cntrl_nodes.append(dir_cntrl) |
| |
| return dir_cntrl_nodes |
| |
| |
| def construct_gpudirs(options, system, ruby_system, network): |
| |
| dir_cntrl_nodes = [] |
| mem_ctrls = [] |
| |
| xor_low_bit = 0 |
| |
| # For an odd number of CPUs, still create the right number of controllers |
| TCC_bits = int(math.log(options.num_tccs, 2)) |
| |
| dir_bits = int(math.log(options.dgpu_num_dirs, 2)) |
| block_size_bits = int(math.log(options.cacheline_size, 2)) |
| numa_bit = block_size_bits + dir_bits - 1 |
| |
| gpu_mem_range = AddrRange(0, size=options.dgpu_mem_size) |
| for i in range(options.dgpu_num_dirs): |
| addr_range = m5.objects.AddrRange( |
| gpu_mem_range.start, |
| size=gpu_mem_range.size(), |
| intlvHighBit=numa_bit, |
| intlvBits=dir_bits, |
| intlvMatch=i, |
| xorHighBit=xor_low_bit, |
| ) |
| |
| dir_cntrl = DirCntrl(noTCCdir=True, TCC_select_num_bits=TCC_bits) |
| dir_cntrl.create(options, [addr_range], ruby_system, system) |
| dir_cntrl.number_of_TBEs = options.num_tbes |
| dir_cntrl.useL3OnWT = False |
| |
| # Connect the Directory controller to the ruby network |
| dir_cntrl.requestFromCores = MessageBuffer(ordered=True) |
| dir_cntrl.requestFromCores.in_port = network.out_port |
| |
| dir_cntrl.responseFromCores = MessageBuffer() |
| dir_cntrl.responseFromCores.in_port = network.out_port |
| |
| dir_cntrl.unblockFromCores = MessageBuffer() |
| dir_cntrl.unblockFromCores.in_port = network.out_port |
| |
| dir_cntrl.probeToCore = MessageBuffer() |
| dir_cntrl.probeToCore.out_port = network.in_port |
| |
| dir_cntrl.responseToCore = MessageBuffer() |
| dir_cntrl.responseToCore.out_port = network.in_port |
| |
| dir_cntrl.triggerQueue = MessageBuffer(ordered=True) |
| dir_cntrl.L3triggerQueue = MessageBuffer(ordered=True) |
| dir_cntrl.requestToMemory = MessageBuffer() |
| dir_cntrl.responseFromMemory = MessageBuffer() |
| |
| dir_cntrl.requestFromDMA = MessageBuffer(ordered=True) |
| dir_cntrl.requestFromDMA.in_port = network.out_port |
| |
| dir_cntrl.responseToDMA = MessageBuffer() |
| dir_cntrl.responseToDMA.out_port = network.in_port |
| |
| dir_cntrl.requestToMemory = MessageBuffer() |
| dir_cntrl.responseFromMemory = MessageBuffer() |
| |
| # Create memory controllers too |
| mem_type = ObjectList.mem_list.get(options.dgpu_mem_type) |
| dram_intf = MemConfig.create_mem_intf( |
| mem_type, |
| gpu_mem_range, |
| i, |
| int(math.log(options.dgpu_num_dirs, 2)), |
| options.cacheline_size, |
| xor_low_bit, |
| ) |
| if issubclass(mem_type, DRAMInterface): |
| mem_ctrl = m5.objects.MemCtrl(dram=dram_intf) |
| else: |
| mem_ctrl = dram_intf |
| |
| mem_ctrl.port = dir_cntrl.memory_out_port |
| mem_ctrl.dram.enable_dram_powerdown = False |
| dir_cntrl.addr_ranges = dram_intf.range |
| |
| # Append |
| exec("system.ruby.gpu_dir_cntrl%d = dir_cntrl" % i) |
| dir_cntrl_nodes.append(dir_cntrl) |
| mem_ctrls.append(mem_ctrl) |
| |
| system.gpu_mem_ctrls = mem_ctrls |
| |
| return dir_cntrl_nodes, mem_ctrls |
| |
| |
| def construct_corepairs(options, system, ruby_system, network): |
| |
| cpu_sequencers = [] |
| cp_cntrl_nodes = [] |
| |
| for i in range((options.num_cpus + 1) // 2): |
| |
| cp_cntrl = CPCntrl() |
| cp_cntrl.create(options, ruby_system, system) |
| |
| exec("ruby_system.cp_cntrl%d = cp_cntrl" % i) |
| # |
| # Add controllers and sequencers to the appropriate lists |
| # |
| cpu_sequencers.extend([cp_cntrl.sequencer, cp_cntrl.sequencer1]) |
| |
| # Connect the CP controllers and the network |
| cp_cntrl.requestFromCore = MessageBuffer() |
| cp_cntrl.requestFromCore.out_port = network.in_port |
| |
| cp_cntrl.responseFromCore = MessageBuffer() |
| cp_cntrl.responseFromCore.out_port = network.in_port |
| |
| cp_cntrl.unblockFromCore = MessageBuffer() |
| cp_cntrl.unblockFromCore.out_port = network.in_port |
| |
| cp_cntrl.probeToCore = MessageBuffer() |
| cp_cntrl.probeToCore.in_port = network.out_port |
| |
| cp_cntrl.responseToCore = MessageBuffer() |
| cp_cntrl.responseToCore.in_port = network.out_port |
| |
| cp_cntrl.mandatoryQueue = MessageBuffer() |
| cp_cntrl.triggerQueue = MessageBuffer(ordered=True) |
| |
| cp_cntrl_nodes.append(cp_cntrl) |
| |
| return (cpu_sequencers, cp_cntrl_nodes) |
| |
| |
| def construct_tcps(options, system, ruby_system, network): |
| |
| tcp_sequencers = [] |
| tcp_cntrl_nodes = [] |
| |
| # For an odd number of CPUs, still create the right number of controllers |
| TCC_bits = int(math.log(options.num_tccs, 2)) |
| |
| for i in range(options.num_compute_units): |
| |
| tcp_cntrl = TCPCntrl( |
| TCC_select_num_bits=TCC_bits, issue_latency=1, number_of_TBEs=2560 |
| ) |
| # TBEs set to max outstanding requests |
| tcp_cntrl.create(options, ruby_system, system) |
| tcp_cntrl.WB = options.WB_L1 |
| tcp_cntrl.disableL1 = options.noL1 |
| tcp_cntrl.L1cache.tagAccessLatency = options.TCP_latency |
| tcp_cntrl.L1cache.dataAccessLatency = options.TCP_latency |
| |
| exec("ruby_system.tcp_cntrl%d = tcp_cntrl" % i) |
| # |
| # Add controllers and sequencers to the appropriate lists |
| # |
| tcp_sequencers.append(tcp_cntrl.coalescer) |
| tcp_cntrl_nodes.append(tcp_cntrl) |
| |
| # Connect the TCP controller to the ruby network |
| tcp_cntrl.requestFromTCP = MessageBuffer(ordered=True) |
| tcp_cntrl.requestFromTCP.out_port = network.in_port |
| |
| tcp_cntrl.responseFromTCP = MessageBuffer(ordered=True) |
| tcp_cntrl.responseFromTCP.out_port = network.in_port |
| |
| tcp_cntrl.unblockFromCore = MessageBuffer() |
| tcp_cntrl.unblockFromCore.out_port = network.in_port |
| |
| tcp_cntrl.probeToTCP = MessageBuffer(ordered=True) |
| tcp_cntrl.probeToTCP.in_port = network.out_port |
| |
| tcp_cntrl.responseToTCP = MessageBuffer(ordered=True) |
| tcp_cntrl.responseToTCP.in_port = network.out_port |
| |
| tcp_cntrl.mandatoryQueue = MessageBuffer() |
| |
| return (tcp_sequencers, tcp_cntrl_nodes) |
| |
| |
| def construct_sqcs(options, system, ruby_system, network): |
| |
| sqc_sequencers = [] |
| sqc_cntrl_nodes = [] |
| |
| # For an odd number of CPUs, still create the right number of controllers |
| TCC_bits = int(math.log(options.num_tccs, 2)) |
| |
| for i in range(options.num_sqc): |
| |
| sqc_cntrl = SQCCntrl(TCC_select_num_bits=TCC_bits) |
| sqc_cntrl.create(options, ruby_system, system) |
| |
| exec("ruby_system.sqc_cntrl%d = sqc_cntrl" % i) |
| # |
| # Add controllers and sequencers to the appropriate lists |
| # |
| sqc_sequencers.append(sqc_cntrl.sequencer) |
| sqc_cntrl_nodes.append(sqc_cntrl) |
| |
| # Connect the SQC controller to the ruby network |
| sqc_cntrl.requestFromSQC = MessageBuffer(ordered=True) |
| sqc_cntrl.requestFromSQC.out_port = network.in_port |
| |
| sqc_cntrl.probeToSQC = MessageBuffer(ordered=True) |
| sqc_cntrl.probeToSQC.in_port = network.out_port |
| |
| sqc_cntrl.responseToSQC = MessageBuffer(ordered=True) |
| sqc_cntrl.responseToSQC.in_port = network.out_port |
| |
| sqc_cntrl.mandatoryQueue = MessageBuffer() |
| |
| return (sqc_sequencers, sqc_cntrl_nodes) |
| |
| |
| def construct_scalars(options, system, ruby_system, network): |
| |
| scalar_sequencers = [] |
| scalar_cntrl_nodes = [] |
| |
| # For an odd number of CPUs, still create the right number of controllers |
| TCC_bits = int(math.log(options.num_tccs, 2)) |
| |
| for i in range(options.num_scalar_cache): |
| scalar_cntrl = SQCCntrl(TCC_select_num_bits=TCC_bits) |
| scalar_cntrl.create(options, ruby_system, system) |
| |
| exec("ruby_system.scalar_cntrl%d = scalar_cntrl" % i) |
| |
| scalar_sequencers.append(scalar_cntrl.sequencer) |
| scalar_cntrl_nodes.append(scalar_cntrl) |
| |
| scalar_cntrl.requestFromSQC = MessageBuffer(ordered=True) |
| scalar_cntrl.requestFromSQC.out_port = network.in_port |
| |
| scalar_cntrl.probeToSQC = MessageBuffer(ordered=True) |
| scalar_cntrl.probeToSQC.in_port = network.out_port |
| |
| scalar_cntrl.responseToSQC = MessageBuffer(ordered=True) |
| scalar_cntrl.responseToSQC.in_port = network.out_port |
| |
| scalar_cntrl.mandatoryQueue = MessageBuffer( |
| buffer_size=options.scalar_buffer_size |
| ) |
| |
| return (scalar_sequencers, scalar_cntrl_nodes) |
| |
| |
| def construct_cmdprocs(options, system, ruby_system, network): |
| |
| cmdproc_sequencers = [] |
| cmdproc_cntrl_nodes = [] |
| |
| # For an odd number of CPUs, still create the right number of controllers |
| TCC_bits = int(math.log(options.num_tccs, 2)) |
| |
| for i in range(options.num_cp): |
| |
| tcp_ID = options.num_compute_units + i |
| sqc_ID = options.num_sqc + i |
| |
| tcp_cntrl = TCPCntrl( |
| TCC_select_num_bits=TCC_bits, issue_latency=1, number_of_TBEs=2560 |
| ) |
| # TBEs set to max outstanding requests |
| tcp_cntrl.createCP(options, ruby_system, system) |
| tcp_cntrl.WB = options.WB_L1 |
| tcp_cntrl.disableL1 = options.noL1 |
| tcp_cntrl.L1cache.tagAccessLatency = options.TCP_latency |
| tcp_cntrl.L1cache.dataAccessLatency = options.TCP_latency |
| |
| exec("ruby_system.tcp_cntrl%d = tcp_cntrl" % tcp_ID) |
| # |
| # Add controllers and sequencers to the appropriate lists |
| # |
| cmdproc_sequencers.append(tcp_cntrl.sequencer) |
| cmdproc_cntrl_nodes.append(tcp_cntrl) |
| |
| # Connect the CP (TCP) controllers to the ruby network |
| tcp_cntrl.requestFromTCP = MessageBuffer(ordered=True) |
| tcp_cntrl.requestFromTCP.out_port = network.in_port |
| |
| tcp_cntrl.responseFromTCP = MessageBuffer(ordered=True) |
| tcp_cntrl.responseFromTCP.out_port = network.in_port |
| |
| tcp_cntrl.unblockFromCore = MessageBuffer(ordered=True) |
| tcp_cntrl.unblockFromCore.out_port = network.in_port |
| |
| tcp_cntrl.probeToTCP = MessageBuffer(ordered=True) |
| tcp_cntrl.probeToTCP.in_port = network.out_port |
| |
| tcp_cntrl.responseToTCP = MessageBuffer(ordered=True) |
| tcp_cntrl.responseToTCP.in_port = network.out_port |
| |
| tcp_cntrl.mandatoryQueue = MessageBuffer() |
| |
| sqc_cntrl = SQCCntrl(TCC_select_num_bits=TCC_bits) |
| sqc_cntrl.create(options, ruby_system, system) |
| |
| exec("ruby_system.sqc_cntrl%d = sqc_cntrl" % sqc_ID) |
| # |
| # Add controllers and sequencers to the appropriate lists |
| # |
| cmdproc_sequencers.append(sqc_cntrl.sequencer) |
| cmdproc_cntrl_nodes.append(sqc_cntrl) |
| |
| return (cmdproc_sequencers, cmdproc_cntrl_nodes) |
| |
| |
| def construct_tccs(options, system, ruby_system, network): |
| |
| tcc_cntrl_nodes = [] |
| |
| for i in range(options.num_tccs): |
| |
| tcc_cntrl = TCCCntrl(l2_response_latency=options.TCC_latency) |
| tcc_cntrl.create(options, ruby_system, system) |
| tcc_cntrl.l2_request_latency = options.gpu_to_dir_latency |
| tcc_cntrl.l2_response_latency = options.TCC_latency |
| tcc_cntrl_nodes.append(tcc_cntrl) |
| tcc_cntrl.WB = options.WB_L2 |
| tcc_cntrl.number_of_TBEs = 2560 * options.num_compute_units |
| # the number_of_TBEs is inclusive of TBEs below |
| |
| # Connect the TCC controllers to the ruby network |
| tcc_cntrl.requestFromTCP = MessageBuffer(ordered=True) |
| tcc_cntrl.requestFromTCP.in_port = network.out_port |
| |
| tcc_cntrl.responseToCore = MessageBuffer(ordered=True) |
| tcc_cntrl.responseToCore.out_port = network.in_port |
| |
| tcc_cntrl.probeFromNB = MessageBuffer() |
| tcc_cntrl.probeFromNB.in_port = network.out_port |
| |
| tcc_cntrl.responseFromNB = MessageBuffer() |
| tcc_cntrl.responseFromNB.in_port = network.out_port |
| |
| tcc_cntrl.requestToNB = MessageBuffer(ordered=True) |
| tcc_cntrl.requestToNB.out_port = network.in_port |
| |
| tcc_cntrl.responseToNB = MessageBuffer() |
| tcc_cntrl.responseToNB.out_port = network.in_port |
| |
| tcc_cntrl.unblockToNB = MessageBuffer() |
| tcc_cntrl.unblockToNB.out_port = network.in_port |
| |
| tcc_cntrl.triggerQueue = MessageBuffer(ordered=True) |
| |
| exec("ruby_system.tcc_cntrl%d = tcc_cntrl" % i) |
| |
| return tcc_cntrl_nodes |
| |
| |
| def create_system( |
| options, full_system, system, dma_devices, bootmem, ruby_system, cpus |
| ): |
| if buildEnv["PROTOCOL"] != "GPU_VIPER": |
| panic("This script requires the GPU_VIPER protocol to be built.") |
| |
| cpu_sequencers = [] |
| |
| # |
| # Must create the individual controllers before the network to ensure the |
| # controller constructors are called before the network constructor |
| # |
| |
| # This is the base crossbar that connects the L3s, Dirs, and cpu/gpu |
| # Clusters |
| crossbar_bw = None |
| mainCluster = None |
| cpuCluster = None |
| gpuCluster = None |
| |
| if hasattr(options, "bw_scalor") and options.bw_scalor > 0: |
| # Assuming a 2GHz clock |
| crossbar_bw = 16 * options.num_compute_units * options.bw_scalor |
| mainCluster = Cluster(intBW=crossbar_bw) |
| cpuCluster = Cluster(extBW=crossbar_bw, intBW=crossbar_bw) |
| gpuCluster = Cluster(extBW=crossbar_bw, intBW=crossbar_bw) |
| else: |
| mainCluster = Cluster(intBW=8) # 16 GB/s |
| cpuCluster = Cluster(extBW=8, intBW=8) # 16 GB/s |
| gpuCluster = Cluster(extBW=8, intBW=8) # 16 GB/s |
| |
| # Create CPU directory controllers |
| dir_cntrl_nodes = construct_dirs( |
| options, system, ruby_system, ruby_system.network |
| ) |
| for dir_cntrl in dir_cntrl_nodes: |
| mainCluster.add(dir_cntrl) |
| |
| # Create CPU core pairs |
| (cp_sequencers, cp_cntrl_nodes) = construct_corepairs( |
| options, system, ruby_system, ruby_system.network |
| ) |
| cpu_sequencers.extend(cp_sequencers) |
| for cp_cntrl in cp_cntrl_nodes: |
| cpuCluster.add(cp_cntrl) |
| |
| # Register CPUs and caches for each CorePair and directory (SE mode only) |
| if not full_system: |
| for i in range((options.num_cpus + 1) // 2): |
| FileSystemConfig.register_cpu( |
| physical_package_id=0, |
| core_siblings=range(options.num_cpus), |
| core_id=i * 2, |
| thread_siblings=[], |
| ) |
| |
| FileSystemConfig.register_cpu( |
| physical_package_id=0, |
| core_siblings=range(options.num_cpus), |
| core_id=i * 2 + 1, |
| thread_siblings=[], |
| ) |
| |
| FileSystemConfig.register_cache( |
| level=0, |
| idu_type="Instruction", |
| size=options.l1i_size, |
| line_size=options.cacheline_size, |
| assoc=options.l1i_assoc, |
| cpus=[i * 2, i * 2 + 1], |
| ) |
| |
| FileSystemConfig.register_cache( |
| level=0, |
| idu_type="Data", |
| size=options.l1d_size, |
| line_size=options.cacheline_size, |
| assoc=options.l1d_assoc, |
| cpus=[i * 2], |
| ) |
| |
| FileSystemConfig.register_cache( |
| level=0, |
| idu_type="Data", |
| size=options.l1d_size, |
| line_size=options.cacheline_size, |
| assoc=options.l1d_assoc, |
| cpus=[i * 2 + 1], |
| ) |
| |
| FileSystemConfig.register_cache( |
| level=1, |
| idu_type="Unified", |
| size=options.l2_size, |
| line_size=options.cacheline_size, |
| assoc=options.l2_assoc, |
| cpus=[i * 2, i * 2 + 1], |
| ) |
| |
| for i in range(options.num_dirs): |
| FileSystemConfig.register_cache( |
| level=2, |
| idu_type="Unified", |
| size=options.l3_size, |
| line_size=options.cacheline_size, |
| assoc=options.l3_assoc, |
| cpus=[n for n in range(options.num_cpus)], |
| ) |
| |
| # Create TCPs |
| (tcp_sequencers, tcp_cntrl_nodes) = construct_tcps( |
| options, system, ruby_system, ruby_system.network |
| ) |
| cpu_sequencers.extend(tcp_sequencers) |
| for tcp_cntrl in tcp_cntrl_nodes: |
| gpuCluster.add(tcp_cntrl) |
| |
| # Create SQCs |
| (sqc_sequencers, sqc_cntrl_nodes) = construct_sqcs( |
| options, system, ruby_system, ruby_system.network |
| ) |
| cpu_sequencers.extend(sqc_sequencers) |
| for sqc_cntrl in sqc_cntrl_nodes: |
| gpuCluster.add(sqc_cntrl) |
| |
| # Create Scalars |
| (scalar_sequencers, scalar_cntrl_nodes) = construct_scalars( |
| options, system, ruby_system, ruby_system.network |
| ) |
| cpu_sequencers.extend(scalar_sequencers) |
| for scalar_cntrl in scalar_cntrl_nodes: |
| gpuCluster.add(scalar_cntrl) |
| |
| # Create command processors |
| (cmdproc_sequencers, cmdproc_cntrl_nodes) = construct_cmdprocs( |
| options, system, ruby_system, ruby_system.network |
| ) |
| cpu_sequencers.extend(cmdproc_sequencers) |
| for cmdproc_cntrl in cmdproc_cntrl_nodes: |
| gpuCluster.add(cmdproc_cntrl) |
| |
| # Create TCCs |
| tcc_cntrl_nodes = construct_tccs( |
| options, system, ruby_system, ruby_system.network |
| ) |
| for tcc_cntrl in tcc_cntrl_nodes: |
| gpuCluster.add(tcc_cntrl) |
| |
| for i, dma_device in enumerate(dma_devices): |
| dma_seq = DMASequencer(version=i, ruby_system=ruby_system) |
| dma_cntrl = DMA_Controller( |
| version=i, dma_sequencer=dma_seq, ruby_system=ruby_system |
| ) |
| exec("system.dma_cntrl%d = dma_cntrl" % i) |
| |
| # IDE doesn't have a .type but seems like everything else does. |
| if not hasattr(dma_device, "type"): |
| exec("system.dma_cntrl%d.dma_sequencer.in_ports = dma_device" % i) |
| elif dma_device.type == "MemTest": |
| exec( |
| "system.dma_cntrl%d.dma_sequencer.in_ports = dma_devices.test" |
| % i |
| ) |
| else: |
| exec( |
| "system.dma_cntrl%d.dma_sequencer.in_ports = dma_device.dma" |
| % i |
| ) |
| |
| dma_cntrl.requestToDir = MessageBuffer(buffer_size=0) |
| dma_cntrl.requestToDir.out_port = ruby_system.network.in_port |
| dma_cntrl.responseFromDir = MessageBuffer(buffer_size=0) |
| dma_cntrl.responseFromDir.in_port = ruby_system.network.out_port |
| dma_cntrl.mandatoryQueue = MessageBuffer(buffer_size=0) |
| gpuCluster.add(dma_cntrl) |
| |
| # Add cpu/gpu clusters to main cluster |
| mainCluster.add(cpuCluster) |
| mainCluster.add(gpuCluster) |
| |
| ruby_system.network.number_of_virtual_networks = 11 |
| |
| return (cpu_sequencers, dir_cntrl_nodes, mainCluster) |