| # Copyright (c) 2010-2015 Advanced Micro Devices, Inc. |
| # All rights reserved. |
| # |
| # For use for simulation and test purposes only |
| # |
| # Redistribution and use in source and binary forms, with or without |
| # modification, are permitted provided that the following conditions are met: |
| # |
| # 1. Redistributions of source code must retain the above copyright notice, |
| # this list of conditions and the following disclaimer. |
| # |
| # 2. Redistributions in binary form must reproduce the above copyright notice, |
| # this list of conditions and the following disclaimer in the documentation |
| # and/or other materials provided with the distribution. |
| # |
| # 3. Neither the name of the copyright holder nor the names of its |
| # contributors may be used to endorse or promote products derived from this |
| # software without specific prior written permission. |
| # |
| # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
| # AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
| # ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE |
| # LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
| # CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
| # SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
| # INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
| # CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
| # ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
| # POSSIBILITY OF SUCH DAMAGE. |
| # |
| # Authors: Lisa Hsu |
| |
| import math |
| import m5 |
| from m5.objects import * |
| from m5.defines import buildEnv |
| from m5.util import addToPath |
| from Ruby import create_topology |
| from Ruby import send_evicts |
| from common import FileSystemConfig |
| |
| addToPath('../') |
| |
| from topologies.Cluster import Cluster |
| from topologies.Crossbar import Crossbar |
| |
| class CntrlBase: |
| _seqs = 0 |
| @classmethod |
| def seqCount(cls): |
| # Use SeqCount not class since we need global count |
| CntrlBase._seqs += 1 |
| return CntrlBase._seqs - 1 |
| |
| _cntrls = 0 |
| @classmethod |
| def cntrlCount(cls): |
| # Use CntlCount not class since we need global count |
| CntrlBase._cntrls += 1 |
| return CntrlBase._cntrls - 1 |
| |
| _version = 0 |
| @classmethod |
| def versionCount(cls): |
| cls._version += 1 # Use count for this particular type |
| return cls._version - 1 |
| |
| class L1DCache(RubyCache): |
| resourceStalls = False |
| def create(self, options): |
| self.size = MemorySize(options.l1d_size) |
| self.assoc = options.l1d_assoc |
| self.replacement_policy = PseudoLRUReplacementPolicy() |
| |
| class L1ICache(RubyCache): |
| resourceStalls = False |
| def create(self, options): |
| self.size = MemorySize(options.l1i_size) |
| self.assoc = options.l1i_assoc |
| self.replacement_policy = PseudoLRUReplacementPolicy() |
| |
| class L2Cache(RubyCache): |
| resourceStalls = False |
| def create(self, options): |
| self.size = MemorySize(options.l2_size) |
| self.assoc = options.l2_assoc |
| self.replacement_policy = PseudoLRUReplacementPolicy() |
| |
| class CPCntrl(CorePair_Controller, CntrlBase): |
| |
| def create(self, options, ruby_system, system): |
| self.version = self.versionCount() |
| |
| self.L1Icache = L1ICache() |
| self.L1Icache.create(options) |
| self.L1D0cache = L1DCache() |
| self.L1D0cache.create(options) |
| self.L1D1cache = L1DCache() |
| self.L1D1cache.create(options) |
| self.L2cache = L2Cache() |
| self.L2cache.create(options) |
| |
| self.sequencer = RubySequencer() |
| self.sequencer.icache_hit_latency = 2 |
| self.sequencer.dcache_hit_latency = 2 |
| self.sequencer.version = self.seqCount() |
| self.sequencer.icache = self.L1Icache |
| self.sequencer.dcache = self.L1D0cache |
| self.sequencer.ruby_system = ruby_system |
| self.sequencer.coreid = 0 |
| self.sequencer.is_cpu_sequencer = True |
| |
| self.sequencer1 = RubySequencer() |
| self.sequencer1.version = self.seqCount() |
| self.sequencer1.icache = self.L1Icache |
| self.sequencer1.dcache = self.L1D1cache |
| self.sequencer1.icache_hit_latency = 2 |
| self.sequencer1.dcache_hit_latency = 2 |
| self.sequencer1.ruby_system = ruby_system |
| self.sequencer1.coreid = 1 |
| self.sequencer1.is_cpu_sequencer = True |
| |
| self.issue_latency = options.cpu_to_dir_latency |
| self.send_evictions = send_evicts(options) |
| |
| self.ruby_system = ruby_system |
| |
| if options.recycle_latency: |
| self.recycle_latency = options.recycle_latency |
| |
| class L3Cache(RubyCache): |
| assoc = 8 |
| dataArrayBanks = 256 |
| tagArrayBanks = 256 |
| |
| def create(self, options, ruby_system, system): |
| self.size = MemorySize(options.l3_size) |
| self.size.value /= options.num_dirs |
| self.dataArrayBanks /= options.num_dirs |
| self.tagArrayBanks /= options.num_dirs |
| self.dataArrayBanks /= options.num_dirs |
| self.tagArrayBanks /= options.num_dirs |
| self.dataAccessLatency = options.l3_data_latency |
| self.tagAccessLatency = options.l3_tag_latency |
| self.resourceStalls = options.no_resource_stalls |
| self.replacement_policy = PseudoLRUReplacementPolicy() |
| |
| class L3Cntrl(L3Cache_Controller, CntrlBase): |
| def create(self, options, ruby_system, system): |
| self.version = self.versionCount() |
| self.L3cache = L3Cache() |
| self.L3cache.create(options, ruby_system, system) |
| |
| self.l3_response_latency = max(self.L3cache.dataAccessLatency, |
| self.L3cache.tagAccessLatency) |
| self.ruby_system = ruby_system |
| |
| if options.recycle_latency: |
| self.recycle_latency = options.recycle_latency |
| |
| def connectWireBuffers(self, req_to_dir, resp_to_dir, l3_unblock_to_dir, |
| req_to_l3, probe_to_l3, resp_to_l3): |
| self.reqToDir = req_to_dir |
| self.respToDir = resp_to_dir |
| self.l3UnblockToDir = l3_unblock_to_dir |
| self.reqToL3 = req_to_l3 |
| self.probeToL3 = probe_to_l3 |
| self.respToL3 = resp_to_l3 |
| |
| class DirCntrl(Directory_Controller, CntrlBase): |
| def create(self, options, dir_ranges, ruby_system, system): |
| self.version = self.versionCount() |
| |
| self.response_latency = 30 |
| |
| self.addr_ranges = dir_ranges |
| self.directory = RubyDirectoryMemory() |
| |
| self.L3CacheMemory = L3Cache() |
| self.L3CacheMemory.create(options, ruby_system, system) |
| |
| self.l3_hit_latency = max(self.L3CacheMemory.dataAccessLatency, |
| self.L3CacheMemory.tagAccessLatency) |
| |
| self.number_of_TBEs = options.num_tbes |
| |
| self.ruby_system = ruby_system |
| |
| if options.recycle_latency: |
| self.recycle_latency = options.recycle_latency |
| |
| self.CPUonly = True |
| |
| def connectWireBuffers(self, req_to_dir, resp_to_dir, l3_unblock_to_dir, |
| req_to_l3, probe_to_l3, resp_to_l3): |
| self.reqToDir = req_to_dir |
| self.respToDir = resp_to_dir |
| self.l3UnblockToDir = l3_unblock_to_dir |
| self.reqToL3 = req_to_l3 |
| self.probeToL3 = probe_to_l3 |
| self.respToL3 = resp_to_l3 |
| |
| def define_options(parser): |
| parser.add_option("--num-subcaches", type="int", default=4) |
| parser.add_option("--l3-data-latency", type="int", default=20) |
| parser.add_option("--l3-tag-latency", type="int", default=15) |
| parser.add_option("--cpu-to-dir-latency", type="int", default=15) |
| parser.add_option("--no-resource-stalls", action="store_false", |
| default=True) |
| parser.add_option("--num-tbes", type="int", default=256) |
| parser.add_option("--l2-latency", type="int", default=50) # load to use |
| |
| def create_system(options, full_system, system, dma_devices, bootmem, |
| ruby_system): |
| if buildEnv['PROTOCOL'] != 'MOESI_AMD_Base': |
| panic("This script requires the MOESI_AMD_Base protocol.") |
| |
| cpu_sequencers = [] |
| |
| # |
| # The ruby network creation expects the list of nodes in the system to |
| # be consistent with the NetDest list. Therefore the l1 controller |
| # nodes must be listed before the directory nodes and directory nodes |
| # before dma nodes, etc. |
| # |
| l1_cntrl_nodes = [] |
| l3_cntrl_nodes = [] |
| dir_cntrl_nodes = [] |
| |
| control_count = 0 |
| |
| # |
| # Must create the individual controllers before the network to ensure |
| # the controller constructors are called before the network constructor |
| # |
| |
| # This is the base crossbar that connects the L3s, Dirs, and cpu |
| # Cluster |
| mainCluster = Cluster(extBW = 512, intBW = 512) # 1 TB/s |
| |
| if options.numa_high_bit: |
| numa_bit = options.numa_high_bit |
| else: |
| # if the numa_bit is not specified, set the directory bits as the |
| # lowest bits above the block offset bits, and the numa_bit as the |
| # highest of those directory bits |
| dir_bits = int(math.log(options.num_dirs, 2)) |
| block_size_bits = int(math.log(options.cacheline_size, 2)) |
| numa_bit = block_size_bits + dir_bits - 1 |
| |
| for i in range(options.num_dirs): |
| dir_ranges = [] |
| for r in system.mem_ranges: |
| addr_range = m5.objects.AddrRange(r.start, size = r.size(), |
| intlvHighBit = numa_bit, |
| intlvBits = dir_bits, |
| intlvMatch = i) |
| dir_ranges.append(addr_range) |
| |
| |
| dir_cntrl = DirCntrl(TCC_select_num_bits = 0) |
| dir_cntrl.create(options, dir_ranges, ruby_system, system) |
| |
| # Connect the Directory controller to the ruby network |
| dir_cntrl.requestFromCores = MessageBuffer(ordered = True) |
| dir_cntrl.requestFromCores.slave = ruby_system.network.master |
| |
| dir_cntrl.responseFromCores = MessageBuffer() |
| dir_cntrl.responseFromCores.slave = ruby_system.network.master |
| |
| dir_cntrl.unblockFromCores = MessageBuffer() |
| dir_cntrl.unblockFromCores.slave = ruby_system.network.master |
| |
| dir_cntrl.probeToCore = MessageBuffer() |
| dir_cntrl.probeToCore.master = ruby_system.network.slave |
| |
| dir_cntrl.responseToCore = MessageBuffer() |
| dir_cntrl.responseToCore.master = ruby_system.network.slave |
| |
| dir_cntrl.triggerQueue = MessageBuffer(ordered = True) |
| dir_cntrl.L3triggerQueue = MessageBuffer(ordered = True) |
| dir_cntrl.responseFromMemory = MessageBuffer() |
| |
| exec("system.dir_cntrl%d = dir_cntrl" % i) |
| dir_cntrl_nodes.append(dir_cntrl) |
| |
| mainCluster.add(dir_cntrl) |
| |
| # Technically this config can support an odd number of cpus, but the top |
| # level config files, such as the ruby_random_tester, will get confused if |
| # the number of cpus does not equal the number of sequencers. Thus make |
| # sure that an even number of cpus is specified. |
| assert((options.num_cpus % 2) == 0) |
| |
| # For an odd number of CPUs, still create the right number of controllers |
| cpuCluster = Cluster(extBW = 512, intBW = 512) # 1 TB/s |
| for i in range((options.num_cpus + 1) // 2): |
| |
| cp_cntrl = CPCntrl() |
| cp_cntrl.create(options, ruby_system, system) |
| |
| exec("system.cp_cntrl%d = cp_cntrl" % i) |
| # |
| # Add controllers and sequencers to the appropriate lists |
| # |
| cpu_sequencers.extend([cp_cntrl.sequencer, cp_cntrl.sequencer1]) |
| |
| # Connect the CP controllers and the network |
| cp_cntrl.requestFromCore = MessageBuffer() |
| cp_cntrl.requestFromCore.master = ruby_system.network.slave |
| |
| cp_cntrl.responseFromCore = MessageBuffer() |
| cp_cntrl.responseFromCore.master = ruby_system.network.slave |
| |
| cp_cntrl.unblockFromCore = MessageBuffer() |
| cp_cntrl.unblockFromCore.master = ruby_system.network.slave |
| |
| cp_cntrl.probeToCore = MessageBuffer() |
| cp_cntrl.probeToCore.slave = ruby_system.network.master |
| |
| cp_cntrl.responseToCore = MessageBuffer() |
| cp_cntrl.responseToCore.slave = ruby_system.network.master |
| |
| cp_cntrl.mandatoryQueue = MessageBuffer() |
| cp_cntrl.triggerQueue = MessageBuffer(ordered = True) |
| |
| cpuCluster.add(cp_cntrl) |
| |
| # Register CPUs and caches for each CorePair and directory (SE mode only) |
| if not full_system: |
| FileSystemConfig.config_filesystem(options) |
| for i in xrange((options.num_cpus + 1) // 2): |
| FileSystemConfig.register_cpu(physical_package_id = 0, |
| core_siblings = |
| xrange(options.num_cpus), |
| core_id = i*2, |
| thread_siblings = []) |
| |
| FileSystemConfig.register_cpu(physical_package_id = 0, |
| core_siblings = |
| xrange(options.num_cpus), |
| core_id = i*2+1, |
| thread_siblings = []) |
| |
| FileSystemConfig.register_cache(level = 0, |
| idu_type = 'Instruction', |
| size = options.l1i_size, |
| line_size = options.cacheline_size, |
| assoc = options.l1i_assoc, |
| cpus = [i*2, i*2+1]) |
| |
| FileSystemConfig.register_cache(level = 0, |
| idu_type = 'Data', |
| size = options.l1d_size, |
| line_size = options.cacheline_size, |
| assoc = options.l1d_assoc, |
| cpus = [i*2]) |
| |
| FileSystemConfig.register_cache(level = 0, |
| idu_type = 'Data', |
| size = options.l1d_size, |
| line_size = options.cacheline_size, |
| assoc = options.l1d_assoc, |
| cpus = [i*2+1]) |
| |
| FileSystemConfig.register_cache(level = 1, |
| idu_type = 'Unified', |
| size = options.l2_size, |
| line_size = options.cacheline_size, |
| assoc = options.l2_assoc, |
| cpus = [i*2, i*2+1]) |
| |
| for i in range(options.num_dirs): |
| FileSystemConfig.register_cache(level = 2, |
| idu_type = 'Unified', |
| size = options.l3_size, |
| line_size = options.cacheline_size, |
| assoc = options.l3_assoc, |
| cpus = [n for n in |
| xrange(options.num_cpus)]) |
| |
| # Assuming no DMA devices |
| assert(len(dma_devices) == 0) |
| |
| # Add cpu/gpu clusters to main cluster |
| mainCluster.add(cpuCluster) |
| |
| ruby_system.network.number_of_virtual_networks = 10 |
| |
| return (cpu_sequencers, dir_cntrl_nodes, mainCluster) |