changeset 6e2408ad4425 in /z/repo/gem5
details: http://repo.gem5.org/gem5?cmd=changeset;node=6e2408ad4425
description:
arm, config: Fixups for the example big.LITTLE(tm) configuration
This patch refactors the configuration file to use a more
object-oriented design.
Change-Id: I44ac2d063c2b5901f385544fb6ce3f259459cb05
Reviewed-by: Andreas Sandberg <[email protected]>
Reviewed-by: Gabor Dozsa <[email protected]>
diffstat:
configs/example/arm/devices.py | 147 +++++++++++++++++++++++++---
configs/example/arm/fs_bigLITTLE.py | 181 ++++++++++-------------------------
2 files changed, 181 insertions(+), 147 deletions(-)
diffs (truncated from 401 to 300 lines):
diff -r 22f08c96bf7f -r 6e2408ad4425 configs/example/arm/devices.py
--- a/configs/example/arm/devices.py Tue Sep 13 23:20:03 2016 -0400
+++ b/configs/example/arm/devices.py Thu Sep 15 18:00:59 2016 +0100
@@ -42,6 +42,7 @@
from m5.objects import *
m5.util.addToPath('../../common')
from Caches import *
+import CpuConfig
class L1I(L1_ICache):
hit_latency = 1
@@ -98,33 +99,104 @@
default = Self.badaddr_responder.pio
+class CpuCluster(SubSystem):
+ def __init__(self, system, num_cpus, cpu_clock, cpu_voltage,
+ cpu_type, l1i_type, l1d_type, wcache_type, l2_type):
+ super(CpuCluster, self).__init__()
+ self._cpu_type = cpu_type
+ self._l1i_type = l1i_type
+ self._l1d_type = l1d_type
+ self._wcache_type = wcache_type
+ self._l2_type = l2_type
+
+ assert num_cpus > 0
+
+ self.voltage_domain = VoltageDomain(voltage=cpu_voltage)
+ self.clk_domain = SrcClockDomain(clock=cpu_clock,
+ voltage_domain=self.voltage_domain)
+
+ self.cpus = [ self._cpu_type(cpu_id=system.numCpus() + idx,
+ clk_domain=self.clk_domain)
+ for idx in range(num_cpus) ]
+
+ for cpu in self.cpus:
+ cpu.createThreads()
+ cpu.createInterruptController()
+ cpu.socket_id = system.numCpuClusters()
+ system.addCpuCluster(self, num_cpus)
+
+ def requireCaches(self):
+ return self._cpu_type.require_caches()
+
+ def memoryMode(self):
+ return self._cpu_type.memory_mode()
+
+ def addL1(self):
+ for cpu in self.cpus:
+ l1i = None if self._l1i_type is None else self._l1i_type()
+ l1d = None if self._l1d_type is None else self._l1d_type()
+ iwc = None if self._wcache_type is None else self._wcache_type()
+ dwc = None if self._wcache_type is None else self._wcache_type()
+ cpu.addPrivateSplitL1Caches(l1i, l1d, iwc, dwc)
+
+ def addL2(self, clk_domain):
+ if self._l2_type is None:
+ return
+ self.toL2Bus = L2XBar(width=64, clk_domain=clk_domain)
+ self.l2 = self._l2_type()
+ for cpu in self.cpus:
+ cpu.connectAllPorts(self.toL2Bus)
+ self.toL2Bus.master = self.l2.cpu_side
+
+ def connectMemSide(self, bus):
+ bus.slave
+ try:
+ self.l2.mem_side = bus.slave
+ except AttributeError:
+ for cpu in self.cpus:
+ cpu.connectAllPorts(bus)
+
+
+class AtomicCluster(CpuCluster):
+ def __init__(self, system, num_cpus, cpu_clock, cpu_voltage="1.0V"):
+ cpu_config = [ CpuConfig.get("atomic"), None, None, None, None ]
+ super(AtomicCluster, self).__init__(system, num_cpus, cpu_clock,
+ cpu_voltage, *cpu_config)
+ def addL1(self):
+ pass
+
+
class SimpleSystem(LinuxArmSystem):
cache_line_size = 64
- voltage_domain = VoltageDomain(voltage="1.0V")
- clk_domain = SrcClockDomain(clock="1GHz",
- voltage_domain=Parent.voltage_domain)
+ def __init__(self, **kwargs):
+ super(SimpleSystem, self).__init__(**kwargs)
- realview = VExpress_GEM5_V1()
+ self.voltage_domain = VoltageDomain(voltage="1.0V")
+ self.clk_domain = SrcClockDomain(clock="1GHz",
+ voltage_domain=Parent.voltage_domain)
- gic_cpu_addr = realview.gic.cpu_addr
- flags_addr = realview.realview_io.pio_addr + 0x30
+ self.realview = VExpress_GEM5_V1()
- membus = MemBus()
+ self.gic_cpu_addr = self.realview.gic.cpu_addr
+ self.flags_addr = self.realview.realview_io.pio_addr + 0x30
- intrctrl = IntrControl()
- terminal = Terminal()
- vncserver = VncServer()
+ self.membus = MemBus()
- iobus = IOXBar()
- # CPUs->PIO
- iobridge = Bridge(delay='50ns')
- # Device DMA -> MEM
- dmabridge = Bridge(delay='50ns', ranges=realview._mem_regions)
+ self.intrctrl = IntrControl()
+ self.terminal = Terminal()
+ self.vncserver = VncServer()
- _pci_devices = 0
- _clusters = []
- _cpus = []
+ self.iobus = IOXBar()
+ # CPUs->PIO
+ self.iobridge = Bridge(delay='50ns')
+ # Device DMA -> MEM
+ self.dmabridge = Bridge(delay='50ns',
+ ranges=self.realview._mem_regions)
+
+ self._pci_devices = 0
+ self._clusters = []
+ self._num_cpus = 0
def attach_pci(self, dev):
dev.pci_bus, dev.pci_dev, dev.pci_func = (0, self._pci_devices + 1, 0)
@@ -142,3 +214,42 @@
self.realview.attachOnChipIO(self.membus, self.iobridge)
self.realview.attachIO(self.iobus)
self.system_port = self.membus.slave
+
+ def numCpuClusters(self):
+ return len(self._clusters)
+
+ def addCpuCluster(self, cpu_cluster, num_cpus):
+ assert cpu_cluster not in self._clusters
+ assert num_cpus > 0
+ self._clusters.append(cpu_cluster)
+ self._num_cpus += num_cpus
+
+ def numCpus(self):
+ return self._num_cpus
+
+ def addCaches(self, need_caches, last_cache_level):
+ if not need_caches:
+ # connect each cluster to the memory hierarchy
+ for cluster in self._clusters:
+ cluster.connectMemSide(self.membus)
+ return
+
+ cluster_mem_bus = self.membus
+ assert last_cache_level >= 1 and last_cache_level <= 3
+ for cluster in self._clusters:
+ cluster.addL1()
+ if last_cache_level > 1:
+ for cluster in self._clusters:
+ cluster.addL2(cluster.clk_domain)
+ if last_cache_level > 2:
+ max_clock_cluster = max(self._clusters,
+ key=lambda c: c.clk_domain.clock[0])
+ self.l3 = L3(clk_domain=max_clock_cluster.clk_domain)
+ self.toL3Bus = L2XBar(width=64)
+ self.toL3Bus.master = self.l3.cpu_side
+ self.l3.mem_side = self.membus.slave
+ cluster_mem_bus = self.toL3Bus
+
+ # connect each cluster to the memory hierarchy
+ for cluster in self._clusters:
+ cluster.connectMemSide(cluster_mem_bus)
diff -r 22f08c96bf7f -r 6e2408ad4425 configs/example/arm/fs_bigLITTLE.py
--- a/configs/example/arm/fs_bigLITTLE.py Tue Sep 13 23:20:03 2016 -0400
+++ b/configs/example/arm/fs_bigLITTLE.py Thu Sep 15 18:00:59 2016 +0100
@@ -60,10 +60,27 @@
default_mem_size= "2GB"
-def createSystem(kernel, mem_mode, bootscript, disks=[]):
+
+class BigCluster(devices.CpuCluster):
+ def __init__(self, system, num_cpus, cpu_clock,
+ cpu_voltage="1.0V"):
+ cpu_config = [ CpuConfig.get("arm_detailed"), devices.L1I, devices.L1D,
+ devices.WalkCache, devices.L2 ]
+ super(BigCluster, self).__init__(system, num_cpus, cpu_clock,
+ cpu_voltage, *cpu_config)
+
+class LittleCluster(devices.CpuCluster):
+ def __init__(self, system, num_cpus, cpu_clock,
+ cpu_voltage="1.0V"):
+ cpu_config = [ CpuConfig.get("minor"), devices.L1I, devices.L1D,
+ devices.WalkCache, devices.L2 ]
+ super(LittleCluster, self).__init__(system, num_cpus, cpu_clock,
+ cpu_voltage, *cpu_config)
+
+
+def createSystem(kernel, bootscript, disks=[]):
sys = devices.SimpleSystem(kernel=SysPaths.binary(kernel),
readfile=bootscript,
- mem_mode=mem_mode,
machine_type="DTOnly")
mem_region = sys.realview._mem_regions[0]
@@ -91,90 +108,6 @@
return sys
-class CpuCluster(SubSystem):
- def addCPUs(self, cpu_config, num_cpus, cpu_clock, cpu_voltage="1.0V"):
- try:
- self._cluster_id
- m5.util.panic("CpuCluster.addCPUs() must be called exactly once")
- except AttributeError:
- pass
-
- assert num_cpus > 0
- system = self._parent
- self._cluster_id = len(system._clusters)
- system._clusters.append(self)
- self._config = cpu_config
-
- self.voltage_domain = VoltageDomain(voltage=cpu_voltage)
- self.clk_domain = SrcClockDomain(clock=cpu_clock,
- voltage_domain=self.voltage_domain)
-
- cpu_class = cpu_config['cpu']
- self.cpus = [ cpu_class(cpu_id=len(system._cpus) + idx,
- clk_domain=self.clk_domain)
- for idx in range(num_cpus) ]
-
- for cpu in self.cpus:
- cpu.createThreads()
- cpu.createInterruptController()
- cpu.socket_id = self._cluster_id
- system._cpus.append(cpu)
-
- def createCache(self, key):
- try:
- return self._config[key]()
- except KeyError:
- return None
-
- def addL1(self):
- self._cluster_id
- for cpu in self.cpus:
- l1i = self.createCache('l1i')
- l1d = self.createCache('l1d')
- iwc = self.createCache('wcache')
- dwc = self.createCache('wcache')
- cpu.addPrivateSplitL1Caches(l1i, l1d, iwc, dwc)
-
- def addL2(self, clk_domain):
- self._cluster_id
- self.toL2Bus = L2XBar(width=64, clk_domain=clk_domain)
- #self.toL2Bus = L2XBar(width=64, clk_domain=clk_domain,
- #snoop_filter=NULL)
- self.l2 = self._config['l2']()
- for cpu in self.cpus:
- cpu.connectAllPorts(self.toL2Bus)
- self.toL2Bus.master = self.l2.cpu_side
-
- def connectMemSide(self, bus):
- self._cluster_id
- bus.slave
- try:
- self.l2.mem_side = bus.slave
- except AttributeError:
- for cpu in self.cpus:
- cpu.connectAllPorts(bus)
-
-
-def addCaches(system, last_cache_level):
- cluster_mem_bus = system.membus
- assert last_cache_level >= 1 and last_cache_level <= 3
- for cluster in system._clusters:
- cluster.addL1()
- if last_cache_level > 1:
- for cluster in system._clusters:
- cluster.addL2(cluster.clk_domain)
- if last_cache_level > 2:
- max_clock_cluster = max(system._clusters,
- key=lambda c: c.clk_domain.clock[0])
- system.l3 = devices.L3(clk_domain=max_clock_cluster.clk_domain)
- system.toL3Bus = L2XBar(width=64)
- system.toL3Bus.master = system.l3.cpu_side
- system.l3.mem_side = system.membus.slave
- cluster_mem_bus = system.toL3Bus
-
- return cluster_mem_bus
-
-
def main():
parser = argparse.ArgumentParser(
_______________________________________________
gem5-dev mailing list
[email protected]
http://m5sim.org/mailman/listinfo/gem5-dev