Hi, I noticed that the default memory size set in Benchmarks.py is 128MB, isn't it too small for reasonable simulations?
Previously when I was using ALPHA_SE, the physmem is set to "2GB", and the simulation ran well. In FS mode, however, if 2GB is used, booting up Linux (with atomic CPU) becomes extremely slow; if 1GB or 512MB is used, I can boot up the OS, start the program and make a checkpoint successfully. However, restoring from the checkpoint directly with detailed CPU (--detailed) gives me "segmentation fault", the interesting thing is: if I restore the checkpoint with atomic CPU and then switch to timing and detailed ones (--standard-switch), the simulation runs well. For the default value 128MB, both --detailed and --standard-switch can run. I am confused by this observation. Am I missing anything here? What is a reasonable memory size in FS mode (say, for PARSEC programs)? Thanks, Lide
_______________________________________________ m5-users mailing list [email protected] http://m5sim.org/cgi-bin/mailman/listinfo/m5-users
