Hi, I'm trying to compare the number of cache misses (D1, I1, and LL) between what Perf gives me (on the hardware itself) vs. what Cachegrind thinks the number of misses should be. The server machine has two Sandy Bridge Intel Xeon E5-2430 CPUs on it, and the PARSEC 3.0 suite (compiled in gcc-serial format, single threaded) is being run through cachegrind for analysis to obtain the number of D1, I1, and LL misses vs. the number of real misses on the hardware obtained by running the same benchmark binaries through Perf and counting D1 load and store misses as well as I1 misses. A ratio of the Perf misses to Cachegrind misses holds to about a factor of 1-2x. However, some benchmarks like ferret have a much higher number of misses on Perf than on Cachegrind. Has anyone else done analysis on Perf results vs. Cachegrind simulated results by running benchmarks on both of these? The machines are also running RedHat 5. Thanks for any information.
------------------------------------------------------------------------------ LogMeIn Rescue: Anywhere, Anytime Remote support for IT. Free Trial Remotely access PCs and mobile devices and provide instant support Improve your efficiency, and focus on delivering more value-add services Discover what IT Professionals Know. Rescue delivers http://p.sf.net/sfu/logmein_12329d2d _______________________________________________ Valgrind-users mailing list Valgrind-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/valgrind-users