Hi, I'm trying to compare the number of cache misses (D1, I1, and LL)
between what Perf gives me (on the hardware itself) vs. what
Cachegrind thinks the number of misses should be. The server machine
has two Sandy Bridge Intel Xeon E5-2430 CPUs on it, and the PARSEC 3.0
suite (compiled in gcc-serial format, single threaded) is being run
through cachegrind for analysis to obtain the number of D1, I1, and LL
misses vs. the number of real misses on the hardware obtained by
running the same benchmark binaries through Perf and counting D1 load
and store misses as well as I1 misses. A ratio of the Perf misses to
Cachegrind misses holds to about a factor of 1-2x. However, some
benchmarks like ferret have a much higher number of misses on Perf
than on Cachegrind. Has anyone else done analysis on Perf results vs.
Cachegrind simulated results by running benchmarks on both of these?
The machines are also running RedHat 5. Thanks for any information.

------------------------------------------------------------------------------
LogMeIn Rescue: Anywhere, Anytime Remote support for IT. Free Trial
Remotely access PCs and mobile devices and provide instant support
Improve your efficiency, and focus on delivering more value-add services
Discover what IT Professionals Know. Rescue delivers
http://p.sf.net/sfu/logmein_12329d2d
_______________________________________________
Valgrind-users mailing list
Valgrind-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/valgrind-users

Reply via email to