FYI, we noticed the below changes on git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master commit d4573c3e1c992668f5dcd57d1c2ced56ae9650b9 ("sched: Improve load balancing in the presence of idle CPUs")
========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/test: nhm-white/unixbench/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/execl commit: dfbca41f347997e57048a53755611c8e2d792924 d4573c3e1c992668f5dcd57d1c2ced56ae9650b9 dfbca41f347997e5 d4573c3e1c992668f5dcd57d1c ---------------- -------------------------- %stddev %change %stddev \ | \ 4725 ± 0% -2.7% 4599 ± 0% unixbench.score 2123335 ± 0% -1.7% 2087061 ± 0% unixbench.time.involuntary_context_switches 99575417 ± 0% -2.3% 97252046 ± 0% unixbench.time.minor_page_faults 317.00 ± 0% -2.2% 310.00 ± 0% unixbench.time.percent_of_cpu_this_job_got 515.93 ± 0% -2.3% 504.21 ± 0% unixbench.time.system_time 450501 ± 0% -4.9% 428319 ± 0% unixbench.time.voluntary_context_switches 301368 ± 0% -11.4% 267086 ± 0% softirqs.SCHED 49172197 ± 0% -10.0% 44274425 ± 0% cpuidle.C1E-NHM.time 613281 ± 0% -17.6% 505485 ± 0% cpuidle.C1E-NHM.usage 232695 ± 5% -12.9% 202734 ± 1% latency_stats.sum.call_rwsem_down_write_failed.unlink_file_vma.free_pgtables.unmap_region.do_munmap.mmap_region.do_mmap_pgoff.vm_mmap_pgoff.SyS_mmap_pgoff.SyS_mmap.system_call_fastpath 470921 ± 2% -10.7% 420710 ± 1% latency_stats.sum.call_rwsem_down_write_failed.vma_adjust.__split_vma.split_vma.mprotect_fixup.SyS_mprotect.system_call_fastpath 42.74 ± 0% -2.2% 41.81 ± 0% turbostat.%Busy 1232 ± 0% -2.2% 1205 ± 0% turbostat.Avg_MHz 45.75 ± 33% +100.0% 91.50 ± 17% sched_debug.cfs_rq[1]:/.load 1788 ± 22% -64.0% 644.25 ± 61% sched_debug.cfs_rq[3]:/.blocked_load_avg 1950 ± 22% -62.9% 724.00 ± 55% sched_debug.cfs_rq[3]:/.tg_load_contrib -315.00 ± -5% +13.1% -356.25 ± -5% sched_debug.cpu#0.nr_uninterruptible 69.00 ± 6% +12.3% 77.50 ± 4% sched_debug.cpu#1.cpu_load[3] 45.75 ± 33% +100.0% 91.50 ± 17% sched_debug.cpu#1.load 449022 ± 6% +15.4% 518171 ± 4% sched_debug.cpu#2.avg_idle 624.00 ± 62% +115.9% 1347 ± 26% sched_debug.cpu#2.curr->pid -403.75 ± -4% +26.9% -512.25 ± -9% sched_debug.cpu#2.nr_uninterruptible -433.00 ± -4% +18.0% -511.00 ±-11% sched_debug.cpu#3.nr_uninterruptible 315.50 ± 6% +31.5% 415.00 ± 9% sched_debug.cpu#4.nr_uninterruptible 399.00 ± 4% +18.2% 471.75 ± 7% sched_debug.cpu#5.nr_uninterruptible 407.50 ± 0% +18.6% 483.25 ± 4% sched_debug.cpu#6.nr_uninterruptible 402.00 ± 8% +20.8% 485.50 ± 2% sched_debug.cpu#7.nr_uninterruptible ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test: lituya/unixbench/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/execl commit: dfbca41f347997e57048a53755611c8e2d792924 d4573c3e1c992668f5dcd57d1c2ced56ae9650b9 dfbca41f347997e5 d4573c3e1c992668f5dcd57d1c ---------------- -------------------------- %stddev %change %stddev \ | \ 10886 ± 0% -5.9% 10249 ± 0% unixbench.score 4700905 ± 0% -4.9% 4468392 ± 0% unixbench.time.involuntary_context_switches 2.16e+08 ± 0% -5.4% 2.044e+08 ± 0% unixbench.time.minor_page_faults 554.50 ± 0% -9.9% 499.50 ± 2% unixbench.time.percent_of_cpu_this_job_got 902.06 ± 0% -8.1% 828.84 ± 0% unixbench.time.system_time 192.66 ± 0% -7.6% 177.94 ± 0% unixbench.time.user_time 2695111 ± 0% -10.9% 2400967 ± 3% unixbench.time.voluntary_context_switches 525929 ± 0% -25.9% 389861 ± 1% softirqs.SCHED 2695111 ± 0% -10.9% 2400967 ± 3% time.voluntary_context_switches 121703 ± 0% -6.6% 113648 ± 2% vmstat.system.cs 26498 ± 0% -6.2% 24852 ± 2% vmstat.system.in 6429895 ± 0% -9.8% 5800792 ± 0% cpuidle.C1-HSW.usage 1927178 ± 0% -20.6% 1529231 ± 1% cpuidle.C1E-HSW.usage 50019600 ± 3% +92.3% 96177564 ± 4% cpuidle.C3-HSW.time 721114 ± 2% +40.6% 1013620 ± 4% cpuidle.C3-HSW.usage 928033 ± 2% +13.8% 1056036 ± 2% cpuidle.C6-HSW.usage 36.80 ± 0% -9.6% 33.28 ± 2% turbostat.%Busy 1214 ± 0% -9.5% 1098 ± 2% turbostat.Avg_MHz 0.07 ± 14% +85.7% 0.13 ± 35% turbostat.Pkg%pc2 3.68 ± 10% +85.5% 6.83 ± 35% turbostat.Pkg%pc6 53.85 ± 0% -4.2% 51.61 ± 2% turbostat.PkgWatt 0.00 ± -1% +Inf% 4318915 ±150% latency_stats.avg.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.new_sync_write.vfs_write.SyS_write.system_call_fastpath 11715 ±101% +3928.9% 471985 ±157% latency_stats.avg.wait_on_page_bit.filemap_fdatawait_range.filemap_fdatawait.sync_inodes_sb.sync_inodes_one_sb.iterate_supers.sys_sync.system_call_fastpath 0.00 ± -1% +Inf% 4717643 ±134% latency_stats.max.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.new_sync_write.vfs_write.SyS_write.system_call_fastpath 91245 ±109% +4137.1% 3866124 ±152% latency_stats.max.wait_on_page_bit.filemap_fdatawait_range.filemap_fdatawait.sync_inodes_sb.sync_inodes_one_sb.iterate_supers.sys_sync.system_call_fastpath 0.00 ± -1% +Inf% 4757271 ±133% latency_stats.sum.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.new_sync_write.vfs_write.SyS_write.system_call_fastpath 19114 ±158% -71.4% 5475 ± 67% latency_stats.sum.sync_inodes_sb.sync_inodes_one_sb.iterate_supers.sys_sync.system_call_fastpath 109615 ± 98% +3435.2% 3875132 ±152% latency_stats.sum.wait_on_page_bit.filemap_fdatawait_range.filemap_fdatawait.sync_inodes_sb.sync_inodes_one_sb.iterate_supers.sys_sync.system_call_fastpath 22954 ± 4% -10.8% 20484 ± 9% sched_debug.cfs_rq[0]:/.exec_clock 306157 ± 0% -13.1% 266065 ± 5% sched_debug.cfs_rq[0]:/.min_vruntime 20641 ± 11% -27.2% 15036 ± 20% sched_debug.cfs_rq[10]:/.avg->runnable_avg_sum 19987 ± 4% -11.2% 17751 ± 6% sched_debug.cfs_rq[10]:/.exec_clock 300899 ± 0% -12.5% 263419 ± 5% sched_debug.cfs_rq[10]:/.min_vruntime 452.50 ± 11% -27.6% 327.75 ± 20% sched_debug.cfs_rq[10]:/.tg_runnable_contrib 300940 ± 0% -11.9% 265047 ± 5% sched_debug.cfs_rq[11]:/.min_vruntime -5222 ±-16% -80.4% -1022 ±-175% sched_debug.cfs_rq[11]:/.spread0 19789 ± 11% -21.2% 15590 ± 16% sched_debug.cfs_rq[12]:/.avg->runnable_avg_sum 20790 ± 3% -14.4% 17801 ± 6% sched_debug.cfs_rq[12]:/.exec_clock 302961 ± 0% -13.0% 263467 ± 5% sched_debug.cfs_rq[12]:/.min_vruntime 432.75 ± 11% -21.3% 340.75 ± 17% sched_debug.cfs_rq[12]:/.tg_runnable_contrib 20451 ± 5% -12.8% 17830 ± 6% sched_debug.cfs_rq[13]:/.exec_clock 302744 ± 1% -12.7% 264381 ± 5% sched_debug.cfs_rq[13]:/.min_vruntime 1.75 ± 47% +171.4% 4.75 ± 40% sched_debug.cfs_rq[13]:/.nr_spread_over 19559 ± 0% -11.0% 17407 ± 4% sched_debug.cfs_rq[14]:/.exec_clock 300081 ± 0% -12.3% 263170 ± 4% sched_debug.cfs_rq[14]:/.min_vruntime -6082 ±-15% -52.3% -2900 ±-60% sched_debug.cfs_rq[14]:/.spread0 300413 ± 0% -11.7% 265326 ± 4% sched_debug.cfs_rq[15]:/.min_vruntime -5751 ±-18% -87.0% -745.73 ±-366% sched_debug.cfs_rq[15]:/.spread0 303798 ± 0% -13.1% 264112 ± 4% sched_debug.cfs_rq[1]:/.min_vruntime 84.00 ±100% +229.5% 276.75 ± 63% sched_debug.cfs_rq[1]:/.utilization_load_avg 302656 ± 0% -12.1% 266065 ± 5% sched_debug.cfs_rq[2]:/.min_vruntime -3502 ±-17% -100.0% -1.05 ±-196837% sched_debug.cfs_rq[2]:/.spread0 19933 ± 3% -13.5% 17238 ± 3% sched_debug.cfs_rq[3]:/.exec_clock 305206 ± 0% -12.8% 265991 ± 4% sched_debug.cfs_rq[3]:/.min_vruntime 20973 ± 6% -24.6% 15805 ± 18% sched_debug.cfs_rq[4]:/.avg->runnable_avg_sum 20068 ± 5% -11.5% 17767 ± 5% sched_debug.cfs_rq[4]:/.exec_clock 305752 ± 0% -12.9% 266399 ± 5% sched_debug.cfs_rq[4]:/.min_vruntime 461.50 ± 6% -25.0% 346.25 ± 18% sched_debug.cfs_rq[4]:/.tg_runnable_contrib 303317 ± 0% -12.6% 264993 ± 5% sched_debug.cfs_rq[5]:/.min_vruntime -2842 ±-35% -62.2% -1073 ±-105% sched_debug.cfs_rq[5]:/.spread0 20814 ± 9% -21.2% 16410 ± 26% sched_debug.cfs_rq[6]:/.avg->runnable_avg_sum 19473 ± 0% -10.9% 17351 ± 4% sched_debug.cfs_rq[6]:/.exec_clock 304159 ± 0% -12.7% 265678 ± 5% sched_debug.cfs_rq[6]:/.min_vruntime 455.75 ± 9% -21.0% 360.00 ± 26% sched_debug.cfs_rq[6]:/.tg_runnable_contrib 304471 ± 0% -11.9% 268359 ± 4% sched_debug.cfs_rq[7]:/.min_vruntime 298485 ± 0% -13.3% 258901 ± 3% sched_debug.cfs_rq[8]:/.min_vruntime 18.00 ± 29% +356.9% 82.25 ± 51% sched_debug.cfs_rq[8]:/.runnable_load_avg 231.00 ± 30% +116.6% 500.25 ± 20% sched_debug.cfs_rq[8]:/.utilization_load_avg 19913 ± 3% -13.2% 17285 ± 4% sched_debug.cfs_rq[9]:/.exec_clock 14.00 ± 35% +250.0% 49.00 ± 39% sched_debug.cfs_rq[9]:/.load 300167 ± 0% -13.1% 260832 ± 4% sched_debug.cfs_rq[9]:/.min_vruntime -1710 ± -2% -23.6% -1306 ±-13% sched_debug.cpu#0.nr_uninterruptible 158575 ± 6% -11.5% 140360 ± 1% sched_debug.cpu#1.ttwu_count 553112 ± 72% -45.1% 303486 ± 5% sched_debug.cpu#10.nr_switches 554861 ± 72% -45.1% 304840 ± 5% sched_debug.cpu#10.sched_count 27.00 ± 10% -20.4% 21.50 ± 13% sched_debug.cpu#12.cpu_load[4] 186234 ± 90% -58.3% 77602 ± 7% sched_debug.cpu#12.ttwu_local 20.50 ± 21% +48.8% 30.50 ± 18% sched_debug.cpu#14.cpu_load[1] 21.50 ± 21% +58.1% 34.00 ± 19% sched_debug.cpu#14.cpu_load[2] 22.75 ± 14% +44.0% 32.75 ± 17% sched_debug.cpu#14.cpu_load[3] 109416 ± 5% -13.7% 94393 ± 4% sched_debug.cpu#14.sched_goidle 130439 ± 2% -12.4% 114236 ± 4% sched_debug.cpu#14.ttwu_count 30.50 ±101% +301.6% 122.50 ± 25% sched_debug.cpu#2.cpu_load[0] 27.25 ± 47% +209.2% 84.25 ± 34% sched_debug.cpu#2.cpu_load[1] 37.50 ± 27% -32.7% 25.25 ± 21% sched_debug.cpu#4.cpu_load[4] 17.25 ± 43% +362.3% 79.75 ± 56% sched_debug.cpu#4.load 57816 ± 15% -15.3% 48950 ± 2% sched_debug.cpu#4.nr_load_updates 1220774 ±121% -73.1% 328817 ± 6% sched_debug.cpu#4.nr_switches 1222807 ±121% -73.0% 330486 ± 6% sched_debug.cpu#4.sched_count 542794 ±133% -78.8% 115250 ± 5% sched_debug.cpu#4.sched_goidle 588622 ±125% -76.2% 140136 ± 7% sched_debug.cpu#4.ttwu_count 522443 ±142% -84.6% 80605 ± 6% sched_debug.cpu#4.ttwu_local 345322 ± 1% -7.4% 319747 ± 4% sched_debug.cpu#7.nr_switches 347310 ± 1% -7.5% 321411 ± 4% sched_debug.cpu#7.sched_count 120552 ± 2% -9.4% 109240 ± 4% sched_debug.cpu#7.sched_goidle 14.00 ± 35% +233.9% 46.75 ± 36% sched_debug.cpu#9.load 136971 ± 10% -15.1% 116346 ± 1% sched_debug.cpu#9.ttwu_count 0.14 ± 57% +411.4% 0.72 ± 73% sched_debug.rt_rq[1]:/.rt_time ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test: lituya/unixbench/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/powersave/execl commit: dfbca41f347997e57048a53755611c8e2d792924 d4573c3e1c992668f5dcd57d1c2ced56ae9650b9 dfbca41f347997e5 d4573c3e1c992668f5dcd57d1c ---------------- -------------------------- %stddev %change %stddev \ | \ 10563 ± 1% -5.8% 9952 ± 0% unixbench.score 4414071 ± 1% -5.2% 4184851 ± 0% unixbench.time.involuntary_context_switches 2.028e+08 ± 1% -5.5% 1.917e+08 ± 0% unixbench.time.minor_page_faults 540.75 ± 1% -7.2% 502.00 ± 0% unixbench.time.percent_of_cpu_this_job_got 882.57 ± 1% -7.4% 816.95 ± 0% unixbench.time.system_time 188.00 ± 1% -7.0% 174.82 ± 0% unixbench.time.user_time 2858074 ± 3% -8.8% 2605521 ± 0% unixbench.time.voluntary_context_switches 511610 ± 1% -25.5% 381032 ± 0% softirqs.SCHED 2858074 ± 3% -8.8% 2605521 ± 0% time.voluntary_context_switches 118783 ± 0% -4.6% 113276 ± 0% vmstat.system.cs 25883 ± 0% -4.3% 24765 ± 0% vmstat.system.in 683238 ± 2% -9.4% 619110 ± 0% latency_stats.hits.call_rwsem_down_write_failed.vma_adjust.__split_vma.split_vma.mprotect_fixup.SyS_mprotect.system_call_fastpath 96234 ± 5% -11.9% 84756 ± 1% latency_stats.hits.wait_on_page_bit_killable.__lock_page_or_retry.filemap_fault.__do_fault.handle_pte_fault.handle_mm_fault.__do_page_fault.do_page_fault.page_fault 191410 ± 5% -13.4% 165713 ± 1% latency_stats.hits.wait_on_page_bit_killable.__lock_page_or_retry.filemap_fault.__do_fault.handle_pte_fault.handle_mm_fault.__do_page_fault.do_page_fault.page_fault.clear_user.padzero.load_elf_binary 36.16 ± 1% -6.9% 33.68 ± 0% turbostat.%Busy 1148 ± 1% -7.1% 1066 ± 0% turbostat.Avg_MHz 50.93 ± 0% -1.4% 50.21 ± 0% turbostat.PkgWatt 1928029 ± 0% -20.0% 1543361 ± 0% cpuidle.C1E-HSW.usage 55003420 ± 8% +78.9% 98419263 ± 1% cpuidle.C3-HSW.time 799437 ± 7% +33.2% 1064466 ± 1% cpuidle.C3-HSW.usage 873657 ± 3% +11.9% 977668 ± 1% cpuidle.C6-HSW.usage 19945 ± 2% -7.9% 18369 ± 0% sched_debug.cfs_rq[13]:/.exec_clock 21825 ± 10% -18.4% 17818 ± 1% sched_debug.cfs_rq[14]:/.avg->runnable_avg_sum 478.50 ± 10% -18.6% 389.50 ± 1% sched_debug.cfs_rq[14]:/.tg_runnable_contrib 942.75 ± 31% -70.3% 280.00 ± 62% sched_debug.cfs_rq[15]:/.blocked_load_avg 958.25 ± 31% -68.3% 303.75 ± 62% sched_debug.cfs_rq[15]:/.tg_load_contrib 20260 ± 7% -9.5% 18339 ± 1% sched_debug.cfs_rq[2]:/.exec_clock 240.25 ± 43% +185.7% 686.50 ± 49% sched_debug.cfs_rq[2]:/.utilization_load_avg 18628 ± 2% -9.2% 16914 ± 5% sched_debug.cfs_rq[4]:/.avg->runnable_avg_sum 405.75 ± 1% -8.7% 370.50 ± 5% sched_debug.cfs_rq[4]:/.tg_runnable_contrib 19914 ± 4% -12.0% 17521 ± 2% sched_debug.cfs_rq[5]:/.exec_clock -2928 ±-101% -146.4% 1357 ± 75% sched_debug.cfs_rq[6]:/.spread0 20113 ± 2% -8.2% 18458 ± 3% sched_debug.cfs_rq[7]:/.exec_clock -2088 ±-11% -19.0% -1691 ± -3% sched_debug.cpu#0.nr_uninterruptible 14.25 ± 90% +452.6% 78.75 ± 66% sched_debug.cpu#1.cpu_load[0] 1057371 ±117% -70.1% 316240 ± 2% sched_debug.cpu#1.nr_switches 1059402 ±116% -70.0% 317990 ± 2% sched_debug.cpu#1.sched_count 472923 ±128% -76.4% 111412 ± 3% sched_debug.cpu#1.sched_goidle 507097 ±122% -72.5% 139349 ± 3% sched_debug.cpu#1.ttwu_count 27.50 ± 12% +53.6% 42.25 ± 23% sched_debug.cpu#11.cpu_load[3] 28.00 ± 11% +29.5% 36.25 ± 12% sched_debug.cpu#11.cpu_load[4] 603068 ± 3% -23.4% 462150 ± 20% sched_debug.cpu#13.avg_idle 436039 ± 23% +29.1% 563133 ± 2% sched_debug.cpu#14.avg_idle 402209 ±116% -81.9% 72660 ± 3% sched_debug.cpu#14.ttwu_local -2215 ±-11% +13.7% -2519 ± -3% sched_debug.cpu#2.nr_uninterruptible 41.25 ± 26% -35.2% 26.75 ± 10% sched_debug.cpu#3.cpu_load[2] 34.00 ±117% +242.6% 116.50 ± 79% sched_debug.cpu#4.load -2266 ±-13% +16.9% -2648 ± -2% sched_debug.cpu#4.nr_uninterruptible 884.50 ± 33% +44.8% 1280 ± 14% sched_debug.cpu#5.curr->pid -2219 ± -7% +11.4% -2472 ± -3% sched_debug.cpu#5.nr_uninterruptible 55415 ± 20% -19.8% 44460 ± 0% sched_debug.cpu#6.nr_load_updates -2350 ± -8% +14.2% -2684 ± -2% sched_debug.cpu#6.nr_uninterruptible 847343 ± 91% -84.0% 135604 ± 2% sched_debug.cpu#6.ttwu_count 784793 ± 99% -90.3% 75880 ± 3% sched_debug.cpu#6.ttwu_local 39.00 ± 14% -30.1% 27.25 ± 23% sched_debug.cpu#7.cpu_load[3] 35.00 ± 11% -22.1% 27.25 ± 12% sched_debug.cpu#7.cpu_load[4] 128907 ± 8% -10.8% 114951 ± 2% sched_debug.cpu#8.ttwu_count 624669 ± 88% -56.2% 273343 ± 0% sched_debug.cpu#9.nr_switches 626375 ± 88% -56.1% 274746 ± 0% sched_debug.cpu#9.sched_count 262109 ±102% -64.9% 92004 ± 1% sched_debug.cpu#9.sched_goidle 2.11 ± 4% +18.9% 2.50 ± 5% sched_debug.rt_rq[0]:/.rt_time ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test: lituya/unixbench/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/powersave/shell1 commit: dfbca41f347997e57048a53755611c8e2d792924 d4573c3e1c992668f5dcd57d1c2ced56ae9650b9 dfbca41f347997e5 d4573c3e1c992668f5dcd57d1c ---------------- -------------------------- %stddev %change %stddev \ | \ 5132876 ± 0% +2.0% 5236374 ± 0% unixbench.time.involuntary_context_switches 1571 ± 0% +1.2% 1591 ± 0% unixbench.time.system_time 795.19 ± 0% +1.3% 805.33 ± 0% unixbench.time.user_time 932469 ± 0% -12.7% 813874 ± 0% softirqs.SCHED 18660 ± 0% +1.7% 18975 ± 0% vmstat.system.in 15466776 ± 1% -5.1% 14675901 ± 0% latency_stats.sum.call_rwsem_down_write_failed.unlink_file_vma.free_pgtables.exit_mmap.mmput.flush_old_exec.load_elf_binary.search_binary_handler.do_execveat_common.SyS_execve.stub_execve 5322533 ± 1% -8.5% 4868555 ± 1% latency_stats.sum.call_rwsem_down_write_failed.vma_adjust.__split_vma.do_munmap.mmap_region.do_mmap_pgoff.vm_mmap_pgoff.SyS_mmap_pgoff.SyS_mmap.system_call_fastpath 21589512 ± 1% -7.4% 19993473 ± 1% latency_stats.sum.call_rwsem_down_write_failed.vma_adjust.__split_vma.split_vma.mprotect_fixup.SyS_mprotect.system_call_fastpath 9791019 ± 1% -8.2% 8988996 ± 1% latency_stats.sum.call_rwsem_down_write_failed.vma_link.mmap_region.do_mmap_pgoff.vm_mmap_pgoff.SyS_mmap_pgoff.SyS_mmap.system_call_fastpath 4.135e+08 ± 0% -3.0% 4.012e+08 ± 0% latency_stats.sum.sigsuspend.SyS_rt_sigsuspend.system_call_fastpath 24904 ± 4% -25.2% 18617 ± 11% sched_debug.cfs_rq[0]:/.blocked_load_avg 8.75 ± 58% +508.6% 53.25 ±106% sched_debug.cfs_rq[0]:/.load 391160 ± 1% -11.4% 346400 ± 1% sched_debug.cfs_rq[0]:/.tg_load_avg 25100 ± 4% -25.2% 18770 ± 11% sched_debug.cfs_rq[0]:/.tg_load_contrib 390600 ± 1% -11.1% 347123 ± 2% sched_debug.cfs_rq[10]:/.tg_load_avg 391242 ± 1% -11.7% 345453 ± 1% sched_debug.cfs_rq[11]:/.tg_load_avg 390931 ± 1% -11.5% 345958 ± 1% sched_debug.cfs_rq[12]:/.tg_load_avg 23834 ± 5% -17.3% 19701 ± 11% sched_debug.cfs_rq[13]:/.blocked_load_avg 390601 ± 1% -11.7% 345060 ± 2% sched_debug.cfs_rq[13]:/.tg_load_avg 24080 ± 5% -17.7% 19823 ± 11% sched_debug.cfs_rq[13]:/.tg_load_contrib 390376 ± 1% -11.5% 345360 ± 2% sched_debug.cfs_rq[14]:/.tg_load_avg 23210 ± 5% -12.3% 20347 ± 4% sched_debug.cfs_rq[15]:/.blocked_load_avg 390137 ± 0% -11.4% 345480 ± 2% sched_debug.cfs_rq[15]:/.tg_load_avg 23345 ± 5% -12.4% 20458 ± 4% sched_debug.cfs_rq[15]:/.tg_load_contrib 390654 ± 1% -11.4% 345993 ± 2% sched_debug.cfs_rq[1]:/.tg_load_avg 24857 ± 5% -16.0% 20876 ± 19% sched_debug.cfs_rq[1]:/.tg_load_contrib 391340 ± 1% -11.6% 346023 ± 2% sched_debug.cfs_rq[2]:/.tg_load_avg 26275 ± 5% -15.6% 22180 ± 19% sched_debug.cfs_rq[3]:/.blocked_load_avg 391666 ± 1% -11.7% 345867 ± 2% sched_debug.cfs_rq[3]:/.tg_load_avg 26369 ± 5% -15.5% 22271 ± 19% sched_debug.cfs_rq[3]:/.tg_load_contrib 391430 ± 1% -11.5% 346427 ± 2% sched_debug.cfs_rq[4]:/.tg_load_avg 391321 ± 1% -11.5% 346235 ± 2% sched_debug.cfs_rq[5]:/.tg_load_avg 25744 ± 3% -17.8% 21156 ± 14% sched_debug.cfs_rq[6]:/.blocked_load_avg 1.00 ± 70% +225.0% 3.25 ± 25% sched_debug.cfs_rq[6]:/.nr_spread_over 389932 ± 1% -11.1% 346764 ± 2% sched_debug.cfs_rq[6]:/.tg_load_avg 25873 ± 3% -17.6% 21329 ± 14% sched_debug.cfs_rq[6]:/.tg_load_contrib 389907 ± 1% -11.0% 346962 ± 2% sched_debug.cfs_rq[7]:/.tg_load_avg 23576 ± 4% -20.0% 18853 ± 10% sched_debug.cfs_rq[8]:/.blocked_load_avg 390564 ± 1% -11.1% 347109 ± 2% sched_debug.cfs_rq[8]:/.tg_load_avg 23775 ± 3% -20.3% 18937 ± 10% sched_debug.cfs_rq[8]:/.tg_load_contrib 391152 ± 1% -11.2% 347502 ± 2% sched_debug.cfs_rq[9]:/.tg_load_avg 450623 ± 16% +28.9% 580722 ± 4% sched_debug.cpu#0.avg_idle 29.75 ± 66% -74.8% 7.50 ± 56% sched_debug.cpu#12.load 540.50 ± 6% -15.5% 456.75 ± 3% sched_debug.cpu#14.nr_uninterruptible 88.25 ± 72% -83.9% 14.25 ± 27% sched_debug.cpu#2.cpu_load[0] 61.00 ± 50% -69.3% 18.75 ± 24% sched_debug.cpu#2.cpu_load[1] 46.00 ± 26% -37.5% 28.75 ± 18% sched_debug.cpu#3.cpu_load[2] 30.00 ± 21% +43.3% 43.00 ± 31% sched_debug.cpu#6.cpu_load[2] 21361 ± 16% +72.5% 36844 ± 39% sched_debug.cpu#8.curr->pid nhm-white: Nehalem Memory: 6G lituya: Grantley Haswell Memory: 16G To reproduce: git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git cd lkp-tests bin/lkp install job.yaml # job file is attached in this email bin/lkp run job.yaml Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. Thanks, Ying Huang
--- LKP_SERVER: inn LKP_CGI_PORT: 80 LKP_CIFS_PORT: 139 testcase: unixbench default-monitors: wait: pre-test uptime: iostat: vmstat: numa-numastat: numa-vmstat: numa-meminfo: proc-vmstat: proc-stat: interval: 10 meminfo: slabinfo: interrupts: lock_stat: latency_stats: softirqs: bdi_dev_mapping: diskstats: nfsstat: cpuidle: cpufreq-stats: turbostat: pmeter: sched_debug: interval: 60 default-watchdogs: oom-killer: watchdog: cpufreq_governor: performance commit: b953c0d234bc72e8489d3bf51a276c5c4ec85345 model: Grantley Haswell nr_cpu: 16 memory: 16G hdd_partitions: swap_partitions: rootfs_partition: unixbench: test: execl queue: cyclic testbox: lituya tbox_group: lituya kconfig: x86_64-rhel enqueue_time: 2015-06-22 13:55:49.804466128 +08:00 user: lkp compiler: gcc-4.9 head_commit: af84726a453b7e9ed818fd24086db2bff72760c6 base_commit: b953c0d234bc72e8489d3bf51a276c5c4ec85345 branch: next/master kernel: "/pkg/linux/x86_64-rhel/gcc-4.9/b953c0d234bc72e8489d3bf51a276c5c4ec85345/vmlinuz-4.1.0" rootfs: debian-x86_64-2015-02-07.cgz result_root: "/result/unixbench/performance-execl/lituya/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/b953c0d234bc72e8489d3bf51a276c5c4ec85345/0" job_file: "/lkp/scheduled/lituya/cyclic_unixbench-performance-execl-x86_64-rhel-CYCLIC_BASE-b953c0d234bc72e8489d3bf51a276c5c4ec85345-0-20150622-107681-1e1d77.yaml" dequeue_time: 2015-06-23 22:23:50.759877002 +08:00 max_uptime: 1891.0800000000002 initrd: "/osimage/debian/debian-x86_64-2015-02-07.cgz" bootloader_append: - root=/dev/ram0 - user=lkp - job=/lkp/scheduled/lituya/cyclic_unixbench-performance-execl-x86_64-rhel-CYCLIC_BASE-b953c0d234bc72e8489d3bf51a276c5c4ec85345-0-20150622-107681-1e1d77.yaml - ARCH=x86_64 - kconfig=x86_64-rhel - branch=next/master - commit=b953c0d234bc72e8489d3bf51a276c5c4ec85345 - BOOT_IMAGE=/pkg/linux/x86_64-rhel/gcc-4.9/b953c0d234bc72e8489d3bf51a276c5c4ec85345/vmlinuz-4.1.0 - max_uptime=1891 - RESULT_ROOT=/result/unixbench/performance-execl/lituya/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/b953c0d234bc72e8489d3bf51a276c5c4ec85345/0 - LKP_SERVER=inn - |2- earlyprintk=ttyS0,115200 systemd.log_level=err debug apic=debug sysrq_always_enabled rcupdate.rcu_cpu_stall_timeout=100 panic=-1 softlockup_panic=1 nmi_watchdog=panic oops=panic load_ramdisk=2 prompt_ramdisk=0 console=ttyS0,115200 console=tty0 vga=normal rw lkp_initrd: "/lkp/lkp/lkp-x86_64.cgz" modules_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/b953c0d234bc72e8489d3bf51a276c5c4ec85345/modules.cgz" bm_initrd: "/osimage/deps/debian-x86_64-2015-02-07.cgz/lkp.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/run-ipconfig.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/turbostat.cgz,/lkp/benchmarks/turbostat.cgz,/lkp/benchmarks/unixbench.cgz" job_state: finished loadavg: 11.20 4.24 1.56 1/217 3673 start_time: '1435069457' end_time: '1435069655' version: "/lkp/lkp/.src-20150623-212413"
echo performance > /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu1/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu10/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu11/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu12/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu13/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu14/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu15/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu2/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu3/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu4/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu5/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu6/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu7/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu8/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu9/cpufreq/scaling_governor ./Run execl