Hi Christoph,

On Sun, Aug 14, 2016 at 06:51:28AM +0800, Fengguang Wu wrote:
Hi Christoph,

On Sun, Aug 14, 2016 at 12:15:08AM +0200, Christoph Hellwig wrote:
Hi Fengguang,

feel free to try this git tree:

  git://git.infradead.org/users/hch/vfs.git iomap-fixes

I just queued some test jobs for it.

% queue -q vip -t ivb44 -b hch-vfs/iomap-fixes aim7-fs-1brd.yaml fs=xfs -r3 -k 
fe9c2c81ed073878768785a985295cbacc349e42 -k 
ca2edab2e1d8f30dda874b7f717c2d4664991e9b -k 
99091700659f4df965e138b38b4fa26a29b7eade

That job file can be found here:

       
https://git.kernel.org/cgit/linux/kernel/git/wfg/lkp-tests.git/tree/jobs/aim7-fs-1brd.yaml

It specifies a matrix of the below atom tests:

       wfg /c/lkp-tests% split-job jobs/aim7-fs-1brd.yaml -s 'fs: xfs'

       jobs/aim7-fs-1brd.yaml => 
./aim7-fs-1brd-1BRD_48G-xfs-disk_src-3000-performance.yaml
       jobs/aim7-fs-1brd.yaml => 
./aim7-fs-1brd-1BRD_48G-xfs-disk_rr-3000-performance.yaml
       jobs/aim7-fs-1brd.yaml => 
./aim7-fs-1brd-1BRD_48G-xfs-disk_rw-3000-performance.yaml
       jobs/aim7-fs-1brd.yaml => 
./aim7-fs-1brd-1BRD_48G-xfs-disk_cp-3000-performance.yaml
       jobs/aim7-fs-1brd.yaml => 
./aim7-fs-1brd-1BRD_48G-xfs-disk_wrt-3000-performance.yaml
       jobs/aim7-fs-1brd.yaml => 
./aim7-fs-1brd-1BRD_48G-xfs-sync_disk_rw-600-performance.yaml
       jobs/aim7-fs-1brd.yaml => 
./aim7-fs-1brd-1BRD_48G-xfs-creat-clo-1500-performance.yaml
       jobs/aim7-fs-1brd.yaml => 
./aim7-fs-1brd-1BRD_48G-xfs-disk_rd-9000-performance.yaml

I got some results now. The several finished aim7 tests have some
performance regressions for commit fe9c2c81 ("xfs: rewrite and
optimize the delalloc write path") comparing to its parent commit
ca2edab2e and their base mainline commit 990917006 ("Merge tag
'nfs-for-4.8-2' of git://git.linux-nfs.org/projects/trondmy/linux-nfs").

wfg@inn ~% compare -g aim7 -ai 99091700659f4df965e138b38b4fa26a29b7eade 
ca2edab2e1d8f30dda874b7f717c2d4664991e9b 
fe9c2c81ed073878768785a985295cbacc349e42
tests: 4
   60      perf-index  fe9c2c81ed073878768785a985295cbacc349e42
   97     power-index  fe9c2c81ed073878768785a985295cbacc349e42

99091700659f4df9  ca2edab2e1d8f30dda874b7f71  fe9c2c81ed073878768785a985  
testcase/testparams/testbox
----------------  --------------------------  --------------------------  
---------------------------
        %stddev     %change         %stddev     %change         %stddev
            \          |                \          |                \
   270459                      272267 ±  3%       -48%     139834 ±  3%  
aim7/1BRD_48G-xfs-disk_cp-3000-performance/ivb44
   473257                      468546               5%     497512        
aim7/1BRD_48G-xfs-disk_rd-9000-performance/ivb44
   360578             -18%     296589             -60%     144974        
aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44
   358701              -6%     335712             -40%     216057        
GEO-MEAN aim7.jobs-per-min

99091700659f4df9  ca2edab2e1d8f30dda874b7f71  fe9c2c81ed073878768785a985
----------------  --------------------------  --------------------------
    48.48                       48.15              36%      65.85        
aim7/1BRD_48G-xfs-disk_cp-3000-performance/ivb44
    89.50                       89.76                       88.75        
aim7/1BRD_48G-xfs-disk_rd-9000-performance/ivb44
    35.78              23%      43.93              76%      63.09        
aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44
    53.75               7%      57.48              33%      71.71        
GEO-MEAN turbostat.%Busy

99091700659f4df9  ca2edab2e1d8f30dda874b7f71  fe9c2c81ed073878768785a985
----------------  --------------------------  --------------------------
     1439                        1431              36%       1964        
aim7/1BRD_48G-xfs-disk_cp-3000-performance/ivb44
     2671                        2674                        2650        
aim7/1BRD_48G-xfs-disk_rd-9000-performance/ivb44
     1057              23%       1303              78%       1883        
aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44
     1595               7%       1708              34%       2139        
GEO-MEAN turbostat.Avg_MHz

99091700659f4df9  ca2edab2e1d8f30dda874b7f71  fe9c2c81ed073878768785a985
----------------  --------------------------  --------------------------
      167                         167               6%        177        
aim7/1BRD_48G-xfs-disk_cp-3000-performance/ivb44
      175                         175                         176        
aim7/1BRD_48G-xfs-disk_rd-9000-performance/ivb44
      150               8%        162              19%        178        
aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44
      164                         168               8%        177        
GEO-MEAN turbostat.PkgWatt

99091700659f4df9  ca2edab2e1d8f30dda874b7f71  fe9c2c81ed073878768785a985
----------------  --------------------------  --------------------------
    10.27                       10.43             -14%       8.79        
aim7/1BRD_48G-xfs-disk_cp-3000-performance/ivb44
     6.85                        6.66                        6.88        
aim7/1BRD_48G-xfs-disk_rd-9000-performance/ivb44
     9.96              14%      11.36              -7%       9.23        
aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44
     8.88               4%       9.24              -7%       8.23        
GEO-MEAN turbostat.RAMWatt

Here are the detailed numbers for each test case. The perf-profile and
latency_stats numbers are sorted by absolute change in each sub-category
now. perf-profile numbers > 5 are all shown.

It may be more pleasant to view the long trace.call.funcs lines with
vim ":set nowrap" option.

aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44

99091700659f4df9 ca2edab2e1d8f30dda874b7f71 fe9c2c81ed073878768785a985 ---------------- -------------------------- -------------------------- %stddev %change %stddev %change %stddev \ | \ | \ 360578 -18% 294351 -60% 144974 aim7.jobs-per-min
    12835             458%      71658             480%      74445        
aim7.time.involuntary_context_switches
      755              50%       1136             373%       3570        
aim7.time.system_time
   155970                      152810              73%     269438        
aim7.time.minor_page_faults
    50.15              22%      61.39             148%     124.39        
aim7.time.elapsed_time
    50.15              22%      61.39             148%     124.39        
aim7.time.elapsed_time.max
   438660                      428601              -7%     407807        
aim7.time.voluntary_context_switches
     2452                        2480               5%       2584        
aim7.time.maximum_resident_set_size
    34293 ±  4%        70%      58129 ± 19%       213%     107483        
interrupts.CAL:Function_call_interrupts
    79.70 ±  6%        16%      92.63 ±  6%        89%     150.33        
uptime.boot
     2890 ±  8%         6%       3077 ±  8%        15%       3329        
uptime.idle
   150186 ±  9%        41%     212090             122%     333727        
softirqs.RCU
   161166               9%     176318              16%     186527        
softirqs.SCHED
   648051              33%     864346             222%    2089349        
softirqs.TIMER
    50.15              22%      61.39             148%     124.39        
time.elapsed_time
    50.15              22%      61.39             148%     124.39        
time.elapsed_time.max
    12835             458%      71658             480%      74445        
time.involuntary_context_switches
   155970                      152810              73%     269438        
time.minor_page_faults
     1563              21%       1898              85%       2895        
time.percent_of_cpu_this_job_got
      755              50%       1136             373%       3570        
time.system_time
  4564660 ±  4%        68%    7651587              79%    8159302        
numa-numastat.node0.numa_foreign
  3929898              81%    7129718              46%    5733813        
numa-numastat.node0.numa_miss
        0                           2 ± 20%                     2        
numa-numastat.node1.other_node
  4569811 ±  4%        68%    7654689              79%    8163206        
numa-numastat.node1.numa_miss
  3935075              81%    7132850              46%    5737410        
numa-numastat.node1.numa_foreign
 34767917               4%   36214694              11%   38627727        
numa-numastat.node1.numa_hit
 34767917               4%   36214691              11%   38627725        
numa-numastat.node1.local_node
    12377 ± 18%      3615%     459790            2848%     364868        
vmstat.io.bo
      119              -8%        110 ±  4%       -16%        101        
vmstat.memory.buff
 18826454             -16%   15748045             -37%   11882562        
vmstat.memory.free
       16              25%         20             106%         33        
vmstat.procs.r
    19407             469%     110509             520%     120350        
vmstat.system.cs
    48215              10%      52977               3%      49819        
vmstat.system.in
   142459             -11%     126667             -23%     109481        
cpuidle.C1-IVT.usage
 29494441 ±  3%       -18%   24206809             -36%   18889149        
cpuidle.C1-IVT.time
  5736732              28%    7315830             525%   35868316        
cpuidle.C1E-IVT.time
    51148               9%      55743              98%     101021        
cpuidle.C1E-IVT.usage
 18347890              27%   23243942              21%   22154105        
cpuidle.C3-IVT.time
    96127               9%     104487             -29%      68552        
cpuidle.C3-IVT.usage
1.525e+09               6%  1.617e+09              41%  2.147e+09        
cpuidle.C6-IVT.time
  1805218              11%    1998052              33%    2397285        
cpuidle.C6-IVT.usage
      286 ± 11%        14%        328 ±  7%       389%       1402        
cpuidle.POLL.usage
  1013526 ± 41%        98%    2003264 ± 20%       272%    3774675        
cpuidle.POLL.time
    35.78              24%      44.22              76%      63.09        
turbostat.%Busy
     1057              24%       1312              78%       1883        
turbostat.Avg_MHz
    34.80              -3%      33.63             -22%      27.18        
turbostat.CPU%c1
     0.34              -5%       0.33             -77%       0.08        
turbostat.CPU%c3
    29.07             -25%      21.82             -67%       9.65        
turbostat.CPU%c6
      118              11%        130              23%        145        
turbostat.CorWatt
     9.39 ± 13%       -19%       7.61 ±  6%       -61%       3.67        
turbostat.Pkg%pc2
     3.04 ± 33%       -49%       1.55 ± 14%       -76%       0.72        
turbostat.Pkg%pc6
      150               9%        164              19%        178        
turbostat.PkgWatt
     9.96              14%      11.34              -7%       9.23        
turbostat.RAMWatt
    18232 ±  8%        -8%      16747 ± 10%        11%      20267        
meminfo.AnonHugePages
    80723                       78330             -24%      61572        
meminfo.CmaFree
  4690642 ± 10%       -15%    3981312             -15%    3983392        
meminfo.DirectMap2M
  1060897             -21%     834807             -22%     828755        
meminfo.Dirty
  2362330              26%    2983603              44%    3391287        
meminfo.Inactive
  2353250              26%    2974520              44%    3382139        
meminfo.Inactive(file)
 19388991             -18%   15966408             -38%   12038822        
meminfo.MemFree
  1186231               4%    1236627              13%    1341728        
meminfo.SReclaimable
   179570               3%     185696              14%     204382        
meminfo.SUnreclaim
  1365802               4%    1422323              13%    1546111        
meminfo.Slab
   318863              10%     352026              16%     368386        
meminfo.Unevictable
     0.00                        0.00                        9.15        
perf-profile.cycles-pp.xfs_file_iomap_begin_delay.isra.9.xfs_file_iomap_begin.iomap_apply.iomap_file_buffered_write.xfs_file_buffered_aio_write
     0.00                        0.00                        8.90        
perf-profile.cycles-pp.xfs_inode_set_eofblocks_tag.xfs_file_iomap_begin_delay.isra.9.xfs_file_iomap_begin.iomap_apply.iomap_file_buffered_write
     0.00                        0.00                        8.61        
perf-profile.cycles-pp._raw_spin_lock.xfs_inode_set_eofblocks_tag.xfs_file_iomap_begin_delay.isra.9.xfs_file_iomap_begin.iomap_apply
     0.00                        0.00                        8.50        
perf-profile.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.xfs_inode_set_eofblocks_tag.xfs_file_iomap_begin_delay.isra.9.xfs_file_iomap_begin
     6.05             -11%       5.42 ±  4%       -15%       5.14        
perf-profile.cycles-pp.hrtimer_interrupt.local_apic_timer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter
     6.54             -11%       5.80 ±  4%       -16%       5.51        
perf-profile.cycles-pp.local_apic_timer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter.call_cpuidle
    16.78              -9%      15.34 ±  9%       -11%      14.90        
perf-profile.cycles-pp.apic_timer_interrupt.cpuidle_enter.call_cpuidle.cpu_startup_entry.start_secondary
    16.51 ±  3%        -9%      14.99 ±  9%       -12%      14.49        
perf-profile.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter.call_cpuidle.cpu_startup_entry
     0.23 ± 23%        20%       0.28 ± 12%      3683%       8.70        
perf-profile.func.cycles-pp.native_queued_spin_lock_slowpath
4.369e+11 ±  4%        20%  5.239e+11              97%  8.601e+11        
perf-stat.branch-instructions
     0.38               5%       0.40             -27%       0.28        
perf-stat.branch-miss-rate
1.678e+09 ±  3%        26%  2.117e+09              44%  2.413e+09        
perf-stat.branch-misses
    42.30              -7%      39.31              -5%      40.38        
perf-stat.cache-miss-rate
6.874e+09 ±  4%        19%   8.21e+09              51%  1.041e+10        
perf-stat.cache-misses
1.625e+10 ±  3%        29%  2.089e+10              59%  2.578e+10        
perf-stat.cache-references
  1017846             588%    7005227            1401%   15273586        
perf-stat.context-switches
2.757e+12 ±  4%        48%  4.092e+12             318%  1.151e+13        
perf-stat.cpu-cycles
   177918              15%     204776              35%     241051        
perf-stat.cpu-migrations
     0.37 ± 14%        60%       0.60 ±  3%        45%       0.54        
perf-stat.dTLB-load-miss-rate
2.413e+09 ± 14%        97%  4.757e+09 ±  4%       149%  6.001e+09        
perf-stat.dTLB-load-misses
6.438e+11              23%  7.893e+11              71%  1.103e+12        
perf-stat.dTLB-loads
     0.06 ± 38%       100%       0.11 ±  6%       207%       0.17        
perf-stat.dTLB-store-miss-rate
2.656e+08 ± 34%       123%   5.91e+08 ±  7%       203%  8.038e+08        
perf-stat.dTLB-store-misses
    45.99 ±  5%         8%      49.56 ± 11%        14%      52.61        
perf-stat.iTLB-load-miss-rate
 45151945                    45832755              72%   77697494        
perf-stat.iTLB-load-misses
 53205262 ±  7%       -10%   47792612 ± 21%        32%   69997751        
perf-stat.iTLB-loads
2.457e+12 ±  4%        16%  2.851e+12              66%  4.084e+12        
perf-stat.instructions
     0.89             -22%       0.70             -60%       0.35        
perf-stat.ipc
   286640               8%     310690              99%     571225        
perf-stat.minor-faults
    29.16               7%      31.25               8%      31.42        
perf-stat.node-load-miss-rate
 4.86e+08 ±  3%       123%  1.084e+09             250%    1.7e+09        
perf-stat.node-load-misses
 1.18e+09             102%  2.385e+09             214%  3.711e+09        
perf-stat.node-loads
    21.51              30%      27.95              62%      34.86        
perf-stat.node-store-miss-rate
1.262e+09              58%  1.989e+09             177%  3.499e+09        
perf-stat.node-store-misses
4.606e+09              11%  5.126e+09              42%  6.539e+09        
perf-stat.node-stores
   286617               8%     310730              99%     571253        
perf-stat.page-faults
  1166432              23%    1429828              42%    1653754        
numa-meminfo.node0.Inactive(file)
  1175123              22%    1434274              41%    1662351        
numa-meminfo.node0.Inactive
   513534             -23%     394773             -24%     392567        
numa-meminfo.node0.Dirty
  9717968             -17%    8082393             -37%    6159862        
numa-meminfo.node0.MemFree
   159470              11%     176717              16%     184229        
numa-meminfo.node0.Unevictable
 23148226               7%   24783802              15%   26706333        
numa-meminfo.node0.MemUsed
   103531 ± 32%       -10%      93669 ± 40%        40%     144469        
numa-meminfo.node0.SUnreclaim
  1187035              30%    1549075              46%    1727751        
numa-meminfo.node1.Inactive
  1186646              30%    1544438              46%    1727201        
numa-meminfo.node1.Inactive(file)
 21000905               3%   21647702              13%   23741428        
numa-meminfo.node1.Active(file)
 21083707               3%   21748741              13%   23822391        
numa-meminfo.node1.Active
   547021             -20%     438525             -21%     433706        
numa-meminfo.node1.Dirty
  9663240             -19%    7870896             -39%    5869977        
numa-meminfo.node1.MemFree
   561241              12%     625903              21%     679671        
numa-meminfo.node1.SReclaimable
   637259 ±  4%        13%     717863 ±  5%        16%     739482        
numa-meminfo.node1.Slab
 23329350               8%   25121687              16%   27122606        
numa-meminfo.node1.MemUsed
   159394              10%     175315              16%     184159        
numa-meminfo.node1.Unevictable
   521615              33%     695562             267%    1916159        
latency_stats.avg.call_rwsem_down_write_failed.do_unlinkat.SyS_unlink.entry_SYSCALL_64_fastpath
   500644              33%     667614             261%    1805608        
latency_stats.avg.call_rwsem_down_write_failed.path_openat.do_filp_open.do_sys_open.SyS_creat.entry_SYSCALL_64_fastpath
     8932 ± 46%       -70%       2717 ±  4%       -95%        464        
latency_stats.avg.wait_on_page_bit.truncate_inode_pages_range.truncate_inode_pages_final.evict.iput.dentry_unlink_inode.__dentry_kill.dput.__fput.____fput.task_work_run.exit_to_usermode_loop
        0                           0                       73327        
latency_stats.hits.wait_on_page_bit.__migration_entry_wait.migration_entry_wait.do_swap_page.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
       43 ± 37%      7923%       3503 ±  4%     31792%      13926        
latency_stats.hits.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agf.xfs_alloc_read_agf.xfs_alloc_fix_freelist.xfs_free_extent_fix_freelist.xfs_free_extent.xfs_trans_free_extent
  1422573              30%    1852368 ±  5%       228%    4672496        
latency_stats.max.call_rwsem_down_write_failed.path_openat.do_filp_open.do_sys_open.SyS_creat.entry_SYSCALL_64_fastpath
  1423130              30%    1851873 ±  5%       228%    4661765        
latency_stats.max.call_rwsem_down_write_failed.do_unlinkat.SyS_unlink.entry_SYSCALL_64_fastpath
      627 ± 66%      3788%      24404 ± 17%      6254%      39883        
latency_stats.max.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agf.xfs_alloc_read_agf.xfs_alloc_fix_freelist.xfs_free_extent_fix_freelist.xfs_free_extent.xfs_trans_free_extent
     3922 ± 18%        56%       6134 ± 29%       634%      28786        
latency_stats.max.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agi.xfs_ialloc_read_agi.xfs_dialloc.xfs_ialloc.xfs_dir_ialloc.xfs_create
        0                           0                       16665        
latency_stats.max.wait_on_page_bit.__migration_entry_wait.migration_entry_wait.do_swap_page.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
 5.15e+10              25%  6.454e+10             220%  1.649e+11        
latency_stats.sum.call_rwsem_down_write_failed.do_unlinkat.SyS_unlink.entry_SYSCALL_64_fastpath
        0                           0                   1.385e+08        
latency_stats.sum.wait_on_page_bit.__migration_entry_wait.migration_entry_wait.do_swap_page.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
 11666476              45%   16905624             755%   99756088        
latency_stats.sum.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agi.xfs_iunlink_remove.xfs_ifree.xfs_inactive_ifree.xfs_inactive.xfs_fs_destroy_inode
     2216 ± 69%     80030%    1775681 ±  4%     3e+06%   67521154        
latency_stats.sum.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agf.xfs_alloc_read_agf.xfs_alloc_fix_freelist.xfs_free_extent_fix_freelist.xfs_free_extent.xfs_trans_free_extent
  1601815              28%    2053992             288%    6213577        
latency_stats.sum.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agi.xfs_ialloc_read_agi.xfs_dialloc.xfs_ialloc.xfs_dir_ialloc.xfs_create
  1774397              20%    2120576             244%    6099374        
latency_stats.sum.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agi.xfs_iunlink.xfs_droplink.xfs_remove.xfs_vn_unlink.vfs_unlink
      628 ±141%       125%       1416 ±  5%     4e+05%    2677036        
latency_stats.sum.xfs_iget.xfs_ialloc.xfs_dir_ialloc.xfs_create.xfs_generic_create.xfs_vn_mknod.xfs_vn_create.path_openat.do_filp_open.do_sys_open.SyS_creat.entry_SYSCALL_64_fastpath
     6087 ± 92%      1277%      83839 ±  3%     11105%     682063        
latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_commit.xfs_vn_update_time.file_update_time.xfs_file_aio_write_checks.xfs_file_buffered_aio_write.xfs_file_write_iter.__vfs_write.vfs_write.SyS_write
        0                           0                      116108        
latency_stats.sum.xlog_grant_head_wait.xlog_grant_head_check.xfs_log_reserve.xfs_trans_reserve.xfs_trans_alloc.xfs_vn_update_time.file_update_time.xfs_file_aio_write_checks.xfs_file_buffered_aio_write.xfs_file_write_iter.__vfs_write.vfs_write
     1212 ± 59%      1842%      23546 ±  7%      4861%      60149        
latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_commit.xfs_vn_update_time.touch_atime.generic_file_read_iter.xfs_file_buffered_aio_read.xfs_file_read_iter.__vfs_read.vfs_read.SyS_read
     1624 ± 22%      1356%      23637 ±  3%      1596%      27545        
latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.__xfs_trans_roll.xfs_trans_roll.xfs_itruncate_extents.xfs_free_eofblocks.xfs_release.xfs_file_release.__fput.____fput.task_work_run
     2068 ± 27%       834%      19319 ± 23%      1125%      25334        
latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.__xfs_trans_roll.xfs_trans_roll.xfs_itruncate_extents.xfs_inactive_truncate.xfs_inactive.xfs_fs_destroy_inode.destroy_inode.evict.iput
        0                           0                       22155        
latency_stats.sum.xlog_grant_head_wait.xlog_grant_head_check.xfs_log_reserve.xfs_trans_reserve.xfs_trans_alloc.xfs_inactive_truncate.xfs_inactive.xfs_fs_destroy_inode.destroy_inode.evict.iput.dentry_unlink_inode
       39 ± 71%     41280%      16414 ± 14%     51951%      20647        
latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.__xfs_trans_roll.xfs_trans_roll.xfs_defer_trans_roll.xfs_defer_finish.xfs_itruncate_extents.xfs_inactive_truncate.xfs_inactive.xfs_fs_destroy_inode.destroy_inode
        0                           0                       15600        
latency_stats.sum.xlog_grant_head_wait.xlog_grant_head_check.xfs_log_reserve.xfs_trans_reserve.xfs_trans_alloc.xfs_inactive_ifree.xfs_inactive.xfs_fs_destroy_inode.destroy_inode.evict.iput.dentry_unlink_inode
       10 ±141%      6795%        689 ± 70%     1e+05%      10637        
latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_commit.xfs_inactive_ifree.xfs_inactive.xfs_fs_destroy_inode.destroy_inode.evict.iput.dentry_unlink_inode.__dentry_kill
       99 ±112%        86%        185 ± 80%      9978%      10011        
latency_stats.sum.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_imap_to_bp.xfs_iunlink_remove.xfs_ifree.xfs_inactive_ifree.xfs_inactive.xfs_fs_destroy_inode
    18232 ±134%       -16%      15260 ± 54%       -40%      10975        
latency_stats.sum.xfs_lock_two_inodes.xfs_remove.xfs_vn_unlink.vfs_unlink.do_unlinkat.SyS_unlink.entry_SYSCALL_64_fastpath
      647 ±  3%       -97%         21 ± 19%        34%        868        
proc-vmstat.kswapd_high_wmark_hit_quickly
     1091             -97%         36 ±  9%        29%       1411        
proc-vmstat.kswapd_low_wmark_hit_quickly
   265066             -21%     208142             -22%     206344        
proc-vmstat.nr_dirty
    20118                       19574             -23%      15432        
proc-vmstat.nr_free_cma
  4844108             -18%    3988031             -38%    3008251        
proc-vmstat.nr_free_pages
   588262              26%     743537              44%     845765        
proc-vmstat.nr_inactive_file
       50 ± 25%       192%        148 ± 15%       103%        103        
proc-vmstat.nr_pages_scanned
   296623               4%     309201              13%     335474        
proc-vmstat.nr_slab_reclaimable
    44880               3%      46405              14%      51078        
proc-vmstat.nr_slab_unreclaimable
    79716              10%      88008              16%      92097        
proc-vmstat.nr_unevictable
      167 ±  9%     9e+06%   14513434           2e+06%    3569348        
proc-vmstat.nr_vmscan_immediate_reclaim
   162380 ± 18%      4392%    7294622            7024%   11567602        
proc-vmstat.nr_written
   588257              26%     743537              44%     845784        
proc-vmstat.nr_zone_inactive_file
    79716              10%      88008              16%      92097        
proc-vmstat.nr_zone_unevictable
   265092             -21%     208154             -22%     206388        
proc-vmstat.nr_zone_write_pending
  8507451 ±  3%        74%   14784261              64%   13918067        
proc-vmstat.numa_foreign
       10 ±  4%                    10 ±  4%     6e+05%      57855        
proc-vmstat.numa_hint_faults
  8507451 ±  3%        74%   14784187              64%   13918067        
proc-vmstat.numa_miss
       72                          72           3e+05%     213175        
proc-vmstat.numa_pte_updates
     1740             -97%         59 ± 12%        33%       2306        
proc-vmstat.pageoutrun
  5322372            1068%   62167111            1024%   59824114        
proc-vmstat.pgactivate
  2816355              27%    3575784              14%    3203214        
proc-vmstat.pgalloc_dma32
 74392338              11%   82333943              14%   84954110        
proc-vmstat.pgalloc_normal
 60958397             -18%   49976330             -26%   45055885        
proc-vmstat.pgdeactivate
   302790               9%     329088              94%     586116        
proc-vmstat.pgfault
 61061205              14%   69758545              18%   72000453        
proc-vmstat.pgfree
   655652 ± 18%      4352%   29190304            6967%   46338056        
proc-vmstat.pgpgout
 60965725             -18%   49983704             -26%   45063375        
proc-vmstat.pgrefill
        2 ± 17%     4e+07%     985929 ±  8%     7e+07%    1952629        
proc-vmstat.pgrotated
    82046 ± 36%     50634%   41625211            5397%    4510385        
proc-vmstat.pgscan_direct
 60128369             -37%   38068394              10%   66306637        
proc-vmstat.pgscan_kswapd
     2030 ± 46%     1e+06%   27038054 ±  3%     78642%    1598733        
proc-vmstat.pgsteal_direct
        0                     2414551 ±  3%               3694833        
proc-vmstat.workingset_activate
        0                     2414551 ±  3%               3694833        
proc-vmstat.workingset_refault
       26 ± 39%     1e+07%    2657286           3e+06%     658792        
numa-vmstat.node0.nr_vmscan_immediate_reclaim
    40449 ± 22%      3135%    1308601 ±  4%      4723%    1950670        
numa-vmstat.node0.nr_written
   291648              22%     357059              42%     413612        
numa-vmstat.node0.nr_zone_inactive_file
   291655              22%     357053              42%     413596        
numa-vmstat.node0.nr_inactive_file
  1542314 ±  5%        77%    2731911              98%    3056411        
numa-vmstat.node0.numa_foreign
  1366073 ±  4%       103%    2766780 ±  3%        68%    2293117        
numa-vmstat.node0.numa_miss
   128634             -23%      99104             -24%      98062        
numa-vmstat.node0.nr_dirty
   128663             -23%      99130             -24%      98051        
numa-vmstat.node0.nr_zone_write_pending
  2424918             -16%    2033425             -37%    1537826        
numa-vmstat.node0.nr_free_pages
 14037168              10%   15473174              20%   16883787        
numa-vmstat.node0.numa_local
 14037172              10%   15473174              20%   16883790        
numa-vmstat.node0.numa_hit
    39867              10%      44022              16%      46058        
numa-vmstat.node0.nr_zone_unevictable
    39867              10%      44022              16%      46058        
numa-vmstat.node0.nr_unevictable
    25871 ± 32%        -9%      23414 ± 40%        40%      36094        
numa-vmstat.node0.nr_slab_unreclaimable
 14851187               6%   15749527              11%   16497187        
numa-vmstat.node0.nr_dirtied
        0                     1225299 ±  4%               2008478        
numa-vmstat.node1.workingset_refault
        0                     1225299 ±  4%               2008478        
numa-vmstat.node1.workingset_activate
       23 ± 35%     1e+07%    2974198 ±  3%     3e+06%     683002        
numa-vmstat.node1.nr_vmscan_immediate_reclaim
    40769 ± 26%      3264%    1371611 ±  3%      5569%    2311374        
numa-vmstat.node1.nr_written
       25 ±  8%       216%         81 ±  3%       356%        117        
numa-vmstat.node1.nr_pages_scanned
   296681              30%     385708              45%     431591        
numa-vmstat.node1.nr_zone_inactive_file
   296681              30%     385709              45%     431591        
numa-vmstat.node1.nr_inactive_file
  5252547                     5401234              13%    5936151        
numa-vmstat.node1.nr_zone_active_file
  5252547                     5401238              13%    5936151        
numa-vmstat.node1.nr_active_file
   136060             -19%     110021             -21%     107114        
numa-vmstat.node1.nr_zone_write_pending
   136060             -19%     110019             -21%     107107        
numa-vmstat.node1.nr_dirty
  1520682 ±  3%        76%    2681012              98%    3008493        
numa-vmstat.node1.numa_miss
  2413468             -18%    1980184             -39%    1466738        
numa-vmstat.node1.nr_free_pages
  1344474 ±  3%       102%    2715690 ±  4%        67%    2245159        
numa-vmstat.node1.numa_foreign
    20160                       19698             -22%      15673        
numa-vmstat.node1.nr_free_cma
 14350439              12%   16005551              27%   18257157        
numa-vmstat.node1.numa_local
 14350440              12%   16005552              27%   18257158        
numa-vmstat.node1.numa_hit
 15381788               9%   16829619              21%   18645441        
numa-vmstat.node1.nr_dirtied
   140354              11%     156202              21%     169950        
numa-vmstat.node1.nr_slab_reclaimable
    39848              10%      43676              16%      46041        
numa-vmstat.node1.nr_zone_unevictable
    39848              10%      43676              16%      46041        
numa-vmstat.node1.nr_unevictable
      377 ±  9%                   370 ±  5%        24%        468        
slabinfo.bdev_cache.active_objs
      377 ±  9%                   370 ±  5%        24%        468        
slabinfo.bdev_cache.num_objs
      389 ± 13%       604%       2737 ± 23%      3371%      13501        
slabinfo.bio-1.active_objs
      389 ± 13%       612%       2770 ± 24%      3441%      13774        
slabinfo.bio-1.num_objs
        7 ± 17%      1039%         83 ± 24%      3623%        273        
slabinfo.bio-1.active_slabs
        7 ± 17%      1039%         83 ± 24%      3623%        273        
slabinfo.bio-1.num_slabs
      978 ±  4%        10%       1075              17%       1144        
slabinfo.blkdev_requests.active_objs
      978 ±  4%        10%       1075              17%       1144        
slabinfo.blkdev_requests.num_objs
 10942119               3%   11286505              13%   12389701        
slabinfo.buffer_head.num_objs
   280566               3%     289397              13%     317684        
slabinfo.buffer_head.active_slabs
   280566               3%     289397              13%     317684        
slabinfo.buffer_head.num_slabs
 10941627                    10693692              11%   12140372        
slabinfo.buffer_head.active_objs
     7436 ±  3%                  7558              20%       8922        
slabinfo.cred_jar.active_objs
     7436 ±  3%                  7558              20%       8922        
slabinfo.cred_jar.num_objs
     4734              85%       8767 ±  8%        60%       7554        
slabinfo.kmalloc-128.num_objs
     4734              78%       8418 ±  8%        45%       6848        
slabinfo.kmalloc-128.active_objs
    17074             -11%      15121             -10%      15379        
slabinfo.kmalloc-256.num_objs
     3105               4%       3216              14%       3527        
slabinfo.kmalloc-4096.num_objs
     3061               4%       3170              12%       3419        
slabinfo.kmalloc-4096.active_objs
    13131 ±  3%        17%      15379              12%      14714        
slabinfo.kmalloc-512.num_objs
     1623 ±  3%                  1664 ±  3%        16%       1889        
slabinfo.mnt_cache.active_objs
     1623 ±  3%                  1664 ±  3%        16%       1889        
slabinfo.mnt_cache.num_objs
     2670               6%       2821              19%       3178        
slabinfo.nsproxy.active_objs
     2670               6%       2821              19%       3178        
slabinfo.nsproxy.num_objs
     2532               5%       2656              17%       2959        
slabinfo.posix_timers_cache.active_objs
     2532               5%       2656              17%       2959        
slabinfo.posix_timers_cache.num_objs
    20689              87%      38595 ± 13%        47%      30452        
slabinfo.radix_tree_node.active_objs
      399              83%        730 ± 13%        47%        587        
slabinfo.radix_tree_node.active_slabs
      399              83%        730 ± 13%        47%        587        
slabinfo.radix_tree_node.num_slabs
    22379              83%      40931 ± 13%        47%      32872        
slabinfo.radix_tree_node.num_objs
     4688                        4706              22%       5712        
slabinfo.sigqueue.active_objs
     4688                        4706              22%       5712        
slabinfo.sigqueue.num_objs
      979 ±  4%         7%       1046 ±  3%       -15%        833        
slabinfo.task_group.active_objs
      979 ±  4%         7%       1046 ±  3%       -15%        833        
slabinfo.task_group.num_objs
     1344               5%       1410              17%       1570        
slabinfo.xfs_btree_cur.active_objs
     1344               5%       1410              17%       1570        
slabinfo.xfs_btree_cur.num_objs
     2500               5%       2632              18%       2946        
slabinfo.xfs_da_state.active_objs
     2500               5%       2632              18%       2946        
slabinfo.xfs_da_state.num_objs
     1299             279%       4917 ± 17%       134%       3035        
slabinfo.xfs_efd_item.num_objs
     1299             278%       4911 ± 17%       126%       2940        
slabinfo.xfs_efd_item.active_objs
     1904 ±  3%         4%       1982              42%       2703        
slabinfo.xfs_inode.num_objs
     1904 ±  3%         4%       1982              39%       2644        
slabinfo.xfs_inode.active_objs
     1659             113%       3538 ± 27%      1360%      24227        
slabinfo.xfs_log_ticket.active_objs
     1659             116%       3588 ± 27%      1369%      24383        
slabinfo.xfs_log_ticket.num_objs
       37             169%         99 ± 29%      1405%        557        
slabinfo.xfs_log_ticket.active_slabs
       37             169%         99 ± 29%      1405%        557        
slabinfo.xfs_log_ticket.num_slabs
     2615              84%       4821 ± 28%      1549%      43132        
slabinfo.xfs_trans.active_objs
     2615              86%       4860 ± 28%      1551%      43171        
slabinfo.xfs_trans.num_objs
       37             162%         97 ± 30%      1614%        634        
slabinfo.xfs_trans.active_slabs
       37             162%         97 ± 30%      1614%        634        
slabinfo.xfs_trans.num_slabs
     3255 ± 12%      9210%     303094           38966%    1271810        
sched_debug.cfs_rq:/.min_vruntime.avg
     8273 ± 10%       382%      39836 ± 17%       309%      33806        
sched_debug.cfs_rq:/.load.avg
      716 ± 34%     28783%     206899           1e+05%    1034000        
sched_debug.cfs_rq:/.min_vruntime.min
     1830 ±  5%      4365%      81731           10579%     195502        
sched_debug.cfs_rq:/.min_vruntime.stddev
     1845 ±  4%      4330%      81754           10503%     195683        
sched_debug.cfs_rq:/.spread0.stddev
    73578 ± 34%      1043%     841209 ± 34%       452%     405848        
sched_debug.cfs_rq:/.load.max
    12.67 ± 35%      3999%     519.25            1979%     263.33        
sched_debug.cfs_rq:/.runnable_load_avg.max
     2.34 ± 33%      4268%     102.01            1854%      45.63        
sched_debug.cfs_rq:/.runnable_load_avg.stddev
    10284 ± 12%      4107%     432665 ±  7%     15350%    1588973        
sched_debug.cfs_rq:/.min_vruntime.max
     1.05 ± 20%      2335%      25.54            1631%      18.15        
sched_debug.cfs_rq:/.runnable_load_avg.avg
    44.06 ± 28%       254%     155.90 ± 16%       310%     180.49        
sched_debug.cfs_rq:/.util_avg.stddev
    15448 ± 19%       831%     143829 ± 22%       422%      80585        
sched_debug.cfs_rq:/.load.stddev
      597 ± 13%       -39%        367 ± 17%       -49%        303        
sched_debug.cfs_rq:/.util_avg.min
     1464 ± 23%       -55%        664 ± 30%       -63%        546        
sched_debug.cfs_rq:/.load_avg.min
     1830 ±  3%       -50%        911 ±  5%       -65%        642        
sched_debug.cfs_rq:/.load_avg.avg
     0.30 ± 13%        22%       0.36 ± 11%        86%       0.56        
sched_debug.cfs_rq:/.nr_running.avg
     2302 ± 11%       -31%       1589             -50%       1157        
sched_debug.cfs_rq:/.load_avg.max
      819 ±  3%        36%       1116              15%        940        
sched_debug.cfs_rq:/.util_avg.max
      728             -14%        630              -9%        664        
sched_debug.cfs_rq:/.util_avg.avg
    73578 ± 34%      1043%     841209 ± 34%       452%     405848        
sched_debug.cpu.load.max
     1.81 ± 11%        77%       3.22             395%       8.98        
sched_debug.cpu.clock.stddev
     1.81 ± 11%        77%       3.22             395%       8.98        
sched_debug.cpu.clock_task.stddev
     8278 ± 10%       379%      39671 ± 18%       305%      33517        
sched_debug.cpu.load.avg
     3600             385%      17452            1023%      40419        
sched_debug.cpu.nr_load_updates.min
     5446             305%      22069             754%      46492        
sched_debug.cpu.nr_load_updates.avg
     8627 ±  5%       217%      27314             517%      53222        
sched_debug.cpu.nr_load_updates.max
     6221 ±  3%      2137%     139191            3486%     223092        
sched_debug.cpu.nr_switches.max
    15.67 ± 40%      3187%     515.00            1579%     263.00        
sched_debug.cpu.cpu_load[0].max
     2.55 ± 33%      3886%     101.45            1697%      45.73        
sched_debug.cpu.cpu_load[0].stddev
    15452 ± 19%       831%     143937 ± 22%       421%      80431        
sched_debug.cpu.load.stddev
     1144             236%       3839             329%       4911        
sched_debug.cpu.nr_load_updates.stddev
    23.67 ± 41%       709%     191.50 ±  6%       637%     174.33        
sched_debug.cpu.nr_uninterruptible.max
      978            7241%      71831 ±  3%     13746%     135493        
sched_debug.cpu.nr_switches.avg
     0.96 ± 19%      2503%      24.95            1720%      17.44        
sched_debug.cpu.cpu_load[0].avg
      957 ±  4%      3406%      33568            3626%      35679        
sched_debug.cpu.nr_switches.stddev
    29644 ± 16%       107%      61350 ±  8%       190%      86111        
sched_debug.cpu.clock.max
    29644 ± 16%       107%      61350 ±  8%       190%      86111        
sched_debug.cpu.clock_task.max
    29640 ± 16%       107%      61344 ±  8%       190%      86096        
sched_debug.cpu.clock.avg
    29640 ± 16%       107%      61344 ±  8%       190%      86096        
sched_debug.cpu.clock_task.avg
    29635 ± 16%       107%      61338 ±  8%       190%      86079        
sched_debug.cpu.clock.min
    29635 ± 16%       107%      61338 ±  8%       190%      86079        
sched_debug.cpu.clock_task.min
      335 ±  4%      7948%      27014           22596%      76183        
sched_debug.cpu.nr_switches.min
     1.62 ± 32%      1784%      30.61 ±  3%      1100%      19.51        
sched_debug.cpu.cpu_load[4].avg
     5.46 ± 15%      2325%     132.40            1031%      61.73        
sched_debug.cpu.nr_uninterruptible.stddev
      424 ± 11%       106%        875 ± 13%       263%       1541        
sched_debug.cpu.curr->pid.avg
     1400             166%       3721             264%       5100        
sched_debug.cpu.curr->pid.max
      610 ±  3%       108%       1269             126%       1380        
sched_debug.cpu.curr->pid.stddev
     0.43 ± 15%         4%       0.45 ± 16%        48%       0.64        
sched_debug.cpu.nr_running.avg
   253789 ± 13%        -5%     241499 ±  3%       -22%     198383        
sched_debug.cpu.avg_idle.stddev
    29638 ± 16%       107%      61339 ±  8%       190%      86079        
sched_debug.cpu_clk
    28529 ± 17%       111%      60238 ±  8%       198%      84957        
sched_debug.ktime
     0.17             -74%       0.04 ±  8%       -83%       0.03        
sched_debug.rt_rq:/.rt_time.avg
     0.85 ±  3%       -74%       0.22 ±  8%       -83%       0.14        
sched_debug.rt_rq:/.rt_time.stddev
     5.14 ± 10%       -75%       1.28 ±  6%       -83%       0.88        
sched_debug.rt_rq:/.rt_time.max
    29638 ± 16%       107%      61339 ±  8%       190%      86079        
sched_debug.sched_clk

aim7/1BRD_48G-xfs-disk_rd-9000-performance/ivb44

99091700659f4df9 ca2edab2e1d8f30dda874b7f71 fe9c2c81ed073878768785a985 ---------------- -------------------------- -------------------------- 473257 468546 5% 497512 aim7.jobs-per-min
   613996              11%     681283              -7%     571701        
aim7.time.involuntary_context_switches
     4914                        4977              -6%       4634        
aim7.time.system_time
   114.83                      115.98              -5%     109.23        
aim7.time.elapsed_time
   114.83                      115.98              -5%     109.23        
aim7.time.elapsed_time.max
    60711 ±  8%        20%      73007              -9%      55449        
aim7.time.voluntary_context_switches
     2509              -6%       2360              -4%       2416        
aim7.time.maximum_resident_set_size
   362268              19%     430263              -8%     332046        
softirqs.RCU
      352 ±  7%       -32%        238             -35%        230        
vmstat.procs.r
        5 ± 16%        80%          9             -40%          3        
vmstat.procs.b
     9584               7%      10255             -10%       8585        
vmstat.system.cs
    20442 ±  5%        38%      28201             -40%      12270        
cpuidle.C1-IVT.usage
     3.95              -3%       3.81               9%       4.29        
turbostat.CPU%c1
     0.81 ± 14%        44%       1.17              28%       1.04        
turbostat.Pkg%pc6
    19711 ±  5%        -7%      18413             -17%      16384        
meminfo.AnonHugePages
  3974485                     3977216              27%    5046310        
meminfo.DirectMap2M
   139742 ±  4%                137012             -17%     116493        
meminfo.DirectMap4k
   244933 ±  4%        -7%     228626              15%     280670        
meminfo.PageTables
    12.47 ± 39%        84%      22.89              64%      20.46        
perf-profile.func.cycles-pp.poll_idle
    57.44 ±  6%       -10%      51.55             -13%      50.13        
perf-profile.func.cycles-pp.intel_idle
     0.20               3%       0.20              -5%       0.19        
perf-stat.branch-miss-rate
5.356e+08               4%  5.552e+08              -6%  5.046e+08        
perf-stat.branch-misses
  1113549               7%    1187535             -15%     951607        
perf-stat.context-switches
 1.48e+13                   1.491e+13              -6%  1.397e+13        
perf-stat.cpu-cycles
   101697 ±  3%         9%     111167              -3%      98319        
perf-stat.cpu-migrations
     0.69 ± 20%       -17%       0.57             139%       1.65        
perf-stat.dTLB-load-miss-rate
3.264e+09 ± 19%       -17%  2.712e+09             148%  8.084e+09        
perf-stat.dTLB-load-misses
4.695e+11                   4.718e+11                   4.818e+11        
perf-stat.dTLB-loads
3.276e+11 ±  3%             3.303e+11               8%  3.528e+11        
perf-stat.dTLB-stores
    56.47 ± 19%        41%      79.48             -58%      23.96        
perf-stat.iTLB-load-miss-rate
 48864487 ±  4%         7%   52183944             -12%   43166037        
perf-stat.iTLB-load-misses
 40455495 ± 41%       -67%   13468883             239%   1.37e+08        
perf-stat.iTLB-loads
    29278 ±  4%        -6%      27480              12%      32844        
perf-stat.instructions-per-iTLB-miss
     0.10                        0.10               5%       0.10        
perf-stat.ipc
    47.16                       46.36                       46.51        
perf-stat.node-store-miss-rate
     6568 ± 44%       -59%       2721             -71%       1916        
numa-meminfo.node0.Shmem
   194395               7%     207086              15%     224164        
numa-meminfo.node0.Active
    10218 ± 24%       -37%       6471             -36%       6494        
numa-meminfo.node0.Mapped
     7496 ± 34%       -97%        204              37%      10278        
numa-meminfo.node0.AnonHugePages
   178888               6%     188799              16%     208213        
numa-meminfo.node0.AnonPages
   179468               6%     191062              17%     209704        
numa-meminfo.node0.Active(anon)
   256890             -15%     219489             -15%     219503        
numa-meminfo.node1.Active
    12213 ± 24%        49%      18208             -50%       6105        
numa-meminfo.node1.AnonHugePages
    45080 ± 23%       -33%      30138              87%      84468        
numa-meminfo.node1.PageTables
   241623             -15%     204604             -16%     203913        
numa-meminfo.node1.Active(anon)
   240637             -15%     204491             -15%     203847        
numa-meminfo.node1.AnonPages
 23782392 ±139%       673%  1.838e+08            -100%          0        
latency_stats.sum.wait_on_page_bit.__migration_entry_wait.migration_entry_wait.do_swap_page.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
    61157 ±  4%        -6%      57187              14%      69751        
proc-vmstat.nr_page_table_pages
     1641 ± 44%       -59%        679             -71%        478        
numa-vmstat.node0.nr_shmem
     2655 ± 23%       -35%       1715             -35%       1726        
numa-vmstat.node0.nr_mapped
    44867               5%      47231              16%      52261        
numa-vmstat.node0.nr_anon_pages
    45014               6%      47793              17%      52636        
numa-vmstat.node0.nr_zone_active_anon
    45014               6%      47793              17%      52636        
numa-vmstat.node0.nr_active_anon
    11300 ± 23%       -33%       7542              88%      21209        
numa-vmstat.node1.nr_page_table_pages
    60581             -16%      51156             -15%      51193        
numa-vmstat.node1.nr_zone_active_anon
    60581             -16%      51156             -15%      51193        
numa-vmstat.node1.nr_active_anon
    60328             -15%      51127             -15%      51174        
numa-vmstat.node1.nr_anon_pages
    13671                       13608              11%      15190        
slabinfo.cred_jar.active_objs
    13707                       13608              11%      15231        
slabinfo.cred_jar.num_objs
    24109                       24386             -11%      21574        
slabinfo.kmalloc-16.active_objs
    24109                       24386             -11%      21574        
slabinfo.kmalloc-16.num_objs
    13709 ±  6%                 13391             -15%      11600        
slabinfo.kmalloc-512.active_objs
    13808 ±  6%                 13454             -16%      11657        
slabinfo.kmalloc-512.num_objs
  1456658               4%    1511260              15%    1675984        
sched_debug.cfs_rq:/.min_vruntime.min
   441613 ±  3%       -28%     316751             -76%     105734        
sched_debug.cfs_rq:/.min_vruntime.stddev
   443999 ±  3%       -28%     318033             -76%     106909        
sched_debug.cfs_rq:/.spread0.stddev
  2657974                     2625551             -19%    2158111        
sched_debug.cfs_rq:/.min_vruntime.max
     0.22 ± 23%        96%       0.43             109%       0.46        
sched_debug.cfs_rq:/.nr_spread_over.stddev
     1.50             100%       3.00             133%       3.50        
sched_debug.cfs_rq:/.nr_spread_over.max
   111.95 ± 26%        15%     128.92             128%     254.81        
sched_debug.cfs_rq:/.exec_clock.stddev
      802               3%        829             -16%        671        
sched_debug.cfs_rq:/.load_avg.min
      874                         879             -11%        780        
sched_debug.cfs_rq:/.load_avg.avg
     1256 ± 17%       -20%       1011             -24%        957        
sched_debug.cfs_rq:/.load_avg.max
     1.33 ± 35%      -100%       0.00             200%       4.00        
sched_debug.cpu.cpu_load[4].min
     4.56 ±  6%       -11%       4.07             -27%       3.33        
sched_debug.cpu.cpu_load[4].stddev
     4.76 ±  3%       -13%       4.14             -30%       3.35        
sched_debug.cpu.cpu_load[3].stddev
    25.17 ± 12%       -26%      18.50             -21%      20.00        
sched_debug.cpu.cpu_load[3].max
    25.67 ±  9%       -32%      17.50             -24%      19.50        
sched_debug.cpu.cpu_load[0].max
     4.67 ±  3%       -17%       3.90             -22%       3.62        
sched_debug.cpu.cpu_load[0].stddev
     4.88             -15%       4.14             -31%       3.39        
sched_debug.cpu.cpu_load[2].stddev
    26.17 ± 10%       -29%      18.50             -25%      19.50        
sched_debug.cpu.cpu_load[2].max
     7265               4%       7556             -12%       6419        
sched_debug.cpu.nr_switches.avg
     9.41 ± 10%                  9.67              21%      11.38        
sched_debug.cpu.cpu_load[1].avg
     9.03 ± 12%         3%       9.32              23%      11.09        
sched_debug.cpu.cpu_load[0].avg
     4140 ±  4%       -11%       3698             -11%       3703        
sched_debug.cpu.nr_switches.stddev
     9.41 ± 10%         3%       9.71              22%      11.49        
sched_debug.cpu.cpu_load[3].avg
     4690                        4821              -9%       4257        
sched_debug.cpu.nr_switches.min
     9.39 ±  9%         3%       9.69              23%      11.52        
sched_debug.cpu.cpu_load[4].avg
     9.43 ± 10%                  9.71              21%      11.44        
sched_debug.cpu.cpu_load[2].avg
    57.92 ± 18%        -4%      55.55             -23%      44.50        
sched_debug.cpu.nr_uninterruptible.stddev
     3002 ±  3%        10%       3288              31%       3919        
sched_debug.cpu.curr->pid.avg
     6666                        6652             -10%       6025        
sched_debug.cpu.curr->pid.max
     1379                        1361             -19%       1118        
sched_debug.cpu.ttwu_local.avg
     1849 ±  3%       -12%       1628             -18%       1517        
sched_debug.cpu.ttwu_local.stddev
     1679 ±  8%         4%       1738             -15%       1423        
sched_debug.cpu.curr->pid.stddev
     1.58 ± 33%       -11%       1.41              65%       2.60        
sched_debug.cpu.nr_running.avg
     1767               6%       1880             -16%       1489        
sched_debug.cpu.ttwu_count.avg
      506 ±  6%       -15%        430             -17%        419        
sched_debug.cpu.ttwu_count.min
     7139               8%       7745             -11%       6355        
sched_debug.cpu.sched_count.avg
     4355               6%       4625             -11%       3884        
sched_debug.cpu.sched_count.min
     4.91 ±  3%       -16%       4.13             -28%       3.52        
sched_debug.cpu.cpu_load[1].stddev
    26.67 ±  9%       -29%      19.00             -27%      19.50        
sched_debug.cpu.cpu_load[1].max
      209 ±  8%        19%        247             -15%        178        
sched_debug.cpu.sched_goidle.avg
     5.67 ± 27%       -12%       5.00              50%       8.50        
sched_debug.cpu.nr_running.max
    36072 ±  7%        70%      61152              17%      42236        
sched_debug.cpu.sched_count.max
     2008              -8%       1847             -18%       1645        
sched_debug.cpu.ttwu_count.stddev
     0.07 ± 19%       -20%       0.06             186%       0.21        
sched_debug.rt_rq:/.rt_time.avg
     0.36 ± 17%       -23%       0.28             142%       0.88        
sched_debug.rt_rq:/.rt_time.stddev
     2.33 ± 15%       -27%       1.70              87%       4.35        
sched_debug.rt_rq:/.rt_time.max

aim7/1BRD_48G-xfs-disk_cp-3000-performance/ivb44

99091700659f4df9 ca2edab2e1d8f30dda874b7f71 fe9c2c81ed073878768785a985 ---------------- -------------------------- -------------------------- 270459 272267 ± 3% -48% 139834 ± 3% aim7.jobs-per-min
    21229 ±  5%                 20896 ±  3%       449%     116516 ±  6%  
aim7.time.involuntary_context_switches
     1461 ±  5%                  1454 ±  5%       174%       3998 ±  3%  
aim7.time.system_time
   155368                      153149             149%     386164        
aim7.time.minor_page_faults
    66.84                       66.41 ±  3%        93%     129.07 ±  3%  
aim7.time.elapsed_time
    66.84                       66.41 ±  3%        93%     129.07 ±  3%  
aim7.time.elapsed_time.max
   328369               3%     339077              96%     644393        
aim7.time.voluntary_context_switches
    49489 ± 11%        -8%      45459              39%      68941 ±  4%  
interrupts.CAL:Function_call_interrupts
    96.62 ±  7%                 97.09              61%     155.12        
uptime.boot
   186640 ± 10%                186707             127%     424522 ±  4%  
softirqs.RCU
   146596                      147043              37%     201373        
softirqs.SCHED
  1005660 ±  3%                991053 ±  4%       118%    2196513        
softirqs.TIMER
    66.84                       66.41 ±  3%        93%     129.07 ±  3%  
time.elapsed_time
    66.84                       66.41 ±  3%        93%     129.07 ±  3%  
time.elapsed_time.max
    21229 ±  5%                 20896 ±  3%       449%     116516 ±  6%  
time.involuntary_context_switches
   155368                      153149             149%     386164        
time.minor_page_faults
     2212                        2215              41%       3112        
time.percent_of_cpu_this_job_got
     1461 ±  5%                  1454 ±  5%       174%       3998 ±  3%  
time.system_time
   328369               3%     339077              96%     644393        
time.voluntary_context_switches
  1197810 ± 16%       -67%     393936 ± 40%       -56%     530668 ± 43%  
numa-numastat.node0.numa_miss
  1196269 ± 16%       -68%     387751 ± 40%       -55%     533013 ± 42%  
numa-numastat.node1.numa_foreign
      112             159%        292 ±  4%       146%        277        
vmstat.memory.buff
 16422228                    16461619             -28%   11832310        
vmstat.memory.free
       22              -3%         22              87%         42 ±  3%  
vmstat.procs.r
    48853                       48768                       50273        
vmstat.system.in
   125202               8%     135626              51%     189515 ±  4%  
cpuidle.C1-IVT.usage
 28088338 ±  3%        11%   31082173              17%   32997314 ±  5%  
cpuidle.C1-IVT.time
  3471814              27%    4422338 ± 15%      2877%  1.034e+08 ±  3%  
cpuidle.C1E-IVT.time
    33353               8%      36128             703%     267725        
cpuidle.C1E-IVT.usage
 11371800               9%   12381174             244%   39113028        
cpuidle.C3-IVT.time
    64048               5%      67490              62%     103940        
cpuidle.C3-IVT.usage
1.637e+09                   1.631e+09              20%  1.959e+09        
cpuidle.C6-IVT.time
  1861259               4%    1931551              19%    2223599        
cpuidle.C6-IVT.usage
      230 ±  9%        42%        326            1631%       3986        
cpuidle.POLL.usage
  1724995 ± 41%        54%    2656939 ± 10%       112%    3662791        
cpuidle.POLL.time
    48.48                       48.15              36%      65.85        
turbostat.%Busy
     1439                        1431              36%       1964        
turbostat.Avg_MHz
    33.28                       33.45             -25%      24.85        
turbostat.CPU%c1
    18.09 ±  3%                 18.24 ±  4%       -49%       9.16        
turbostat.CPU%c6
      134                         133               8%        144        
turbostat.CorWatt
     5.39 ± 17%         4%       5.63 ±  8%       -34%       3.54        
turbostat.Pkg%pc2
     2.97 ± 44%       -17%       2.48 ± 32%       -70%       0.91 ± 22%  
turbostat.Pkg%pc6
      167                         167               6%        177        
turbostat.PkgWatt
    10.27                       10.43             -14%       8.79        
turbostat.RAMWatt
 44376005            -100%     205734            -100%     214640        
meminfo.Active
 44199835            -100%      30412            -100%      30241        
meminfo.Active(file)
   103029 ±  3%        27%     130507 ±  6%        29%     133114 ±  8%  
meminfo.CmaFree
   124701 ±  4%                123685 ± 14%        16%     144180 ±  3%  
meminfo.DirectMap4k
     7886 ±  4%                  7993 ±  5%       144%      19231 ±  7%  
meminfo.Dirty
  2472446            1791%   46747572            1976%   51320420        
meminfo.Inactive
  2463353            1797%   46738477            1983%   51311261        
meminfo.Inactive(file)
 16631615                    16664565             -28%   11936074        
meminfo.MemFree
4.125e+11              -5%  3.927e+11             103%   8.36e+11        
perf-stat.branch-instructions
     0.41             -20%       0.33             -43%       0.23        
perf-stat.branch-miss-rate
1.671e+09             -23%   1.28e+09              16%  1.946e+09        
perf-stat.branch-misses
7.138e+09              -3%  6.917e+09              23%  8.746e+09        
perf-stat.cache-misses
2.036e+10              -4%  1.956e+10              22%  2.476e+10        
perf-stat.cache-references
   821470               4%     851532              88%    1548125 ±  3%  
perf-stat.context-switches
 4.93e+12 ±  3%        -4%  4.755e+12 ±  4%       154%   1.25e+13        
perf-stat.cpu-cycles
   125073               4%     129993             167%     333599        
perf-stat.cpu-migrations
3.595e+09 ± 16%       -19%  2.895e+09 ± 17%        39%  4.987e+09 ± 10%  
perf-stat.dTLB-load-misses
6.411e+11                   6.339e+11 ±  3%        57%  1.004e+12        
perf-stat.dTLB-loads
     0.06 ±  3%       -42%       0.04              87%       0.12 ±  3%  
perf-stat.dTLB-store-miss-rate
2.738e+08             -39%  1.675e+08              64%  4.502e+08 ±  5%  
perf-stat.dTLB-store-misses
4.321e+11               5%  4.552e+11             -12%   3.81e+11 ±  8%  
perf-stat.dTLB-stores
2.343e+12              -5%  2.229e+12              67%  3.918e+12        
perf-stat.instructions
    46162 ± 41%                 46733 ±  3%        55%      71500        
perf-stat.instructions-per-iTLB-miss
     0.48 ±  4%                  0.47 ±  5%       -34%       0.31        
perf-stat.ipc
   325877                      322934             115%     699924        
perf-stat.minor-faults
    42.88               3%      44.33                       43.65        
perf-stat.node-load-miss-rate
9.499e+08                   9.578e+08              66%  1.581e+09        
perf-stat.node-load-misses
1.266e+09              -5%  1.203e+09              61%   2.04e+09        
perf-stat.node-loads
    39.17                       40.00               8%      42.12        
perf-stat.node-store-miss-rate
3.198e+09               4%  3.318e+09              36%  4.344e+09        
perf-stat.node-store-misses
4.966e+09                   4.977e+09              20%  5.968e+09        
perf-stat.node-stores
   325852                      322963             115%     699918        
perf-stat.page-faults
 21719324            -100%      15215 ±  3%      -100%      14631        
numa-meminfo.node0.Active(file)
  1221037            1806%   23278263            1969%   25269114        
numa-meminfo.node0.Inactive(file)
  1223564            1803%   23286857            1965%   25269597        
numa-meminfo.node0.Inactive
 21811771            -100%     102448            -100%     104424        
numa-meminfo.node0.Active
     2971 ± 13%        -8%       2734 ±  3%       157%       7626 ±  4%  
numa-meminfo.node0.Dirty
  8476780                     8356206             -27%    6162743        
numa-meminfo.node0.MemFree
   617361                      611434              11%     687829        
numa-meminfo.node0.SReclaimable
  1249068            1779%   23471025            1985%   26046948        
numa-meminfo.node1.Inactive
  1242501            1789%   23470523            1996%   26038272        
numa-meminfo.node1.Inactive(file)
 22500867            -100%      15202 ±  4%      -100%      15613        
numa-meminfo.node1.Active(file)
 22584509            -100%     103192 ±  6%      -100%     109976        
numa-meminfo.node1.Active
     4814 ± 13%                  4957 ±  5%       135%      11335        
numa-meminfo.node1.Dirty
  8132889                     8297084 ±  3%       -29%    5777419 ±  3%  
numa-meminfo.node1.MemFree
    83641 ±  7%         5%      87990 ±  7%        13%      94363        
numa-meminfo.node1.Active(anon)
    82877 ±  7%         4%      86528 ±  6%        13%      93620        
numa-meminfo.node1.AnonPages
        0                           0                      842360 ±100%  
latency_stats.avg.call_rwsem_down_write_failed.do_unlinkat.SyS_unlink.do_syscall_64.return_from_SYSCALL_64
      113 ±173%       232%        376 ±100%     2e+05%     203269 ±  4%  
latency_stats.hits.wait_on_page_bit.__migration_entry_wait.migration_entry_wait.do_swap_page.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
     5245 ± 14%                  5325 ±  3%       535%      33286 ± 23%  
latency_stats.max.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agi.xfs_ialloc_read_agi.xfs_dialloc.xfs_ialloc.xfs_dir_ialloc.xfs_create
     1133 ±173%       113%       2416 ±100%      1351%      16434        
latency_stats.max.wait_on_page_bit.__migration_entry_wait.migration_entry_wait.do_swap_page.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
        0                           0                      842360 ±100%  
latency_stats.max.call_rwsem_down_write_failed.do_unlinkat.SyS_unlink.do_syscall_64.return_from_SYSCALL_64
     7813 ± 13%       -33%       5197 ±  9%       403%      39305 ± 18%  
latency_stats.max.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agi.xfs_iunlink_remove.xfs_ifree.xfs_inactive_ifree.xfs_inactive.xfs_fs_destroy_inode
     5271 ± 13%        -3%       5091 ±  5%       288%      20467        
latency_stats.max.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agi.xfs_iunlink.xfs_droplink.xfs_remove.xfs_vn_unlink.vfs_unlink
    10369 ± 17%       -41%       6086 ± 21%       -96%        385 ±100%  
latency_stats.max.wait_on_page_bit.truncate_inode_pages_range.truncate_inode_pages_final.evict.iput.dentry_unlink_inode.__dentry_kill.dput.__fput.____fput.task_work_run.exit_to_usermode_loop
    94417 ±173%       556%     619712 ±100%     3e+05%  3.061e+08 ±  5%  
latency_stats.sum.wait_on_page_bit.__migration_entry_wait.migration_entry_wait.do_swap_page.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
 22126648 ±  4%              22776886            1311%  3.123e+08 ±  7%  
latency_stats.sum.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agi.xfs_iunlink_remove.xfs_ifree.xfs_inactive_ifree.xfs_inactive.xfs_fs_destroy_inode
     2536 ±117%       -98%         48 ± 43%      2059%      54765 ±100%  
latency_stats.sum.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agf.xfs_alloc_read_agf.xfs_alloc_fix_freelist.xfs_free_extent_fix_freelist.xfs_free_extent.xfs_trans_free_extent
  1702264 ±  3%         5%    1790192             509%   10359205 ±  6%  
latency_stats.sum.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agi.xfs_ialloc_read_agi.xfs_dialloc.xfs_ialloc.xfs_dir_ialloc.xfs_create
  1180839 ±  3%         5%    1238547             453%    6527115 ±  5%  
latency_stats.sum.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agi.xfs_iunlink.xfs_droplink.xfs_remove.xfs_vn_unlink.vfs_unlink
      467 ±173%       680%       3644 ±  4%     7e+05%    3196407 ±  3%  
latency_stats.sum.xfs_iget.xfs_ialloc.xfs_dir_ialloc.xfs_create.xfs_generic_create.xfs_vn_mknod.xfs_vn_create.path_openat.do_filp_open.do_sys_open.SyS_creat.entry_SYSCALL_64_fastpath
        0                           0                      842360 ±100%  
latency_stats.sum.call_rwsem_down_write_failed.do_unlinkat.SyS_unlink.do_syscall_64.return_from_SYSCALL_64
   159018 ± 43%       -49%      81514 ± 19%       -99%        999 ±100%  
latency_stats.sum.wait_on_page_bit.truncate_inode_pages_range.truncate_inode_pages_final.evict.iput.dentry_unlink_inode.__dentry_kill.dput.__fput.____fput.task_work_run.exit_to_usermode_loop
     1084 ±  5%        21%       1313 ±  3%        86%       2018        
proc-vmstat.kswapd_high_wmark_hit_quickly
     1817 ±  3%        38%       2511 ±  3%       175%       4989        
proc-vmstat.kswapd_low_wmark_hit_quickly
 11055004            -100%       7603            -100%       7559        
proc-vmstat.nr_active_file
     1993                        2013 ±  4%       128%       4553 ±  5%  
proc-vmstat.nr_dirty
    25746 ±  3%        26%      32494 ±  6%        29%      33319 ±  8%  
proc-vmstat.nr_free_cma
  4152484                     4162399             -28%    2984494        
proc-vmstat.nr_free_pages
   615907            1798%   11688190            1983%   12827366        
proc-vmstat.nr_inactive_file
 11055042            -100%       7603            -100%       7559        
proc-vmstat.nr_zone_active_file
   615904            1798%   11688234            1983%   12827434        
proc-vmstat.nr_zone_inactive_file
     2016 ±  3%                  2025 ±  4%       127%       4582 ±  4%  
proc-vmstat.nr_zone_write_pending
     2912 ±  3%        32%       3834 ±  3%       141%       7009        
proc-vmstat.pageoutrun
  5380414            -100%       2502 ±  3%      -100%       2602 ±  3%  
proc-vmstat.pgactivate
 61925072            -100%          0            -100%          0        
proc-vmstat.pgdeactivate
   348105                      343315             108%     723517        
proc-vmstat.pgfault
 61932469            -100%          0            -100%          0        
proc-vmstat.pgrefill
  5432311            -100%       3802 ±  3%      -100%       3657        
numa-vmstat.node0.nr_zone_active_file
  5432276            -100%       3802 ±  3%      -100%       3657        
numa-vmstat.node0.nr_active_file
   305236            1802%    5806215            1969%    6314975        
numa-vmstat.node0.nr_zone_inactive_file
   305239            1802%    5806170            1969%    6314910        
numa-vmstat.node0.nr_inactive_file
      748 ±  7%       -20%        597 ± 10%       114%       1602        
numa-vmstat.node0.nr_dirty
      775 ±  7%       -21%        610 ± 12%       112%       1642        
numa-vmstat.node0.nr_zone_write_pending
  2116796                     2102494 ±  3%       -27%    1543100        
numa-vmstat.node0.nr_free_pages
   154392                      152538              11%     171898        
numa-vmstat.node0.nr_slab_reclaimable
   310642            1784%    5853811            1995%    6507801        
numa-vmstat.node1.nr_zone_inactive_file
   310642            1784%    5853814            1995%    6507801        
numa-vmstat.node1.nr_inactive_file
  5627293            -100%       3799 ±  4%      -100%       3903        
numa-vmstat.node1.nr_zone_active_file
  5627293            -100%       3799 ±  4%      -100%       3903        
numa-vmstat.node1.nr_active_file
     1206 ± 16%        14%       1373             129%       2758 ± 10%  
numa-vmstat.node1.nr_zone_write_pending
     1205 ± 16%        14%       1373             129%       2757 ± 10%  
numa-vmstat.node1.nr_dirty
  2031121                     2088592 ±  3%       -29%    1446172 ±  3%  
numa-vmstat.node1.nr_free_pages
    25743 ±  3%        27%      32608 ±  7%        30%      33415 ±  8%  
numa-vmstat.node1.nr_free_cma
    20877 ±  7%         6%      22077 ±  6%        13%      23620        
numa-vmstat.node1.nr_zone_active_anon
    20877 ±  7%         6%      22077 ±  6%        13%      23620        
numa-vmstat.node1.nr_active_anon
    20684 ±  7%         5%      21709 ±  6%        13%      23431        
numa-vmstat.node1.nr_anon_pages
     4687                        4704              11%       5205 ±  3%  
slabinfo.kmalloc-128.num_objs
     4687                        4704              11%       5205 ±  3%  
slabinfo.kmalloc-128.active_objs
     1401             -19%       1142               8%       1516 ±  6%  
slabinfo.xfs_efd_item.num_objs
     1401             -19%       1142               8%       1516 ±  6%  
slabinfo.xfs_efd_item.active_objs
     1725 ±  5%        -8%       1589             -12%       1518        
slabinfo.xfs_inode.num_objs
     1725 ±  5%        -8%       1589             -12%       1518        
slabinfo.xfs_inode.active_objs
   382810 ±  4%                383813 ±  3%       301%    1535378        
sched_debug.cfs_rq:/.min_vruntime.avg
   249011 ±  6%                245840 ±  3%       420%    1294704        
sched_debug.cfs_rq:/.min_vruntime.min
   105216                      106278              79%     188096        
sched_debug.cfs_rq:/.min_vruntime.stddev
   105260                      106358              79%     188314        
sched_debug.cfs_rq:/.spread0.stddev
     9414 ±  4%                  9361 ±  4%       230%      31092        
sched_debug.cfs_rq:/.exec_clock.min
   541056 ±  9%                540188 ±  3%       236%    1820030        
sched_debug.cfs_rq:/.min_vruntime.max
   150.87 ± 11%       -21%     119.80 ± 10%        34%     202.73 ±  7%  
sched_debug.cfs_rq:/.util_avg.stddev
    13783                       13656             170%      37192        
sched_debug.cfs_rq:/.exec_clock.avg
    17625                       17508             141%      42564        
sched_debug.cfs_rq:/.exec_clock.max
  3410.74 ±  3%               3458.30              38%    4706.14        
sched_debug.cfs_rq:/.exec_clock.stddev
      732 ± 11%        11%        809 ±  3%       -34%        480 ±  7%  
sched_debug.cfs_rq:/.load_avg.min
      844 ±  8%         7%        901             -33%        569 ±  4%  
sched_debug.cfs_rq:/.load_avg.avg
     0.41 ±  7%        11%       0.46 ± 11%        21%       0.50 ±  5%  
sched_debug.cfs_rq:/.nr_running.avg
     1339 ±  5%                  1338             -32%        909        
sched_debug.cfs_rq:/.load_avg.max
     0.53 ±  4%        -4%       0.51              32%       0.70        
sched_debug.cfs_rq:/.nr_spread_over.avg
     0.50                        0.50              33%       0.67        
sched_debug.cfs_rq:/.nr_spread_over.min
   355.00 ± 26%       -67%     118.75 ±  4%       -82%      64.83 ± 20%  
sched_debug.cpu.cpu_load[4].max
    18042                       17697             135%      42380        
sched_debug.cpu.nr_load_updates.min
    51.83 ± 22%       -66%      17.44             -78%      11.18 ±  5%  
sched_debug.cpu.cpu_load[4].stddev
    22708                       22546             111%      47986        
sched_debug.cpu.nr_load_updates.avg
    29633 ±  7%        -7%      27554              83%      54243        
sched_debug.cpu.nr_load_updates.max
    48.83 ± 29%       -65%      16.91 ± 29%       -73%      13.34 ± 13%  
sched_debug.cpu.cpu_load[3].stddev
   329.25 ± 34%       -65%     113.75 ± 30%       -76%      79.67 ± 28%  
sched_debug.cpu.cpu_load[3].max
    17106              14%      19541 ± 19%        34%      22978 ±  6%  
sched_debug.cpu.nr_switches.max
     1168 ±  4%        -3%       1131 ±  4%       144%       2846 ± 21%  
sched_debug.cpu.ttwu_local.max
     3826 ±  3%                  3766              17%       4487        
sched_debug.cpu.nr_load_updates.stddev
    19.73 ± 12%        -4%      18.86 ± 14%        59%      31.42 ±  8%  
sched_debug.cpu.nr_uninterruptible.avg
   149.75 ±  8%                150.00 ± 11%        42%     212.50        
sched_debug.cpu.nr_uninterruptible.max
    98147 ± 34%                 97985 ± 42%        59%     156085 ±  8%  
sched_debug.cpu.avg_idle.min
     8554 ±  3%         4%       8896 ±  5%        62%      13822        
sched_debug.cpu.nr_switches.avg
     2582 ±  3%        11%       2857 ± 11%        19%       3083 ±  3%  
sched_debug.cpu.nr_switches.stddev
    60029 ±  9%                 60817 ±  7%        44%      86205        
sched_debug.cpu.clock.max
    60029 ±  9%                 60817 ±  7%        44%      86205        
sched_debug.cpu.clock_task.max
    60020 ±  9%                 60807 ±  7%        44%      86188        
sched_debug.cpu.clock.avg
    60020 ±  9%                 60807 ±  7%        44%      86188        
sched_debug.cpu.clock_task.avg
    60008 ±  9%                 60793 ±  7%        44%      86169        
sched_debug.cpu.clock.min
    60008 ±  9%                 60793 ±  7%        44%      86169        
sched_debug.cpu.clock_task.min
    18.36 ±  7%       -37%      11.60 ±  5%       -33%      12.21        
sched_debug.cpu.cpu_load[3].avg
     5577 ±  6%         3%       5772 ±  6%        81%      10121        
sched_debug.cpu.nr_switches.min
    19.14 ±  3%       -36%      12.24             -36%      12.33        
sched_debug.cpu.cpu_load[4].avg
    17.21 ± 14%       -31%      11.90 ± 18%       -27%      12.56 ±  6%  
sched_debug.cpu.cpu_load[2].avg
    83.49 ±  7%         5%      87.64 ±  3%        17%      97.56 ±  4%  
sched_debug.cpu.nr_uninterruptible.stddev
     3729                        3735              18%       4409 ± 13%  
sched_debug.cpu.curr->pid.max
      374 ±  9%        -4%        360 ±  9%       157%        962        
sched_debug.cpu.ttwu_local.min
      665                         671             122%       1479        
sched_debug.cpu.ttwu_local.avg
      196 ±  7%         5%        207 ±  8%        88%        369 ± 14%  
sched_debug.cpu.ttwu_local.stddev
     1196 ±  4%         5%       1261 ±  6%        11%       1333 ± 10%  
sched_debug.cpu.curr->pid.stddev
     0.45 ±  7%        17%       0.53 ± 16%        29%       0.58 ± 16%  
sched_debug.cpu.nr_running.avg
     6738 ± 16%         8%       7296 ± 20%        52%      10236        
sched_debug.cpu.ttwu_count.max
     3952 ±  4%         5%       4150 ±  5%        75%       6917        
sched_debug.cpu.ttwu_count.avg
      913              22%       1117 ± 18%        42%       1302 ±  3%  
sched_debug.cpu.sched_goidle.stddev
     2546 ±  4%         4%       2653 ±  7%        89%       4816        
sched_debug.cpu.ttwu_count.min
     5301 ±  6%        36%       7190 ± 33%        61%       8513 ±  8%  
sched_debug.cpu.sched_goidle.max
     4683 ± 16%        14%       5355 ± 25%        52%       7125        
sched_debug.cpu.sched_count.stddev
     8262 ±  3%         6%       8746 ±  7%        68%      13912        
sched_debug.cpu.sched_count.avg
     5139 ±  5%         4%       5362 ±  6%        90%       9773        
sched_debug.cpu.sched_count.min
     2088 ±  6%         7%       2229 ±  5%        55%       3242        
sched_debug.cpu.sched_goidle.min
     3258 ±  4%         6%       3445 ±  6%        44%       4706        
sched_debug.cpu.sched_goidle.avg
    37088 ± 17%        12%      41540 ± 23%        60%      59447        
sched_debug.cpu.sched_count.max
     1007 ±  7%        13%       1139 ± 14%        38%       1386 ±  3%  
sched_debug.cpu.ttwu_count.stddev
   262591 ±  4%        -3%     253748 ±  4%       -11%     232974        
sched_debug.cpu.avg_idle.stddev
    60009 ±  9%                 60795 ±  7%        44%      86169        
sched_debug.cpu_clk
    58763 ±  9%                 59673 ±  7%        45%      85068        
sched_debug.ktime
    60009 ±  9%                 60795 ±  7%        44%      86169        
sched_debug.sched_clk

aim7/1BRD_48G-xfs-creat-clo-1500-performance/ivb44

99091700659f4df9 fe9c2c81ed073878768785a985 ---------------- -------------------------- 69789 5% 73162 aim7.jobs-per-min
    81603              -7%      75897 ±  5%  
aim7.time.involuntary_context_switches
     3825              -6%       3583        aim7.time.system_time
   129.08              -5%     123.16        aim7.time.elapsed_time
   129.08              -5%     123.16        aim7.time.elapsed_time.max
     2536              -4%       2424        aim7.time.maximum_resident_set_size
     3145             131%       7253 ± 20%  numa-numastat.node1.numa_miss
     3145             131%       7253 ± 20%  numa-numastat.node1.numa_foreign
     7059               4%       7362        vmstat.system.cs
  7481848              40%   10487336 ±  8%  cpuidle.C1-IVT.time
  1491314              75%    2607219 ± 10%  cpuidle.POLL.time
       67              10%         73 ±  4%  turbostat.CoreTmp
       66              12%         73 ±  4%  turbostat.PkgTmp
  5025792             -21%    3973802        meminfo.DirectMap2M
    49098              12%      54859        meminfo.PageTables
     3.94              97%       7.76 ± 18%  
perf-profile.cycles-pp.poll_idle.cpuidle_enter_state.cpuidle_enter.call_cpuidle.cpu_startup_entry
    11.88             -24%       8.99 ± 14%  
perf-profile.cycles-pp.apic_timer_interrupt.cpuidle_enter.call_cpuidle.cpu_startup_entry.start_secondary
    11.63             -25%       8.78 ± 13%  
perf-profile.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter.call_cpuidle.cpu_startup_entry
8.412e+11              -7%   7.83e+11        perf-stat.branch-instructions
     0.30                        0.29        perf-stat.branch-miss-rate
2.495e+09              -8%  2.292e+09        perf-stat.branch-misses
4.277e+09              -6%  4.003e+09        perf-stat.cache-misses
1.396e+10              -5%  1.327e+10        perf-stat.cache-references
1.224e+13              -8%   1.12e+13        perf-stat.cpu-cycles
     0.58             -57%       0.25 ± 16%  perf-stat.dTLB-load-miss-rate
5.407e+09             -60%  2.175e+09 ± 18%  perf-stat.dTLB-load-misses
9.243e+11              -6%  8.708e+11        perf-stat.dTLB-loads
     0.17             -58%       0.07 ±  4%  perf-stat.dTLB-store-miss-rate
4.368e+08             -50%  2.177e+08 ±  3%  perf-stat.dTLB-store-misses
2.549e+11              19%  3.041e+11        perf-stat.dTLB-stores
3.737e+12              -6%  3.498e+12        perf-stat.instructions
     0.31                        0.31        perf-stat.ipc
   439716                      426816        perf-stat.minor-faults
2.164e+09              -7%  2.012e+09        perf-stat.node-load-misses
2.417e+09              -7%  2.259e+09        perf-stat.node-loads
 1.24e+09              -3%  1.198e+09        perf-stat.node-store-misses
1.556e+09              -4%  1.501e+09        perf-stat.node-stores
   439435                      426823        perf-stat.page-faults
    51452              14%      58403 ±  8%  numa-meminfo.node0.Active(anon)
    10472             -36%       6692 ± 45%  numa-meminfo.node1.Shmem
     7665              74%      13316        numa-meminfo.node1.PageTables
     6724             144%      16416 ± 43%  
latency_stats.avg.perf_event_alloc.SYSC_perf_event_open.SyS_perf_event_open.entry_SYSCALL_64_fastpath
     6724             144%      16416 ± 43%  
latency_stats.max.perf_event_alloc.SYSC_perf_event_open.SyS_perf_event_open.entry_SYSCALL_64_fastpath
     6724             144%      16416 ± 43%  
latency_stats.sum.perf_event_alloc.SYSC_perf_event_open.SyS_perf_event_open.entry_SYSCALL_64_fastpath
    12237              12%      13693        proc-vmstat.nr_page_table_pages
    12824              14%      14578 ±  8%  
numa-vmstat.node0.nr_zone_active_anon
    12824              14%      14578 ±  8%  numa-vmstat.node0.nr_active_anon
     2618             -36%       1672 ± 45%  numa-vmstat.node1.nr_shmem
    17453              24%      21726 ±  6%  numa-vmstat.node1.numa_miss
     1909              74%       3323        
numa-vmstat.node1.nr_page_table_pages
    17453              24%      21726 ±  6%  numa-vmstat.node1.numa_foreign
      922              24%       1143 ±  6%  
slabinfo.blkdev_requests.active_objs
      922              24%       1143 ±  6%  slabinfo.blkdev_requests.num_objs
      569              21%        686 ± 11%  
slabinfo.file_lock_cache.active_objs
      569              21%        686 ± 11%  slabinfo.file_lock_cache.num_objs
     9.07              16%      10.56 ±  9%  
sched_debug.cfs_rq:/.runnable_load_avg.avg
    18406             -14%      15835 ± 10%  sched_debug.cfs_rq:/.load.stddev
     0.67             150%       1.67 ± 43%  
sched_debug.cfs_rq:/.nr_spread_over.max
      581             -11%        517 ±  4%  sched_debug.cfs_rq:/.load_avg.min
      659             -10%        596 ±  4%  sched_debug.cfs_rq:/.load_avg.avg
      784             -12%        692 ±  4%  sched_debug.cfs_rq:/.load_avg.max
    18086             -12%      15845 ±  9%  sched_debug.cpu.load.stddev
    18.72             -17%      15.49 ±  8%  
sched_debug.cpu.nr_uninterruptible.avg
    69.33              42%      98.67 ±  7%  
sched_debug.cpu.nr_uninterruptible.max
   317829             -12%     280218 ±  4%  sched_debug.cpu.avg_idle.min
     9.80              18%      11.54 ± 10%  sched_debug.cpu.cpu_load[1].avg
     8.91              15%      10.28 ±  9%  sched_debug.cpu.cpu_load[0].avg
     9.53              22%      11.64 ± 10%  sched_debug.cpu.cpu_load[3].avg
     7083              11%       7853        sched_debug.cpu.nr_switches.min
     9.73              22%      11.90 ±  7%  sched_debug.cpu.cpu_load[4].avg
     9.68              20%      11.59 ± 11%  sched_debug.cpu.cpu_load[2].avg
    24.59              49%      36.53 ± 17%  
sched_debug.cpu.nr_uninterruptible.stddev
     1176              12%       1319 ±  4%  sched_debug.cpu.curr->pid.avg
      373              35%        502 ±  6%  sched_debug.cpu.ttwu_local.min
     3644              13%       4120 ±  3%  sched_debug.cpu.ttwu_count.min
     4855              13%       5463 ±  6%  sched_debug.cpu.sched_goidle.max
     7019              10%       7745        sched_debug.cpu.sched_count.min
     2305              10%       2529 ±  3%  sched_debug.cpu.sched_goidle.min
     0.00             -19%       0.00 ±  7%  sched_debug.cpu.next_balance.stddev
     0.68             -17%       0.57 ± 11%  sched_debug.cpu.nr_running.stddev
     0.05              27%       0.06 ± 14%  
sched_debug.rt_rq:/.rt_nr_running.stddev

Thanks,
Fengguang

Reply via email to