FYI, we noticed the below changes on

https://github.com/0day-ci/linux 
Vladimir-Davydov/vmpressure-Fix-subtree-pressure-detection/20160128-003153
commit 78d1788350477246d516496d8d7684fa80ef7f18 ("vmpressure: Fix subtree 
pressure detection")


=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
  
gcc-4.9/performance/x86_64-rhel/debian-x86_64-2015-02-07.cgz/300s/ivb43/lru-file-mmap-read/vm-scalability

commit: 
  v4.5-rc1
  78d1788350477246d516496d8d7684fa80ef7f18

        v4.5-rc1 78d1788350477246d516496d8d 
---------------- -------------------------- 
         %stddev     %change         %stddev
             \          |                \  
    195486 ±  1%    +274.1%     731248 ±  2%  
vm-scalability.time.involuntary_context_switches
      2939 ± 21%     +37.5%       4041 ± 12%  numa-vmstat.node1.nr_anon_pages
    897378 ±  1%     +17.7%    1056357 ±  2%  softirqs.RCU
      4293 ±  4%     +82.5%       7835 ±  2%  vmstat.system.cs
    120.40 ± 22%     +41.6%     170.50 ± 19%  cpuidle.C3-IVT.usage
 4.723e+08 ±  4%     +13.8%  5.375e+08 ±  5%  cpuidle.C6-IVT.time
     93336 ± 31%    +611.8%     664396 ± 88%  
latency_stats.max.pipe_wait.pipe_read.__vfs_read.vfs_read.SyS_read.entry_SYSCALL_64_fastpath
    371333 ±166%     -70.1%     110861 ±117%  
latency_stats.sum.pipe_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
    195486 ±  1%    +274.1%     731248 ±  2%  time.involuntary_context_switches
    717.20 ± 14%     +20.3%     862.50 ±  6%  time.major_page_faults
      2724 ± 60%     -93.6%     174.25 ±173%  numa-meminfo.node0.AnonHugePages
      1630 ±122%    +222.2%       5254 ± 16%  numa-meminfo.node1.AnonHugePages
     11760 ± 21%     +37.6%      16177 ± 12%  numa-meminfo.node1.AnonPages
      1.01 ±  4%     +26.5%       1.28 ± 10%  
perf-profile.cycles-pp.__add_to_page_cache_locked.add_to_page_cache_lru.mpage_readpages.xfs_vm_readpages.__do_page_cache_readahead
      0.90 ± 56%     -72.4%       0.25 ± 95%  
perf-profile.cycles-pp.__alloc_pages_nodemask.alloc_pages_current.__page_cache_alloc.__do_page_cache_readahead.filemap_fault
      0.90 ± 56%     -72.5%       0.25 ± 95%  
perf-profile.cycles-pp.__page_cache_alloc.__do_page_cache_readahead.filemap_fault.xfs_filemap_fault.__do_fault
      2.18 ±  2%     +15.9%       2.52 ±  7%  
perf-profile.cycles-pp.__remove_mapping.shrink_page_list.shrink_inactive_list.shrink_zone_memcg.shrink_zone
      1.04 ±  4%     +16.8%       1.21 ±  7%  
perf-profile.cycles-pp.__rmqueue.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_current.__page_cache_alloc
      1.08 ±  6%     +18.1%       1.28 ±  5%  
perf-profile.cycles-pp._raw_spin_lock_irqsave.__remove_mapping.shrink_page_list.shrink_inactive_list.shrink_zone_memcg
      0.90 ± 56%     -72.4%       0.25 ± 95%  
perf-profile.cycles-pp.alloc_pages_current.__page_cache_alloc.__do_page_cache_readahead.filemap_fault.xfs_filemap_fault
      0.77 ± 20%     +47.0%       1.14 ±  9%  
perf-profile.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.__remove_mapping.shrink_page_list.shrink_inactive_list
      1.38 ±  2%     +15.8%       1.60 ±  7%  
perf-profile.cycles-pp.rmap_walk.try_to_unmap.shrink_page_list.shrink_inactive_list.shrink_zone_memcg
      2.17 ±  2%     -14.7%       1.85 ±  9%  
perf-profile.cycles-pp.shrink_inactive_list.shrink_zone_memcg.shrink_zone.kswapd.kthread
      2.19 ±  2%     -14.5%       1.87 ±  9%  
perf-profile.cycles-pp.shrink_zone.kswapd.kthread.ret_from_fork
      2.18 ±  1%     -14.8%       1.86 ±  9%  
perf-profile.cycles-pp.shrink_zone_memcg.shrink_zone.kswapd.kthread.ret_from_fork
      1.46 ±  2%     +19.5%       1.75 ± 10%  
perf-profile.cycles-pp.try_to_unmap.shrink_page_list.shrink_inactive_list.shrink_zone_memcg.shrink_zone
     50.20 ±133%    +273.0%     187.25 ±109%  sched_debug.cfs_rq:/.load.5
     56.60 ± 11%     +97.4%     111.75 ± 55%  sched_debug.cfs_rq:/.load_avg.0
     17.60 ±  9%     +25.0%      22.00 ± 14%  sched_debug.cfs_rq:/.load_avg.19
     47.00 ± 33%     -54.8%      21.25 ± 13%  sched_debug.cfs_rq:/.load_avg.23
     21.60 ± 30%    +139.6%      51.75 ± 63%  sched_debug.cfs_rq:/.load_avg.30
     20.80 ± 30%     +50.2%      31.25 ± 26%  sched_debug.cfs_rq:/.load_avg.36
     16.40 ±  2%     +18.9%      19.50 ± 11%  
sched_debug.cfs_rq:/.runnable_load_avg.14
     16685 ±278%    -374.2%     -45753 ±-25%  sched_debug.cfs_rq:/.spread0.12
    119599 ± 47%     -71.8%      33762 ±121%  sched_debug.cfs_rq:/.spread0.27
    106418 ± 42%    -267.6%    -178399 ±-221%  sched_debug.cfs_rq:/.spread0.28
     56797 ± 81%    -136.5%     -20757 ±-159%  sched_debug.cfs_rq:/.spread0.3
     63128 ± 63%    -157.9%     -36549 ±-295%  sched_debug.cfs_rq:/.spread0.46
     57759 ±100%    -676.1%    -332737 ±-164%  sched_debug.cfs_rq:/.spread0.8
    151767 ± 32%     -40.2%      90684 ± 14%  sched_debug.cfs_rq:/.spread0.max
    855.00 ±  1%      +7.5%     918.75 ±  7%  sched_debug.cfs_rq:/.util_avg.20
     16.40 ±  2%     +18.9%      19.50 ± 11%  sched_debug.cpu.cpu_load[0].14
     16.40 ±  2%     +18.9%      19.50 ± 11%  sched_debug.cpu.cpu_load[1].14
     16.40 ±  2%     +18.9%      19.50 ± 11%  sched_debug.cpu.cpu_load[2].14
     16.40 ±  2%     +18.9%      19.50 ± 11%  sched_debug.cpu.cpu_load[3].14
     16.40 ±  2%     +18.9%      19.50 ± 11%  sched_debug.cpu.cpu_load[4].14
      1305 ±  1%     +16.6%       1521 ± 13%  sched_debug.cpu.curr->pid.23
     50.20 ±133%    +273.0%     187.25 ±109%  sched_debug.cpu.load.5
      0.00 ±  0%      +Inf%       1.00 ±  0%  sched_debug.cpu.nr_running.5
     10459 ± 30%     +86.1%      19462 ± 13%  sched_debug.cpu.nr_switches.12
     11042 ±113%    +169.5%      29759 ± 31%  sched_debug.cpu.nr_switches.14
     17516 ± 69%     +85.9%      32567 ± 46%  sched_debug.cpu.nr_switches.18
      8287 ± 69%    +271.4%      30775 ± 71%  sched_debug.cpu.nr_switches.21
     16525 ± 59%     +80.6%      29850 ± 41%  sched_debug.cpu.nr_switches.23
      8306 ± 36%     +92.8%      16016 ±  8%  sched_debug.cpu.nr_switches.24
      9271 ± 64%    +149.7%      23153 ± 30%  sched_debug.cpu.nr_switches.25
      6094 ± 49%    +243.8%      20951 ± 31%  sched_debug.cpu.nr_switches.26
      6569 ± 32%    +194.9%      19373 ± 17%  sched_debug.cpu.nr_switches.30
     10034 ±111%    +295.8%      39717 ± 41%  sched_debug.cpu.nr_switches.31
      7417 ± 70%    +205.4%      22657 ± 35%  sched_debug.cpu.nr_switches.38
      9204 ± 69%    +125.4%      20746 ± 26%  sched_debug.cpu.nr_switches.42
      8742 ± 68%    +172.8%      23845 ± 14%  sched_debug.cpu.nr_switches.47
     11267 ± 75%    +311.6%      46375 ± 68%  sched_debug.cpu.nr_switches.7
     14373 ±  4%     +74.8%      25131 ±  3%  sched_debug.cpu.nr_switches.avg
     68237 ± 15%     +30.3%      88895 ± 10%  sched_debug.cpu.nr_switches.max
      1792 ± 11%    +610.9%      12741 ±  4%  sched_debug.cpu.nr_switches.min
     -1.00 ±-244%    -200.0%       1.00 ±212%  
sched_debug.cpu.nr_uninterruptible.21
    -29.60 ±-22%     -31.7%     -20.21 ±-29%  
sched_debug.cpu.nr_uninterruptible.min
     11570 ± 27%     +75.8%      20344 ± 14%  sched_debug.cpu.sched_count.12
     18187 ± 52%     +67.7%      30504 ± 40%  sched_debug.cpu.sched_count.23
      8874 ± 33%     +83.7%      16301 ±  7%  sched_debug.cpu.sched_count.24
      9847 ± 60%    +137.9%      23423 ± 29%  sched_debug.cpu.sched_count.25
      7299 ± 38%    +190.8%      21225 ± 30%  sched_debug.cpu.sched_count.26
      7156 ± 30%    +179.8%      20022 ± 16%  sched_debug.cpu.sched_count.30
     10606 ±105%    +279.7%      40278 ± 41%  sched_debug.cpu.sched_count.31
      7977 ± 66%    +187.5%      22938 ± 35%  sched_debug.cpu.sched_count.38
     10713 ± 47%    +138.5%      25554 ± 54%  sched_debug.cpu.sched_count.39
     25911 ± 63%     +76.9%      45826 ± 39%  sched_debug.cpu.sched_count.4
      9758 ± 66%    +115.6%      21039 ± 26%  sched_debug.cpu.sched_count.42
      9252 ± 64%    +160.6%      24110 ± 14%  sched_debug.cpu.sched_count.47
     12153 ± 69%    +296.0%      48126 ± 65%  sched_debug.cpu.sched_count.7
    110528 ±  0%      +9.1%     120637 ±  0%  sched_debug.cpu.sched_count.avg
      2364 ±  8%    +455.3%      13129 ±  5%  sched_debug.cpu.sched_count.min
    816.60 ± 32%     -40.5%     485.75 ± 16%  sched_debug.cpu.sched_goidle.17
    383.20 ± 22%     +56.1%     598.00 ± 22%  sched_debug.cpu.sched_goidle.23
    204.20 ± 12%    +210.7%     634.50 ± 86%  sched_debug.cpu.sched_goidle.25
    190.60 ± 26%     +54.6%     294.75 ± 12%  sched_debug.cpu.sched_goidle.31
      4752 ± 36%    +107.1%       9844 ± 12%  sched_debug.cpu.ttwu_count.12
      5428 ±115%    +178.2%      15102 ± 31%  sched_debug.cpu.ttwu_count.14
      8902 ± 73%     +80.8%      16095 ± 47%  sched_debug.cpu.ttwu_count.18
      3698 ± 62%    +315.1%      15351 ± 75%  sched_debug.cpu.ttwu_count.21
      8371 ± 57%     +80.6%      15121 ± 41%  sched_debug.cpu.ttwu_count.23
      4569 ± 33%     +85.9%       8496 ± 10%  sched_debug.cpu.ttwu_count.24
      4884 ± 56%    +152.6%      12339 ± 27%  sched_debug.cpu.ttwu_count.25
      3187 ± 39%    +258.7%      11434 ± 23%  sched_debug.cpu.ttwu_count.26
      3513 ± 31%    +183.5%       9962 ± 19%  sched_debug.cpu.ttwu_count.30
      5462 ± 97%    +283.9%      20970 ± 41%  sched_debug.cpu.ttwu_count.31
      5605 ± 43%     +85.8%      10415 ± 34%  sched_debug.cpu.ttwu_count.34
      4011 ± 66%    +190.7%      11661 ± 33%  sched_debug.cpu.ttwu_count.38
      4305 ± 41%    +183.7%      12215 ± 59%  sched_debug.cpu.ttwu_count.39
      4828 ± 73%    +125.9%      10906 ± 25%  sched_debug.cpu.ttwu_count.42
      4560 ± 74%    +168.1%      12227 ± 12%  sched_debug.cpu.ttwu_count.47
      5663 ± 78%    +313.7%      23431 ± 67%  sched_debug.cpu.ttwu_count.7
      7326 ±  4%     +74.8%      12808 ±  3%  sched_debug.cpu.ttwu_count.avg
     34268 ± 15%     +29.6%      44398 ±  9%  sched_debug.cpu.ttwu_count.max
    792.03 ± 20%    +698.7%       6325 ±  5%  sched_debug.cpu.ttwu_count.min
      3951 ± 45%    +119.7%       8681 ± 13%  sched_debug.cpu.ttwu_local.12
      4585 ±137%    +205.0%      13987 ± 33%  sched_debug.cpu.ttwu_local.14
      8045 ± 80%     +88.6%      15173 ± 48%  sched_debug.cpu.ttwu_local.18
      2941 ± 79%    +386.9%      14319 ± 79%  sched_debug.cpu.ttwu_local.21
      7262 ± 67%     +95.1%      14172 ± 44%  sched_debug.cpu.ttwu_local.23
      3991 ± 40%    +100.4%       8000 ± 10%  sched_debug.cpu.ttwu_local.24
      4237 ± 68%    +162.6%      11128 ± 29%  sched_debug.cpu.ttwu_local.25
      2570 ± 41%    +292.2%      10081 ± 30%  sched_debug.cpu.ttwu_local.26
      2891 ± 40%    +215.8%       9132 ± 21%  sched_debug.cpu.ttwu_local.30
      4665 ±113%    +318.2%      19509 ± 43%  sched_debug.cpu.ttwu_local.31
      4533 ± 50%    +115.0%       9748 ± 35%  sched_debug.cpu.ttwu_local.34
      4131 ± 56%    +145.5%      10144 ± 47%  sched_debug.cpu.ttwu_local.37
      3349 ± 82%    +225.3%      10897 ± 34%  sched_debug.cpu.ttwu_local.38
      3630 ± 48%    +219.9%      11611 ± 63%  sched_debug.cpu.ttwu_local.39
      4138 ± 82%    +140.4%       9947 ± 29%  sched_debug.cpu.ttwu_local.42
      3356 ± 78%    +238.4%      11359 ± 14%  sched_debug.cpu.ttwu_local.47
      4866 ± 91%    +360.6%      22415 ± 71%  sched_debug.cpu.ttwu_local.7
      6406 ±  5%     +84.0%      11787 ±  4%  sched_debug.cpu.ttwu_local.avg
     33023 ± 15%     +29.7%      42825 ± 11%  sched_debug.cpu.ttwu_local.max
    450.27 ± 20%   +1199.8%       5852 ±  4%  sched_debug.cpu.ttwu_local.min
      1.58 ±  4%     -55.1%       0.71 ± 99%  sched_debug.rt_rq:/.rt_time.12

ivb43: Ivytown Ivy Bridge-EP
Memory: 64G

                  vm-scalability.time.involuntary_context_switches

  800000 ++-----------------------------------------------------------------+
         |                   O       O  O                  O  O       O  O  |
  700000 ++                    O  O        O O  O  O  O O        O  O       O
         |                                                                  |
  600000 O+                                                                 |
         |       O  O                                                       |
  500000 ++ O  O       O  O                                                 |
         |                                                                  |
  400000 ++                                                                 |
         |                                                                  |
  300000 ++                                                                 |
         |                                                                  |
  200000 *+.*..*.*..*..*..*..*.*..*..*..*..*.*..*..*..*.*..*..*..*..*.*     |
         |                                                                  |
  100000 ++-----------------------------------------------------------------+

        [*] bisect-good sample
        [O] bisect-bad  sample

To reproduce:

        git clone 
git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
        cd lkp-tests
        bin/lkp install job.yaml  # job file is attached in this email
        bin/lkp run     job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Ying Huang
---
LKP_SERVER: inn
LKP_CGI_PORT: 80
LKP_CIFS_PORT: 139
testcase: vm-scalability
default-monitors:
  wait: activate-monitor
  kmsg: 
  uptime: 
  iostat: 
  vmstat: 
  numa-numastat: 
  numa-vmstat: 
  numa-meminfo: 
  proc-vmstat: 
  proc-stat:
    interval: 10
  meminfo: 
  slabinfo: 
  interrupts: 
  lock_stat: 
  latency_stats: 
  softirqs: 
  bdi_dev_mapping: 
  diskstats: 
  nfsstat: 
  cpuidle: 
  cpufreq-stats: 
  turbostat: 
  pmeter: 
  sched_debug:
    interval: 60
cpufreq_governor: performance
default-watchdogs:
  oom-killer: 
  watchdog: 
commit: 78d1788350477246d516496d8d7684fa80ef7f18
model: Ivytown Ivy Bridge-EP
nr_cpu: 48
memory: 64G
nr_hdd_partitions: 0
hdd_partitions: 
rootfs_partition: 
swap_partitions: 
category: benchmark
perf-profile: 
runtime: 300s
size: 
vm-scalability:
  test: lru-file-mmap-read
queue: bisect
testbox: ivb43
tbox_group: ivb43
kconfig: x86_64-rhel
enqueue_time: 2016-01-31 04:22:27.102563621 +08:00
id: 2542d63be154b791122544fcefff3f1d6d23a064
user: lkp
compiler: gcc-4.9
head_commit: 5edef856d8228f4fc81476cd9ff36693ad0ffccb
base_commit: 92e963f50fc74041b5e9e744c330dca48e04f08d
branch: linux-devel/devel-hourly-2016013009
rootfs: debian-x86_64-2015-02-07.cgz
result_root: 
"/result/vm-scalability/performance-300s-lru-file-mmap-read/ivb43/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/78d1788350477246d516496d8d7684fa80ef7f18/0"
job_file: 
"/lkp/scheduled/ivb43/bisect_vm-scalability-performance-300s-lru-file-mmap-read-debian-x86_64-2015-02-07.cgz-x86_64-rhel-78d1788350477246d516496d8d7684fa80ef7f18-20160131-120638-crvdne-0.yaml"
max_uptime: 1500
initrd: "/osimage/debian/debian-x86_64-2015-02-07.cgz"
bootloader_append:
- root=/dev/ram0
- user=lkp
- 
job=/lkp/scheduled/ivb43/bisect_vm-scalability-performance-300s-lru-file-mmap-read-debian-x86_64-2015-02-07.cgz-x86_64-rhel-78d1788350477246d516496d8d7684fa80ef7f18-20160131-120638-crvdne-0.yaml
- ARCH=x86_64
- kconfig=x86_64-rhel
- branch=linux-devel/devel-hourly-2016013009
- commit=78d1788350477246d516496d8d7684fa80ef7f18
- 
BOOT_IMAGE=/pkg/linux/x86_64-rhel/gcc-4.9/78d1788350477246d516496d8d7684fa80ef7f18/vmlinuz-4.5.0-rc1-00001-g78d1788
- max_uptime=1500
- 
RESULT_ROOT=/result/vm-scalability/performance-300s-lru-file-mmap-read/ivb43/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/78d1788350477246d516496d8d7684fa80ef7f18/0
- LKP_SERVER=inn
- |2-


  earlyprintk=ttyS0,115200 systemd.log_level=err
  debug apic=debug sysrq_always_enabled rcupdate.rcu_cpu_stall_timeout=100
  panic=-1 softlockup_panic=1 nmi_watchdog=panic oops=panic load_ramdisk=2 
prompt_ramdisk=0
  console=ttyS0,115200 console=tty0 vga=normal

  rw
lkp_initrd: "/lkp/lkp/lkp-x86_64.cgz"
modules_initrd: 
"/pkg/linux/x86_64-rhel/gcc-4.9/78d1788350477246d516496d8d7684fa80ef7f18/modules.cgz"
bm_initrd: 
"/osimage/deps/debian-x86_64-2015-02-07.cgz/lkp.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/run-ipconfig.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/turbostat.cgz,/lkp/benchmarks/turbostat.cgz,/lkp/benchmarks/vm-scalability.cgz"
linux_headers_initrd: 
"/pkg/linux/x86_64-rhel/gcc-4.9/78d1788350477246d516496d8d7684fa80ef7f18/linux-headers.cgz"
repeat_to: 2
kernel: 
"/pkg/linux/x86_64-rhel/gcc-4.9/78d1788350477246d516496d8d7684fa80ef7f18/vmlinuz-4.5.0-rc1-00001-g78d1788"
dequeue_time: 2016-01-31 04:37:47.136664651 +08:00
job_state: finished
loadavg: 23.80 27.32 13.15 2/525 8525
start_time: '1454186311'
end_time: '1454186622'
version: "/lkp/lkp/.src-20160127-223853"

Attachment: reproduce.sh
Description: Bourne shell script

Reply via email to