[LKP] [vmstat] c3916839649: -63.9% will-it-scale.time.involuntary_context_switches, -17.6% will-it-scale.time.voluntary_context_switches, no primary result change

From: Huang Ying
Date: Sun Feb 15 2015 - 02:55:32 EST


FYI, we noticed the below changes on

git://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git master
commit c3916839649ddb3c0d541ce346360c4197898d52 ("vmstat: do not use deferrable delayed work for vmstat_update")


testbox/testcase/testparams: wsm/will-it-scale/performance-malloc1

516245a5b8e7c56e c3916839649ddb3c0d541ce346
---------------- --------------------------
1213 Â 1% -63.9% 438 Â 6% will-it-scale.time.involuntary_context_switches
2978260 Â 4% -17.6% 2453255 Â 3% will-it-scale.time.voluntary_context_switches
242 Â 0% +3.3% 250 Â 1% will-it-scale.time.system_time
1274586 Â 5% -84.6% 196270 Â 2% cpuidle.C6-NHM.usage
140438 Â 18% -72.4% 38803 Â 14% cpuidle.C1-NHM.usage
33943 Â 8% -72.9% 9203 Â 5% sched_debug.cfs_rq[5]:/.exec_clock
114238 Â 19% +72.3% 196836 Â 17% sched_debug.cpu#4.ttwu_local
308423 Â 12% -68.9% 95778 Â 1% sched_debug.cfs_rq[5]:/.min_vruntime
146477 Â 10% +48.7% 217860 Â 17% sched_debug.cpu#4.ttwu_count
64938 Â 16% -26.8% 47543 Â 4% sched_debug.cpu#0.ttwu_count
14242451 Â 14% -65.4% 4921778 Â 15% cpuidle.C1-NHM.time
37 Â 14% -52.7% 17 Â 31% sched_debug.cpu#5.cpu_load[4]
1213 Â 1% -63.9% 438 Â 6% time.involuntary_context_switches
61803 Â 13% -25.4% 46093 Â 4% sched_debug.cpu#0.sched_goidle
40 Â 13% -52.8% 19 Â 33% sched_debug.cpu#5.cpu_load[3]
46 Â 17% -52.2% 22 Â 39% sched_debug.cpu#5.cpu_load[2]
79 Â 10% +59.3% 126 Â 32% sched_debug.cpu#6.cpu_load[4]
73 Â 10% +54.9% 113 Â 33% sched_debug.cpu#6.cpu_load[3]
48956 Â 5% -46.1% 26390 Â 12% sched_debug.cpu#5.nr_load_updates
54 Â 26% -50.0% 27 Â 48% sched_debug.cpu#5.cpu_load[1]
327 Â 10% -44.1% 182 Â 8% sched_debug.cfs_rq[5]:/.tg_runnable_contrib
15032 Â 10% -43.9% 8426 Â 8% sched_debug.cfs_rq[5]:/.avg->runnable_avg_sum
32594 Â 9% +78.4% 58157 Â 0% sched_debug.cfs_rq[11]:/.exec_clock
285793 Â 11% +73.6% 496180 Â 3% sched_debug.cfs_rq[11]:/.min_vruntime
83 Â 15% -30.4% 57 Â 11% sched_debug.cpu#10.load
29 Â 12% +72.4% 50 Â 5% sched_debug.cpu#11.cpu_load[4]
31 Â 9% +59.8% 50 Â 6% sched_debug.cpu#11.cpu_load[3]
48 Â 15% +41.5% 69 Â 11% sched_debug.cfs_rq[9]:/.runnable_load_avg
68 Â 25% +41.9% 96 Â 20% sched_debug.cfs_rq[9]:/.load
81 Â 17% -26.5% 59 Â 12% sched_debug.cfs_rq[10]:/.load
521 Â 9% +43.4% 748 Â 6% cpuidle.POLL.usage
14974 Â 12% +46.0% 21864 Â 1% sched_debug.cfs_rq[11]:/.avg->runnable_avg_sum
327 Â 12% +45.8% 476 Â 1% sched_debug.cfs_rq[11]:/.tg_runnable_contrib
58930 Â 6% +41.2% 83196 Â 1% sched_debug.cpu#11.nr_load_updates
5326 Â 13% -18.5% 4343 Â 4% meminfo.AnonHugePages
222147 Â 7% +47.1% 326809 Â 30% sched_debug.cfs_rq[6]:/.spread0
149897 Â 12% +44.3% 216276 Â 17% sched_debug.cpu#4.sched_goidle
127451 Â 13% -24.2% 96598 Â 4% sched_debug.cpu#0.nr_switches
127599 Â 13% -24.2% 96740 Â 4% sched_debug.cpu#0.sched_count
36 Â 8% +42.1% 51 Â 8% sched_debug.cpu#11.cpu_load[2]
301690 Â 12% +43.9% 434016 Â 17% sched_debug.cpu#4.nr_switches
301787 Â 12% +43.8% 434096 Â 17% sched_debug.cpu#4.sched_count
116070 Â 3% -25.4% 86585 Â 9% sched_debug.cpu#11.ttwu_count
1824349 Â 4% +29.4% 2361351 Â 4% cpuidle.C3-NHM.usage
10615 Â 13% -17.4% 8766 Â 13% sched_debug.cfs_rq[3]:/.avg->runnable_avg_sum
230 Â 13% -17.4% 190 Â 13% sched_debug.cfs_rq[3]:/.tg_runnable_contrib
61 Â 14% -19.4% 49 Â 12% sched_debug.cfs_rq[10]:/.runnable_load_avg
99092 Â 10% -16.8% 82477 Â 0% sched_debug.cpu#10.sched_goidle
1175 Â 12% +18.8% 1397 Â 11% sched_debug.cpu#11.curr->pid
199498 Â 10% -16.9% 165794 Â 0% sched_debug.cpu#10.sched_count
199295 Â 10% -16.9% 165611 Â 0% sched_debug.cpu#10.nr_switches
326090 Â 12% -11.6% 288127 Â 9% sched_debug.cfs_rq[2]:/.min_vruntime
6.955e+08 Â 1% +25.2% 8.707e+08 Â 1% cpuidle.C3-NHM.time
2978260 Â 4% -17.6% 2453255 Â 3% time.voluntary_context_switches
1.06e+08 Â 3% -21.3% 83401246 Â 8% cpuidle.C1E-NHM.time
109454 Â 4% -19.9% 87649 Â 8% sched_debug.cpu#6.ttwu_count
48 Â 3% +17.6% 56 Â 10% sched_debug.cpu#9.cpu_load[2]
139443 Â 5% +15.4% 160850 Â 4% proc-vmstat.thp_fault_alloc
139441 Â 5% +15.4% 160848 Â 4% proc-vmstat.thp_split
220724 Â 2% -20.1% 176362 Â 9% sched_debug.cpu#11.nr_switches
220846 Â 2% -20.1% 176481 Â 9% sched_debug.cpu#11.sched_count
109644 Â 2% -20.0% 87693 Â 9% sched_debug.cpu#11.sched_goidle
46 Â 2% +17.3% 54 Â 8% sched_debug.cpu#9.cpu_load[3]
79492 Â 1% +12.9% 89782 Â 4% softirqs.SCHED
100431 Â 8% -17.2% 83150 Â 9% sched_debug.cpu#9.ttwu_count
726690 Â 2% +11.9% 813169 Â 5% sched_debug.cpu#4.avg_idle
91979 Â 0% +12.9% 103877 Â 12% sched_debug.cfs_rq[6]:/.exec_clock
201606 Â 5% -11.7% 177929 Â 5% sched_debug.cpu#6.nr_switches
201773 Â 5% -11.7% 178108 Â 5% sched_debug.cpu#6.sched_count
100423 Â 5% -11.8% 88575 Â 5% sched_debug.cpu#6.sched_goidle
40821 Â 2% +15.8% 47259 Â 5% softirqs.RCU
44 Â 3% +16.9% 52 Â 8% sched_debug.cpu#9.cpu_load[4]
9.56 Â 4% -11.6% 8.45 Â 1% time.user_time
306985 Â 12% -12.0% 270022 Â 4% sched_debug.cfs_rq[1]:/.min_vruntime
1.544e+09 Â 1% -10.0% 1.389e+09 Â 0% cpuidle.C6-NHM.time
115366 Â 1% +11.3% 128364 Â 10% sched_debug.cpu#6.nr_load_updates
22.81 Â 1% +10.8% 25.26 Â 1% turbostat.CPU%c3
25858 Â 3% -13.3% 22427 Â 2% vmstat.system.cs

wsm: Westmere
Memory: 6G




time.involuntary_context_switches

1300 ++-------------------------------------------------------------------+
1200 *+*.. .*. .*. .*. .*.*.*..*.*
| .*. *.*.*..*.*.*.. .*..* *.*..*.*.*. *.*..* |
1100 ++ *.* *.* |
1000 ++ |
| |
900 ++ |
800 ++ |
700 ++ |
| |
600 ++ |
500 ++ O O O |
O O O O O O O O O O O O O |
400 ++O O O O O |
300 ++-------------------------------------------------------------------+


cpuidle.C3-NHM.time

9.5e+08 ++----------------------------------------------------------------+
| |
9e+08 ++ O |
O O O O O O O O O O |
| O O O O O O O O O O |
8.5e+08 ++ |
| |
8e+08 ++ |
| |
7.5e+08 ++ |
| .* *.*. .*. *. |
| .*.* *.*. + + *..* *.*. + * |
7e+08 *+ .*. + + * *..*.* + .*.*.*. .*..*. .*
| * * *. * * |
6.5e+08 ++----------------------------------------------------------------+


cpuidle.C6-NHM.time

1.6e+09 ++---------------------------------------------------------------+
| *. *. * |
|.*.*.. *.*. : * .* .* * : *.*. ::|
1.55e+09 *+ + *. : : * + .* + + + : *.*..* : :|
| * *. : : + *..* * *..* + : *
| * * * |
1.5e+09 ++ |
| |
1.45e+09 ++ |
| |
| O O |
1.4e+09 ++ O O |
O O O O O O O O O O O |
| O O O O O |
1.35e+09 ++----------------------------O----------------------------------+


cpuidle.C6-NHM.usage

1.4e+06 ++--------------------------------------------------------------*-+
*.*. *.* .* .* *
1.2e+06 ++ *..*. + + .*.*. .*. .*. .*. + .*.*. .*.*. .*.*. |
| * *. * *. *.*.* * *. * |
1e+06 ++ |
| |
800000 ++ |
| |
600000 ++ |
| |
400000 ++ |
| |
200000 O+O O O O O O O O O O O O O O O O O O O O |
| |
0 ++----------------------------------------------------------------+


will-it-scale.time.involuntary_context_switches

1300 ++-------------------------------------------------------------------+
1200 *+*.. .*. .*. .*. .*.*.*..*.*
| .*. *.*.*..*.*.*.. .*..* *.*..*.*.*. *.*..* |
1100 ++ *.* *.* |
1000 ++ |
| |
900 ++ |
800 ++ |
700 ++ |
| |
600 ++ |
500 ++ O O O |
O O O O O O O O O O O O O |
400 ++O O O O O |
300 ++-------------------------------------------------------------------+


[*] bisect-good sample
[O] bisect-bad sample

To reproduce:

apt-get install ruby
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/setup-local job.yaml # the job file attached in this email
bin/run-local job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Huang, Ying

---
testcase: will-it-scale
default-monitors:
wait: pre-test
uptime:
iostat:
vmstat:
numa-numastat:
numa-vmstat:
numa-meminfo:
proc-vmstat:
proc-stat:
meminfo:
slabinfo:
interrupts:
lock_stat:
latency_stats:
softirqs:
bdi_dev_mapping:
diskstats:
nfsstat:
cpuidle:
cpufreq-stats:
turbostat:
pmeter:
sched_debug:
interval: 10
default_watchdogs:
watch-oom:
watchdog:
cpufreq_governor: performance
commit: 7496ad672d0ed7177fefec9c70ece8a70bbd2dc1
model: Westmere
memory: 6G
nr_hdd_partitions: 1
hdd_partitions:
swap_partitions:
rootfs_partition:
netconsole_port: 6667
perf-profile:
freq: 800
will-it-scale:
test: malloc1
testbox: wsm
tbox_group: wsm
kconfig: x86_64-rhel
enqueue_time: 2015-02-12 07:25:08.935926504 +08:00
head_commit: 7496ad672d0ed7177fefec9c70ece8a70bbd2dc1
base_commit: bfa76d49576599a4b9f9b7a71f23d73d6dcff735
branch: next/master
kernel: "/kernel/x86_64-rhel/7496ad672d0ed7177fefec9c70ece8a70bbd2dc1/vmlinuz-3.19.0-next-20150212-g7496ad6"
user: lkp
queue: cyclic
rootfs: debian-x86_64-2015-02-07.cgz
result_root: "/result/wsm/will-it-scale/performance-malloc1/debian-x86_64-2015-02-07.cgz/x86_64-rhel/7496ad672d0ed7177fefec9c70ece8a70bbd2dc1/0"
job_file: "/lkp/scheduled/wsm/cyclic_will-it-scale-performance-malloc1-x86_64-rhel-HEAD-7496ad672d0ed7177fefec9c70ece8a70bbd2dc1-0-20150212-31440-ix9pa1.yaml"
dequeue_time: 2015-02-12 16:35:45.778268771 +08:00
nr_cpu: "$(nproc)"
job_state: finished
loadavg: 8.37 4.91 2.02 1/158 5629
start_time: '1423730169'
end_time: '1423730474'
version: "/lkp/lkp/.src-20150212-162408"
./runtest.py malloc1 32 both 1 6 9 12
_______________________________________________
LKP mailing list
LKP@xxxxxxxxxxxxxxx