[lkp] [sched/fair] 98d8fd8126: -20.8% hackbench.throughput

From: kernel test robot
Date: Wed Sep 23 2015 - 22:01:03 EST


FYI, we noticed the below changes on

https://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git sched/core
commit 98d8fd8126676f7ba6e133e65b2ca4b17989d32c ("sched/fair: Initialize task load and utilization before placing task on rq")


=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_threads/mode/ipc:
lkp-ws02/hackbench/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/1600%/process/pipe

commit:
231678b768da07d19ab5683a39eeb0c250631d02
98d8fd8126676f7ba6e133e65b2ca4b17989d32c

231678b768da07d1 98d8fd8126676f7ba6e133e65b
---------------- --------------------------
%stddev %change %stddev
\ | \
188818 ± 1% -20.8% 149585 ± 1% hackbench.throughput
81712173 ± 4% +211.8% 2.548e+08 ± 1% hackbench.time.involuntary_context_switches
21611286 ± 0% -19.2% 17453366 ± 1% hackbench.time.minor_page_faults
2226 ± 0% +1.3% 2255 ± 0% hackbench.time.percent_of_cpu_this_job_got
12445 ± 0% +2.1% 12704 ± 0% hackbench.time.system_time
2.494e+08 ± 3% +118.5% 5.448e+08 ± 1% hackbench.time.voluntary_context_switches
1097790 ± 0% +50.6% 1653664 ± 1% softirqs.RCU
554877 ± 3% +137.8% 1319318 ± 1% vmstat.system.cs
89017 ± 4% +187.8% 256235 ± 1% vmstat.system.in
1.312e+08 ± 1% -16.0% 1.102e+08 ± 4% numa-numastat.node0.local_node
1.312e+08 ± 1% -16.0% 1.102e+08 ± 4% numa-numastat.node0.numa_hit
1.302e+08 ± 1% -34.9% 84785305 ± 5% numa-numastat.node1.local_node
1.302e+08 ± 1% -34.9% 84785344 ± 5% numa-numastat.node1.numa_hit
302.00 ± 1% -19.2% 244.00 ± 1% time.file_system_outputs
81712173 ± 4% +211.8% 2.548e+08 ± 1% time.involuntary_context_switches
21611286 ± 0% -19.2% 17453366 ± 1% time.minor_page_faults
2.494e+08 ± 3% +118.5% 5.448e+08 ± 1% time.voluntary_context_switches
92.88 ± 0% +1.3% 94.13 ± 0% turbostat.%Busy
2675 ± 0% +1.8% 2723 ± 0% turbostat.Avg_MHz
4.44 ± 1% -24.9% 3.34 ± 2% turbostat.CPU%c1
0.98 ± 2% -32.2% 0.66 ± 3% turbostat.CPU%c3
2.79e+08 ± 4% -25.2% 2.086e+08 ± 6% cpuidle.C1-NHM.time
1.235e+08 ± 4% -28.6% 88251264 ± 7% cpuidle.C1E-NHM.time
243525 ± 4% -21.9% 190252 ± 8% cpuidle.C1E-NHM.usage
1.819e+08 ± 2% -25.8% 1.35e+08 ± 1% cpuidle.C3-NHM.time
260585 ± 1% -20.4% 207474 ± 2% cpuidle.C3-NHM.usage
266207 ± 1% -39.4% 161453 ± 3% cpuidle.C6-NHM.usage
493467 ± 0% +26.5% 624337 ± 0% meminfo.Active
395397 ± 0% +33.0% 525811 ± 0% meminfo.Active(anon)
372719 ± 1% +34.2% 500207 ± 1% meminfo.AnonPages
4543041 ± 1% +37.5% 6248687 ± 1% meminfo.Committed_AS
185265 ± 1% +16.3% 215373 ± 0% meminfo.KernelStack
302233 ± 1% +37.1% 414289 ± 1% meminfo.PageTables
333827 ± 0% +18.6% 396038 ± 0% meminfo.SUnreclaim
380340 ± 0% +16.6% 443518 ± 0% meminfo.Slab
51154 ±143% -100.0% 5.00 ±100% latency_stats.avg.call_rwsem_down_write_failed.copy_process._do_fork.SyS_clone.entry_SYSCALL_64_fastpath
0.00 ± -1% +Inf% 30679 ±100% latency_stats.avg.proc_cgroup_show.proc_single_show.seq_read.__vfs_read.vfs_read.SyS_read.entry_SYSCALL_64_fastpath
7795 ±100% +1304.6% 109497 ± 93% latency_stats.max.blk_execute_rq.scsi_execute.scsi_execute_req_flags.ses_recv_diag.[ses].ses_get_page2_descriptor.[ses].ses_get_power_status.[ses].ses_enclosure_data_process.[ses].ses_intf_add.[ses].class_interface_register.scsi_register_interface.0xffffffffa0006013.do_one_initcall
297190 ±117% -100.0% 23.00 ±100% latency_stats.max.call_rwsem_down_write_failed.copy_process._do_fork.SyS_clone.entry_SYSCALL_64_fastpath
0.00 ± -1% +Inf% 97905 ±109% latency_stats.max.proc_cgroup_show.proc_single_show.seq_read.__vfs_read.vfs_read.SyS_read.entry_SYSCALL_64_fastpath
12901 ±131% -78.9% 2717 ±135% latency_stats.max.wait_on_page_bit.wait_on_page_read.do_read_cache_page.read_cache_page_gfp.btrfs_scan_one_device.[btrfs].btrfs_control_ioctl.[btrfs].do_vfs_ioctl.SyS_ioctl.entry_SYSCALL_64_fastpath
392778 ±128% -100.0% 75.50 ±100% latency_stats.sum.call_rwsem_down_write_failed.copy_process._do_fork.SyS_clone.entry_SYSCALL_64_fastpath
13678 ± 75% -68.1% 4368 ± 67% latency_stats.sum.flush_work.__cancel_work_timer.cancel_delayed_work_sync.disk_block_events.__blkdev_get.blkdev_get.blkdev_get_by_path.btrfs_scan_one_device.[btrfs].btrfs_control_ioctl.[btrfs].do_vfs_ioctl.SyS_ioctl.entry_SYSCALL_64_fastpath
19088 ±101% -100.0% 8.67 ±110% latency_stats.sum.path_openat.do_filp_open.do_sys_open.SyS_open.entry_SYSCALL_64_fastpath
0.00 ± -1% +Inf% 139824 ±104% latency_stats.sum.proc_cgroup_show.proc_single_show.seq_read.__vfs_read.vfs_read.SyS_read.entry_SYSCALL_64_fastpath
17538 ± 95% -71.0% 5094 ± 72% latency_stats.sum.wait_on_page_bit.wait_on_page_read.do_read_cache_page.read_cache_page_gfp.btrfs_scan_one_device.[btrfs].btrfs_control_ioctl.[btrfs].do_vfs_ioctl.SyS_ioctl.entry_SYSCALL_64_fastpath
99221 ± 1% +32.5% 131458 ± 0% proc-vmstat.nr_active_anon
93610 ± 1% +33.5% 124954 ± 0% proc-vmstat.nr_anon_pages
11604 ± 1% +16.4% 13503 ± 0% proc-vmstat.nr_kernel_stack
75770 ± 1% +36.6% 103508 ± 1% proc-vmstat.nr_page_table_pages
83664 ± 0% +18.6% 99265 ± 0% proc-vmstat.nr_slab_unreclaimable
2.615e+08 ± 1% -25.4% 1.95e+08 ± 1% proc-vmstat.numa_hit
2.615e+08 ± 1% -25.4% 1.95e+08 ± 1% proc-vmstat.numa_local
2151 ± 9% +33.7% 2875 ± 7% proc-vmstat.pgactivate
53318467 ± 1% -16.3% 44611139 ± 4% proc-vmstat.pgalloc_dma32
2.124e+08 ± 1% -27.6% 1.538e+08 ± 2% proc-vmstat.pgalloc_normal
21951016 ± 0% -17.9% 18028538 ± 1% proc-vmstat.pgfault
2.656e+08 ± 1% -25.3% 1.983e+08 ± 1% proc-vmstat.pgfree
231019 ± 6% +19.0% 274944 ± 3% numa-meminfo.node0.Active
182158 ± 8% +24.4% 226602 ± 4% numa-meminfo.node0.Active(anon)
4049 ± 0% -100.0% 0.00 ± -1% numa-meminfo.node0.AnonHugePages
171061 ± 5% +28.8% 220397 ± 3% numa-meminfo.node0.AnonPages
253455 ± 4% -5.4% 239791 ± 4% numa-meminfo.node0.FilePages
9402 ± 3% -90.3% 915.75 ± 39% numa-meminfo.node0.Inactive(anon)
10253 ± 0% -50.8% 5041 ± 1% numa-meminfo.node0.Mapped
131300 ± 9% +25.1% 164194 ± 7% numa-meminfo.node0.PageTables
20469 ± 54% -65.5% 7058 ±145% numa-meminfo.node0.Shmem
235929 ± 5% +43.0% 337441 ± 1% numa-meminfo.node1.Active
186748 ± 8% +53.8% 287257 ± 1% numa-meminfo.node1.Active(anon)
175187 ± 5% +52.9% 267826 ± 5% numa-meminfo.node1.AnonPages
1249 ± 39% +676.4% 9697 ± 4% numa-meminfo.node1.Inactive(anon)
105601 ± 9% +39.4% 147194 ± 10% numa-meminfo.node1.KernelStack
5032 ± 1% +103.5% 10238 ± 0% numa-meminfo.node1.Mapped
1028697 ± 5% +40.4% 1444560 ± 5% numa-meminfo.node1.MemUsed
147371 ± 7% +62.4% 239296 ± 7% numa-meminfo.node1.PageTables
185026 ± 7% +43.2% 264909 ± 10% numa-meminfo.node1.SUnreclaim
209508 ± 7% +38.5% 290116 ± 9% numa-meminfo.node1.Slab
45770 ± 7% +24.9% 57169 ± 3% numa-vmstat.node0.nr_active_anon
42981 ± 4% +29.4% 55616 ± 3% numa-vmstat.node0.nr_anon_pages
63378 ± 4% -5.4% 59946 ± 4% numa-vmstat.node0.nr_file_pages
2351 ± 3% -90.3% 228.25 ± 38% numa-vmstat.node0.nr_inactive_anon
2589 ± 1% -51.6% 1253 ± 0% numa-vmstat.node0.nr_mapped
32990 ± 8% +25.6% 41423 ± 7% numa-vmstat.node0.nr_page_table_pages
5131 ± 54% -65.6% 1763 ±145% numa-vmstat.node0.nr_shmem
64745848 ± 2% -13.8% 55814423 ± 2% numa-vmstat.node0.numa_hit
64743896 ± 2% -13.9% 55752339 ± 2% numa-vmstat.node0.numa_local
1951 ± 91% +3081.4% 62084 ± 1% numa-vmstat.node0.numa_other
45977 ± 8% +57.0% 72172 ± 1% numa-vmstat.node1.nr_active_anon
43078 ± 7% +56.1% 67261 ± 3% numa-vmstat.node1.nr_anon_pages
313.50 ± 40% +673.4% 2424 ± 4% numa-vmstat.node1.nr_inactive_anon
6558 ± 11% +39.9% 9175 ± 8% numa-vmstat.node1.nr_kernel_stack
1262 ± 1% +102.2% 2552 ± 0% numa-vmstat.node1.nr_mapped
36358 ± 9% +65.2% 60055 ± 5% numa-vmstat.node1.nr_page_table_pages
45984 ± 9% +43.9% 66189 ± 8% numa-vmstat.node1.nr_slab_unreclaimable
64599981 ± 2% -34.3% 42454481 ± 5% numa-vmstat.node1.numa_hit
64534349 ± 2% -34.2% 42449235 ± 5% numa-vmstat.node1.numa_local
65632 ± 2% -92.0% 5245 ± 23% numa-vmstat.node1.numa_other
148962 ± 0% +32.1% 196766 ± 2% slabinfo.anon_vma.active_objs
3066 ± 0% +32.5% 4062 ± 1% slabinfo.anon_vma.active_slabs
156402 ± 0% +32.5% 207216 ± 1% slabinfo.anon_vma.num_objs
3066 ± 0% +32.5% 4062 ± 1% slabinfo.anon_vma.num_slabs
15321 ± 0% +14.6% 17563 ± 1% slabinfo.files_cache.active_objs
16470 ± 0% +15.1% 18958 ± 1% slabinfo.files_cache.num_objs
8808 ± 0% +17.6% 10359 ± 1% slabinfo.kmalloc-1024.active_objs
9268 ± 0% +16.1% 10758 ± 0% slabinfo.kmalloc-1024.num_objs
22656 ± 1% +9.9% 24899 ± 2% slabinfo.kmalloc-128.num_objs
31867 ± 0% +11.6% 35548 ± 0% slabinfo.kmalloc-192.active_objs
32775 ± 0% +11.4% 36522 ± 0% slabinfo.kmalloc-192.num_objs
15221 ± 0% +23.1% 18731 ± 0% slabinfo.kmalloc-256.active_objs
16380 ± 0% +19.4% 19557 ± 0% slabinfo.kmalloc-256.num_objs
308147 ± 0% +33.0% 409879 ± 2% slabinfo.kmalloc-64.active_objs
6591 ± 1% +17.9% 7770 ± 1% slabinfo.kmalloc-64.active_slabs
421883 ± 1% +17.9% 497347 ± 1% slabinfo.kmalloc-64.num_objs
6591 ± 1% +17.9% 7770 ± 1% slabinfo.kmalloc-64.num_slabs
482.75 ± 11% +39.7% 674.50 ± 7% slabinfo.kmem_cache_node.active_objs
495.50 ± 10% +38.7% 687.25 ± 7% slabinfo.kmem_cache_node.num_objs
9328 ± 0% +29.1% 12045 ± 2% slabinfo.mm_struct.active_objs
612.00 ± 0% +28.6% 787.00 ± 1% slabinfo.mm_struct.active_slabs
10411 ± 0% +28.6% 13390 ± 1% slabinfo.mm_struct.num_objs
612.00 ± 0% +28.6% 787.00 ± 1% slabinfo.mm_struct.num_slabs
12765 ± 1% +15.7% 14765 ± 1% slabinfo.sighand_cache.active_objs
861.75 ± 1% +18.4% 1020 ± 0% slabinfo.sighand_cache.active_slabs
12933 ± 1% +18.4% 15308 ± 0% slabinfo.sighand_cache.num_objs
861.75 ± 1% +18.4% 1020 ± 0% slabinfo.sighand_cache.num_slabs
14455 ± 1% +11.8% 16167 ± 1% slabinfo.signal_cache.active_objs
14698 ± 1% +14.2% 16779 ± 1% slabinfo.signal_cache.num_objs
11628 ± 1% +16.6% 13563 ± 1% slabinfo.task_struct.active_objs
3899 ± 1% +18.5% 4620 ± 1% slabinfo.task_struct.active_slabs
11699 ± 1% +18.5% 13861 ± 1% slabinfo.task_struct.num_objs
3899 ± 1% +18.5% 4620 ± 1% slabinfo.task_struct.num_slabs
224907 ± 0% +34.2% 301780 ± 2% slabinfo.vm_area_struct.active_objs
5290 ± 0% +34.9% 7135 ± 2% slabinfo.vm_area_struct.active_slabs
232815 ± 0% +34.8% 313951 ± 2% slabinfo.vm_area_struct.num_objs
5290 ± 0% +34.9% 7135 ± 2% slabinfo.vm_area_struct.num_slabs
0.30 ± 89% +2528.3% 7.88 ± 13% perf-profile.cpu-cycles.__account_scheduler_latency.enqueue_entity.enqueue_task_fair.enqueue_task.activate_task
0.02 ± 74% +17735.0% 2.97 ± 19% perf-profile.cpu-cycles.__kernel_text_address.print_context_stack.dump_trace.save_stack_trace_tsk.__account_scheduler_latency
0.02 ±-5000% +21150.0% 4.25 ± 44% perf-profile.cpu-cycles.__schedule.schedule.pipe_wait.pipe_read.__vfs_read
0.00 ± -1% +Inf% 0.64 ± 58% perf-profile.cpu-cycles.__schedule.schedule.prepare_exit_to_usermode.syscall_return_slowpath.int_ret_from_sys_call
7.24 ± 19% +141.5% 17.49 ± 14% perf-profile.cpu-cycles.__vfs_read.vfs_read.sys_read.entry_SYSCALL_64_fastpath
1.36 ± 84% +926.2% 13.92 ± 16% perf-profile.cpu-cycles.__wake_up_common.__wake_up_sync_key.pipe_write.__vfs_write.vfs_write
1.46 ±107% +981.3% 15.76 ± 17% perf-profile.cpu-cycles.__wake_up_sync_key.pipe_write.__vfs_write.vfs_write.sys_write
0.03 ±-3333% +3200.0% 0.99 ± 28% perf-profile.cpu-cycles.activate_task.ttwu_do_activate.sched_ttwu_pending.scheduler_ipi.smp_reschedule_interrupt
0.64 ± 86% +1399.6% 9.65 ± 13% perf-profile.cpu-cycles.activate_task.ttwu_do_activate.try_to_wake_up.default_wake_function.autoremove_wake_function
1.28 ± 86% +966.4% 13.62 ± 16% perf-profile.cpu-cycles.autoremove_wake_function.__wake_up_common.__wake_up_sync_key.pipe_write.__vfs_write
1.26 ± 86% +968.6% 13.50 ± 16% perf-profile.cpu-cycles.default_wake_function.autoremove_wake_function.__wake_up_common.__wake_up_sync_key.pipe_write
0.01 ± 0% +3200.0% 0.33 ±140% perf-profile.cpu-cycles.do_wait.sys_wait4.entry_SYSCALL_64_fastpath
0.16 ± 96% +4396.9% 7.20 ± 12% perf-profile.cpu-cycles.dump_trace.save_stack_trace_tsk.__account_scheduler_latency.enqueue_entity.enqueue_task_fair
0.48 ± 86% +1877.1% 9.49 ± 14% perf-profile.cpu-cycles.enqueue_entity.enqueue_task_fair.enqueue_task.activate_task.ttwu_do_activate
0.03 ±-3333% +3175.0% 0.98 ± 28% perf-profile.cpu-cycles.enqueue_task.activate_task.ttwu_do_activate.sched_ttwu_pending.scheduler_ipi
0.64 ± 86% +1396.9% 9.63 ± 13% perf-profile.cpu-cycles.enqueue_task.activate_task.ttwu_do_activate.try_to_wake_up.default_wake_function
0.02 ±-5000% +4162.5% 0.85 ± 27% perf-profile.cpu-cycles.enqueue_task_fair.enqueue_task.activate_task.ttwu_do_activate.sched_ttwu_pending
0.59 ± 86% +1473.7% 9.29 ± 13% perf-profile.cpu-cycles.enqueue_task_fair.enqueue_task.activate_task.ttwu_do_activate.try_to_wake_up
111.99 ± 2% -23.5% 85.65 ± 7% perf-profile.cpu-cycles.entry_SYSCALL_64_fastpath
0.00 ± -1% +Inf% 0.77 ± 51% perf-profile.cpu-cycles.int_ret_from_sys_call
3.68 ± 31% +309.3% 15.06 ± 17% perf-profile.cpu-cycles.pipe_read.__vfs_read.vfs_read.sys_read.entry_SYSCALL_64_fastpath
0.03 ±-3333% +19350.0% 5.83 ± 38% perf-profile.cpu-cycles.pipe_wait.pipe_read.__vfs_read.vfs_read.sys_read
10.52 ± 23% +109.9% 22.09 ± 18% perf-profile.cpu-cycles.pipe_write.__vfs_write.vfs_write.sys_write.entry_SYSCALL_64_fastpath
0.00 ± -1% +Inf% 0.68 ± 56% perf-profile.cpu-cycles.prepare_exit_to_usermode.syscall_return_slowpath.int_ret_from_sys_call
0.10 ± 96% +6080.0% 6.18 ± 13% perf-profile.cpu-cycles.print_context_stack.dump_trace.save_stack_trace_tsk.__account_scheduler_latency.enqueue_entity
0.16 ± 97% +4329.6% 7.24 ± 13% perf-profile.cpu-cycles.save_stack_trace_tsk.__account_scheduler_latency.enqueue_entity.enqueue_task_fair.enqueue_task
0.02 ±-5000% +22300.0% 4.48 ± 43% perf-profile.cpu-cycles.schedule.pipe_wait.pipe_read.__vfs_read.vfs_read
0.00 ± -1% +Inf% 0.62 ± 62% perf-profile.cpu-cycles.schedule.prepare_exit_to_usermode.syscall_return_slowpath.int_ret_from_sys_call
41.93 ± 3% -21.8% 32.80 ± 4% perf-profile.cpu-cycles.sys_read.entry_SYSCALL_64_fastpath
0.01 ± 0% +3225.0% 0.33 ±140% perf-profile.cpu-cycles.sys_wait4.entry_SYSCALL_64_fastpath
65.72 ± 3% -25.2% 49.18 ± 9% perf-profile.cpu-cycles.sys_write.entry_SYSCALL_64_fastpath
0.00 ± -1% +Inf% 0.72 ± 52% perf-profile.cpu-cycles.syscall_return_slowpath.int_ret_from_sys_call
1.28 ± 86% +961.1% 13.58 ± 16% perf-profile.cpu-cycles.try_to_wake_up.default_wake_function.autoremove_wake_function.__wake_up_common.__wake_up_sync_key
0.04 ±-2500% +2512.5% 1.04 ± 28% perf-profile.cpu-cycles.ttwu_do_activate.constprop.85.sched_ttwu_pending.scheduler_ipi.smp_reschedule_interrupt.reschedule_interrupt
0.70 ± 88% +1343.2% 10.10 ± 13% perf-profile.cpu-cycles.ttwu_do_activate.constprop.85.try_to_wake_up.default_wake_function.autoremove_wake_function.__wake_up_common
34.35 ± 3% -16.2% 28.77 ± 6% perf-profile.cpu-cycles.vfs_read.sys_read.entry_SYSCALL_64_fastpath
58.47 ± 4% -26.7% 42.85 ± 9% perf-profile.cpu-cycles.vfs_write.sys_write.entry_SYSCALL_64_fastpath
0.01 ± 70% +4475.0% 0.30 ±139% perf-profile.cpu-cycles.wait_consider_task.do_wait.sys_wait4.entry_SYSCALL_64_fastpath
121.25 ± 40% -43.9% 68.00 ± 4% sched_debug.cfs_rq[0]:/.load_avg
7372993 ± 2% +59.5% 11756998 ± 6% sched_debug.cfs_rq[0]:/.min_vruntime
2526 ± 4% -36.8% 1596 ± 10% sched_debug.cfs_rq[0]:/.tg_load_avg
96.00 ± 8% -29.2% 68.00 ± 4% sched_debug.cfs_rq[0]:/.tg_load_avg_contrib
91.75 ± 23% -40.1% 55.00 ± 12% sched_debug.cfs_rq[10]:/.load_avg
8761646 ± 6% +202.7% 26523962 ± 11% sched_debug.cfs_rq[10]:/.min_vruntime
8.00 ± 56% +1043.8% 91.50 ± 95% sched_debug.cfs_rq[10]:/.nr_spread_over
1375813 ± 32% +967.5% 14686694 ± 19% sched_debug.cfs_rq[10]:/.spread0
2218 ± 8% -22.1% 1729 ± 9% sched_debug.cfs_rq[10]:/.tg_load_avg
92.25 ± 23% -40.1% 55.25 ± 12% sched_debug.cfs_rq[10]:/.tg_load_avg_contrib
83.00 ± 8% -28.6% 59.25 ± 26% sched_debug.cfs_rq[11]:/.load_avg
8281705 ± 5% +221.0% 26587558 ± 12% sched_debug.cfs_rq[11]:/.min_vruntime
12.75 ± 76% +488.2% 75.00 ± 57% sched_debug.cfs_rq[11]:/.nr_spread_over
895534 ± 36% +1546.8% 14747704 ± 20% sched_debug.cfs_rq[11]:/.spread0
2181 ± 7% -24.0% 1657 ± 7% sched_debug.cfs_rq[11]:/.tg_load_avg
83.00 ± 8% -28.6% 59.25 ± 26% sched_debug.cfs_rq[11]:/.tg_load_avg_contrib
7597123 ± 4% +64.9% 12525605 ± 7% sched_debug.cfs_rq[12]:/.min_vruntime
2144 ± 8% -18.9% 1739 ± 8% sched_debug.cfs_rq[12]:/.tg_load_avg
7974727 ± 1% +76.7% 14092249 ± 8% sched_debug.cfs_rq[13]:/.min_vruntime
587700 ± 33% +282.6% 2248438 ± 24% sched_debug.cfs_rq[13]:/.spread0
2161 ± 8% -18.0% 1771 ± 6% sched_debug.cfs_rq[13]:/.tg_load_avg
9111334 ± 6% +49.2% 13594919 ± 7% sched_debug.cfs_rq[14]:/.min_vruntime
2155 ± 10% -17.9% 1768 ± 7% sched_debug.cfs_rq[14]:/.tg_load_avg
91.00 ± 11% -28.3% 65.25 ± 24% sched_debug.cfs_rq[15]:/.load_avg
7755071 ± 5% +80.3% 13985714 ± 7% sched_debug.cfs_rq[15]:/.min_vruntime
9.75 ± 15% +520.5% 60.50 ±128% sched_debug.cfs_rq[15]:/.nr_spread_over
367395 ± 78% +481.9% 2137747 ± 16% sched_debug.cfs_rq[15]:/.spread0
2175 ± 12% -22.1% 1694 ± 7% sched_debug.cfs_rq[15]:/.tg_load_avg
91.00 ± 11% -28.3% 65.25 ± 24% sched_debug.cfs_rq[15]:/.tg_load_avg_contrib
7790185 ± 3% +81.0% 14103194 ± 6% sched_debug.cfs_rq[16]:/.min_vruntime
402239 ± 43% +460.5% 2254357 ± 12% sched_debug.cfs_rq[16]:/.spread0
2196 ± 9% -22.8% 1694 ± 6% sched_debug.cfs_rq[16]:/.tg_load_avg
8545892 ± 6% +64.3% 14041885 ± 7% sched_debug.cfs_rq[17]:/.min_vruntime
1157569 ± 38% +89.4% 2192052 ± 15% sched_debug.cfs_rq[17]:/.spread0
2168 ± 9% -21.2% 1709 ± 3% sched_debug.cfs_rq[17]:/.tg_load_avg
8083785 ± 3% +194.3% 23786635 ± 12% sched_debug.cfs_rq[18]:/.min_vruntime
7.25 ± 69% +286.2% 28.00 ± 51% sched_debug.cfs_rq[18]:/.nr_spread_over
695207 ± 68% +1616.3% 11932029 ± 23% sched_debug.cfs_rq[18]:/.spread0
2134 ± 10% -18.9% 1731 ± 5% sched_debug.cfs_rq[18]:/.tg_load_avg
123.00 ± 23% -50.8% 60.50 ± 18% sched_debug.cfs_rq[19]:/.load_avg
8843043 ± 4% +212.8% 27657482 ± 14% sched_debug.cfs_rq[19]:/.min_vruntime
11.50 ± 72% +1156.5% 144.50 ± 89% sched_debug.cfs_rq[19]:/.nr_spread_over
1454237 ± 28% +986.6% 15802087 ± 22% sched_debug.cfs_rq[19]:/.spread0
2121 ± 10% -18.7% 1724 ± 7% sched_debug.cfs_rq[19]:/.tg_load_avg
121.00 ± 21% -50.0% 60.50 ± 18% sched_debug.cfs_rq[19]:/.tg_load_avg_contrib
101.50 ± 14% -37.2% 63.75 ± 8% sched_debug.cfs_rq[1]:/.load_avg
8066420 ± 2% +67.1% 13476103 ± 7% sched_debug.cfs_rq[1]:/.min_vruntime
689384 ± 53% +147.3% 1704860 ± 22% sched_debug.cfs_rq[1]:/.spread0
2514 ± 4% -39.0% 1533 ± 6% sched_debug.cfs_rq[1]:/.tg_load_avg
101.75 ± 14% -37.3% 63.75 ± 8% sched_debug.cfs_rq[1]:/.tg_load_avg_contrib
91.00 ± 8% -23.1% 70.00 ± 10% sched_debug.cfs_rq[20]:/.load_avg
9057239 ± 7% +200.9% 27252092 ± 13% sched_debug.cfs_rq[20]:/.min_vruntime
1668188 ± 44% +822.9% 15396001 ± 23% sched_debug.cfs_rq[20]:/.spread0
2153 ± 9% -20.7% 1707 ± 5% sched_debug.cfs_rq[20]:/.tg_load_avg
91.00 ± 8% -24.2% 69.00 ± 11% sched_debug.cfs_rq[20]:/.tg_load_avg_contrib
8505030 ± 8% +224.1% 27566380 ± 13% sched_debug.cfs_rq[21]:/.min_vruntime
15.50 ± 51% +575.8% 104.75 ± 42% sched_debug.cfs_rq[21]:/.nr_spread_over
1115371 ± 70% +1308.4% 15708882 ± 21% sched_debug.cfs_rq[21]:/.spread0
2129 ± 9% -19.2% 1720 ± 5% sched_debug.cfs_rq[21]:/.tg_load_avg
8352879 ± 2% +232.1% 27739662 ± 13% sched_debug.cfs_rq[22]:/.min_vruntime
13.00 ± 45% +1930.8% 264.00 ± 38% sched_debug.cfs_rq[22]:/.nr_spread_over
962764 ± 8% +1549.5% 15880365 ± 22% sched_debug.cfs_rq[22]:/.spread0
2119 ± 8% -14.5% 1811 ± 10% sched_debug.cfs_rq[22]:/.tg_load_avg
8119642 ± 4% +241.9% 27759824 ± 13% sched_debug.cfs_rq[23]:/.min_vruntime
729257 ± 37% +2080.1% 15898645 ± 21% sched_debug.cfs_rq[23]:/.spread0
2087 ± 7% -16.0% 1753 ± 3% sched_debug.cfs_rq[23]:/.tg_load_avg
101.75 ± 19% -31.7% 69.50 ± 10% sched_debug.cfs_rq[2]:/.load_avg
9427522 ± 14% +44.3% 13605129 ± 7% sched_debug.cfs_rq[2]:/.min_vruntime
2441 ± 7% -35.2% 1583 ± 10% sched_debug.cfs_rq[2]:/.tg_load_avg
102.50 ± 19% -32.2% 69.50 ± 10% sched_debug.cfs_rq[2]:/.tg_load_avg_contrib
7664612 ± 6% +76.4% 13520491 ± 4% sched_debug.cfs_rq[3]:/.min_vruntime
283759 ±142% +509.0% 1728055 ± 23% sched_debug.cfs_rq[3]:/.spread0
2355 ± 8% -32.8% 1582 ± 12% sched_debug.cfs_rq[3]:/.tg_load_avg
118.75 ± 20% -33.7% 78.75 ± 12% sched_debug.cfs_rq[4]:/.load_avg
7770292 ± 5% +73.0% 13442540 ± 6% sched_debug.cfs_rq[4]:/.min_vruntime
388453 ±139% +322.2% 1640216 ± 19% sched_debug.cfs_rq[4]:/.spread0
2286 ± 8% -29.9% 1603 ± 10% sched_debug.cfs_rq[4]:/.tg_load_avg
119.00 ± 20% -33.2% 79.50 ± 12% sched_debug.cfs_rq[4]:/.tg_load_avg_contrib
41.00 ± 12% +72.0% 70.50 ± 58% sched_debug.cfs_rq[5]:/.load
8361817 ± 5% +59.9% 13374083 ± 7% sched_debug.cfs_rq[5]:/.min_vruntime
2265 ± 8% -29.0% 1608 ± 10% sched_debug.cfs_rq[5]:/.tg_load_avg
8064101 ± 5% +170.9% 21848536 ± 12% sched_debug.cfs_rq[6]:/.min_vruntime
12.25 ± 48% +81.6% 22.25 ± 28% sched_debug.cfs_rq[6]:/.nr_spread_over
680647 ± 89% +1373.8% 10031232 ± 26% sched_debug.cfs_rq[6]:/.spread0
2298 ± 8% -29.7% 1615 ± 8% sched_debug.cfs_rq[6]:/.tg_load_avg
94.25 ± 16% -38.2% 58.25 ± 19% sched_debug.cfs_rq[7]:/.load_avg
8303387 ± 6% +218.5% 26442227 ± 14% sched_debug.cfs_rq[7]:/.min_vruntime
40.25 ± 9% -25.5% 30.00 ± 17% sched_debug.cfs_rq[7]:/.runnable_load_avg
919200 ± 58% +1490.1% 14616571 ± 24% sched_debug.cfs_rq[7]:/.spread0
2277 ± 7% -28.1% 1638 ± 12% sched_debug.cfs_rq[7]:/.tg_load_avg
94.50 ± 16% -38.4% 58.25 ± 19% sched_debug.cfs_rq[7]:/.tg_load_avg_contrib
93.50 ± 19% -38.2% 57.75 ± 18% sched_debug.cfs_rq[8]:/.load_avg
8657132 ± 6% +206.7% 26552197 ± 12% sched_debug.cfs_rq[8]:/.min_vruntime
10.00 ± 49% +2720.0% 282.00 ± 56% sched_debug.cfs_rq[8]:/.nr_spread_over
1272282 ± 40% +1057.2% 14722281 ± 21% sched_debug.cfs_rq[8]:/.spread0
2256 ± 8% -25.2% 1688 ± 9% sched_debug.cfs_rq[8]:/.tg_load_avg
88.25 ± 18% -33.7% 58.50 ± 18% sched_debug.cfs_rq[8]:/.tg_load_avg_contrib
89.25 ± 10% -43.4% 50.50 ± 18% sched_debug.cfs_rq[9]:/.load_avg
8573840 ± 11% +212.1% 26757495 ± 13% sched_debug.cfs_rq[9]:/.min_vruntime
13.00 ± 70% +909.6% 131.25 ± 46% sched_debug.cfs_rq[9]:/.nr_spread_over
1188401 ± 86% +1155.7% 14923175 ± 23% sched_debug.cfs_rq[9]:/.spread0
2235 ± 7% -27.0% 1630 ± 9% sched_debug.cfs_rq[9]:/.tg_load_avg
89.25 ± 10% -43.4% 50.50 ± 18% sched_debug.cfs_rq[9]:/.tg_load_avg_contrib
13660 ± 26% +25.6% 17164 ± 7% sched_debug.cpu#0.curr->pid
25.75 ± 28% +564.1% 171.00 ± 31% sched_debug.cpu#0.nr_running
6234824 ± 3% +79.9% 11214928 ± 11% sched_debug.cpu#0.nr_switches
92.25 ± 41% +101.4% 185.75 ± 37% sched_debug.cpu#0.nr_uninterruptible
10264120 ± 2% +48.8% 15270454 ± 8% sched_debug.cpu#0.sched_count
49574 ± 5% -12.4% 43430 ± 6% sched_debug.cpu#0.sched_goidle
5147188 ± 4% +79.7% 9249436 ± 12% sched_debug.cpu#0.ttwu_count
2269312 ± 2% +62.5% 3688685 ± 10% sched_debug.cpu#0.ttwu_local
23.25 ± 28% +589.2% 160.25 ± 36% sched_debug.cpu#1.nr_running
6569750 ± 2% +83.6% 12058832 ± 9% sched_debug.cpu#1.nr_switches
6570052 ± 2% +83.6% 12059572 ± 9% sched_debug.cpu#1.sched_count
4992425 ± 1% +109.0% 10435243 ± 3% sched_debug.cpu#1.ttwu_count
2463897 ± 1% +74.8% 4307284 ± 9% sched_debug.cpu#1.ttwu_local
13.00 ± 44% +303.8% 52.50 ± 42% sched_debug.cpu#10.nr_running
6572956 ± 2% +196.2% 19469907 ± 4% sched_debug.cpu#10.nr_switches
6573272 ± 2% +196.2% 19471457 ± 4% sched_debug.cpu#10.sched_count
5113245 ± 2% +125.5% 11531340 ± 2% sched_debug.cpu#10.ttwu_count
2449382 ± 2% +146.9% 6046615 ± 4% sched_debug.cpu#10.ttwu_local
500000 ± 0% +14.1% 570712 ± 8% sched_debug.cpu#11.max_idle_balance_cost
15.00 ± 46% +246.7% 52.00 ± 42% sched_debug.cpu#11.nr_running
6631320 ± 2% +189.1% 19172684 ± 4% sched_debug.cpu#11.nr_switches
6631668 ± 2% +189.1% 19174234 ± 4% sched_debug.cpu#11.sched_count
5054950 ± 2% +120.6% 11152494 ± 4% sched_debug.cpu#11.ttwu_count
2405487 ± 2% +145.7% 5910910 ± 3% sched_debug.cpu#11.ttwu_local
12.00 ± 59% +791.7% 107.00 ± 75% sched_debug.cpu#12.nr_running
6356857 ± 4% +95.9% 12451675 ± 8% sched_debug.cpu#12.nr_switches
134.25 ± 46% +58.7% 213.00 ± 22% sched_debug.cpu#12.nr_uninterruptible
6357220 ± 4% +95.9% 12452542 ± 8% sched_debug.cpu#12.sched_count
46934 ± 6% -16.9% 38993 ± 10% sched_debug.cpu#12.sched_goidle
5089230 ± 6% +99.1% 10134621 ± 4% sched_debug.cpu#12.ttwu_count
2416053 ± 2% +79.9% 4346652 ± 6% sched_debug.cpu#12.ttwu_local
6657066 ± 5% +86.1% 12387203 ± 9% sched_debug.cpu#13.nr_switches
94.50 ± 71% +109.8% 198.25 ± 19% sched_debug.cpu#13.nr_uninterruptible
6657360 ± 5% +86.1% 12387844 ± 9% sched_debug.cpu#13.sched_count
5089824 ± 1% +103.3% 10347591 ± 6% sched_debug.cpu#13.ttwu_count
2613812 ± 2% +77.8% 4646155 ± 10% sched_debug.cpu#13.ttwu_local
14.25 ± 64% +761.4% 122.75 ± 73% sched_debug.cpu#14.nr_running
7217227 ± 7% +73.5% 12520898 ± 7% sched_debug.cpu#14.nr_switches
-109.00 ±-154% -226.6% 138.00 ± 34% sched_debug.cpu#14.nr_uninterruptible
7217548 ± 7% +73.5% 12521622 ± 7% sched_debug.cpu#14.sched_count
4933024 ± 2% +99.8% 9853790 ± 5% sched_debug.cpu#14.ttwu_count
2627711 ± 3% +76.7% 4643465 ± 5% sched_debug.cpu#14.ttwu_local
11.50 ± 88% +995.7% 126.00 ± 75% sched_debug.cpu#15.nr_running
6705165 ± 4% +91.4% 12831218 ± 9% sched_debug.cpu#15.nr_switches
41.50 ± 82% +256.0% 147.75 ± 27% sched_debug.cpu#15.nr_uninterruptible
6705518 ± 4% +91.4% 12831891 ± 9% sched_debug.cpu#15.sched_count
5124902 ± 2% +102.0% 10351785 ± 4% sched_debug.cpu#15.ttwu_count
2537246 ± 2% +84.4% 4679721 ± 9% sched_debug.cpu#15.ttwu_local
59.75 ± 72% +116.7% 129.50 ± 63% sched_debug.cpu#16.load
12.00 ± 91% +991.7% 131.00 ± 75% sched_debug.cpu#16.nr_running
6807914 ± 3% +88.7% 12847644 ± 6% sched_debug.cpu#16.nr_switches
35.75 ±243% +416.8% 184.75 ± 28% sched_debug.cpu#16.nr_uninterruptible
6808195 ± 3% +88.7% 12848273 ± 6% sched_debug.cpu#16.sched_count
4965300 ± 5% +109.5% 10400978 ± 2% sched_debug.cpu#16.ttwu_count
2587259 ± 3% +84.3% 4769137 ± 4% sched_debug.cpu#16.ttwu_local
7343797 ± 4% +71.0% 12556479 ± 9% sched_debug.cpu#17.nr_switches
-21.50 ±-291% -637.2% 115.50 ± 9% sched_debug.cpu#17.nr_uninterruptible
7344102 ± 4% +71.0% 12557075 ± 9% sched_debug.cpu#17.sched_count
48302 ± 10% -21.2% 38075 ± 9% sched_debug.cpu#17.sched_goidle
4860214 ± 2% +105.2% 9973186 ± 3% sched_debug.cpu#17.ttwu_count
2631813 ± 1% +77.3% 4667413 ± 6% sched_debug.cpu#17.ttwu_local
10.50 ± 70% +361.9% 48.50 ± 34% sched_debug.cpu#18.nr_running
6423142 ± 0% +193.8% 18871804 ± 6% sched_debug.cpu#18.nr_switches
6423521 ± 0% +193.8% 18873844 ± 6% sched_debug.cpu#18.sched_count
4996106 ± 2% +103.7% 10174733 ± 5% sched_debug.cpu#18.ttwu_count
2472857 ± 2% +120.8% 5460849 ± 4% sched_debug.cpu#18.ttwu_local
13.00 ± 54% +267.3% 47.75 ± 47% sched_debug.cpu#19.nr_running
6685332 ± 4% +198.9% 19980393 ± 5% sched_debug.cpu#19.nr_switches
-66.25 ±-66% +202.6% -200.50 ± -2% sched_debug.cpu#19.nr_uninterruptible
6685659 ± 4% +198.9% 19981845 ± 5% sched_debug.cpu#19.sched_count
4916266 ± 4% +151.1% 12346570 ± 5% sched_debug.cpu#19.ttwu_count
2554700 ± 4% +163.4% 6729723 ± 4% sched_debug.cpu#19.ttwu_local
13552 ± 10% +39.4% 18891 ± 15% sched_debug.cpu#2.curr->pid
17.25 ± 67% +784.1% 152.50 ± 42% sched_debug.cpu#2.nr_running
7014128 ± 5% +72.9% 12125114 ± 8% sched_debug.cpu#2.nr_switches
7014454 ± 5% +72.9% 12125842 ± 8% sched_debug.cpu#2.sched_count
4929757 ± 3% +102.3% 9971509 ± 2% sched_debug.cpu#2.ttwu_count
2473376 ± 3% +75.9% 4350629 ± 7% sched_debug.cpu#2.ttwu_local
9.50 ± 58% +365.8% 44.25 ± 48% sched_debug.cpu#20.nr_running
7094564 ± 7% +180.5% 19900502 ± 5% sched_debug.cpu#20.nr_switches
-22.50 ±-193% +866.7% -217.50 ±-35% sched_debug.cpu#20.nr_uninterruptible
7094941 ± 7% +180.5% 19901947 ± 5% sched_debug.cpu#20.sched_count
4847005 ± 2% +150.6% 12148790 ± 4% sched_debug.cpu#20.ttwu_count
2596984 ± 4% +162.1% 6806600 ± 5% sched_debug.cpu#20.ttwu_local
8.50 ± 50% +400.0% 42.50 ± 52% sched_debug.cpu#21.nr_running
6734635 ± 6% +197.1% 20005787 ± 4% sched_debug.cpu#21.nr_switches
6734978 ± 6% +197.1% 20007174 ± 4% sched_debug.cpu#21.sched_count
4954934 ± 2% +152.5% 12510106 ± 7% sched_debug.cpu#21.ttwu_count
2548363 ± 3% +169.5% 6867282 ± 4% sched_debug.cpu#21.ttwu_local
10.00 ± 53% +365.0% 46.50 ± 40% sched_debug.cpu#22.nr_running
6793937 ± 1% +192.2% 19850213 ± 4% sched_debug.cpu#22.nr_switches
6794279 ± 1% +192.2% 19851667 ± 4% sched_debug.cpu#22.sched_count
4999277 ± 2% +147.2% 12359089 ± 5% sched_debug.cpu#22.ttwu_count
2575092 ± 1% +159.1% 6671652 ± 5% sched_debug.cpu#22.ttwu_local
9.50 ± 47% +355.3% 43.25 ± 39% sched_debug.cpu#23.nr_running
6760476 ± 3% +194.8% 19928574 ± 4% sched_debug.cpu#23.nr_switches
6760836 ± 3% +194.8% 19929942 ± 4% sched_debug.cpu#23.sched_count
5057550 ± 0% +142.4% 12258087 ± 4% sched_debug.cpu#23.ttwu_count
2590172 ± 1% +159.7% 6726524 ± 4% sched_debug.cpu#23.ttwu_local
17.00 ± 59% +764.7% 147.00 ± 46% sched_debug.cpu#3.nr_running
6553148 ± 3% +89.1% 12389631 ± 9% sched_debug.cpu#3.nr_switches
-2.50 ±-3542% -7430.0% 183.25 ± 22% sched_debug.cpu#3.nr_uninterruptible
6553515 ± 3% +89.1% 12390332 ± 9% sched_debug.cpu#3.sched_count
5061548 ± 3% +105.1% 10380529 ± 5% sched_debug.cpu#3.ttwu_count
2374084 ± 3% +82.0% 4321429 ± 9% sched_debug.cpu#3.ttwu_local
822869 ± 10% -12.4% 720748 ± 11% sched_debug.cpu#4.avg_idle
14.50 ± 66% +886.2% 143.00 ± 52% sched_debug.cpu#4.nr_running
6627944 ± 4% +85.8% 12313003 ± 8% sched_debug.cpu#4.nr_switches
6628260 ± 4% +85.8% 12313670 ± 8% sched_debug.cpu#4.sched_count
5029009 ± 4% +105.9% 10353607 ± 4% sched_debug.cpu#4.ttwu_count
2417262 ± 3% +79.5% 4339802 ± 7% sched_debug.cpu#4.ttwu_local
41.25 ± 13% +72.1% 71.00 ± 59% sched_debug.cpu#5.load
17.50 ± 68% +691.4% 138.50 ± 47% sched_debug.cpu#5.nr_running
6997533 ± 3% +73.8% 12164617 ± 8% sched_debug.cpu#5.nr_switches
78.00 ± 48% +184.9% 222.25 ± 21% sched_debug.cpu#5.nr_uninterruptible
6997845 ± 3% +73.8% 12165273 ± 8% sched_debug.cpu#5.sched_count
4969036 ± 2% +94.9% 9684195 ± 1% sched_debug.cpu#5.ttwu_count
2483310 ± 3% +71.0% 4247364 ± 5% sched_debug.cpu#5.ttwu_local
831260 ± 12% -17.7% 683789 ± 9% sched_debug.cpu#6.avg_idle
16.75 ± 36% +334.3% 72.75 ± 45% sched_debug.cpu#6.nr_running
6163396 ± 2% +169.8% 16626723 ± 4% sched_debug.cpu#6.nr_switches
54.50 ±162% -311.5% -115.25 ±-34% sched_debug.cpu#6.nr_uninterruptible
6163891 ± 2% +169.8% 16629352 ± 4% sched_debug.cpu#6.sched_count
5128410 ± 3% +77.0% 9074888 ± 4% sched_debug.cpu#6.ttwu_count
2309853 ± 3% +87.4% 4328580 ± 2% sched_debug.cpu#6.ttwu_local
40.25 ± 9% -26.1% 29.75 ± 18% sched_debug.cpu#7.cpu_load[0]
40.25 ± 9% -28.6% 28.75 ± 22% sched_debug.cpu#7.cpu_load[1]
40.00 ± 8% -28.8% 28.50 ± 24% sched_debug.cpu#7.cpu_load[2]
39.75 ± 8% -28.9% 28.25 ± 25% sched_debug.cpu#7.cpu_load[3]
39.75 ± 8% -28.3% 28.50 ± 24% sched_debug.cpu#7.cpu_load[4]
19.25 ± 41% +231.2% 63.75 ± 48% sched_debug.cpu#7.nr_running
6566332 ± 5% +186.9% 18836315 ± 5% sched_debug.cpu#7.nr_switches
15.50 ±158% -1111.3% -156.75 ±-42% sched_debug.cpu#7.nr_uninterruptible
6566642 ± 5% +186.9% 18837866 ± 5% sched_debug.cpu#7.sched_count
5014516 ± 5% +127.1% 11386703 ± 3% sched_debug.cpu#7.ttwu_count
2467246 ± 2% +138.0% 5872007 ± 3% sched_debug.cpu#7.ttwu_local
6802328 ± 5% +181.3% 19136656 ± 4% sched_debug.cpu#8.nr_switches
-76.25 ±-113% +200.0% -228.75 ±-38% sched_debug.cpu#8.nr_uninterruptible
6802722 ± 5% +181.3% 19138285 ± 4% sched_debug.cpu#8.sched_count
4959323 ± 2% +128.0% 11305581 ± 3% sched_debug.cpu#8.ttwu_count
2437353 ± 1% +146.6% 6011685 ± 3% sched_debug.cpu#8.ttwu_local
500000 ± 0% +14.7% 573462 ± 10% sched_debug.cpu#9.max_idle_balance_cost
10.75 ± 56% +409.3% 54.75 ± 57% sched_debug.cpu#9.nr_running
6544270 ± 6% +188.6% 18884484 ± 4% sched_debug.cpu#9.nr_switches
55.50 ±103% -369.8% -149.75 ±-45% sched_debug.cpu#9.nr_uninterruptible
6544607 ± 6% +188.6% 18885996 ± 4% sched_debug.cpu#9.sched_count
5030467 ± 4% +122.3% 11181045 ± 5% sched_debug.cpu#9.ttwu_count
2417027 ± 4% +138.4% 5762968 ± 4% sched_debug.cpu#9.ttwu_local
0.40 ±172% -99.8% 0.00 ± 85% sched_debug.rt_rq[16]:/.rt_time
0.00 ± 49% +73814.5% 0.80 ±100% sched_debug.rt_rq[9]:/.rt_time


lkp-ws02: Westmere-EP
Memory: 16G




hackbench.time.involuntary_context_switches

3e+08 ++------------------------O---------------------------------------+
O O O O O |
2.5e+08 ++ O O O O O
| O O O O |
| O O O O |
2e+08 ++ O |
| O |
1.5e+08 ++ |
| |
1e+08 ++ |
| .*...*..*..*...*..* |
*..*...*..*..*...*.. .*.. ..*..*. |
5e+07 ++ *. *. |
| |
0 ++----------------------------------------------------------------+

vmstat.system.in

300000 ++-----------------------------------------------------------------+
| O O O |
O O O O O |
250000 ++ O O O O
| O |
| O O O O O O |
200000 ++ O O |
| |
150000 ++ |
| |
| |
100000 ++ |
| .*... .*..*...*..*..*...* |
*..*...*..*. *.. .*... .*...*. |
50000 ++------------------*------*---------------------------------------+

[*] bisect-good sample
[O] bisect-bad sample

To reproduce:

git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml

Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Ying Huang
---
LKP_SERVER: inn
LKP_CGI_PORT: 80
LKP_CIFS_PORT: 139
testcase: hackbench
default-monitors:
wait: activate-monitor
kmsg:
uptime:
iostat:
vmstat:
numa-numastat:
numa-vmstat:
numa-meminfo:
proc-vmstat:
proc-stat:
interval: 10
meminfo:
slabinfo:
interrupts:
lock_stat:
latency_stats:
softirqs:
bdi_dev_mapping:
diskstats:
nfsstat:
cpuidle:
cpufreq-stats:
turbostat:
pmeter:
sched_debug:
interval: 60
cpufreq_governor:
default-watchdogs:
oom-killer:
watchdog:
commit: c35a95644687030b27ac6a1b65e65b6d64f33684
model: Westmere-EP
memory: 16G
nr_hdd_partitions: 11
hdd_partitions: "/dev/disk/by-id/scsi-35000c500*-part1"
swap_partitions:
rootfs_partition: "/dev/disk/by-id/ata-WDC_WD1002FAEX-00Z3A0_WD-WCATR5408564-part3"
category: benchmark
nr_threads: 1600%
perf-profile:
freq: 800
hackbench:
mode: process
ipc: pipe
queue: cyclic
testbox: lkp-ws02
tbox_group: lkp-ws02
kconfig: x86_64-rhel
enqueue_time: 2015-09-19 04:55:45.485270869 +08:00
id: 2b587f5fe0998b1d0fd21b2dd529834423980d71
user: lkp
compiler: gcc-4.9
head_commit: c35a95644687030b27ac6a1b65e65b6d64f33684
base_commit: 1f93e4a96c9109378204c147b3eec0d0e8100fde
branch: linux-devel/devel-hourly-2015092113
kernel: "/pkg/linux/x86_64-rhel/gcc-4.9/c35a95644687030b27ac6a1b65e65b6d64f33684/vmlinuz-4.3.0-rc2-01116-gc35a956"
rootfs: debian-x86_64-2015-02-07.cgz
result_root: "/result/hackbench/1600%-process-pipe/lkp-ws02/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/c35a95644687030b27ac6a1b65e65b6d64f33684/0"
job_file: "/lkp/scheduled/lkp-ws02/cyclic_hackbench-1600%-process-pipe-x86_64-rhel-CYCLIC_HEAD-c35a95644687030b27ac6a1b65e65b6d64f33684-20150919-123271-uyuzvi-0.yaml"
dequeue_time: 2015-09-21 14:42:42.666998088 +08:00
nr_cpu: "$(nproc)"
max_uptime: 2400
initrd: "/osimage/debian/debian-x86_64-2015-02-07.cgz"
bootloader_append:
- root=/dev/ram0
- user=lkp
- job=/lkp/scheduled/lkp-ws02/cyclic_hackbench-1600%-process-pipe-x86_64-rhel-CYCLIC_HEAD-c35a95644687030b27ac6a1b65e65b6d64f33684-20150919-123271-uyuzvi-0.yaml
- ARCH=x86_64
- kconfig=x86_64-rhel
- branch=linux-devel/devel-hourly-2015092113
- commit=c35a95644687030b27ac6a1b65e65b6d64f33684
- BOOT_IMAGE=/pkg/linux/x86_64-rhel/gcc-4.9/c35a95644687030b27ac6a1b65e65b6d64f33684/vmlinuz-4.3.0-rc2-01116-gc35a956
- max_uptime=2400
- RESULT_ROOT=/result/hackbench/1600%-process-pipe/lkp-ws02/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/c35a95644687030b27ac6a1b65e65b6d64f33684/0
- LKP_SERVER=inn
- |-
ipmi_watchdog.start_now=1

earlyprintk=ttyS0,115200 systemd.log_level=err
debug apic=debug sysrq_always_enabled rcupdate.rcu_cpu_stall_timeout=100
panic=-1 softlockup_panic=1 nmi_watchdog=panic oops=panic load_ramdisk=2 prompt_ramdisk=0
console=ttyS0,115200 console=tty0 vga=normal

rw
lkp_initrd: "/lkp/lkp/lkp-x86_64.cgz"
modules_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/c35a95644687030b27ac6a1b65e65b6d64f33684/modules.cgz"
bm_initrd: "/osimage/deps/debian-x86_64-2015-02-07.cgz/lkp.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/run-ipconfig.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/turbostat.cgz,/lkp/benchmarks/turbostat.cgz"
job_state: finished
loadavg: 3971.50 4324.77 2450.49 1/325 16949
start_time: '1442817814'
end_time: '1442818421'
version: "/lkp/lkp/.src-20150921-141129"
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875
/usr/bin/hackbench -g 384 --process --pipe -l 1875