Re: [LKP] [lkp] [dcache_{readdir, dir_lseek}() users] 4e82901cd6: reaim.jobs_per_min -49.1% regression

From: Huang\, Ying
Date: Thu Jun 02 2016 - 02:28:43 EST


"Huang, Ying" <ying.huang@xxxxxxxxx> writes:

> Al Viro <viro@xxxxxxxxxxxxxxxxxx> writes:
>
>> On Tue, May 31, 2016 at 04:15:15PM +0800, kernel test robot wrote:
>>>
>>>
>>> FYI, we noticed reaim.jobs_per_min -49.1% regression due to commit:
>>>
>>> commit 4e82901cd6d1af21ae232ae835c36d8230c809e8 ("dcache_{readdir,dir_lseek}() users: switch to ->iterate_shared")
>>> https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
>>>
>>> in testcase: reaim
>>> on test machine: lkp-hsx04: 144 threads Brickland Haswell-EX with 512G memory
>>> with following parameters: cpufreq_governor=performance/iterations=4/nr_task=1600%/test=fserver
>>
>> [snip]
>>
>> Is there any way to get the profiles?
>
> Sorry, our perf-profile support is broken after upgrading perf-profile
> recently. We will restore it ASAP and send back to you the perf profile
> results.

Here is the comparison result with perf profile information. You can
find it via searching 'perf-profile'.

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/nr_task/iterations/test:
lkp-hsx04/reaim/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/1600%/4/fserver

commit:
3125d2650cae97d8f313ab696cd0ed66916e767a
4e82901cd6d1af21ae232ae835c36d8230c809e8

3125d2650cae97d8 4e82901cd6d1af21ae232ae835
---------------- --------------------------
%stddev %change %stddev
\ | \
479.79 Â 3% +2515.6% 12549 Â 2% reaim.child_systime
379.45 Â 0% +13.1% 429.26 Â 0% reaim.child_utime
176186 Â 10% -44.8% 97342 Â 1% reaim.jobs_per_min
76.47 Â 10% -44.8% 42.25 Â 1% reaim.jobs_per_min_child
71.15 Â 5% +22.7% 87.33 Â 0% reaim.jti
178407 Â 10% -44.1% 99667 Â 2% reaim.max_jobs_per_min
80.07 Â 9% +79.3% 143.55 Â 1% reaim.parent_time
28.40 Â 14% -57.1% 12.19 Â 6% reaim.std_dev_percent
17.87 Â 0% -10.7% 15.95 Â 4% reaim.std_dev_time
352.92 Â 0% +23.9% 437.31 Â 1% reaim.time.elapsed_time
352.92 Â 0% +23.9% 437.31 Â 1% reaim.time.elapsed_time.max
1.895e+08 Â 10% -30.0% 1.327e+08 Â 0% reaim.time.minor_page_faults
1052 Â 7% +746.1% 8903 Â 1% reaim.time.percent_of_cpu_this_job_got
2073 Â 7% +1715.8% 37650 Â 2% reaim.time.system_time
1644 Â 10% -21.7% 1287 Â 0% reaim.time.user_time
8298349 Â 13% -48.2% 4296696 Â 5% reaim.time.voluntary_context_switches
404.21 Â 1% +20.7% 487.87 Â 2% uptime.boot
53381 Â 1% -67.3% 17448 Â 7% uptime.idle
4916499 Â 10% -30.7% 3408447 Â 0% softirqs.NET_RX
668035 Â 15% +367.2% 3121347 Â 2% softirqs.RCU
629095 Â 3% +29.8% 816830 Â 1% softirqs.SCHED
3444643 Â 3% +542.4% 22127503 Â 2% softirqs.TIMER
941847 Â 0% +30.3% 1227512 Â 0% vmstat.memory.cache
1.00 Â 0% +36366.7% 364.67 Â 19% vmstat.procs.b
11.67 Â 17% +1437.1% 179.33 Â 2% vmstat.procs.r
60855 Â 11% -55.9% 26830 Â 4% vmstat.system.cs
12673 Â 7% +739.8% 106430 Â 1% vmstat.system.in
352.92 Â 0% +23.9% 437.31 Â 1% time.elapsed_time
352.92 Â 0% +23.9% 437.31 Â 1% time.elapsed_time.max
1.895e+08 Â 10% -30.0% 1.327e+08 Â 0% time.minor_page_faults
1052 Â 7% +746.1% 8903 Â 1% time.percent_of_cpu_this_job_got
2073 Â 7% +1715.8% 37650 Â 2% time.system_time
1644 Â 10% -21.7% 1287 Â 0% time.user_time
8298349 Â 13% -48.2% 4296696 Â 5% time.voluntary_context_switches
5.486e+08 Â 17% +80.6% 9.91e+08 Â 12% cpuidle.C1-HSW.time
2.765e+09 Â 22% -67.7% 8.921e+08 Â 11% cpuidle.C1E-HSW.time
3551940 Â 13% -86.8% 470084 Â 11% cpuidle.C1E-HSW.usage
3.63e+08 Â 3% +384.0% 1.757e+09 Â 12% cpuidle.C3-HSW.time
439297 Â 1% +54.1% 677092 Â 8% cpuidle.C3-HSW.usage
4.368e+10 Â 1% -58.1% 1.83e+10 Â 1% cpuidle.C6-HSW.time
5469354 Â 9% -56.8% 2361450 Â 3% cpuidle.C6-HSW.usage
1.179e+08 Â 11% +119.6% 2.589e+08 Â 14% cpuidle.POLL.time
988.00 Â 12% +199.9% 2963 Â 8% cpuidle.POLL.usage
48184543 Â 17% -41.6% 28158682 Â 9% numa-numastat.node0.local_node
48196904 Â 17% -41.5% 28174412 Â 9% numa-numastat.node0.numa_hit
12361 Â 17% +27.3% 15730 Â 4% numa-numastat.node0.other_node
43586002 Â 10% -20.6% 34591781 Â 8% numa-numastat.node1.local_node
43592187 Â 10% -20.6% 34599930 Â 8% numa-numastat.node1.numa_hit
46908094 Â 5% -27.4% 34074424 Â 8% numa-numastat.node2.local_node
46917366 Â 5% -27.3% 34087462 Â 8% numa-numastat.node2.numa_hit
44223596 Â 11% -30.1% 30903064 Â 8% numa-numastat.node3.local_node
44237505 Â 11% -30.1% 30916293 Â 8% numa-numastat.node3.numa_hit
7.67 Â 7% +753.7% 65.45 Â 1% turbostat.%Busy
223.00 Â 7% +749.2% 1893 Â 1% turbostat.Avg_MHz
21.89 Â 5% +14.3% 25.02 Â 2% turbostat.CPU%c1
0.90 Â 6% +94.8% 1.75 Â 12% turbostat.CPU%c3
69.55 Â 0% -88.8% 7.79 Â 0% turbostat.CPU%c6
42.00 Â 1% +30.2% 54.67 Â 3% turbostat.CoreTmp
5.33 Â 10% -84.4% 0.83 Â 9% turbostat.Pkg%pc2
0.60 Â 31% -75.7% 0.15 Â 58% turbostat.Pkg%pc6
46.67 Â 4% +25.7% 58.67 Â 0% turbostat.PkgTmp
262.91 Â 0% +83.7% 483.07 Â 0% turbostat.PkgWatt
233.44 Â 0% +2.2% 238.47 Â 0% turbostat.RAMWatt
699933 Â 2% +22.7% 859108 Â 1% meminfo.Active
596870 Â 2% +26.9% 757230 Â 1% meminfo.Active(anon)
584065 Â 3% +11.8% 652888 Â 1% meminfo.AnonPages
508898 Â 0% +16.6% 593287 Â 0% meminfo.Cached
2321912 Â 2% +18.4% 2749325 Â 4% meminfo.Committed_AS
5969346 Â 0% +17.7% 7024098 Â 0% meminfo.DirectMap2M
10525 Â 0% +13.7% 11967 Â 1% meminfo.Inactive(anon)
57595 Â 1% +248.1% 200467 Â 2% meminfo.KernelStack
18442 Â 2% +12.2% 20693 Â 0% meminfo.Mapped
112198 Â 1% +23.7% 138840 Â 0% meminfo.PageTables
84558 Â 0% +21.2% 102517 Â 1% meminfo.SReclaimable
348470 Â 0% +53.1% 533390 Â 1% meminfo.SUnreclaim
24174 Â 9% +348.7% 108462 Â 2% meminfo.Shmem
433028 Â 0% +46.9% 635908 Â 1% meminfo.Slab
149068 Â 2% +27.1% 189400 Â 1% proc-vmstat.nr_active_anon
145815 Â 3% +11.9% 163138 Â 1% proc-vmstat.nr_anon_pages
127230 Â 0% +16.7% 148491 Â 0% proc-vmstat.nr_file_pages
2634 Â 0% +14.1% 3006 Â 1% proc-vmstat.nr_inactive_anon
0.00 Â 0% +Inf% 462.33 Â 15% proc-vmstat.nr_isolated_anon
3623 Â 1% +244.2% 12473 Â 2% proc-vmstat.nr_kernel_stack
4620 Â 2% +12.5% 5197 Â 1% proc-vmstat.nr_mapped
27939 Â 1% +24.3% 34731 Â 0% proc-vmstat.nr_page_table_pages
6049 Â 9% +351.0% 27285 Â 2% proc-vmstat.nr_shmem
21138 Â 0% +21.4% 25663 Â 1% proc-vmstat.nr_slab_reclaimable
87101 Â 0% +53.1% 133353 Â 1% proc-vmstat.nr_slab_unreclaimable
0.00 Â 0% +Inf% 870680 Â 2% proc-vmstat.numa_hint_faults
0.00 Â 0% +Inf% 406936 Â 3% proc-vmstat.numa_hint_faults_local
1.829e+08 Â 11% -30.2% 1.278e+08 Â 0% proc-vmstat.numa_hit
1.829e+08 Â 11% -30.2% 1.277e+08 Â 0% proc-vmstat.numa_local
41727 Â 0% +20.2% 50144 Â 0% proc-vmstat.numa_other
0.00 Â 0% +Inf% 258503 Â 3% proc-vmstat.numa_pages_migrated
0.00 Â 0% +Inf% 1389875 Â 2% proc-vmstat.numa_pte_updates
6374 Â 13% +254.1% 22575 Â 2% proc-vmstat.pgactivate
766567 Â 17% -39.9% 460866 Â 9% proc-vmstat.pgalloc_dma32
1.862e+08 Â 11% -29.1% 1.32e+08 Â 0% proc-vmstat.pgalloc_normal
1.906e+08 Â 10% -29.7% 1.339e+08 Â 0% proc-vmstat.pgfault
1.869e+08 Â 11% -29.2% 1.323e+08 Â 0% proc-vmstat.pgfree
0.00 Â 0% +Inf% 2076 Â 4% proc-vmstat.pgmigrate_fail
0.00 Â 0% +Inf% 258503 Â 3% proc-vmstat.pgmigrate_success
26527 Â 23% +104.1% 54152 Â 16% numa-meminfo.node0.KernelStack
3090 Â 3% +80.5% 5578 Â 50% numa-meminfo.node0.Mapped
921890 Â 1% +9.2% 1007082 Â 4% numa-meminfo.node0.MemUsed
108135 Â 14% +22.1% 132019 Â 12% numa-meminfo.node0.SUnreclaim
692.67 Â 32% +562.7% 4590 Â 86% numa-meminfo.node0.Shmem
185924 Â 7% +41.3% 262685 Â 19% numa-meminfo.node1.Active
159207 Â 8% +49.7% 238288 Â 21% numa-meminfo.node1.Active(anon)
151747 Â 6% +15.6% 175383 Â 12% numa-meminfo.node1.AnonPages
10565 Â 74% +348.1% 47347 Â 9% numa-meminfo.node1.KernelStack
801350 Â 7% +17.1% 938412 Â 6% numa-meminfo.node1.MemUsed
22990 Â 15% +34.3% 30874 Â 7% numa-meminfo.node1.SReclaimable
82401 Â 25% +58.8% 130874 Â 5% numa-meminfo.node1.SUnreclaim
105391 Â 16% +53.5% 161749 Â 4% numa-meminfo.node1.Slab
12599 Â 48% +352.5% 57016 Â 2% numa-meminfo.node2.KernelStack
773326 Â 2% +19.0% 920096 Â 1% numa-meminfo.node2.MemUsed
28753 Â 26% +60.8% 46232 Â 17% numa-meminfo.node2.PageTables
85146 Â 14% +75.2% 149181 Â 1% numa-meminfo.node2.SUnreclaim
108938 Â 14% +62.1% 176621 Â 3% numa-meminfo.node2.Slab
159543 Â 16% +44.1% 229926 Â 26% numa-meminfo.node3.Active
134722 Â 20% +51.8% 204566 Â 29% numa-meminfo.node3.Active(anon)
7948 Â 42% +427.7% 41941 Â 12% numa-meminfo.node3.KernelStack
5114 Â 56% +50.8% 7714 Â 38% numa-meminfo.node3.Mapped
707514 Â 8% +24.1% 878069 Â 7% numa-meminfo.node3.MemUsed
72837 Â 13% +66.8% 121492 Â 9% numa-meminfo.node3.SUnreclaim
3674 Â 98% +970.8% 39342 Â121% numa-meminfo.node3.Shmem
90421 Â 9% +60.3% 144902 Â 6% numa-meminfo.node3.Slab
1658 Â 23% +104.0% 3384 Â 16% numa-vmstat.node0.nr_kernel_stack
772.00 Â 3% +81.2% 1399 Â 50% numa-vmstat.node0.nr_mapped
172.67 Â 32% +565.3% 1148 Â 86% numa-vmstat.node0.nr_shmem
27026 Â 14% +22.1% 32987 Â 12% numa-vmstat.node0.nr_slab_unreclaimable
22869693 Â 20% -47.9% 11904640 Â 10% numa-vmstat.node0.numa_hit
22823515 Â 20% -48.1% 11856155 Â 10% numa-vmstat.node0.numa_local
39771 Â 8% +50.4% 59823 Â 20% numa-vmstat.node1.nr_active_anon
37907 Â 6% +16.6% 44186 Â 12% numa-vmstat.node1.nr_anon_pages
0.00 Â 0% +Inf% 122.33 Â 8% numa-vmstat.node1.nr_isolated_anon
664.00 Â 75% +340.8% 2927 Â 8% numa-vmstat.node1.nr_kernel_stack
5747 Â 15% +33.9% 7698 Â 7% numa-vmstat.node1.nr_slab_reclaimable
20592 Â 25% +58.6% 32652 Â 5% numa-vmstat.node1.nr_slab_unreclaimable
20571250 Â 12% -26.6% 15099745 Â 9% numa-vmstat.node1.numa_hit
20512586 Â 12% -26.7% 15039547 Â 9% numa-vmstat.node1.numa_local
787.00 Â 48% +351.6% 3554 Â 2% numa-vmstat.node2.nr_kernel_stack
7141 Â 26% +62.1% 11577 Â 15% numa-vmstat.node2.nr_page_table_pages
21277 Â 14% +75.3% 37298 Â 1% numa-vmstat.node2.nr_slab_unreclaimable
21913451 Â 6% -31.2% 15076742 Â 12% numa-vmstat.node2.numa_hit
21818774 Â 6% -31.3% 14979744 Â 12% numa-vmstat.node2.numa_local
33642 Â 20% +52.6% 51332 Â 29% numa-vmstat.node3.nr_active_anon
0.00 Â 0% +Inf% 117.00 Â 18% numa-vmstat.node3.nr_isolated_anon
489.33 Â 43% +431.1% 2599 Â 13% numa-vmstat.node3.nr_kernel_stack
1277 Â 56% +50.7% 1925 Â 38% numa-vmstat.node3.nr_mapped
917.00 Â 98% +969.0% 9803 Â121% numa-vmstat.node3.nr_shmem
18198 Â 13% +66.8% 30362 Â 9% numa-vmstat.node3.nr_slab_unreclaimable
20763416 Â 12% -36.4% 13211498 Â 9% numa-vmstat.node3.numa_hit
20664485 Â 12% -36.5% 13114307 Â 9% numa-vmstat.node3.numa_local
112783 Â 5% -9.2% 102353 Â 0% slabinfo.anon_vma.active_objs
57839 Â 1% +46.3% 84628 Â 3% slabinfo.cred_jar.active_objs
1386 Â 1% +53.5% 2127 Â 3% slabinfo.cred_jar.active_slabs
58232 Â 1% +53.5% 89381 Â 3% slabinfo.cred_jar.num_objs
1386 Â 1% +53.5% 2127 Â 3% slabinfo.cred_jar.num_slabs
106001 Â 0% +66.1% 176110 Â 2% slabinfo.dentry.active_objs
2615 Â 0% +62.0% 4235 Â 2% slabinfo.dentry.active_slabs
109861 Â 0% +62.0% 177924 Â 2% slabinfo.dentry.num_objs
2615 Â 0% +62.0% 4235 Â 2% slabinfo.dentry.num_slabs
47974 Â 2% +580.1% 326281 Â 5% slabinfo.kmalloc-256.active_objs
1400 Â 0% +282.7% 5358 Â 4% slabinfo.kmalloc-256.active_slabs
89624 Â 0% +282.6% 342936 Â 4% slabinfo.kmalloc-256.num_objs
1400 Â 0% +282.7% 5358 Â 4% slabinfo.kmalloc-256.num_slabs
19286 Â 4% +28.4% 24755 Â 4% slabinfo.kmalloc-512.active_objs
19445 Â 4% +27.9% 24875 Â 4% slabinfo.kmalloc-512.num_objs
8541 Â 1% +11.8% 9546 Â 1% slabinfo.mm_struct.active_objs
18599 Â 2% +279.3% 70553 Â 2% slabinfo.pid.active_objs
290.00 Â 2% +281.3% 1105 Â 2% slabinfo.pid.active_slabs
18599 Â 2% +280.7% 70800 Â 2% slabinfo.pid.num_objs
290.00 Â 2% +281.3% 1105 Â 2% slabinfo.pid.num_slabs
27990 Â 1% -13.1% 24330 Â 1% slabinfo.proc_inode_cache.active_objs
28081 Â 0% -13.4% 24330 Â 1% slabinfo.proc_inode_cache.num_objs
15869 Â 0% +72.3% 27341 Â 3% slabinfo.radix_tree_node.active_objs
283.00 Â 0% +72.6% 488.33 Â 3% slabinfo.radix_tree_node.active_slabs
15869 Â 0% +72.6% 27391 Â 3% slabinfo.radix_tree_node.num_objs
283.00 Â 0% +72.6% 488.33 Â 3% slabinfo.radix_tree_node.num_slabs
38727 Â 0% +14.9% 44516 Â 3% slabinfo.shmem_inode_cache.active_objs
799.67 Â 0% +15.0% 919.67 Â 3% slabinfo.shmem_inode_cache.active_slabs
39214 Â 0% +15.0% 45090 Â 3% slabinfo.shmem_inode_cache.num_objs
799.67 Â 0% +15.0% 919.67 Â 3% slabinfo.shmem_inode_cache.num_slabs
8815 Â 2% +20.7% 10642 Â 1% slabinfo.sighand_cache.active_objs
610.67 Â 2% +17.6% 718.33 Â 1% slabinfo.sighand_cache.active_slabs
9167 Â 2% +17.6% 10778 Â 1% slabinfo.sighand_cache.num_objs
610.67 Â 2% +17.6% 718.33 Â 1% slabinfo.sighand_cache.num_slabs
16257 Â 1% +16.1% 18868 Â 2% slabinfo.signal_cache.active_objs
16613 Â 1% +17.4% 19512 Â 2% slabinfo.signal_cache.num_objs
4120 Â 1% +209.5% 12750 Â 2% slabinfo.task_struct.active_objs
1428 Â 0% +201.3% 4304 Â 2% slabinfo.task_struct.active_slabs
4287 Â 0% +201.2% 12913 Â 2% slabinfo.task_struct.num_objs
1428 Â 0% +201.3% 4304 Â 2% slabinfo.task_struct.num_slabs
9460 Â 5% -13.0% 8228 Â 0% slabinfo.tw_sock_TCP.active_objs
9460 Â 5% -13.0% 8228 Â 0% slabinfo.tw_sock_TCP.num_objs
0.00 Â -1% +Inf% 40.39 Â 9% perf-profile.____fput.task_work_run.exit_to_usermode_loop.syscall_return_slowpath.entry_SYSCALL_64_fastpath
3.16 Â 18% -100.0% 0.00 Â -1% perf-profile.__do_page_fault.do_page_fault.page_fault
1.15 Â 10% -100.0% 0.00 Â -1% perf-profile.__do_softirq.irq_exit.scheduler_ipi.smp_reschedule_interrupt.reschedule_interrupt
0.00 Â -1% +Inf% 40.39 Â 9% perf-profile.__fput.____fput.task_work_run.exit_to_usermode_loop.syscall_return_slowpath
0.95 Â 12% -100.0% 0.00 Â -1% perf-profile.__libc_fork
1.57 Â 14% -100.0% 0.00 Â -1% perf-profile._do_fork.sys_clone.do_syscall_64.return_from_SYSCALL_64
0.00 Â -1% +Inf% 3.69 Â 8% perf-profile._raw_spin_lock.dcache_dir_close.__fput.____fput.task_work_run
0.00 Â -1% +Inf% 47.51 Â 16% perf-profile._raw_spin_lock.dcache_readdir.iterate_dir.sys_getdents.entry_SYSCALL_64_fastpath
18.57 Â 13% -95.4% 0.86 Â 8% perf-profile._raw_spin_lock.iterate_dir.sys_getdents.entry_SYSCALL_64_fastpath
0.00 Â -1% +Inf% 26.77 Â 10% perf-profile._raw_spin_trylock.dput.dcache_dir_close.__fput.____fput
1.50 Â 9% -100.0% 0.00 Â -1% perf-profile.call_cpuidle.cpu_startup_entry.rest_init.start_kernel.x86_64_start_reservations
23.62 Â 0% -88.3% 2.77 Â 5% perf-profile.call_cpuidle.cpu_startup_entry.start_secondary
9.82 Â 3% -100.0% 0.00 Â -1% perf-profile.call_rwsem_down_write_failed.down_write.dcache_dir_lseek.sys_lseek.entry_SYSCALL_64_fastpath
20.75 Â 2% -100.0% 0.00 Â -1% perf-profile.call_rwsem_down_write_failed.down_write.iterate_dir.sys_getdents.entry_SYSCALL_64_fastpath
0.77 Â 36% -100.0% 0.00 Â -1% perf-profile.copy_process.part.30._do_fork.sys_clone.do_syscall_64.return_from_SYSCALL_64
1.51 Â 9% -100.0% 0.00 Â -1% perf-profile.cpu_startup_entry.rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel
24.62 Â 1% -88.6% 2.81 Â 4% perf-profile.cpu_startup_entry.start_secondary
0.00 Â -1% +Inf% 5.70 Â 70% perf-profile.cpu_stopper_thread.smpboot_thread_fn.kthread.ret_from_fork
1.49 Â 10% -100.0% 0.00 Â -1% perf-profile.cpuidle_enter.call_cpuidle.cpu_startup_entry.rest_init.start_kernel
23.61 Â 0% -88.3% 2.77 Â 5% perf-profile.cpuidle_enter.call_cpuidle.cpu_startup_entry.start_secondary
23.40 Â 0% -88.2% 2.76 Â 4% perf-profile.cpuidle_enter_state.cpuidle_enter.call_cpuidle.cpu_startup_entry.start_secondary
0.00 Â -1% +Inf% 40.38 Â 9% perf-profile.dcache_dir_close.__fput.____fput.task_work_run.exit_to_usermode_loop
9.91 Â 3% -100.0% 0.00 Â -1% perf-profile.dcache_dir_lseek.sys_lseek.entry_SYSCALL_64_fastpath
1.27 Â 6% +3652.1% 47.78 Â 15% perf-profile.dcache_readdir.iterate_dir.sys_getdents.entry_SYSCALL_64_fastpath
2.84 Â 4% -100.0% 0.00 Â -1% perf-profile.do_execveat_common.isra.38.sys_execve.do_syscall_64.return_from_SYSCALL_64
1.67 Â 12% -100.0% 0.00 Â -1% perf-profile.do_execveat_common.isra.38.sys_execve.do_syscall_64.return_from_SYSCALL_64.execve
1.50 Â 14% -100.0% 0.00 Â -1% perf-profile.do_exit.do_group_exit.sys_exit_group.entry_SYSCALL_64_fastpath
1.22 Â 4% -100.0% 0.00 Â -1% perf-profile.do_filp_open.do_sys_open.sys_creat.entry_SYSCALL_64_fastpath
1.50 Â 14% -100.0% 0.00 Â -1% perf-profile.do_group_exit.sys_exit_group.entry_SYSCALL_64_fastpath
1.20 Â 29% -100.0% 0.00 Â -1% perf-profile.do_munmap.sys_brk.entry_SYSCALL_64_fastpath
3.19 Â 18% -100.0% 0.00 Â -1% perf-profile.do_page_fault.page_fault
1.33 Â 4% -100.0% 0.00 Â -1% perf-profile.do_sys_open.sys_creat.entry_SYSCALL_64_fastpath
4.59 Â 7% -96.1% 0.18 Â141% perf-profile.do_syscall_64.return_from_SYSCALL_64
1.68 Â 13% -100.0% 0.00 Â -1% perf-profile.do_syscall_64.return_from_SYSCALL_64.execve
9.83 Â 3% -100.0% 0.00 Â -1% perf-profile.down_write.dcache_dir_lseek.sys_lseek.entry_SYSCALL_64_fastpath
20.77 Â 2% -100.0% 0.00 Â -1% perf-profile.down_write.iterate_dir.sys_getdents.entry_SYSCALL_64_fastpath
0.00 Â -1% +Inf% 36.20 Â 9% perf-profile.dput.dcache_dir_close.__fput.____fput.task_work_run
58.93 Â 2% +52.4% 89.80 Â 4% perf-profile.entry_SYSCALL_64_fastpath
1.69 Â 12% -100.0% 0.00 Â -1% perf-profile.execve
1.12 Â 13% -100.0% 0.00 Â -1% perf-profile.exit_mmap.mmput.do_exit.do_group_exit.sys_exit_group
0.00 Â -1% +Inf% 40.41 Â 9% perf-profile.exit_to_usermode_loop.syscall_return_slowpath.entry_SYSCALL_64_fastpath
0.90 Â 14% -100.0% 0.00 Â -1% perf-profile.filemap_map_pages.handle_pte_fault.handle_mm_fault.__do_page_fault.do_page_fault
2.82 Â 19% -100.0% 0.00 Â -1% perf-profile.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
2.45 Â 18% -100.0% 0.00 Â -1% perf-profile.handle_pte_fault.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
21.93 Â 5% -88.0% 2.64 Â 3% perf-profile.intel_idle.cpuidle_enter_state.cpuidle_enter.call_cpuidle.cpu_startup_entry
1.18 Â 11% -100.0% 0.00 Â -1% perf-profile.irq_exit.scheduler_ipi.smp_reschedule_interrupt.reschedule_interrupt.cpuidle_enter
1.17 Â 12% -100.0% 0.00 Â -1% perf-profile.load_elf_binary.search_binary_handler.do_execveat_common.isra.38.sys_execve.do_syscall_64
0.00 Â -1% +Inf% 3.88 Â 8% perf-profile.lockref_put_return.dput.dcache_dir_close.__fput.____fput
1.12 Â 13% -100.0% 0.00 Â -1% perf-profile.mmput.do_exit.do_group_exit.sys_exit_group.entry_SYSCALL_64_fastpath
0.00 Â -1% +Inf% 5.69 Â 70% perf-profile.multi_cpu_stop.cpu_stopper_thread.smpboot_thread_fn.kthread.ret_from_fork
0.00 Â -1% +Inf% 47.42 Â 16% perf-profile.native_queued_spin_lock_slowpath._raw_spin_lock.dcache_readdir.iterate_dir.sys_getdents
3.94 Â 20% -100.0% 0.00 Â -1% perf-profile.osq_lock.rwsem_down_write_failed.call_rwsem_down_write_failed.down_write.dcache_dir_lseek
6.91 Â 20% -100.0% 0.00 Â -1% perf-profile.osq_lock.rwsem_down_write_failed.call_rwsem_down_write_failed.down_write.iterate_dir
3.21 Â 18% -100.0% 0.00 Â -1% perf-profile.page_fault
1.20 Â 4% -100.0% 0.00 Â -1% perf-profile.path_openat.do_filp_open.do_sys_open.sys_creat.entry_SYSCALL_64_fastpath
1.18 Â 11% -100.0% 0.00 Â -1% perf-profile.reschedule_interrupt.cpuidle_enter.call_cpuidle.cpu_startup_entry.rest_init
1.51 Â 9% -100.0% 0.00 Â -1% perf-profile.rest_init.start_kernel.x86_64_start_reservations.x86_64_start_kernel
4.60 Â 7% -96.1% 0.18 Â141% perf-profile.return_from_SYSCALL_64
1.68 Â 13% -100.0% 0.00 Â -1% perf-profile.return_from_SYSCALL_64.execve
1.14 Â 10% -100.0% 0.00 Â -1% perf-profile.run_rebalance_domains.__do_softirq.irq_exit.scheduler_ipi.smp_reschedule_interrupt
9.82 Â 3% -100.0% 0.00 Â -1% perf-profile.rwsem_down_write_failed.call_rwsem_down_write_failed.down_write.dcache_dir_lseek.sys_lseek
20.75 Â 2% -100.0% 0.00 Â -1% perf-profile.rwsem_down_write_failed.call_rwsem_down_write_failed.down_write.iterate_dir.sys_getdents
5.79 Â 8% -100.0% 0.00 Â -1% perf-profile.rwsem_spin_on_owner.rwsem_down_write_failed.call_rwsem_down_write_failed.down_write.dcache_dir_lseek
13.64 Â 13% -100.0% 0.00 Â -1% perf-profile.rwsem_spin_on_owner.rwsem_down_write_failed.call_rwsem_down_write_failed.down_write.iterate_dir
1.18 Â 11% -100.0% 0.00 Â -1% perf-profile.scheduler_ipi.smp_reschedule_interrupt.reschedule_interrupt.cpuidle_enter.call_cpuidle
1.17 Â 12% -100.0% 0.00 Â -1% perf-profile.search_binary_handler.do_execveat_common.isra.38.sys_execve.do_syscall_64.return_from_SYSCALL_64
1.18 Â 11% -100.0% 0.00 Â -1% perf-profile.smp_reschedule_interrupt.reschedule_interrupt.cpuidle_enter.call_cpuidle.cpu_startup_entry
0.17 Â141% +3328.0% 5.71 Â 70% perf-profile.smpboot_thread_fn.kthread.ret_from_fork
1.51 Â 9% -100.0% 0.00 Â -1% perf-profile.start_kernel.x86_64_start_reservations.x86_64_start_kernel
24.70 Â 1% -88.6% 2.81 Â 4% perf-profile.start_secondary
1.43 Â 29% -100.0% 0.00 Â -1% perf-profile.sys_brk.entry_SYSCALL_64_fastpath
1.60 Â 14% -100.0% 0.00 Â -1% perf-profile.sys_clone.do_syscall_64.return_from_SYSCALL_64
1.35 Â 4% -100.0% 0.00 Â -1% perf-profile.sys_creat.entry_SYSCALL_64_fastpath
2.88 Â 4% -100.0% 0.00 Â -1% perf-profile.sys_execve.do_syscall_64.return_from_SYSCALL_64
1.68 Â 13% -100.0% 0.00 Â -1% perf-profile.sys_execve.do_syscall_64.return_from_SYSCALL_64.execve
1.50 Â 14% -100.0% 0.00 Â -1% perf-profile.sys_exit_group.entry_SYSCALL_64_fastpath
9.92 Â 3% -100.0% 0.00 Â -1% perf-profile.sys_lseek.entry_SYSCALL_64_fastpath
0.00 Â -1% +Inf% 40.41 Â 9% perf-profile.syscall_return_slowpath.entry_SYSCALL_64_fastpath
0.00 Â -1% +Inf% 40.39 Â 9% perf-profile.task_work_run.exit_to_usermode_loop.syscall_return_slowpath.entry_SYSCALL_64_fastpath
0.83 Â 28% -100.0% 0.00 Â -1% perf-profile.unmap_region.do_munmap.sys_brk.entry_SYSCALL_64_fastpath
1.51 Â 9% -100.0% 0.00 Â -1% perf-profile.x86_64_start_kernel
1.51 Â 9% -100.0% 0.00 Â -1% perf-profile.x86_64_start_reservations.x86_64_start_kernel
829.37 Â 93% +4.5e+05% 3712300 Â 15% sched_debug.cfs_rq:/.MIN_vruntime.avg
119429 Â 93% +19512.3% 23422909 Â 9% sched_debug.cfs_rq:/.MIN_vruntime.max
9917 Â 93% +71917.5% 7142609 Â 8% sched_debug.cfs_rq:/.MIN_vruntime.stddev
3.21 Â 4% +6575.4% 214.12 Â 40% sched_debug.cfs_rq:/.load.avg
109.00 Â 15% +7665.8% 8464 Â122% sched_debug.cfs_rq:/.load.max
16.09 Â 11% +5426.3% 889.45 Â 93% sched_debug.cfs_rq:/.load.stddev
15.28 Â 40% +1359.4% 223.01 Â 38% sched_debug.cfs_rq:/.load_avg.avg
219.94 Â 11% +3713.8% 8388 Â124% sched_debug.cfs_rq:/.load_avg.max
39.68 Â 28% +2129.6% 884.70 Â 93% sched_debug.cfs_rq:/.load_avg.stddev
829.37 Â 93% +4.5e+05% 3712300 Â 15% sched_debug.cfs_rq:/.max_vruntime.avg
119429 Â 93% +19512.3% 23422909 Â 9% sched_debug.cfs_rq:/.max_vruntime.max
9917 Â 93% +71917.5% 7142609 Â 8% sched_debug.cfs_rq:/.max_vruntime.stddev
814966 Â 17% +2450.9% 20789003 Â 0% sched_debug.cfs_rq:/.min_vruntime.avg
1390384 Â 22% +1767.6% 25967271 Â 1% sched_debug.cfs_rq:/.min_vruntime.max
347827 Â 5% +4440.8% 15793966 Â 2% sched_debug.cfs_rq:/.min_vruntime.min
388051 Â 31% +499.6% 2326901 Â 6% sched_debug.cfs_rq:/.min_vruntime.stddev
0.06 Â 38% +936.7% 0.65 Â 15% sched_debug.cfs_rq:/.nr_running.avg
1.00 Â 0% +100.0% 2.00 Â 10% sched_debug.cfs_rq:/.nr_running.max
0.21 Â 4% +174.3% 0.58 Â 9% sched_debug.cfs_rq:/.nr_running.stddev
1.56 Â 9% +12856.6% 202.00 Â 42% sched_debug.cfs_rq:/.runnable_load_avg.avg
66.06 Â 16% +12219.4% 8137 Â125% sched_debug.cfs_rq:/.runnable_load_avg.max
8.02 Â 11% +10565.5% 855.17 Â 95% sched_debug.cfs_rq:/.runnable_load_avg.stddev
-505397 Â-39% +598.7% -3531421 Â -4% sched_debug.cfs_rq:/.spread0.avg
70987 Â 50% +2221.6% 1648031 Â 24% sched_debug.cfs_rq:/.spread0.max
-973130 Â-32% +776.3% -8527672 Â -5% sched_debug.cfs_rq:/.spread0.min
388593 Â 31% +499.0% 2327750 Â 6% sched_debug.cfs_rq:/.spread0.stddev
82.02 Â 47% +514.1% 503.68 Â 13% sched_debug.cfs_rq:/.util_avg.avg
825.72 Â 2% +23.0% 1015 Â 0% sched_debug.cfs_rq:/.util_avg.max
148.13 Â 6% +157.7% 381.70 Â 5% sched_debug.cfs_rq:/.util_avg.stddev
1000000 Â 0% +42.9% 1428578 Â 2% sched_debug.cpu.avg_idle.max
45692 Â 5% +171.0% 123825 Â 29% sched_debug.cpu.avg_idle.min
200593 Â 1% +29.5% 259832 Â 1% sched_debug.cpu.clock.avg
200617 Â 1% +29.5% 259876 Â 1% sched_debug.cpu.clock.max
200562 Â 1% +29.5% 259755 Â 1% sched_debug.cpu.clock.min
14.70 Â 14% +120.1% 32.35 Â 37% sched_debug.cpu.clock.stddev
200593 Â 1% +29.5% 259832 Â 1% sched_debug.cpu.clock_task.avg
200617 Â 1% +29.5% 259876 Â 1% sched_debug.cpu.clock_task.max
200562 Â 1% +29.5% 259755 Â 1% sched_debug.cpu.clock_task.min
14.70 Â 14% +120.1% 32.35 Â 37% sched_debug.cpu.clock_task.stddev
4.43 Â 4% +4542.0% 205.65 Â 41% sched_debug.cpu.cpu_load[0].avg
232.83 Â 3% +3447.2% 8259 Â123% sched_debug.cpu.cpu_load[0].max
27.55 Â 0% +3070.6% 873.34 Â 93% sched_debug.cpu.cpu_load[0].stddev
3.30 Â 10% +6122.5% 205.26 Â 41% sched_debug.cpu.cpu_load[1].avg
174.61 Â 16% +4605.4% 8216 Â124% sched_debug.cpu.cpu_load[1].max
19.61 Â 12% +4317.9% 866.29 Â 94% sched_debug.cpu.cpu_load[1].stddev
2.61 Â 15% +7760.5% 204.97 Â 41% sched_debug.cpu.cpu_load[2].avg
132.22 Â 22% +6105.3% 8204 Â124% sched_debug.cpu.cpu_load[2].max
14.56 Â 17% +5826.3% 863.13 Â 94% sched_debug.cpu.cpu_load[2].stddev
2.20 Â 16% +9215.8% 204.90 Â 41% sched_debug.cpu.cpu_load[3].avg
99.11 Â 21% +8168.8% 8195 Â124% sched_debug.cpu.cpu_load[3].max
11.30 Â 17% +7522.6% 861.37 Â 94% sched_debug.cpu.cpu_load[3].stddev
2.05 Â 16% +9928.6% 205.14 Â 41% sched_debug.cpu.cpu_load[4].avg
77.39 Â 15% +10472.4% 8181 Â124% sched_debug.cpu.cpu_load[4].max
9.48 Â 14% +8964.4% 859.30 Â 95% sched_debug.cpu.cpu_load[4].stddev
4243 Â 62% +725.9% 35049 Â 11% sched_debug.cpu.curr->pid.avg
13719 Â 17% +114.2% 29382 Â 10% sched_debug.cpu.curr->pid.stddev
3.31 Â 6% +6370.5% 214.11 Â 40% sched_debug.cpu.load.avg
106.78 Â 13% +7827.4% 8464 Â122% sched_debug.cpu.load.max
16.22 Â 11% +5384.4% 889.45 Â 93% sched_debug.cpu.load.stddev
500000 Â 0% +53.0% 764948 Â 1% sched_debug.cpu.max_idle_balance_cost.max
0.00 Â 0% +Inf% 36894 Â 17% sched_debug.cpu.max_idle_balance_cost.stddev
0.00 Â 8% +339.0% 0.00 Â 37% sched_debug.cpu.next_balance.stddev
60875 Â 2% +180.1% 170518 Â 0% sched_debug.cpu.nr_load_updates.avg
78714 Â 5% +145.4% 193190 Â 0% sched_debug.cpu.nr_load_updates.max
44765 Â 3% +231.7% 148501 Â 0% sched_debug.cpu.nr_load_updates.min
10794 Â 7% +15.2% 12434 Â 2% sched_debug.cpu.nr_load_updates.stddev
0.08 Â 62% +1154.1% 1.06 Â 18% sched_debug.cpu.nr_running.avg
1.83 Â 64% +254.5% 6.50 Â 5% sched_debug.cpu.nr_running.max
0.29 Â 42% +294.9% 1.16 Â 9% sched_debug.cpu.nr_running.stddev
62835 Â 12% -44.3% 35017 Â 4% sched_debug.cpu.nr_switches.avg
127178 Â 11% -39.9% 76437 Â 8% sched_debug.cpu.nr_switches.max
22006 Â 13% -24.2% 16683 Â 12% sched_debug.cpu.nr_switches.min
33982 Â 12% -63.7% 12339 Â 4% sched_debug.cpu.nr_switches.stddev
10.66 Â 10% +12.4% 11.98 Â 4% sched_debug.cpu.nr_uninterruptible.avg
399.50 Â 52% +237.2% 1346 Â 1% sched_debug.cpu.nr_uninterruptible.max
-386.89 Â-56% +266.6% -1418 Â-14% sched_debug.cpu.nr_uninterruptible.min
270.10 Â 64% +153.9% 685.84 Â 4% sched_debug.cpu.nr_uninterruptible.stddev
200566 Â 1% +29.5% 259792 Â 1% sched_debug.cpu_clk
198070 Â 1% +30.0% 257505 Â 1% sched_debug.ktime
0.00 Â 27% +2092.7% 0.02 Â 26% sched_debug.rt_rq:/.rt_time.avg
0.02 Â 19% +3170.6% 0.65 Â 15% sched_debug.rt_rq:/.rt_time.max
0.00 Â 18% +3180.2% 0.09 Â 17% sched_debug.rt_rq:/.rt_time.stddev
200566 Â 1% +29.5% 259792 Â 1% sched_debug.sched_clk

Best Regards,
Huang, Ying