[lkp] [fsnotify] 8f2f3eb59d: -4.0% will-it-scale.per_thread_ops

From: kernel test robot
Date: Sat Aug 22 2015 - 19:33:23 EST


FYI, we noticed the below changes on

git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
commit 8f2f3eb59dff4ec538de55f2e0592fec85966aab ("fsnotify: fix oops in fsnotify_clear_marks_by_group_flags()")


=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test:
lkp-sbx04/will-it-scale/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/read1

commit:
447f6a95a9c80da7faaec3e66e656eab8f262640
8f2f3eb59dff4ec538de55f2e0592fec85966aab

447f6a95a9c80da7 8f2f3eb59dff4ec538de55f2e0
---------------- --------------------------
%stddev %change %stddev
\ | \
1844687 ± 0% -4.0% 1770899 ± 0% will-it-scale.per_thread_ops
283.69 ± 0% +9.5% 310.64 ± 0% will-it-scale.time.user_time
4576 ± 3% -7.3% 4242 ± 6% will-it-scale.time.voluntary_context_switches
7211 ± 10% +54.0% 11101 ± 18% cpuidle.C1E-SNB.usage
10636 ± 36% +69.3% 18003 ± 36% numa-meminfo.node1.Shmem
1.07 ± 4% -13.1% 0.93 ± 9% perf-profile.cpu-cycles.selinux_file_permission.security_file_permission.rw_verify_area.vfs_read.sys_read
4576 ± 3% -7.3% 4242 ± 6% time.voluntary_context_switches
526.75 ±104% -94.2% 30.50 ± 98% numa-numastat.node1.other_node
1540 ± 35% -74.2% 398.00 ± 90% numa-numastat.node2.other_node
32344 ± 5% +7.4% 34722 ± 4% numa-vmstat.node0.numa_other
2658 ± 36% +69.3% 4500 ± 36% numa-vmstat.node1.nr_shmem
935792 ±136% +4247.3% 40682138 ±141% latency_stats.avg.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
935792 ±136% +4247.3% 40682138 ±141% latency_stats.max.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
935792 ±136% +4247.3% 40682138 ±141% latency_stats.sum.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
12893 ± 2% -9.1% 11716 ± 1% slabinfo.kmalloc-192.active_objs
1653 ± 9% -10.3% 1483 ± 5% slabinfo.mnt_cache.active_objs
1653 ± 9% -10.3% 1483 ± 5% slabinfo.mnt_cache.num_objs
1.75 ± 47% -81.0% 0.33 ±141% sched_debug.cfs_rq[10]:/.nr_spread_over
-343206 ±-27% -73.2% -91995 ±-170% sched_debug.cfs_rq[14]:/.spread0
533.25 ± 82% -81.5% 98.75 ± 42% sched_debug.cfs_rq[18]:/.blocked_load_avg
541.75 ± 82% -81.3% 101.25 ± 41% sched_debug.cfs_rq[18]:/.tg_load_contrib
-1217705 ± -5% -30.2% -850080 ±-15% sched_debug.cfs_rq[26]:/.spread0
89722 ± 9% +9.8% 98495 ± 10% sched_debug.cfs_rq[32]:/.exec_clock
101180 ±132% +180.8% 284154 ± 30% sched_debug.cfs_rq[35]:/.spread0
37332 ±473% +725.2% 308082 ± 59% sched_debug.cfs_rq[38]:/.spread0
32054 ±502% +981.6% 346689 ± 39% sched_debug.cfs_rq[39]:/.spread0
1.00 ±100% +100.0% 2.00 ± 50% sched_debug.cfs_rq[42]:/.nr_spread_over
-125980 ±-218% -307.1% 260875 ± 46% sched_debug.cfs_rq[42]:/.spread0
-111501 ±-102% -288.7% 210354 ± 94% sched_debug.cfs_rq[45]:/.spread0
-173363 ±-34% -221.0% 209775 ± 94% sched_debug.cfs_rq[47]:/.spread0
-302090 ±-43% -121.8% 65953 ±322% sched_debug.cfs_rq[4]:/.spread0
-490175 ±-18% -41.1% -288722 ±-31% sched_debug.cfs_rq[50]:/.spread0
-594948 ±-10% -59.7% -239840 ±-33% sched_debug.cfs_rq[51]:/.spread0
1.00 ±100% +6050.0% 61.50 ±141% sched_debug.cfs_rq[53]:/.blocked_load_avg
10.50 ± 8% +614.3% 75.00 ±122% sched_debug.cfs_rq[53]:/.tg_load_contrib
-596043 ±-10% -49.0% -304277 ±-36% sched_debug.cfs_rq[54]:/.spread0
10.00 ± 0% +2062.5% 216.25 ± 40% sched_debug.cfs_rq[56]:/.tg_load_contrib
17.75 ±173% +1302.8% 249.00 ± 26% sched_debug.cfs_rq[60]:/.blocked_load_avg
-809633 ± -9% -36.2% -516886 ±-23% sched_debug.cfs_rq[60]:/.spread0
28.00 ±109% +828.6% 260.00 ± 25% sched_debug.cfs_rq[60]:/.tg_load_contrib
277.75 ± 95% -86.3% 38.00 ±171% sched_debug.cfs_rq[7]:/.blocked_load_avg
293.25 ± 90% -81.8% 53.50 ±121% sched_debug.cfs_rq[7]:/.tg_load_contrib
17.50 ± 2% -28.6% 12.50 ± 34% sched_debug.cpu#0.cpu_load[2]
17.00 ± 4% -25.0% 12.75 ± 35% sched_debug.cpu#0.cpu_load[3]
2907 ± 12% +195.9% 8603 ± 63% sched_debug.cpu#0.sched_goidle
16.50 ± 3% -9.1% 15.00 ± 0% sched_debug.cpu#1.cpu_load[2]
16.50 ± 3% -7.6% 15.25 ± 2% sched_debug.cpu#1.cpu_load[3]
5595 ± 26% -36.4% 3557 ± 11% sched_debug.cpu#11.nr_switches
6885 ± 92% -76.2% 1639 ± 40% sched_debug.cpu#11.ttwu_count
1350 ± 34% -55.0% 608.00 ± 14% sched_debug.cpu#11.ttwu_local
17892 ± 74% -78.3% 3877 ± 18% sched_debug.cpu#12.nr_switches
1288 ± 27% -49.8% 647.50 ± 37% sched_debug.cpu#12.ttwu_local
1405 ± 22% -52.7% 664.50 ± 23% sched_debug.cpu#13.ttwu_local
1.25 ±182% -440.0% -4.25 ±-50% sched_debug.cpu#17.nr_uninterruptible
1976 ± 5% -10.0% 1779 ± 0% sched_debug.cpu#18.curr->pid
983.75 ± 8% +101.6% 1983 ± 32% sched_debug.cpu#18.ttwu_local
-0.25 ±-911% +2300.0% -6.00 ±-28% sched_debug.cpu#21.nr_uninterruptible
2979 ± 49% +159.6% 7734 ± 75% sched_debug.cpu#22.ttwu_count
1111 ± 21% +127.6% 2528 ± 32% sched_debug.cpu#22.ttwu_local
1.00 ±141% -275.0% -1.75 ±-84% sched_debug.cpu#25.nr_uninterruptible
14419 ± 54% -58.2% 6022 ± 84% sched_debug.cpu#25.ttwu_count
14395 ± 70% +252.4% 50729 ± 39% sched_debug.cpu#28.nr_switches
-4.75 ±-17% -115.8% 0.75 ±218% sched_debug.cpu#30.nr_uninterruptible
2335 ±115% -76.6% 547.25 ± 18% sched_debug.cpu#34.ttwu_count
1258 ± 25% -43.3% 713.75 ± 11% sched_debug.cpu#35.nr_switches
1409 ± 23% -39.6% 851.75 ± 9% sched_debug.cpu#35.sched_count
969.50 ± 69% -68.8% 302.00 ± 38% sched_debug.cpu#35.ttwu_count
382.00 ± 37% -66.0% 130.00 ± 14% sched_debug.cpu#35.ttwu_local
808.75 ± 18% +28.3% 1037 ± 15% sched_debug.cpu#38.nr_switches
948.50 ± 16% +23.2% 1168 ± 13% sched_debug.cpu#38.sched_count
70695 ± 2% +6.2% 75047 ± 4% sched_debug.cpu#41.nr_load_updates
1269 ± 13% +55.3% 1970 ± 25% sched_debug.cpu#46.nr_switches
3.25 ± 93% -76.9% 0.75 ±197% sched_debug.cpu#46.nr_uninterruptible
1375 ± 12% +51.1% 2078 ± 23% sched_debug.cpu#46.sched_count
3958 ± 97% +462.9% 22281 ± 25% sched_debug.cpu#50.ttwu_count
457.25 ± 26% +64.3% 751.25 ± 28% sched_debug.cpu#53.ttwu_local
753041 ± 3% -11.1% 669815 ± 5% sched_debug.cpu#58.avg_idle
-1.75 ±-142% -257.1% 2.75 ± 64% sched_debug.cpu#59.nr_uninterruptible
2581 ± 27% +1426.4% 39408 ± 57% sched_debug.cpu#60.nr_switches
2632 ± 27% +1400.2% 39495 ± 57% sched_debug.cpu#60.sched_count
34156 ± 94% -94.8% 1776 ± 15% sched_debug.cpu#61.nr_switches
34250 ± 94% -94.7% 1825 ± 15% sched_debug.cpu#61.sched_count
16821 ± 96% -95.4% 768.50 ± 11% sched_debug.cpu#61.sched_goidle
8128 ±146% -91.7% 676.00 ± 10% sched_debug.cpu#61.ttwu_count

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test:
ivb42/will-it-scale/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/readseek1

commit:
447f6a95a9c80da7faaec3e66e656eab8f262640
8f2f3eb59dff4ec538de55f2e0592fec85966aab

447f6a95a9c80da7 8f2f3eb59dff4ec538de55f2e0
---------------- --------------------------
%stddev %change %stddev
\ | \
1915464 ± 0% -2.4% 1869344 ± 0% will-it-scale.per_thread_ops
473.17 ± 0% +6.9% 505.66 ± 0% will-it-scale.time.user_time
0.20 ± 5% -49.4% 0.10 ± 35% turbostat.Pkg%pc6
3.38 ± 0% +34.0% 4.53 ± 1% perf-profile.cpu-cycles.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_file_read_iter.__vfs_read
7.42 ± 0% +16.3% 8.62 ± 1% perf-profile.cpu-cycles.find_lock_entry.shmem_getpage_gfp.shmem_file_read_iter.__vfs_read.vfs_read
0.57 ± 6% +72.2% 0.99 ± 6% perf-profile.cpu-cycles.radix_tree_lookup_slot.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_file_read_iter
10.58 ± 0% +11.4% 11.79 ± 1% perf-profile.cpu-cycles.shmem_getpage_gfp.shmem_file_read_iter.__vfs_read.vfs_read.sys_read
30.50 ±150% +1140.2% 378.25 ± 49% sched_debug.cfs_rq[22]:/.blocked_load_avg
44.75 ±103% +788.3% 397.50 ± 46% sched_debug.cfs_rq[22]:/.tg_load_contrib
89.50 ±159% +300.3% 358.25 ± 75% sched_debug.cfs_rq[2]:/.blocked_load_avg
115.75 ±123% +231.7% 384.00 ± 70% sched_debug.cfs_rq[2]:/.tg_load_contrib
0.50 ±100% +750.0% 4.25 ± 67% sched_debug.cfs_rq[32]:/.nr_spread_over
499.50 ± 44% -98.2% 9.00 ±101% sched_debug.cfs_rq[40]:/.blocked_load_avg
505.50 ± 44% -95.2% 24.50 ± 73% sched_debug.cfs_rq[40]:/.tg_load_contrib
421.00 ± 56% -85.7% 60.25 ±109% sched_debug.cfs_rq[42]:/.blocked_load_avg
428.75 ± 56% -80.4% 84.00 ± 86% sched_debug.cfs_rq[42]:/.tg_load_contrib
8053 ± 2% +13.4% 9132 ± 5% sched_debug.cfs_rq[47]:/.avg->runnable_avg_sum
175.25 ± 2% +12.7% 197.50 ± 5% sched_debug.cfs_rq[47]:/.tg_runnable_contrib
0.25 ±173% +1500.0% 4.00 ± 77% sched_debug.cfs_rq[8]:/.nr_spread_over
90.75 ± 13% -23.1% 69.75 ± 15% sched_debug.cpu#0.cpu_load[2]
97.00 ± 15% -28.4% 69.50 ± 16% sched_debug.cpu#0.cpu_load[3]
99.50 ± 14% -27.6% 72.00 ± 18% sched_debug.cpu#0.cpu_load[4]
-10.25 ±-14% -73.2% -2.75 ±-180% sched_debug.cpu#1.nr_uninterruptible
8173 ±106% -78.9% 1722 ± 35% sched_debug.cpu#10.nr_switches
3896 ±112% -81.3% 727.50 ± 36% sched_debug.cpu#10.sched_goidle
515.00 ± 40% -47.2% 271.75 ± 49% sched_debug.cpu#10.ttwu_local
2.00 ± 81% -325.0% -4.50 ±-77% sched_debug.cpu#11.nr_uninterruptible
3818 ± 39% -58.2% 1598 ± 68% sched_debug.cpu#15.ttwu_local
0.50 ±331% -650.0% -2.75 ±-74% sched_debug.cpu#16.nr_uninterruptible
12671 ± 30% -58.4% 5270 ± 46% sched_debug.cpu#20.ttwu_count
2285 ± 70% -57.0% 983.50 ± 25% sched_debug.cpu#20.ttwu_local
2722 ± 79% -72.9% 738.75 ± 51% sched_debug.cpu#21.ttwu_local
-2.50 ±-72% -200.0% 2.50 ± 82% sched_debug.cpu#23.nr_uninterruptible
1183 ± 31% +188.4% 3413 ± 22% sched_debug.cpu#24.nr_switches
1384 ± 45% +148.4% 3438 ± 22% sched_debug.cpu#24.sched_count
318.50 ± 54% +347.5% 1425 ± 21% sched_debug.cpu#24.ttwu_local
5255 ± 46% -60.2% 2090 ± 54% sched_debug.cpu#25.nr_switches
5276 ± 46% -59.9% 2114 ± 54% sched_debug.cpu#25.sched_count
1893 ± 42% -66.9% 627.00 ± 75% sched_debug.cpu#25.ttwu_local
1.25 ±142% +240.0% 4.25 ± 45% sched_debug.cpu#27.nr_uninterruptible
0.75 ±272% -322.2% -1.67 ±-28% sched_debug.cpu#31.nr_uninterruptible
1977 ±140% -86.5% 267.25 ± 10% sched_debug.cpu#32.sched_goidle
7.67 ± 78% -122.8% -1.75 ±-84% sched_debug.cpu#34.nr_uninterruptible
3642 ± 37% +205.0% 11108 ± 53% sched_debug.cpu#39.nr_switches
1250 ± 51% +292.0% 4902 ± 52% sched_debug.cpu#39.sched_goidle
3.00 ± 0% +216.7% 9.50 ± 30% sched_debug.cpu#45.cpu_load[0]
3.50 ± 24% +121.4% 7.75 ± 10% sched_debug.cpu#45.cpu_load[1]
3.25 ± 13% +123.1% 7.25 ± 11% sched_debug.cpu#45.cpu_load[2]
3.25 ± 13% +92.3% 6.25 ± 23% sched_debug.cpu#45.cpu_load[3]
3.00 ± 0% +91.7% 5.75 ± 22% sched_debug.cpu#45.cpu_load[4]
1593 ± 19% +63.6% 2605 ± 30% sched_debug.cpu#47.curr->pid
365.75 ± 39% +254.6% 1297 ± 98% sched_debug.cpu#6.ttwu_local
8717 ± 80% -78.7% 1856 ± 45% sched_debug.cpu#8.nr_switches
3992 ± 85% -80.5% 778.50 ± 51% sched_debug.cpu#8.sched_goidle
6221 ±128% -83.9% 998.75 ± 44% sched_debug.cpu#8.ttwu_count
722.00 ± 71% -69.5% 220.25 ± 5% sched_debug.cpu#8.ttwu_local
0.25 ±173% +321.4% 1.05 ± 5% sched_debug.rt_rq[12]:/.rt_time
0.04 ±173% +311.0% 0.17 ± 8% sched_debug.rt_rq[13]:/.rt_time


lkp-sbx04: Sandy Bridge-EX
Memory: 64G

ivb42: Ivytown Ivy Bridge-EP
Memory: 64G




will-it-scale.time.user_time

325 ++--------------------------------------------------------------------+
320 ++ O O O |
| O O O O O O O O |
315 O+ O O |
310 ++ O O O O
| O |
305 ++ |
300 ++ |
295 ++ |
| *.. |
290 ++ .*.. ..*...*.. .. . |
285 *+..*... .. . .*...*. . . *...*...*..*...*... ..*.. |
| * *. * *. .|
280 ++ *
275 ++--------------------------------------------------------------------+


[*] bisect-good sample
[O] bisect-bad sample

To reproduce:

git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Ying Huang
---
LKP_SERVER: inn
LKP_CGI_PORT: 80
LKP_CIFS_PORT: 139
testcase: will-it-scale
default-monitors:
wait: activate-monitor
kmsg:
uptime:
iostat:
vmstat:
numa-numastat:
numa-vmstat:
numa-meminfo:
proc-vmstat:
proc-stat:
interval: 10
meminfo:
slabinfo:
interrupts:
lock_stat:
latency_stats:
softirqs:
bdi_dev_mapping:
diskstats:
nfsstat:
cpuidle:
cpufreq-stats:
turbostat:
pmeter:
sched_debug:
interval: 60
cpufreq_governor: performance
default-watchdogs:
oom-killer:
watchdog:
commit: dd2384a75d1c046faf068a6352732a204814b86d
model: Sandy Bridge-EX
nr_cpu: 64
memory: 64G
nr_ssd_partitions: 4
ssd_partitions: "/dev/disk/by-id/ata-INTEL_SSDSC2CW240A3_CVCV20430*-part1"
swap_partitions:
category: benchmark
perf-profile:
freq: 800
will-it-scale:
test: read1
queue: cyclic
testbox: lkp-sbx04
tbox_group: lkp-sbx04
kconfig: x86_64-rhel
enqueue_time: 2015-08-08 06:51:04.467682345 +08:00
id: 7543484f1eea88c654299222e83a89fb3f8fbd44
user: lkp
compiler: gcc-4.9
head_commit: dd2384a75d1c046faf068a6352732a204814b86d
base_commit: 733db573a6451681b60e7372d2862de09d6eb04e
branch: linus/master
kernel: "/pkg/linux/x86_64-rhel/gcc-4.9/dd2384a75d1c046faf068a6352732a204814b86d/vmlinuz-4.2.0-rc5-00156-gdd2384a"
rootfs: debian-x86_64-2015-02-07.cgz
result_root: "/result/will-it-scale/performance-read1/lkp-sbx04/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/dd2384a75d1c046faf068a6352732a204814b86d/0"
job_file: "/lkp/scheduled/lkp-sbx04/cyclic_will-it-scale-performance-read1-x86_64-rhel-CYCLIC_HEAD-dd2384a75d1c046faf068a6352732a204814b86d-20150808-9771-1mploli-0.yaml"
dequeue_time: 2015-08-08 18:18:22.936002643 +08:00
max_uptime: 1500
initrd: "/osimage/debian/debian-x86_64-2015-02-07.cgz"
bootloader_append:
- root=/dev/ram0
- user=lkp
- job=/lkp/scheduled/lkp-sbx04/cyclic_will-it-scale-performance-read1-x86_64-rhel-CYCLIC_HEAD-dd2384a75d1c046faf068a6352732a204814b86d-20150808-9771-1mploli-0.yaml
- ARCH=x86_64
- kconfig=x86_64-rhel
- branch=linus/master
- commit=dd2384a75d1c046faf068a6352732a204814b86d
- BOOT_IMAGE=/pkg/linux/x86_64-rhel/gcc-4.9/dd2384a75d1c046faf068a6352732a204814b86d/vmlinuz-4.2.0-rc5-00156-gdd2384a
- max_uptime=1500
- RESULT_ROOT=/result/will-it-scale/performance-read1/lkp-sbx04/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/dd2384a75d1c046faf068a6352732a204814b86d/0
- LKP_SERVER=inn
- |2-


earlyprintk=ttyS0,115200 systemd.log_level=err
debug apic=debug sysrq_always_enabled rcupdate.rcu_cpu_stall_timeout=100
panic=-1 softlockup_panic=1 nmi_watchdog=panic oops=panic load_ramdisk=2 prompt_ramdisk=0
console=ttyS0,115200 console=tty0 vga=normal

rw
lkp_initrd: "/lkp/lkp/lkp-x86_64.cgz"
modules_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/dd2384a75d1c046faf068a6352732a204814b86d/modules.cgz"
bm_initrd: "/osimage/deps/debian-x86_64-2015-02-07.cgz/lkp.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/run-ipconfig.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/turbostat.cgz,/lkp/benchmarks/turbostat.cgz,/lkp/benchmarks/will-it-scale.cgz"
job_state: finished
loadavg: 47.44 21.45 8.37 1/628 11393
start_time: '1439029153'
end_time: '1439029462'
version: "/lkp/lkp/.src-20150807-183152"
echo performance > /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu1/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu10/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu11/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu12/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu13/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu14/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu15/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu16/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu17/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu18/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu19/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu2/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu20/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu21/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu22/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu23/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu24/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu25/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu26/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu27/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu28/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu29/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu3/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu30/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu31/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu32/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu33/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu34/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu35/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu36/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu37/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu38/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu39/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu4/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu40/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu41/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu42/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu43/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu44/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu45/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu46/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu47/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu48/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu49/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu5/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu50/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu51/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu52/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu53/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu54/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu55/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu56/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu57/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu58/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu59/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu6/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu60/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu61/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu62/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu63/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu7/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu8/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu9/cpufreq/scaling_governor
./runtest.py read1 16 both 1 8 16 24 32 48 64