FYI, we noticed the below changes on git://git.kernel.org/pub/scm/linux/kernel/git/luto/linux.git x86/entry commit 62c79204783e188291d880f23d49c02d8c8f498b ("x86/entry/64: When returning via SYSRET, POP regs instead of using MOV")
========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/load/test: lkp-a06/aim7/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/4000/misc_rtns_1 commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 108044 ± 0% +5.9% 114416 ± 0% aim7.jobs-per-min 223.70 ± 0% -5.6% 211.11 ± 0% aim7.time.elapsed_time 223.70 ± 0% -5.6% 211.11 ± 0% aim7.time.elapsed_time.max 2113772 ± 0% -83.5% 349283 ± 2% aim7.time.involuntary_context_switches 599.10 ± 0% -2.2% 585.90 ± 0% aim7.time.system_time 203.87 ± 1% -4.7% 194.25 ± 0% aim7.time.user_time 2113772 ± 0% -83.5% 349283 ± 2% time.involuntary_context_switches ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/load/test: lkp-a06/aim7/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/4000/new_raph commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 184357 ± 0% +10.8% 204299 ± 0% aim7.jobs-per-min 131.71 ± 0% -9.8% 118.82 ± 0% aim7.time.elapsed_time 131.71 ± 0% -9.8% 118.82 ± 0% aim7.time.elapsed_time.max 2193692 ± 0% -92.5% 164787 ± 0% aim7.time.involuntary_context_switches 18.86 ± 2% -68.5% 5.94 ± 4% aim7.time.system_time 435.63 ± 0% -2.7% 424.04 ± 0% aim7.time.user_time 58879 ± 2% -19.5% 47402 ± 2% aim7.time.voluntary_context_switches 131.71 ± 0% -9.8% 118.82 ± 0% time.elapsed_time 131.71 ± 0% -9.8% 118.82 ± 0% time.elapsed_time.max 2193692 ± 0% -92.5% 164787 ± 0% time.involuntary_context_switches 18.86 ± 2% -68.5% 5.94 ± 4% time.system_time 58879 ± 2% -19.5% 47402 ± 2% time.voluntary_context_switches ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/load/test: lkp-a06/aim7/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/4000/pipe_cpy commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 240551 ± 0% +13.9% 273904 ± 2% aim7.jobs-per-min 101.42 ± 0% -12.2% 89.04 ± 2% aim7.time.elapsed_time 101.42 ± 0% -12.2% 89.04 ± 2% aim7.time.elapsed_time.max 1981819 ± 0% -93.0% 138430 ± 3% aim7.time.involuntary_context_switches 278.03 ± 0% -6.5% 259.97 ± 2% aim7.time.system_time 54936 ± 1% -19.6% 44147 ± 2% aim7.time.voluntary_context_switches 101.42 ± 0% -12.2% 89.04 ± 2% time.elapsed_time 101.42 ± 0% -12.2% 89.04 ± 2% time.elapsed_time.max 1981819 ± 0% -93.0% 138430 ± 3% time.involuntary_context_switches 54936 ± 1% -19.6% 44147 ± 2% time.voluntary_context_switches ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/iterations/nr_threads/disk/fs/filesize/test_size/sync_method/nr_directories/nr_files_per_directory: lkp-sb02/fsmark/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/1x/32t/1HDD/btrfs/9B/400M/fsyncBeforeClose/16d/256fpd commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 2554844 ± 1% -13.1% 2221395 ± 1% fsmark.app_overhead 223789 ± 0% -57.2% 95777 ± 6% fsmark.time.involuntary_context_switches 32.50 ± 1% +5.4% 34.25 ± 1% fsmark.time.percent_of_cpu_this_job_got 2118350 ± 0% +5.2% 2229178 ± 0% fsmark.time.voluntary_context_switches 223789 ± 0% -57.2% 95777 ± 6% time.involuntary_context_switches 39831 ± 0% -9.9% 35903 ± 1% softirqs.SCHED 59651 ± 1% -8.7% 54478 ± 0% softirqs.TIMER 64594 ± 0% -69.9% 19433 ± 0% vmstat.system.cs 25161 ± 0% -94.2% 1465 ± 3% vmstat.system.in 4366124 ± 0% -90.0% 434486 ± 4% cpuidle.C1-SNB.usage 17993557 ± 4% +17.6% 21163962 ± 3% cpuidle.C1E-SNB.time 42154 ± 3% +15.4% 48646 ± 6% cpuidle.C1E-SNB.usage 9261281 ± 5% +28.4% 11895838 ± 4% cpuidle.C3-SNB.time 6639 ± 2% +14.5% 7601 ± 3% cpuidle.C3-SNB.usage 9194 ± 2% -97.7% 208.25 ± 6% cpuidle.POLL.usage 11.44 ± 0% -20.0% 9.14 ± 0% turbostat.%Busy 333.25 ± 0% -21.0% 263.25 ± 1% turbostat.Avg_MHz 1.32 ± 5% +28.4% 1.70 ± 4% turbostat.CPU%c3 11.04 ± 1% -13.0% 9.61 ± 1% turbostat.CorWatt 1.47 ± 3% +26.7% 1.86 ± 4% turbostat.Pkg%pc2 1.19 ± 8% +58.6% 1.89 ± 5% turbostat.Pkg%pc3 19.86 ± 1% +35.2% 26.86 ± 1% turbostat.Pkg%pc6 14.61 ± 0% -9.9% 13.16 ± 1% turbostat.PkgWatt 2143346 ± 57% -100.0% 0.00 ± -1% latency_stats.avg.btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 30524 ± 3% +20.3% 36728 ± 1% latency_stats.hits.btrfs_tree_lock.[btrfs].btrfs_lock_root_node.[btrfs].btrfs_search_slot.[btrfs].btrfs_insert_empty_items.[btrfs].btrfs_insert_delayed_items.[btrfs].btrfs_commit_inode_delayed_items.[btrfs].btrfs_log_inode.[btrfs].btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync 18837 ± 4% +56.0% 29390 ± 7% latency_stats.hits.btrfs_tree_lock.[btrfs].btrfs_lock_root_node.[btrfs].btrfs_search_slot.[btrfs].btrfs_insert_empty_items.[btrfs].copy_items.[btrfs].btrfs_log_inode.[btrfs].btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync 12096 ± 6% +46.5% 17718 ± 4% latency_stats.hits.btrfs_tree_lock.[btrfs].btrfs_lock_root_node.[btrfs].btrfs_search_slot.[btrfs].drop_objectid_items.[btrfs].btrfs_log_inode.[btrfs].btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 62486 ± 1% +19.6% 74731 ± 4% latency_stats.hits.btrfs_tree_read_lock.[btrfs].btrfs_read_lock_root_node.[btrfs].btrfs_search_slot.[btrfs].btrfs_insert_empty_items.[btrfs].copy_items.[btrfs].btrfs_log_inode.[btrfs].btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync 52038 ± 2% +16.7% 60742 ± 3% latency_stats.hits.btrfs_tree_read_lock.[btrfs].btrfs_read_lock_root_node.[btrfs].btrfs_search_slot.[btrfs].drop_objectid_items.[btrfs].btrfs_log_inode.[btrfs].btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 3558104 ± 57% -100.0% 0.00 ± -1% latency_stats.max.btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 4.111e+08 ± 57% -100.0% 0.00 ± -1% latency_stats.sum.btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 1403516 ± 2% +31.1% 1840040 ± 2% latency_stats.sum.btrfs_tree_lock.[btrfs].btrfs_lock_root_node.[btrfs].btrfs_search_slot.[btrfs].btrfs_insert_empty_items.[btrfs].btrfs_insert_delayed_items.[btrfs].btrfs_commit_inode_delayed_items.[btrfs].btrfs_log_inode.[btrfs].btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync 3349730 ± 2% +19.6% 4005849 ± 2% latency_stats.sum.btrfs_tree_lock.[btrfs].btrfs_lock_root_node.[btrfs].btrfs_search_slot.[btrfs].btrfs_insert_empty_items.[btrfs].btrfs_new_inode.[btrfs].btrfs_create.[btrfs].vfs_create.path_openat.do_filp_open.do_sys_open.SyS_open.entry_SYSCALL_64_fastpath 178060 ± 6% +74.3% 310377 ± 11% latency_stats.sum.btrfs_tree_lock.[btrfs].btrfs_lock_root_node.[btrfs].btrfs_search_slot.[btrfs].btrfs_insert_empty_items.[btrfs].copy_items.[btrfs].btrfs_log_inode.[btrfs].btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync 1322666 ± 4% +27.8% 1690982 ± 1% latency_stats.sum.btrfs_tree_lock.[btrfs].btrfs_lock_root_node.[btrfs].btrfs_search_slot.[btrfs].btrfs_insert_empty_items.[btrfs].insert_with_overflow.[btrfs].btrfs_insert_dir_item.[btrfs].btrfs_add_link.[btrfs].btrfs_create.[btrfs].vfs_create.path_openat.do_filp_open.do_sys_open 42794 ± 5% +63.1% 69810 ± 11% latency_stats.sum.btrfs_tree_lock.[btrfs].btrfs_lock_root_node.[btrfs].btrfs_search_slot.[btrfs].btrfs_truncate_inode_items.[btrfs].btrfs_log_inode.[btrfs].btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 122546 ± 7% +60.4% 196523 ± 7% latency_stats.sum.btrfs_tree_lock.[btrfs].btrfs_lock_root_node.[btrfs].btrfs_search_slot.[btrfs].drop_objectid_items.[btrfs].btrfs_log_inode.[btrfs].btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 24192 ± 3% +29.8% 31401 ± 6% latency_stats.sum.btrfs_tree_lock.[btrfs].btrfs_search_slot.[btrfs].btrfs_insert_empty_items.[btrfs].copy_items.[btrfs].btrfs_log_inode.[btrfs].btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 697815 ± 1% +11.1% 775306 ± 2% latency_stats.sum.btrfs_tree_lock.[btrfs].btrfs_search_slot.[btrfs].btrfs_lookup_inode.[btrfs].__btrfs_update_delayed_inode.[btrfs].btrfs_commit_inode_delayed_items.[btrfs].btrfs_log_inode.[btrfs].btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync 566092 ± 2% +29.3% 732008 ± 7% latency_stats.sum.btrfs_tree_read_lock.[btrfs].btrfs_read_lock_root_node.[btrfs].btrfs_search_slot.[btrfs].btrfs_insert_empty_items.[btrfs].copy_items.[btrfs].btrfs_log_inode.[btrfs].btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync 283356 ± 4% +33.1% 377256 ± 7% latency_stats.sum.btrfs_tree_read_lock.[btrfs].btrfs_read_lock_root_node.[btrfs].btrfs_search_slot.[btrfs].btrfs_truncate_inode_items.[btrfs].btrfs_log_inode.[btrfs].btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 525841 ± 2% +24.2% 653227 ± 4% latency_stats.sum.btrfs_tree_read_lock.[btrfs].btrfs_read_lock_root_node.[btrfs].btrfs_search_slot.[btrfs].drop_objectid_items.[btrfs].btrfs_log_inode.[btrfs].btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 759116 ± 2% -9.4% 687617 ± 1% latency_stats.sum.btrfs_tree_read_lock.[btrfs].btrfs_search_slot.[btrfs].btrfs_lookup_dir_item.[btrfs].btrfs_lookup_dentry.[btrfs].btrfs_lookup.[btrfs].lookup_real.path_openat.do_filp_open.do_sys_open.SyS_open.entry_SYSCALL_64_fastpath 12244 ± 8% -11.0% 10892 ± 3% sched_debug.cfs_rq[0]:/.avg->runnable_avg_sum 8078 ± 6% +20.1% 9700 ± 5% sched_debug.cfs_rq[0]:/.exec_clock 12997 ± 8% +25.1% 16258 ± 2% sched_debug.cfs_rq[0]:/.min_vruntime 269.00 ± 8% -11.3% 238.50 ± 3% sched_debug.cfs_rq[0]:/.tg_runnable_contrib 553.11 ±128% -403.8% -1680 ±-72% sched_debug.cfs_rq[2]:/.spread0 28.75 ± 35% -43.5% 16.25 ± 42% sched_debug.cfs_rq[3]:/.nr_spread_over 42568 ± 29% -52.2% 20355 ± 4% sched_debug.cpu#0.nr_load_updates 4438150 ± 51% -92.6% 328982 ± 4% sched_debug.cpu#0.nr_switches 4438484 ± 51% -92.6% 329332 ± 4% sched_debug.cpu#0.sched_count 2060550 ± 53% -96.5% 72130 ± 7% sched_debug.cpu#0.sched_goidle 2323810 ± 48% -87.5% 290944 ± 5% sched_debug.cpu#0.ttwu_count 2136144 ± 53% -97.7% 48220 ± 2% sched_debug.cpu#0.ttwu_local 1705455 ±133% -80.1% 339746 ± 5% sched_debug.cpu#1.nr_switches 1705520 ±133% -80.1% 339819 ± 5% sched_debug.cpu#1.sched_count 739354 ±148% -89.5% 77619 ± 11% sched_debug.cpu#1.sched_goidle 926939 ±122% -74.2% 238908 ± 6% sched_debug.cpu#1.ttwu_count 742260 ±155% -95.8% 31432 ± 1% sched_debug.cpu#1.ttwu_local 968.25 ± 11% -89.3% 104.00 ±102% sched_debug.cpu#2.nr_uninterruptible 63733 ± 9% +21.4% 77348 ± 12% sched_debug.cpu#2.sched_goidle 33525 ± 3% -13.7% 28932 ± 1% sched_debug.cpu#2.ttwu_local 1264 ± 17% -101.5% -18.50 ±-378% sched_debug.cpu#3.nr_uninterruptible ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/iterations/nr_threads/disk/fs/filesize/test_size/sync_method/nr_directories/nr_files_per_directory: nhm4/fsmark/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/1x/32t/1HDD/btrfs/9B/400M/fsyncBeforeClose/16d/256fpd commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 3294133 ± 2% -6.2% 3088824 ± 2% fsmark.app_overhead 463365 ± 1% -51.7% 223905 ± 7% fsmark.time.involuntary_context_switches 140.50 ± 1% +8.2% 152.00 ± 1% fsmark.time.percent_of_cpu_this_job_got 213.09 ± 1% +7.2% 228.37 ± 1% fsmark.time.system_time 4278018 ± 1% +3.6% 4432123 ± 1% fsmark.time.voluntary_context_switches 463365 ± 1% -51.7% 223905 ± 7% time.involuntary_context_switches 7.75 ± 5% +18.4% 9.17 ± 0% turbostat.CPU%c6 5214507 ± 0% -70.1% 1561193 ± 2% cpuidle.C1-NHM.usage 23195 ± 5% -97.2% 641.50 ± 5% cpuidle.POLL.usage 96711 ± 1% -51.9% 46555 ± 0% vmstat.system.cs 30013 ± 1% -87.8% 3649 ± 3% vmstat.system.in 5154 ± 4% -10.8% 4599 ± 3% slabinfo.btrfs_extent_buffer.active_objs 5154 ± 4% -10.8% 4599 ± 3% slabinfo.btrfs_extent_buffer.num_objs 1674 ± 7% -13.2% 1453 ± 5% slabinfo.buffer_head.active_objs 1680 ± 7% -13.1% 1459 ± 5% slabinfo.buffer_head.num_objs 1145192 ± 70% -100.0% 0.00 ± -1% latency_stats.avg.btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 137982 ± 2% +73.7% 239720 ± 75% latency_stats.hits.btrfs_tree_lock.[btrfs].btrfs_search_slot.[btrfs].btrfs_insert_empty_items.[btrfs].insert_with_overflow.[btrfs].btrfs_insert_dir_item.[btrfs].btrfs_add_link.[btrfs].btrfs_create.[btrfs].vfs_create.path_openat.do_filp_open.do_sys_open.SyS_open 1540569 ± 70% -100.0% 0.00 ± -1% latency_stats.max.btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 1.741e+08 ± 70% -100.0% 0.00 ± -1% latency_stats.sum.btrfs_log_inode_parent.[btrfs].btrfs_log_dentry_safe.[btrfs].btrfs_sync_file.[btrfs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 2427124 ± 2% +60.9% 3906177 ± 78% latency_stats.sum.btrfs_tree_lock.[btrfs].btrfs_search_slot.[btrfs].btrfs_insert_empty_items.[btrfs].insert_with_overflow.[btrfs].btrfs_insert_dir_item.[btrfs].btrfs_add_link.[btrfs].btrfs_create.[btrfs].vfs_create.path_openat.do_filp_open.do_sys_open.SyS_open 47047 ± 9% +15.6% 54373 ± 2% sched_debug.cfs_rq[0]:/.min_vruntime 23933 ± 18% +24.7% 29834 ± 2% sched_debug.cfs_rq[0]:/.tg_load_avg 788.25 ± 74% +312.2% 3249 ± 37% sched_debug.cfs_rq[1]:/.blocked_load_avg -297.49 ±-1471% +989.4% -3240 ±-11% sched_debug.cfs_rq[1]:/.spread0 23934 ± 18% +24.7% 29840 ± 2% sched_debug.cfs_rq[1]:/.tg_load_avg 815.00 ± 73% +303.0% 3284 ± 37% sched_debug.cfs_rq[1]:/.tg_load_contrib 42646 ± 8% +20.4% 51331 ± 2% sched_debug.cfs_rq[2]:/.min_vruntime 23848 ± 17% +23.9% 29555 ± 4% sched_debug.cfs_rq[2]:/.tg_load_avg 4349 ± 24% -41.6% 2541 ± 40% sched_debug.cfs_rq[3]:/.blocked_load_avg 119.25 ± 42% -86.6% 16.00 ±141% sched_debug.cfs_rq[3]:/.load 133.50 ± 32% -93.0% 9.33 ±141% sched_debug.cfs_rq[3]:/.runnable_load_avg 23820 ± 17% +23.6% 29446 ± 3% sched_debug.cfs_rq[3]:/.tg_load_avg 4486 ± 24% -41.7% 2617 ± 42% sched_debug.cfs_rq[3]:/.tg_load_contrib 23822 ± 17% +23.5% 29431 ± 3% sched_debug.cfs_rq[4]:/.tg_load_avg 43836 ± 2% +10.4% 48383 ± 2% sched_debug.cfs_rq[5]:/.min_vruntime 23750 ± 17% +23.9% 29419 ± 3% sched_debug.cfs_rq[5]:/.tg_load_avg 41692 ± 4% +15.9% 48332 ± 3% sched_debug.cfs_rq[6]:/.min_vruntime 23751 ± 17% +23.8% 29398 ± 3% sched_debug.cfs_rq[6]:/.tg_load_avg 42923 ± 6% +13.2% 48585 ± 3% sched_debug.cfs_rq[7]:/.min_vruntime 23749 ± 17% +23.7% 29373 ± 3% sched_debug.cfs_rq[7]:/.tg_load_avg 253.00 ± 69% -180.0% -202.50 ±-55% sched_debug.cpu#1.nr_uninterruptible 150.00 ± 43% -81.6% 27.67 ±141% sched_debug.cpu#3.cpu_load[0] 88.50 ± 32% -80.8% 17.00 ±136% sched_debug.cpu#3.cpu_load[1] 59.50 ± 30% -70.2% 17.75 ± 92% sched_debug.cpu#3.cpu_load[2] 46.50 ± 34% -53.2% 21.75 ± 55% sched_debug.cpu#3.cpu_load[3] 1208 ± 19% -91.8% 99.33 ±141% sched_debug.cpu#3.curr->pid 127.50 ± 49% -87.5% 16.00 ±141% sched_debug.cpu#3.load 1447 ± 34% -50.6% 715.00 ± 21% sched_debug.cpu#4.nr_uninterruptible 1029 ± 18% -38.7% 631.00 ± 10% sched_debug.cpu#5.nr_uninterruptible 1244821 ± 70% -70.6% 366365 ± 3% sched_debug.cpu#6.nr_switches 1136 ± 14% -60.6% 447.75 ± 15% sched_debug.cpu#6.nr_uninterruptible 1244958 ± 70% -70.6% 366498 ± 3% sched_debug.cpu#6.sched_count 845.25 ± 17% -30.1% 591.00 ± 7% sched_debug.cpu#7.nr_uninterruptible ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/iterations/nr_threads/disk/fs/filesize/test_size/sync_method/nr_directories/nr_files_per_directory: nhm4/fsmark/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/1x/32t/1HDD/f2fs/16MB/60G/fsyncBeforeClose/16d/256fpd commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 98844 ± 2% -75.2% 24496 ± 1% fsmark.time.involuntary_context_switches 98844 ± 2% -75.2% 24496 ± 1% time.involuntary_context_switches 22613 ± 0% -78.7% 4810 ± 0% vmstat.system.cs 9686 ± 0% -93.5% 631.00 ± 2% vmstat.system.in 1.703e+08 ± 3% -21.2% 1.342e+08 ± 5% cpuidle.C1-NHM.time 4972587 ± 0% -88.7% 560540 ± 2% cpuidle.C1-NHM.usage 2092 ± 5% -91.8% 171.75 ± 6% cpuidle.POLL.usage 3.44 ± 0% -11.9% 3.04 ± 0% turbostat.%Busy 112.00 ± 0% -12.9% 97.50 ± 0% turbostat.Avg_MHz 12.01 ± 1% -15.8% 10.12 ± 3% turbostat.CPU%c1 42309 ±125% -92.6% 3129 ± 72% latency_stats.avg.wait_on_page_bit.f2fs_wait_on_page_writeback.[f2fs].f2fs_wait_on_page_writeback.[f2fs].get_dnode_of_data.[f2fs].f2fs_reserve_block.[f2fs].f2fs_write_begin.[f2fs].generic_perform_write.__generic_file_write_iter.generic_file_write_iter.f2fs_file_write_iter.[f2fs].__vfs_write.vfs_write 153873 ±159% -90.8% 14192 ± 51% latency_stats.max.alloc_nid.[f2fs].f2fs_new_inode.[f2fs].f2fs_create.[f2fs].vfs_create.path_openat.do_filp_open.do_sys_open.SyS_open.entry_SYSCALL_64_fastpath 1458401 ±133% -69.6% 444016 ± 59% latency_stats.sum.alloc_nid.[f2fs].f2fs_new_inode.[f2fs].f2fs_create.[f2fs].vfs_create.path_openat.do_filp_open.do_sys_open.SyS_open.entry_SYSCALL_64_fastpath 20021 ± 29% +6098.0% 1240911 ±137% latency_stats.sum.wait_on_page_bit.f2fs_wait_on_page_writeback.[f2fs].f2fs_wait_on_page_writeback.[f2fs].__f2fs_add_link.[f2fs].f2fs_create.[f2fs].vfs_create.path_openat.do_filp_open.do_sys_open.SyS_open.entry_SYSCALL_64_fastpath 13943 ±132% -53.5% 6479 ±100% latency_stats.sum.wait_on_page_bit.find_data_page.[f2fs].f2fs_find_entry.[f2fs].f2fs_lookup.[f2fs].lookup_real.path_openat.do_filp_open.do_sys_open.SyS_open.entry_SYSCALL_64_fastpath 4129 ± 27% -34.1% 2721 ± 21% sched_debug.cfs_rq[0]:/.avg->runnable_avg_sum 89.00 ± 27% -34.0% 58.75 ± 22% sched_debug.cfs_rq[0]:/.tg_runnable_contrib 30.50 ± 36% -59.0% 12.50 ± 74% sched_debug.cfs_rq[3]:/.runnable_load_avg 3952 ± 34% -41.6% 2306 ± 39% sched_debug.cfs_rq[6]:/.avg->runnable_avg_sum 85.25 ± 34% -41.9% 49.50 ± 40% sched_debug.cfs_rq[6]:/.tg_runnable_contrib 3552 ± 36% -44.5% 1971 ± 23% sched_debug.cfs_rq[7]:/.avg->runnable_avg_sum 207.00 ± 47% -72.9% 56.00 ± 93% sched_debug.cfs_rq[7]:/.load 76.50 ± 36% -45.1% 42.00 ± 23% sched_debug.cfs_rq[7]:/.tg_runnable_contrib 3.67 ± 89% +445.5% 20.00 ± 23% sched_debug.cpu#0.cpu_load[0] 4.75 ±102% +163.2% 12.50 ± 6% sched_debug.cpu#0.cpu_load[1] 16.50 ± 53% -78.8% 3.50 ±140% sched_debug.cpu#1.cpu_load[1] 12.25 ± 49% -77.6% 2.75 ± 39% sched_debug.cpu#1.cpu_load[4] 33933 ± 41% -54.9% 15287 ± 1% sched_debug.cpu#2.nr_load_updates 3238629 ± 74% -96.6% 108909 ± 12% sched_debug.cpu#2.nr_switches 3238752 ± 74% -96.6% 109056 ± 12% sched_debug.cpu#2.sched_count 1570598 ± 75% -97.8% 35193 ± 18% sched_debug.cpu#2.sched_goidle 1582367 ± 76% -98.6% 21962 ± 2% sched_debug.cpu#2.ttwu_local 3380681 ± 73% -96.1% 132270 ± 13% sched_debug.cpu#6.nr_switches 3380800 ± 73% -96.1% 132424 ± 13% sched_debug.cpu#6.sched_count 1648415 ± 74% -97.0% 50035 ± 17% sched_debug.cpu#6.sched_goidle 203.25 ± 47% -72.4% 56.00 ± 93% sched_debug.cpu#7.load ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/iterations/nr_threads/disk/fs/filesize/test_size/sync_method/nr_directories/nr_files_per_directory: nhm4/fsmark/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/1x/32t/1HDD/f2fs/5K/400M/fsyncBeforeClose/16d/256fpd commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 35663 ± 8% -44.4% 19828 ± 0% fsmark.time.involuntary_context_switches 13.00 ± 0% +7.7% 14.00 ± 0% fsmark.time.percent_of_cpu_this_job_got 645071 ± 0% +1.9% 657186 ± 0% fsmark.time.voluntary_context_switches 35663 ± 8% -44.4% 19828 ± 0% time.involuntary_context_switches 69916 ± 0% -83.1% 11789 ± 0% vmstat.system.cs 30442 ± 0% -96.3% 1137 ± 1% vmstat.system.in 3.49 ± 0% -37.3% 2.19 ± 0% turbostat.%Busy 110.50 ± 1% -42.8% 63.25 ± 0% turbostat.Avg_MHz 41.09 ± 1% -11.4% 36.40 ± 0% turbostat.CPU%c1 7.23 ± 4% +22.1% 8.83 ± 5% turbostat.CPU%c6 20352 ± 83% -59.3% 8281 ± 49% latency_stats.sum.call_rwsem_down_read_failed.f2fs_mkdir.[f2fs].vfs_mkdir.SyS_mkdir.entry_SYSCALL_64_fastpath 6985436 ± 1% +13.7% 7940904 ± 3% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].get_node_page.[f2fs].f2fs_write_begin.[f2fs].generic_perform_write.__generic_file_write_iter.generic_file_write_iter.f2fs_file_write_iter.[f2fs].__vfs_write.vfs_write.SyS_write 3675763 ± 1% +12.9% 4149821 ± 4% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].get_node_page.[f2fs].get_dnode_of_data.[f2fs].f2fs_reserve_block.[f2fs].f2fs_write_begin.[f2fs].generic_perform_write.__generic_file_write_iter.generic_file_write_iter.f2fs_file_write_iter.[f2fs].__vfs_write 5179210 ± 1% +13.6% 5882087 ± 2% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].get_node_page.[f2fs].get_dnode_of_data.[f2fs].f2fs_reserve_block.[f2fs].get_new_data_page.[f2fs].__f2fs_add_link.[f2fs].f2fs_create.[f2fs].vfs_create.path_openat.do_filp_open 412246 ± 0% +8.4% 446769 ± 2% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].write_data_page.[f2fs].do_write_data_page.[f2fs].f2fs_write_data_page.[f2fs].__f2fs_writepage.[f2fs].write_cache_pages.f2fs_write_data_pages.[f2fs].do_writepages.__filemap_fdatawrite_range.filemap_write_and_wait_range.f2fs_sync_file.[f2fs] 2.715e+08 ± 2% -11.1% 2.414e+08 ± 1% cpuidle.C1-NHM.time 4779213 ± 0% -93.2% 324049 ± 2% cpuidle.C1-NHM.usage 2.285e+08 ± 2% +13.2% 2.587e+08 ± 2% cpuidle.C6-NHM.time 105567 ± 1% +13.5% 119864 ± 2% cpuidle.C6-NHM.usage 31639 ± 11% -87.1% 4072 ± 57% cpuidle.POLL.time 8304 ± 10% -98.8% 101.00 ± 7% cpuidle.POLL.usage -1817 ± -4% +37.1% -2491 ±-25% sched_debug.cfs_rq[3]:/.spread0 1313 ± 3% +12.7% 1480 ± 3% sched_debug.cfs_rq[5]:/.exec_clock -1878 ± -5% +16.6% -2190 ± -1% sched_debug.cpu#0.nr_uninterruptible 375233 ±118% -90.3% 36371 ± 4% sched_debug.cpu#3.ttwu_count 503266 ± 30% +63.0% 820371 ± 8% sched_debug.cpu#5.avg_idle 15.50 ± 29% -59.7% 6.25 ± 70% sched_debug.cpu#5.cpu_load[3] 13.75 ± 18% -47.3% 7.25 ± 44% sched_debug.cpu#5.cpu_load[4] 743.00 ± 7% -18.3% 607.25 ± 5% sched_debug.cpu#5.nr_uninterruptible 774.25 ± 15% -21.1% 610.50 ± 6% sched_debug.cpu#6.nr_uninterruptible 23789 ± 58% -59.0% 9764 ± 2% sched_debug.cpu#7.nr_load_updates ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/iterations/nr_threads/disk/fs/filesize/test_size/sync_method/nr_directories/nr_files_per_directory: nhm4/fsmark/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/1x/32t/1HDD/f2fs/8K/400M/fsyncBeforeClose/16d/256fpd commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 28399 ± 12% -57.0% 12208 ± 1% fsmark.time.involuntary_context_switches 12.50 ± 4% +12.0% 14.00 ± 0% fsmark.time.percent_of_cpu_this_job_got 400598 ± 0% +1.9% 408272 ± 0% fsmark.time.voluntary_context_switches 104101 ± 0% -88.9% 11556 ± 0% vmstat.system.cs 47821 ± 0% -97.6% 1125 ± 0% vmstat.system.in 27507 ± 3% -7.3% 25490 ± 3% meminfo.Active(anon) 27135 ± 3% -7.4% 25120 ± 3% meminfo.AnonPages 13576 ± 0% -11.3% 12048 ± 0% meminfo.Mapped 6878 ± 3% -7.4% 6370 ± 3% proc-vmstat.nr_active_anon 6796 ± 3% -7.5% 6288 ± 3% proc-vmstat.nr_anon_pages 3392 ± 0% -11.2% 3011 ± 0% proc-vmstat.nr_mapped 19880 ± 6% -17.3% 16446 ± 4% softirqs.RCU 17934 ± 2% -16.1% 15052 ± 4% softirqs.SCHED 29061 ± 3% -13.6% 25095 ± 6% softirqs.TIMER 28399 ± 12% -57.0% 12208 ± 1% time.involuntary_context_switches 11.65 ± 1% +12.1% 13.06 ± 0% time.system_time 0.49 ± 2% +20.7% 0.60 ± 3% time.user_time 1.821e+08 ± 1% -17.5% 1.502e+08 ± 2% cpuidle.C1-NHM.time 4650494 ± 0% -95.7% 199638 ± 1% cpuidle.C1-NHM.usage 1.391e+08 ± 1% +14.3% 1.59e+08 ± 1% cpuidle.C6-NHM.time 61904 ± 1% +20.1% 74370 ± 1% cpuidle.C6-NHM.usage 34499 ± 12% -90.5% 3270 ±113% cpuidle.POLL.time 8367 ± 16% -99.2% 64.00 ± 9% cpuidle.POLL.usage 4.26 ± 0% -48.8% 2.18 ± 0% turbostat.%Busy 139.00 ± 0% -54.7% 63.00 ± 0% turbostat.Avg_MHz 3255 ± 0% -11.1% 2892 ± 0% turbostat.Bzy_MHz 43.70 ± 1% -16.5% 36.48 ± 1% turbostat.CPU%c1 44.77 ± 1% +18.3% 52.95 ± 1% turbostat.CPU%c3 7.27 ± 2% +15.3% 8.38 ± 3% turbostat.CPU%c6 8144 ± 20% +112.7% 17320 ± 59% latency_stats.sum.call_rwsem_down_read_failed.f2fs_write_node_page.[f2fs].sync_node_pages.[f2fs].f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 161602 ± 5% +27.7% 206366 ± 6% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].f2fs_write_node_page.[f2fs].sync_node_pages.[f2fs].f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 1263747 ± 3% +19.4% 1508845 ± 4% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].new_node_page.[f2fs].new_inode_page.[f2fs].init_inode_metadata.[f2fs].__f2fs_add_link.[f2fs].f2fs_create.[f2fs].vfs_create.path_openat.do_filp_open.do_sys_open.SyS_open 662165 ± 2% +17.5% 777739 ± 2% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].get_node_page.[f2fs].f2fs_convert_inline_inode.[f2fs].f2fs_write_begin.[f2fs].generic_perform_write.__generic_file_write_iter.generic_file_write_iter.f2fs_file_write_iter.[f2fs].__vfs_write.vfs_write 3956363 ± 1% +22.3% 4840014 ± 1% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].get_node_page.[f2fs].f2fs_write_begin.[f2fs].generic_perform_write.__generic_file_write_iter.generic_file_write_iter.f2fs_file_write_iter.[f2fs].__vfs_write.vfs_write.SyS_write 436691 ± 1% +15.4% 504042 ± 0% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].get_node_page.[f2fs].get_dnode_of_data.[f2fs].do_write_data_page.[f2fs].f2fs_write_data_page.[f2fs].__f2fs_writepage.[f2fs].write_cache_pages.f2fs_write_data_pages.[f2fs].do_writepages.__filemap_fdatawrite_range 2101052 ± 1% +20.7% 2536800 ± 2% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].get_node_page.[f2fs].get_dnode_of_data.[f2fs].f2fs_reserve_block.[f2fs].f2fs_write_begin.[f2fs].generic_perform_write.__generic_file_write_iter.generic_file_write_iter.f2fs_file_write_iter.[f2fs].__vfs_write 2931560 ± 2% +21.5% 3562381 ± 2% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].get_node_page.[f2fs].get_dnode_of_data.[f2fs].f2fs_reserve_block.[f2fs].get_new_data_page.[f2fs].__f2fs_add_link.[f2fs].f2fs_create.[f2fs].vfs_create.path_openat.do_filp_open 882333 ± 1% +18.1% 1042172 ± 1% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].get_node_page.[f2fs].get_dnode_of_data.[f2fs].get_read_data_page.[f2fs].find_data_page.[f2fs].f2fs_find_entry.[f2fs].f2fs_lookup.[f2fs].lookup_real.path_openat.do_filp_open 86763 ± 3% +24.8% 108290 ± 2% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].get_node_page.[f2fs].update_inode_page.[f2fs].__f2fs_add_link.[f2fs].f2fs_create.[f2fs].vfs_create.path_openat.do_filp_open.do_sys_open.SyS_open 1293620 ± 2% +18.0% 1526759 ± 2% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].read_node_page.[f2fs].get_node_page.[f2fs].update_inode_page.[f2fs].f2fs_write_end.[f2fs].generic_perform_write.__generic_file_write_iter.generic_file_write_iter.f2fs_file_write_iter.[f2fs].__vfs_write.vfs_write 241754 ± 1% +14.7% 277276 ± 3% latency_stats.sum.call_rwsem_down_read_failed.get_node_info.[f2fs].write_data_page.[f2fs].do_write_data_page.[f2fs].f2fs_write_data_page.[f2fs].__f2fs_writepage.[f2fs].write_cache_pages.f2fs_write_data_pages.[f2fs].do_writepages.__filemap_fdatawrite_range.filemap_write_and_wait_range.f2fs_sync_file.[f2fs] 285314 ± 4% +31.5% 375080 ± 6% latency_stats.sum.call_rwsem_down_read_failed.need_dentry_mark.[f2fs].sync_node_pages.[f2fs].f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 298289 ± 4% +36.2% 406142 ± 7% latency_stats.sum.call_rwsem_down_read_failed.need_inode_block_update.[f2fs].f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 1885703 ± 2% +17.4% 2213119 ± 2% latency_stats.sum.call_rwsem_down_write_failed.get_node_info.[f2fs].new_node_page.[f2fs].new_inode_page.[f2fs].init_inode_metadata.[f2fs].__f2fs_add_link.[f2fs].f2fs_create.[f2fs].vfs_create.path_openat.do_filp_open.do_sys_open.SyS_open 198811 ± 3% +34.9% 268119 ± 3% latency_stats.sum.call_rwsem_down_write_failed.set_node_addr.[f2fs].f2fs_write_node_page.[f2fs].sync_node_pages.[f2fs].f2fs_sync_file.[f2fs].vfs_fsync_range.do_fsync.SyS_fsync.entry_SYSCALL_64_fastpath 910989 ± 2% +17.8% 1072972 ± 2% latency_stats.sum.call_rwsem_down_write_failed.set_node_addr.[f2fs].new_node_page.[f2fs].new_inode_page.[f2fs].init_inode_metadata.[f2fs].__f2fs_add_link.[f2fs].f2fs_create.[f2fs].vfs_create.path_openat.do_filp_open.do_sys_open.SyS_open 363528 ± 0% +3.4% 375985 ± 0% latency_stats.sum.pipe_wait.pipe_read.__vfs_read.vfs_read.SyS_read.entry_SYSCALL_64_fastpath 9590 ± 9% -20.1% 7663 ± 5% sched_debug.cfs_rq[0]:/.avg->runnable_avg_sum 211.25 ± 9% -21.1% 166.75 ± 6% sched_debug.cfs_rq[0]:/.tg_runnable_contrib 7864 ± 14% -22.2% 6117 ± 10% sched_debug.cfs_rq[1]:/.avg->runnable_avg_sum 172.00 ± 13% -22.7% 133.00 ± 11% sched_debug.cfs_rq[1]:/.tg_runnable_contrib 6243 ± 10% +32.8% 8292 ± 12% sched_debug.cfs_rq[4]:/.avg->runnable_avg_sum 1910 ± 14% +19.0% 2273 ± 4% sched_debug.cfs_rq[4]:/.min_vruntime 136.50 ± 11% +32.8% 181.25 ± 13% sched_debug.cfs_rq[4]:/.tg_runnable_contrib 726.92 ± 6% +13.6% 825.51 ± 5% sched_debug.cfs_rq[6]:/.exec_clock 1844 ± 9% +17.6% 2169 ± 6% sched_debug.cfs_rq[6]:/.min_vruntime 8.00 ± 45% +209.4% 24.75 ± 11% sched_debug.cpu#0.cpu_load[3] 9.00 ± 22% +141.7% 21.75 ± 8% sched_debug.cpu#0.cpu_load[4] -701.75 ±-39% +56.6% -1098 ± -2% sched_debug.cpu#0.nr_uninterruptible 546261 ± 90% -85.3% 80139 ± 3% sched_debug.cpu#0.ttwu_count 483716 ±101% -95.6% 21310 ± 2% sched_debug.cpu#0.ttwu_local 547901 ±166% -96.5% 19272 ± 14% sched_debug.cpu#3.ttwu_count 24.00 ± 59% -64.6% 8.50 ± 55% sched_debug.cpu#4.cpu_load[3] 20.50 ± 32% -57.3% 8.75 ± 38% sched_debug.cpu#4.cpu_load[4] 13541 ± 61% -62.1% 5134 ± 3% sched_debug.cpu#4.nr_load_updates 1533535 ± 99% -97.1% 43964 ± 29% sched_debug.cpu#4.nr_switches 1533560 ± 99% -97.1% 43984 ± 29% sched_debug.cpu#4.sched_count 760621 ± 99% -97.5% 19182 ± 34% sched_debug.cpu#4.sched_goidle 750180 ±102% -99.8% 1305 ± 5% sched_debug.cpu#4.ttwu_local 414.75 ± 6% -23.3% 318.25 ± 5% sched_debug.cpu#6.nr_uninterruptible 20.25 ± 30% -56.8% 8.75 ± 74% sched_debug.cpu#7.cpu_load[4] ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/nr_threads/iterations/samples: lituya/ftq/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/100%/20x/100000ss commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 7640 ± 1% +5.5% 8056 ± 0% ftq.counts 0.17 ± 11% -82.1% 0.03 ± 6% ftq.stddev 1806627 ± 4% -98.6% 25037 ± 32% ftq.time.involuntary_context_switches 1437 ± 2% +5.6% 1518 ± 0% ftq.time.percent_of_cpu_this_job_got 546.95 ± 0% +3.5% 566.01 ± 0% ftq.time.user_time 16653 ± 0% -13.2% 14452 ± 0% meminfo.Mapped 4163 ± 0% -13.3% 3611 ± 0% proc-vmstat.nr_mapped 9329 ± 4% -17.0% 7746 ± 9% softirqs.SCHED 1.20 ± 2% -52.7% 0.57 ± 12% turbostat.CPU%c1 330.20 ± 5% -7.2% 306.54 ± 4% uptime.idle 96844 ± 0% -97.0% 2925 ± 4% vmstat.system.cs 61693 ± 0% -76.0% 14816 ± 1% vmstat.system.in 1806627 ± 4% -98.6% 25037 ± 32% time.involuntary_context_switches 2.15 ± 8% -50.9% 1.06 ± 1% time.system_time 820.50 ± 33% +45.4% 1192 ± 3% time.voluntary_context_switches 1857689 ± 17% -85.0% 278914 ± 73% cpuidle.C1-HSW.time 110169 ± 31% -99.3% 786.25 ± 4% cpuidle.C1-HSW.usage 779.50 ± 69% +78.4% 1390 ± 4% cpuidle.C3-HSW.usage 2663 ± 41% -48.2% 1380 ± 1% cpuidle.C6-HSW.usage 193.50 ± 95% -99.6% 0.75 ±173% cpuidle.POLL.time 125.00 ± 94% -99.8% 0.25 ±173% cpuidle.POLL.usage 6.25 ± 17% -36.0% 4.00 ± 17% sched_debug.cfs_rq[0]:/.nr_spread_over 118.31 ± 19% +50.0% 177.51 ± 18% sched_debug.cfs_rq[11]:/.exec_clock 1410 ± 23% -32.8% 948.03 ± 23% sched_debug.cfs_rq[14]:/.min_vruntime 16706 ± 38% -72.2% 4641 ± 40% sched_debug.cfs_rq[15]:/.avg->runnable_avg_sum 367.50 ± 38% -72.2% 102.00 ± 41% sched_debug.cfs_rq[15]:/.tg_runnable_contrib 6553 ± 29% -34.8% 4269 ± 12% sched_debug.cfs_rq[2]:/.avg->runnable_avg_sum 143.25 ± 30% -35.1% 93.00 ± 12% sched_debug.cfs_rq[2]:/.tg_runnable_contrib 1304 ± 17% +43.7% 1873 ± 28% sched_debug.cfs_rq[3]:/.min_vruntime 3651 ±139% +234.7% 12223 ± 54% sched_debug.cfs_rq[6]:/.blocked_load_avg 3651 ±139% +240.9% 12449 ± 53% sched_debug.cfs_rq[6]:/.tg_load_contrib 1316 ± 70% +500.7% 7907 ± 51% sched_debug.cfs_rq[8]:/.blocked_load_avg 1330 ± 71% +494.4% 7907 ± 51% sched_debug.cfs_rq[8]:/.tg_load_contrib 111.43 ± 6% -14.1% 95.77 ± 7% sched_debug.cfs_rq[9]:/.exec_clock 238565 ± 59% +271.1% 885357 ± 15% sched_debug.cpu#0.avg_idle 5.75 ± 85% +152.2% 14.50 ± 55% sched_debug.cpu#0.cpu_load[3] 3.25 ± 95% +1215.4% 42.75 ± 78% sched_debug.cpu#1.cpu_load[1] 5.25 ± 78% -119.0% -1.00 ±-300% sched_debug.cpu#10.nr_uninterruptible 2.50 ± 34% +230.0% 8.25 ± 40% sched_debug.cpu#12.nr_uninterruptible 468.25 ± 12% -21.3% 368.50 ± 8% sched_debug.cpu#13.sched_goidle 1.25 ±131% -420.0% -4.00 ±-20% sched_debug.cpu#14.nr_uninterruptible 73.25 ± 4% +39.6% 102.25 ± 13% sched_debug.cpu#14.ttwu_local -3.50 ±-76% -228.6% 4.50 ± 57% sched_debug.cpu#15.nr_uninterruptible 319447 ± 52% +169.7% 861647 ± 13% sched_debug.cpu#2.avg_idle 3181 ± 38% -65.7% 1090 ± 20% sched_debug.cpu#4.nr_load_updates 787.75 ± 15% -23.5% 602.50 ± 6% sched_debug.cpu#5.sched_goidle 4.25 ± 45% +494.1% 25.25 ± 41% sched_debug.cpu#8.cpu_load[3] 9.00 ± 54% +200.0% 27.00 ± 27% sched_debug.cpu#8.cpu_load[4] 2045 ± 34% -41.3% 1201 ± 14% sched_debug.cpu#8.nr_switches 2048 ± 33% -41.2% 1203 ± 14% sched_debug.cpu#8.sched_count 789.00 ± 24% -47.5% 414.25 ± 26% sched_debug.cpu#8.sched_goidle 20.59 ± 92% -74.0% 5.36 ± 12% sched_debug.rt_rq[0]:/.rt_time ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/nr_threads/iterations/samples: lituya/ftq/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/powersave/100%/20x/100000ss commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 7580 ± 0% +6.1% 8044 ± 0% ftq.counts 0.18 ± 7% -82.1% 0.03 ± 8% ftq.stddev 1761823 ± 3% -98.3% 29477 ± 38% ftq.time.involuntary_context_switches 1454 ± 1% +4.3% 1517 ± 0% ftq.time.percent_of_cpu_this_job_got 547.59 ± 0% +3.4% 566.04 ± 0% ftq.time.user_time 562.25 ± 23% +96.8% 1106 ± 5% ftq.time.voluntary_context_switches 16809 ± 0% -13.8% 14482 ± 0% meminfo.Mapped 4198 ± 0% -13.7% 3623 ± 0% proc-vmstat.nr_mapped 484.00 ± 6% +18.2% 572.00 ± 7% slabinfo.blkdev_requests.active_objs 484.00 ± 6% +18.2% 572.00 ± 7% slabinfo.blkdev_requests.num_objs 1.17 ± 2% -50.9% 0.57 ± 8% turbostat.CPU%c1 0.21 ±127% +215.5% 0.66 ± 33% turbostat.RAMWatt 96716 ± 0% -96.9% 2979 ± 10% vmstat.system.cs 61578 ± 0% -75.7% 14933 ± 1% vmstat.system.in 1761823 ± 3% -98.3% 29477 ± 38% time.involuntary_context_switches 1.52 ± 5% -28.3% 1.09 ± 2% time.system_time 562.25 ± 23% +96.8% 1106 ± 5% time.voluntary_context_switches 1633723 ± 55% -89.4% 172951 ± 3% cpuidle.C1-HSW.time 124268 ± 20% -99.4% 713.50 ± 8% cpuidle.C1-HSW.usage 104432 ± 22% +110.9% 220296 ± 63% cpuidle.C1E-HSW.time 552.75 ± 33% +33.6% 738.50 ± 3% cpuidle.C1E-HSW.usage 23.00 ± 43% -100.0% 0.00 ± 0% cpuidle.POLL.time 11.25 ± 51% -100.0% 0.00 ± 0% cpuidle.POLL.usage 49106 ± 3% +32.3% 64986 ± 8% sched_debug.cfs_rq[0]:/.tg_load_avg 47692 ± 6% +35.4% 64593 ± 8% sched_debug.cfs_rq[10]:/.tg_load_avg 47692 ± 6% +35.4% 64593 ± 8% sched_debug.cfs_rq[11]:/.tg_load_avg 47692 ± 6% +35.4% 64593 ± 8% sched_debug.cfs_rq[12]:/.tg_load_avg 47692 ± 6% +35.4% 64593 ± 8% sched_debug.cfs_rq[13]:/.tg_load_avg 860.75 ±116% +570.6% 5772 ± 87% sched_debug.cfs_rq[14]:/.blocked_load_avg 47692 ± 6% +35.4% 64593 ± 8% sched_debug.cfs_rq[14]:/.tg_load_avg 860.75 ±116% +617.8% 6178 ± 92% sched_debug.cfs_rq[14]:/.tg_load_contrib 2268 ± 20% -42.9% 1296 ± 24% sched_debug.cfs_rq[15]:/.min_vruntime 47692 ± 6% +35.4% 64593 ± 8% sched_debug.cfs_rq[15]:/.tg_load_avg 48510 ± 5% +34.0% 64986 ± 8% sched_debug.cfs_rq[1]:/.tg_load_avg 48510 ± 5% +33.9% 64934 ± 8% sched_debug.cfs_rq[2]:/.tg_load_avg 48510 ± 5% +33.9% 64934 ± 8% sched_debug.cfs_rq[3]:/.tg_load_avg 48510 ± 5% +33.9% 64934 ± 8% sched_debug.cfs_rq[4]:/.tg_load_avg 48494 ± 5% +33.9% 64916 ± 8% sched_debug.cfs_rq[5]:/.tg_load_avg 48494 ± 5% +33.9% 64916 ± 8% sched_debug.cfs_rq[6]:/.tg_load_avg 48455 ± 5% +33.9% 64905 ± 8% sched_debug.cfs_rq[7]:/.tg_load_avg 0.50 ±100% +300.0% 2.00 ± 0% sched_debug.cfs_rq[8]:/.nr_spread_over 48404 ± 5% +34.1% 64905 ± 8% sched_debug.cfs_rq[8]:/.tg_load_avg 47692 ± 6% +35.5% 64637 ± 8% sched_debug.cfs_rq[9]:/.tg_load_avg 12.00 ± 27% -52.1% 5.75 ± 43% sched_debug.cpu#1.nr_uninterruptible 347.00 ± 7% +42.0% 492.75 ± 20% sched_debug.cpu#10.sched_goidle 518.00 ± 8% +24.0% 642.50 ± 6% sched_debug.cpu#11.sched_goidle 9.25 ± 62% +278.4% 35.00 ± 67% sched_debug.cpu#12.cpu_load[4] 5.75 ± 74% -100.0% 0.00 ± 0% sched_debug.cpu#14.nr_uninterruptible 1.00 ± 70% +2800.0% 29.00 ± 88% sched_debug.cpu#15.cpu_load[2] 3.00 ± 91% +608.3% 21.25 ± 62% sched_debug.cpu#15.cpu_load[3] 4.00 ±117% +275.0% 15.00 ± 54% sched_debug.cpu#15.cpu_load[4] 3.00 ±131% +91.7% 5.75 ±125% sched_debug.cpu#2.nr_uninterruptible 2443 ± 35% -46.9% 1296 ± 14% sched_debug.cpu#3.nr_load_updates 6257 ± 21% -47.1% 3309 ± 36% sched_debug.cpu#3.nr_switches 6262 ± 21% -47.1% 3315 ± 36% sched_debug.cpu#3.sched_count 2859 ± 24% -54.1% 1313 ± 35% sched_debug.cpu#3.sched_goidle 3022 ± 19% -46.9% 1604 ± 19% sched_debug.cpu#3.ttwu_count 1491 ± 55% -78.3% 323.00 ± 69% sched_debug.cpu#3.ttwu_local 994307 ± 0% -27.9% 716731 ± 21% sched_debug.cpu#6.avg_idle ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/nr_threads/iterations/samples: lituya/fwq/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/100%/20x/100000ss commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 0.11 ± 3% -43.2% 0.06 ± 1% fwq.stddev 3230702 ± 1% -86.7% 430932 ± 3% fwq.time.involuntary_context_switches 159977 ± 11% -25.2% 119739 ± 0% latency_stats.sum.do_wait.SyS_wait4.entry_SYSCALL_64_fastpath 3230702 ± 1% -86.7% 430932 ± 3% time.involuntary_context_switches 1.28 ± 35% -32.7% 0.86 ± 2% time.system_time 0.21 ± 42% -70.6% 0.06 ± 17% turbostat.CPU%c1 0.73 ± 23% -31.1% 0.51 ± 15% turbostat.CPU%c6 16514 ± 0% -79.4% 3404 ± 2% vmstat.system.cs 23255 ± 0% -28.0% 16754 ± 0% vmstat.system.in 1081329 ± 53% -93.7% 67884 ± 33% cpuidle.C1-HSW.time 45064 ± 0% -99.4% 278.25 ± 22% cpuidle.C1-HSW.usage 64741983 ± 8% -39.7% 39054828 ± 14% cpuidle.C6-HSW.time 2214 ± 19% -45.3% 1211 ± 11% cpuidle.C6-HSW.usage 4.50 ± 74% -100.0% 0.00 ± -1% cpuidle.POLL.time 2.50 ± 60% -100.0% 0.00 ± -1% cpuidle.POLL.usage 152.75 ±127% +449.4% 839.25 ± 60% sched_debug.cfs_rq[10]:/.blocked_load_avg 204.75 ± 94% +339.8% 900.50 ± 56% sched_debug.cfs_rq[10]:/.tg_load_contrib -73145 ±-135% -141.7% 30484 ± 48% sched_debug.cfs_rq[14]:/.spread0 51.50 ± 0% +14.1% 58.75 ± 9% sched_debug.cpu#0.cpu_load[1] 51.75 ± 1% +13.0% 58.50 ± 7% sched_debug.cpu#0.cpu_load[2] 51.75 ± 1% +11.1% 57.50 ± 6% sched_debug.cpu#0.cpu_load[3] 11.25 ± 9% -64.4% 4.00 ± 63% sched_debug.cpu#11.nr_uninterruptible -7.50 ±-35% -153.3% 4.00 ± 93% sched_debug.cpu#2.nr_uninterruptible 1419 ± 26% +119.7% 3117 ± 22% sched_debug.cpu#3.sched_goidle 124974 ± 61% -72.4% 34432 ±108% sched_debug.cpu#5.nr_switches 4.50 ± 71% -177.8% -3.50 ±-123% sched_debug.cpu#5.nr_uninterruptible 125045 ± 61% -72.4% 34489 ±108% sched_debug.cpu#5.sched_count 47153 ± 75% -52.2% 22527 ±124% sched_debug.cpu#5.ttwu_count 33719 ± 76% -70.3% 10005 ±131% sched_debug.cpu#5.ttwu_local 53.75 ± 4% +18.1% 63.50 ± 18% sched_debug.cpu#7.cpu_load[1] 5925 ± 34% +483.7% 34588 ± 82% sched_debug.cpu#8.sched_count 56.25 ± 8% -10.2% 50.50 ± 0% sched_debug.cpu#9.cpu_load[1] 58.75 ± 10% -13.2% 51.00 ± 1% sched_debug.cpu#9.cpu_load[2] 61.50 ± 14% -16.3% 51.50 ± 2% sched_debug.cpu#9.cpu_load[3] 63.00 ± 17% -18.3% 51.50 ± 2% sched_debug.cpu#9.cpu_load[4] ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/nr_threads/iterations/samples: lituya/fwq/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/powersave/100%/20x/100000ss commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 0.11 ± 2% -41.0% 0.06 ± 1% fwq.stddev 3249189 ± 1% -86.8% 428991 ± 8% fwq.time.involuntary_context_switches 0.19 ± 24% -58.1% 0.08 ± 10% turbostat.CPU%c1 3249189 ± 1% -86.8% 428991 ± 8% time.involuntary_context_switches 1.81 ± 12% -48.6% 0.93 ± 2% time.system_time 16527 ± 1% -79.2% 3430 ± 3% vmstat.system.cs 23258 ± 0% -28.0% 16746 ± 0% vmstat.system.in 41773 ± 2% -99.2% 318.75 ± 13% cpuidle.C1-HSW.usage 61756882 ± 15% -30.4% 43002599 ± 8% cpuidle.C6-HSW.time 2117 ± 9% -33.5% 1407 ± 3% cpuidle.C6-HSW.usage 66.00 ± 26% -21.6% 51.75 ± 5% sched_debug.cfs_rq[0]:/.load 0.75 ± 57% +633.3% 5.50 ± 58% sched_debug.cfs_rq[13]:/.nr_spread_over 181.50 ±139% +229.6% 598.25 ± 44% sched_debug.cfs_rq[15]:/.blocked_load_avg 239.00 ±104% +175.8% 659.25 ± 42% sched_debug.cfs_rq[15]:/.tg_load_contrib 893.00 ± 0% +14.5% 1022 ± 12% sched_debug.cfs_rq[1]:/.utilization_load_avg -66095 ±-174% -144.5% 29384 ± 74% sched_debug.cfs_rq[4]:/.spread0 -48436 ±-144% -146.3% 22425 ± 36% sched_debug.cfs_rq[5]:/.spread0 66.00 ± 26% -21.6% 51.75 ± 5% sched_debug.cpu#0.load 807.00 ± 0% +7.3% 866.00 ± 6% sched_debug.cpu#11.curr->pid 4.25 ± 76% -84.3% 0.67 ±604% sched_debug.cpu#14.nr_uninterruptible 370327 ±158% -95.9% 15153 ± 80% sched_debug.cpu#15.nr_switches 370400 ±158% -95.9% 15239 ± 80% sched_debug.cpu#15.sched_count 186416 ±158% -97.0% 5614 ± 81% sched_debug.cpu#15.ttwu_count -4.00 ±-81% -187.5% 3.50 ± 24% sched_debug.cpu#5.nr_uninterruptible 0.00 ± 71% +10603.7% 0.19 ± 13% sched_debug.rt_rq[2]:/.rt_time ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/runtime: wsm/ku-latency/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/300s commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 1207 ± 4% -27.5% 876.00 ± 22% proc-vmstat.pgactivate 26029 ± 2% -21.2% 20524 ± 2% softirqs.SCHED 12.52 ± 5% -19.2% 10.12 ± 9% turbostat.CPU%c1 46.75 ± 2% -7.0% 43.50 ± 3% turbostat.CoreTmp 31128 ± 0% -92.1% 2449 ± 1% vmstat.system.cs 15692 ± 0% -92.3% 1214 ± 0% vmstat.system.in 5760097 ±125% -100.0% 0.00 ± -1% latency_stats.avg.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath 5760097 ±125% -100.0% 0.00 ± -1% latency_stats.max.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath 5760097 ±125% -100.0% 0.00 ± -1% latency_stats.sum.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath 55981055 ± 4% -65.3% 19428998 ± 21% cpuidle.C1-NHM.time 4373782 ± 0% -99.5% 22463 ± 9% cpuidle.C1-NHM.usage 312.50 ± 78% -72.3% 86.50 ± 14% cpuidle.C1E-NHM.usage 2690 ± 28% -37.2% 1689 ± 16% cpuidle.C3-NHM.usage 1071 ± 23% -80.9% 204.25 ± 63% cpuidle.POLL.usage 97.00 ± 45% -83.8% 15.67 ± 70% sched_debug.cfs_rq[10]:/.runnable_load_avg 425.75 ± 34% -74.4% 109.00 ± 70% sched_debug.cfs_rq[10]:/.utilization_load_avg -6567 ±-65% -1590.9% 97909 ±167% sched_debug.cfs_rq[11]:/.spread0 85.25 ± 74% +2038.1% 1822 ± 58% sched_debug.cfs_rq[2]:/.blocked_load_avg 217733 ± 95% -57.1% 93407 ±164% sched_debug.cfs_rq[2]:/.min_vruntime 209.25 ± 66% +813.3% 1911 ± 58% sched_debug.cfs_rq[2]:/.tg_load_contrib 2123 ± 49% -68.8% 662.75 ±100% sched_debug.cfs_rq[4]:/.blocked_load_avg 86.00 ± 53% -100.0% 0.00 ± -1% sched_debug.cfs_rq[4]:/.load 58.50 ± 51% -100.0% 0.00 ± -1% sched_debug.cfs_rq[4]:/.runnable_load_avg 2197 ± 48% -69.5% 671.25 ± 98% sched_debug.cfs_rq[4]:/.tg_load_contrib 300.25 ± 48% -100.0% 0.00 ± -1% sched_debug.cfs_rq[4]:/.utilization_load_avg 3463 ± 55% -48.3% 1789 ± 4% sched_debug.cfs_rq[6]:/.min_vruntime 1227502 ± 99% -98.3% 20386 ± 42% sched_debug.cpu#0.ttwu_count 62332 ± 85% -77.9% 13783 ± 3% sched_debug.cpu#1.nr_load_updates 14318 ± 18% +37.4% 19675 ± 19% sched_debug.cpu#1.ttwu_count 4361 ± 11% -20.1% 3483 ± 7% sched_debug.cpu#1.ttwu_local 146.25 ± 33% -82.2% 26.00 ± 13% sched_debug.cpu#10.cpu_load[0] 79.00 ± 36% -81.4% 14.67 ± 21% sched_debug.cpu#10.cpu_load[1] 44.00 ± 36% -83.5% 7.25 ± 58% sched_debug.cpu#10.cpu_load[2] 25.50 ± 37% -77.5% 5.75 ± 33% sched_debug.cpu#10.cpu_load[3] 15.00 ± 41% -70.0% 4.50 ± 24% sched_debug.cpu#10.cpu_load[4] 89157 ± 69% -47.0% 47258 ±126% sched_debug.cpu#2.nr_load_updates 9.50 ± 33% -76.3% 2.25 ±148% sched_debug.cpu#4.cpu_load[3] 7.25 ± 36% -82.8% 1.25 ±131% sched_debug.cpu#4.cpu_load[4] 1030 ± 36% -100.0% 0.00 ± -1% sched_debug.cpu#4.curr->pid 86.00 ± 53% -100.0% 0.00 ± -1% sched_debug.cpu#4.load 850245 ± 5% +11.7% 949380 ± 3% sched_debug.cpu#5.avg_idle 3578 ± 8% -48.9% 1830 ± 64% sched_debug.cpu#5.ttwu_local 925565 ± 7% -17.7% 761616 ± 6% sched_debug.cpu#6.avg_idle 649243 ± 95% -98.0% 12905 ± 31% sched_debug.cpu#6.ttwu_count 2529 ± 49% -63.9% 911.75 ± 83% sched_debug.cpu#9.ttwu_local ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/runtime/nr_threads/cluster/test: lkp-t410/netperf/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/300s/200%/cs-localhost/TCP_SENDFILE commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 4404 ± 0% +6.5% 4689 ± 0% netperf.Throughput_Mbps 1902815 ± 4% -98.0% 38389 ± 9% netperf.time.involuntary_context_switches 208.00 ± 0% +2.3% 212.75 ± 0% netperf.time.percent_of_cpu_this_job_got 610.08 ± 0% +1.6% 619.92 ± 0% netperf.time.system_time 291566 ± 0% +12.0% 326523 ± 0% netperf.time.voluntary_context_switches 1.03 ± 4% +10.4% 1.14 ± 3% perf-profile.cpu-cycles.rw_verify_area.do_splice_to.splice_direct_to_actor.do_splice_direct.do_sendfile 2545 ± 2% -17.3% 2103 ± 1% proc-vmstat.pgactivate 0.21 ± 3% -95.2% 0.01 ± 0% turbostat.CPU%c1 0.22 ± 5% +146.0% 0.54 ± 3% turbostat.CPU%c6 21522 ± 1% -76.5% 5056 ± 0% vmstat.system.cs 11704 ± 1% -62.7% 4363 ± 0% vmstat.system.in 1902815 ± 4% -98.0% 38389 ± 9% time.involuntary_context_switches 17.50 ± 0% +19.8% 20.97 ± 0% time.user_time 291566 ± 0% +12.0% 326523 ± 0% time.voluntary_context_switches 291365 ± 0% +12.0% 326363 ± 0% latency_stats.hits.sk_stream_wait_memory.tcp_sendpage.inet_sendpage.kernel_sendpage.sock_sendpage.pipe_to_sendpage.__splice_from_pipe.splice_from_pipe.generic_splice_sendpage.direct_splice_actor.splice_direct_to_actor.do_splice_direct 975206 ± 4% -61.5% 374990 ± 1% latency_stats.hits.sk_wait_data.tcp_recvmsg.inet_recvmsg.sock_recvmsg.SYSC_recvfrom.SyS_recvfrom.entry_SYSCALL_64_fastpath 52419094 ± 1% +7.2% 56170711 ± 0% latency_stats.sum.sk_stream_wait_memory.tcp_sendpage.inet_sendpage.kernel_sendpage.sock_sendpage.pipe_to_sendpage.__splice_from_pipe.splice_from_pipe.generic_splice_sendpage.direct_splice_actor.splice_direct_to_actor.do_splice_direct 1.009e+09 ± 0% +5.3% 1.062e+09 ± 0% latency_stats.sum.sk_wait_data.tcp_recvmsg.inet_recvmsg.sock_recvmsg.SYSC_recvfrom.SyS_recvfrom.entry_SYSCALL_64_fastpath 794782 ± 8% -99.4% 4498 ± 40% cpuidle.C1-NHM.time 104769 ± 2% -99.8% 166.25 ± 31% cpuidle.C1-NHM.usage 482872 ± 14% -32.6% 325328 ± 28% cpuidle.C3-NHM.time 4632646 ± 1% +44.6% 6697379 ± 1% cpuidle.C6-NHM.time 337.25 ± 8% +35.4% 456.75 ± 9% cpuidle.C6-NHM.usage 1156 ± 15% -100.0% 0.00 ± 0% cpuidle.POLL.time 180.50 ± 17% -100.0% 0.00 ± 0% cpuidle.POLL.usage 505.75 ± 19% -25.4% 377.25 ± 7% sched_debug.cfs_rq[0]:/.load 443.50 ± 5% -13.0% 386.00 ± 6% sched_debug.cfs_rq[0]:/.runnable_load_avg 6167 ± 13% +18.1% 7285 ± 16% sched_debug.cfs_rq[0]:/.tg_load_avg 6057 ± 13% +17.6% 7121 ± 15% sched_debug.cfs_rq[1]:/.tg_load_avg 431.50 ± 14% +228.2% 1416 ± 5% sched_debug.cfs_rq[2]:/.blocked_load_avg 33.75 ± 11% -18.5% 27.50 ± 11% sched_debug.cfs_rq[2]:/.nr_spread_over 818.50 ± 6% +118.3% 1786 ± 5% sched_debug.cfs_rq[2]:/.tg_load_contrib 301.25 ± 14% +38.5% 417.25 ± 16% sched_debug.cfs_rq[3]:/.load 5609 ± 10% +21.4% 6807 ± 16% sched_debug.cfs_rq[3]:/.tg_load_avg 607848 ± 32% +35.5% 823868 ± 6% sched_debug.cpu#0.avg_idle 435380 ± 40% -52.0% 208833 ± 35% sched_debug.cpu#0.nr_switches 435948 ± 40% -52.0% 209079 ± 35% sched_debug.cpu#0.sched_count 1118781 ±122% -82.2% 198660 ± 26% sched_debug.cpu#1.nr_switches 1119297 ±122% -82.2% 198883 ± 26% sched_debug.cpu#1.sched_count 649029 ±121% -77.9% 143430 ± 19% sched_debug.cpu#1.ttwu_count 594549 ±120% -77.7% 132447 ± 20% sched_debug.cpu#1.ttwu_local 273528 ± 94% +214.0% 858787 ± 11% sched_debug.cpu#2.avg_idle 2797014 ± 51% -92.5% 209149 ± 21% sched_debug.cpu#2.nr_switches 2797284 ± 51% -92.5% 209354 ± 21% sched_debug.cpu#2.sched_count 1635953 ± 51% -90.9% 148709 ± 15% sched_debug.cpu#2.ttwu_count 1457289 ± 50% -90.5% 137841 ± 15% sched_debug.cpu#2.ttwu_local 331.25 ± 11% +26.0% 417.50 ± 15% sched_debug.cpu#3.load 617159 ± 28% -70.0% 185105 ± 33% sched_debug.cpu#3.nr_switches 617564 ± 28% -70.0% 185318 ± 33% sched_debug.cpu#3.sched_count 271272 ± 20% -49.9% 135909 ± 21% sched_debug.cpu#3.ttwu_count 254215 ± 19% -50.3% 126227 ± 22% sched_debug.cpu#3.ttwu_local ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test/cluster: wsm/netpipe/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/tcp/cs-localhost commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 9.39 ± 1% -7.6% 8.68 ± 1% netpipe.less_8K_usec.avg 17994 ± 3% -19.9% 14406 ± 5% softirqs.RCU 0.29 ± 4% -13.7% 0.25 ± 6% time.user_time 106713 ± 1% -67.9% 34294 ± 4% vmstat.system.cs 37772 ± 0% -95.5% 1693 ± 0% vmstat.system.in 87490550 ± 4% -54.1% 40120043 ± 2% cpuidle.C1-NHM.time 6115692 ± 1% -72.2% 1698584 ± 4% cpuidle.C1-NHM.usage 2583 ± 5% -93.4% 170.00 ± 23% cpuidle.POLL.usage 16.39 ± 3% -10.9% 14.59 ± 2% turbostat.%Busy 585.25 ± 2% -9.7% 528.25 ± 2% turbostat.Avg_MHz 26.28 ± 2% -25.9% 19.47 ± 2% turbostat.CPU%c1 56.33 ± 0% +14.6% 64.54 ± 1% turbostat.CPU%c6 2320 ± 6% -10.0% 2087 ± 2% sched_debug.cfs_rq[0]:/.tg->runnable_avg 2328 ± 5% -9.6% 2104 ± 2% sched_debug.cfs_rq[10]:/.tg->runnable_avg 2329 ± 5% -9.4% 2109 ± 2% sched_debug.cfs_rq[11]:/.tg->runnable_avg 8.50 ± 50% -64.7% 3.00 ±-33% sched_debug.cfs_rq[1]:/.nr_spread_over 3012 ±687% +900.1% 30128 ±106% sched_debug.cfs_rq[1]:/.spread0 2326 ± 6% -10.2% 2088 ± 2% sched_debug.cfs_rq[1]:/.tg->runnable_avg 2326 ± 6% -10.0% 2094 ± 2% sched_debug.cfs_rq[2]:/.tg->runnable_avg 2327 ± 6% -9.9% 2096 ± 2% sched_debug.cfs_rq[3]:/.tg->runnable_avg 2330 ± 6% -10.0% 2096 ± 2% sched_debug.cfs_rq[4]:/.tg->runnable_avg 2326 ± 6% -9.8% 2097 ± 2% sched_debug.cfs_rq[5]:/.tg->runnable_avg 2326 ± 6% -9.6% 2102 ± 2% sched_debug.cfs_rq[6]:/.tg->runnable_avg 34300 ± 99% -96.8% 1105 ± 16% sched_debug.cfs_rq[7]:/.min_vruntime 2326 ± 6% -9.6% 2102 ± 2% sched_debug.cfs_rq[7]:/.tg->runnable_avg 2326 ± 6% -9.6% 2103 ± 2% sched_debug.cfs_rq[8]:/.tg->runnable_avg 2328 ± 5% -9.7% 2103 ± 2% sched_debug.cfs_rq[9]:/.tg->runnable_avg 29.00 ±141% +330.2% 124.75 ± 65% sched_debug.cpu#2.cpu_load[1] 520151 ±141% -87.1% 66855 ± 50% sched_debug.cpu#2.ttwu_count 52023 ± 71% -88.3% 6101 ± 92% sched_debug.cpu#6.nr_switches 52035 ± 71% -88.3% 6112 ± 92% sched_debug.cpu#6.sched_count 25756 ± 72% -89.0% 2837 ±100% sched_debug.cpu#6.sched_goidle ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/group: lkp-t410/piglit/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/igt-044 commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 83.99 ± 0% -14.7% 71.66 ± 0% piglit.time.elapsed_time 83.99 ± 0% -14.7% 71.66 ± 0% piglit.time.elapsed_time.max 3034690 ± 0% -97.2% 85516 ± 2% piglit.time.involuntary_context_switches 352.25 ± 0% +8.9% 383.75 ± 0% piglit.time.percent_of_cpu_this_job_got 285.25 ± 0% -5.8% 268.60 ± 0% piglit.time.user_time 16.75 ± 13% -25.4% 12.50 ± 12% vmstat.procs.r 85019 ± 0% -86.1% 11835 ± 3% vmstat.system.cs 42758 ± 0% -88.2% 5029 ± 0% vmstat.system.in 83.99 ± 0% -14.7% 71.66 ± 0% time.elapsed_time 83.99 ± 0% -14.7% 71.66 ± 0% time.elapsed_time.max 3034690 ± 0% -97.2% 85516 ± 2% time.involuntary_context_switches 11.15 ± 0% -40.4% 6.65 ± 0% time.system_time ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/nr_threads/blocksize: lkp-sb02/pigz/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/100%/512K commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 65777721 ± 0% +1.5% 66758347 ± 0% pigz.throughput 4731453 ± 0% -95.3% 223426 ± 2% pigz.time.involuntary_context_switches 17391 ± 1% +2.3% 17800 ± 0% pigz.time.minor_page_faults 394.00 ± 0% +1.0% 398.00 ± 0% pigz.time.percent_of_cpu_this_job_got 1170 ± 0% +1.3% 1185 ± 0% pigz.time.user_time 244236 ± 0% +2.8% 250954 ± 0% pigz.time.voluntary_context_switches 34446 ± 32% -99.6% 129.50 ± 60% latency_stats.sum.pipe_read.__vfs_read.vfs_read.SyS_read.entry_SYSCALL_64_fastpath 13004 ± 2% -6.2% 12197 ± 4% meminfo.AnonHugePages 4731453 ± 0% -95.3% 223426 ± 2% time.involuntary_context_switches 13.98 ± 0% -13.6% 12.07 ± 0% time.system_time 0.06 ± 9% -81.8% 0.01 ± 0% turbostat.CPU%c1 0.14 ± 3% +27.3% 0.18 ± 2% turbostat.CPU%c6 33869 ± 0% -89.4% 3581 ± 2% vmstat.system.cs 19774 ± 0% -76.7% 4609 ± 0% vmstat.system.in 591398 ± 9% -94.5% 32236 ± 4% cpuidle.C1-SNB.time 59726 ± 10% -97.7% 1370 ± 84% cpuidle.C1-SNB.usage 2880102 ± 1% +17.3% 3377630 ± 0% cpuidle.C6-SNB.time 76.75 ± 10% -100.0% 0.00 ± 0% cpuidle.POLL.time 43.25 ± 18% -100.0% 0.00 ± 0% cpuidle.POLL.usage 0.74 ± 33% -59.1% 0.30 ± 75% perf-profile.cpu-cycles.call_timer_fn.run_timer_softirq.__do_softirq.irq_exit.smp_apic_timer_interrupt 1.14 ± 72% -100.0% 0.00 ± -1% perf-profile.cpu-cycles.copy_process.part.27._do_fork.sys_clone.entry_SYSCALL_64_fastpath 0.00 ± -1% +Inf% 0.56 ± 56% perf-profile.cpu-cycles.copy_process.part.28._do_fork.sys_clone.entry_SYSCALL_64_fastpath 0.95 ± 21% +33.9% 1.27 ± 3% perf-profile.cpu-cycles.free_hot_cold_page.put_page.anon_pipe_buf_release.pipe_read.__vfs_read 0.51 ± 43% +243.1% 1.75 ± 34% perf-profile.cpu-cycles.kthread.ret_from_fork 0.26 ± 66% +316.3% 1.08 ± 39% perf-profile.cpu-cycles.rcu_nocb_kthread.kthread.ret_from_fork 0.51 ± 43% +243.1% 1.75 ± 34% perf-profile.cpu-cycles.ret_from_fork 198.25 ± 10% +26.1% 250.00 ± 9% sched_debug.cfs_rq[1]:/.runnable_load_avg 26484 ± 49% -107.5% -1995 ±-124% sched_debug.cfs_rq[2]:/.spread0 26370 ± 45% -105.4% -1415 ±-116% sched_debug.cfs_rq[3]:/.spread0 5582478 ± 33% -97.5% 142112 ± 16% sched_debug.cpu#0.nr_switches 5582876 ± 33% -97.4% 142504 ± 16% sched_debug.cpu#0.sched_count 40451 ± 39% -87.2% 5176 ± 26% sched_debug.cpu#0.sched_goidle 2803309 ± 33% -97.2% 77183 ± 14% sched_debug.cpu#0.ttwu_count 2781915 ± 33% -98.0% 55288 ± 16% sched_debug.cpu#0.ttwu_local 2354598 ± 80% -93.5% 153545 ± 7% sched_debug.cpu#1.nr_switches 2354718 ± 80% -93.5% 153683 ± 7% sched_debug.cpu#1.sched_count 1185125 ± 79% -93.2% 80529 ± 10% sched_debug.cpu#1.ttwu_count 1164849 ± 81% -94.8% 60184 ± 14% sched_debug.cpu#1.ttwu_local 3834 ± 21% -35.2% 2483 ± 8% sched_debug.cpu#2.sched_goidle ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/test: nhm-white/unixbench/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/shell1 commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 2494800 ± 0% -6.0% 2345617 ± 0% unixbench.time.involuntary_context_switches 1.085e+09 ± 0% -2.4% 1.059e+09 ± 0% latency_stats.sum.do_wait.SyS_wait4.entry_SYSCALL_64_fastpath 52548 ± 5% -33.6% 34875 ± 0% vmstat.system.cs 14831 ± 9% -59.3% 6034 ± 0% vmstat.system.in 47.37 ± 0% -2.0% 46.41 ± 0% turbostat.%Busy 1360 ± 0% -2.9% 1321 ± 0% turbostat.Avg_MHz 7.69 ± 0% +16.3% 8.94 ± 3% turbostat.CPU%c6 3.51 ± 0% +31.3% 4.61 ± 6% turbostat.Pkg%pc6 65744708 ± 2% -50.6% 32458256 ± 0% cpuidle.C1-NHM.time 3629185 ± 14% -88.0% 435580 ± 2% cpuidle.C1-NHM.usage 4.728e+08 ± 0% +13.3% 5.358e+08 ± 2% cpuidle.C6-NHM.time 602583 ± 0% +11.6% 672654 ± 0% cpuidle.C6-NHM.usage 141592 ± 20% -75.6% 34556 ± 25% cpuidle.POLL.time 14080 ± 14% -91.1% 1252 ± 2% cpuidle.POLL.usage 18338 ± 14% +16.4% 21343 ± 4% sched_debug.cfs_rq[2]:/.blocked_load_avg 18489 ± 14% +16.1% 21467 ± 4% sched_debug.cfs_rq[2]:/.tg_load_contrib 146020 ± 5% -7.3% 135370 ± 0% sched_debug.cpu#0.nr_load_updates 3000951 ± 78% -78.9% 633028 ± 2% sched_debug.cpu#0.nr_switches 3001380 ± 78% -78.9% 633584 ± 2% sched_debug.cpu#0.sched_count 1354868 ± 82% -82.8% 233428 ± 3% sched_debug.cpu#0.sched_goidle 1426794 ± 82% -83.9% 230146 ± 1% sched_debug.cpu#0.ttwu_count 1287294 ± 92% -93.0% 90490 ± 0% sched_debug.cpu#0.ttwu_local 7751 ± 16% +37.1% 10628 ± 6% sched_debug.cpu#1.curr->pid 685476 ± 64% -65.6% 235696 ± 4% sched_debug.cpu#1.ttwu_count 535746 ± 83% -83.4% 88784 ± 1% sched_debug.cpu#1.ttwu_local 69.00 ± 4% -15.9% 58.00 ± 13% sched_debug.cpu#2.cpu_load[2] 69.00 ± 3% -17.8% 56.75 ± 10% sched_debug.cpu#2.cpu_load[3] 68.25 ± 4% -17.2% 56.50 ± 10% sched_debug.cpu#2.cpu_load[4] -69.25 ±-14% -33.2% -46.25 ±-20% sched_debug.cpu#3.nr_uninterruptible 7500 ± 19% +62.1% 12161 ± 17% sched_debug.cpu#4.curr->pid 219010 ± 3% -8.1% 201264 ± 3% sched_debug.cpu#4.sched_goidle 93751 ± 5% -7.3% 86883 ± 0% sched_debug.cpu#4.ttwu_local 137580 ± 4% -5.7% 129702 ± 0% sched_debug.cpu#5.nr_load_updates 2993388 ± 79% -80.6% 581607 ± 1% sched_debug.cpu#5.nr_switches 2993731 ± 79% -80.6% 581990 ± 1% sched_debug.cpu#5.sched_count 1373366 ± 84% -85.1% 204662 ± 1% sched_debug.cpu#5.sched_goidle 1296037 ± 93% -93.0% 90170 ± 0% sched_debug.cpu#5.ttwu_local 78.50 ± 28% -43.0% 44.75 ± 45% sched_debug.cpu#7.nr_uninterruptible 0.01 ± 59% +21923.0% 1.25 ±171% sched_debug.rt_rq[3]:/.rt_time 0.00 ± 58% +7898.3% 0.38 ±169% sched_debug.rt_rq[7]:/.rt_time ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/test: nhm-white/unixbench/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/shell8 commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 5856190 ± 1% -25.5% 4360557 ± 0% unixbench.time.involuntary_context_switches 5856190 ± 1% -25.5% 4360557 ± 0% time.involuntary_context_switches 3.14 ± 3% +29.0% 4.05 ± 2% turbostat.Pkg%pc6 61840 ± 2% -26.2% 45627 ± 0% vmstat.system.cs 18904 ± 3% -44.2% 10543 ± 0% vmstat.system.in 44596831 ± 2% -34.1% 29385365 ± 1% cpuidle.C1-NHM.time 1853539 ± 8% -87.6% 229126 ± 1% cpuidle.C1-NHM.usage 246889 ± 17% -90.1% 24414 ± 38% cpuidle.POLL.time 31710 ± 8% -96.8% 1025 ± 3% cpuidle.POLL.usage 24168 ± 12% +18.9% 28748 ± 4% sched_debug.cfs_rq[0]:/.blocked_load_avg 24325 ± 13% +18.8% 28899 ± 3% sched_debug.cfs_rq[0]:/.tg_load_contrib 91.50 ± 13% +31.7% 120.50 ± 21% sched_debug.cfs_rq[1]:/.runnable_load_avg 83.00 ± 7% +29.2% 107.25 ± 14% sched_debug.cfs_rq[5]:/.runnable_load_avg 70.75 ± 18% +31.1% 92.75 ± 11% sched_debug.cfs_rq[7]:/.runnable_load_avg 7677 ± 34% +46.0% 11210 ± 10% sched_debug.cpu#2.curr->pid 389892 ± 17% +27.3% 496226 ± 13% sched_debug.cpu#4.avg_idle 6249 ± 27% +87.1% 11693 ± 7% sched_debug.cpu#6.curr->pid ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test: wsm/will-it-scale/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/getppid1 commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 18473072 ± 3% -11.3% 16379556 ± 0% will-it-scale.per_process_ops 18119608 ± 1% -9.5% 16402159 ± 0% will-it-scale.per_thread_ops 0.74 ± 7% +18.5% 0.87 ± 2% will-it-scale.scalability 628.18 ± 0% +4.4% 655.83 ± 0% will-it-scale.time.system_time 434.65 ± 0% -6.3% 407.30 ± 0% will-it-scale.time.user_time 22243 ± 4% -3.3% 21501 ± 4% meminfo.AnonPages 5560 ± 4% -3.3% 5374 ± 4% proc-vmstat.nr_anon_pages 23934 ± 4% -22.1% 18653 ± 3% softirqs.SCHED 18.03 ± 7% +14.8% 20.69 ± 0% turbostat.CPU%c6 1802 ± 2% +21.0% 2181 ± 4% slabinfo.kmalloc-512.active_objs 1802 ± 2% +22.4% 2206 ± 3% slabinfo.kmalloc-512.num_objs 30351 ± 0% -93.9% 1837 ± 0% vmstat.system.cs 21766 ± 0% -65.9% 7418 ± 0% vmstat.system.in 49070159 ± 12% -85.0% 7354447 ± 24% cpuidle.C1-NHM.time 4289833 ± 8% -97.5% 107531 ± 5% cpuidle.C1-NHM.usage 1606 ± 7% -27.1% 1171 ± 10% cpuidle.POLL.usage 21.60 ± 1% -22.8% 16.68 ± 0% perf-profile.cpu-cycles.entry_SYSCALL_64 30.96 ± 2% -22.8% 23.90 ± 0% perf-profile.cpu-cycles.entry_SYSCALL_64_after_swapgs 44.96 ± 2% +25.8% 56.58 ± 2% perf-profile.cpu-cycles.entry_SYSCALL_64_fastpath 9.80 ± 16% +156.6% 25.16 ± 2% perf-profile.cpu-cycles.pid_vnr.entry_SYSCALL_64_fastpath 6.40 ± 13% -50.6% 3.16 ± 4% perf-profile.cpu-cycles.pid_vnr.sys_getppid.entry_SYSCALL_64_fastpath 11.04 ± 7% +15.1% 12.70 ± 2% perf-profile.cpu-cycles.sys_getppid.entry_SYSCALL_64_fastpath 13943 ± 10% +25.4% 17491 ± 6% sched_debug.cfs_rq[0]:/.tg_load_avg 13886 ± 10% +25.1% 17373 ± 6% sched_debug.cfs_rq[10]:/.tg_load_avg 13888 ± 10% +24.9% 17352 ± 6% sched_debug.cfs_rq[11]:/.tg_load_avg 2.25 ± 19% +66.7% 3.75 ± 22% sched_debug.cfs_rq[1]:/.nr_spread_over 13949 ± 10% +25.1% 17452 ± 6% sched_debug.cfs_rq[1]:/.tg_load_avg 13949 ± 10% +25.1% 17456 ± 6% sched_debug.cfs_rq[2]:/.tg_load_avg 13948 ± 10% +25.1% 17447 ± 6% sched_debug.cfs_rq[3]:/.tg_load_avg 13929 ± 10% +25.2% 17441 ± 6% sched_debug.cfs_rq[4]:/.tg_load_avg 13931 ± 10% +24.9% 17406 ± 6% sched_debug.cfs_rq[5]:/.tg_load_avg 13929 ± 10% +24.8% 17379 ± 6% sched_debug.cfs_rq[6]:/.tg_load_avg 13910 ± 10% +24.9% 17379 ± 6% sched_debug.cfs_rq[7]:/.tg_load_avg 13890 ± 10% +25.1% 17371 ± 6% sched_debug.cfs_rq[8]:/.tg_load_avg 13888 ± 10% +25.1% 17370 ± 6% sched_debug.cfs_rq[9]:/.tg_load_avg 144.75 ± 52% -47.8% 75.50 ± 31% sched_debug.cpu#1.cpu_load[0] 128.25 ± 38% -43.3% 72.75 ± 14% sched_debug.cpu#1.cpu_load[1] 106.25 ± 25% -33.4% 70.75 ± 7% sched_debug.cpu#1.cpu_load[2] 90.25 ± 15% -23.0% 69.50 ± 4% sched_debug.cpu#1.cpu_load[3] 82.00 ± 7% -15.9% 69.00 ± 3% sched_debug.cpu#1.cpu_load[4] 1005269 ±166% -98.4% 16346 ± 25% sched_debug.cpu#1.nr_switches 1005304 ±166% -98.4% 16385 ± 25% sched_debug.cpu#1.sched_count 499712 ±168% -98.6% 6885 ± 37% sched_debug.cpu#1.sched_goidle 1.50 ±145% +83.3% 2.75 ±150% sched_debug.cpu#10.nr_uninterruptible 2154 ± 14% -38.3% 1330 ± 32% sched_debug.cpu#3.curr->pid 1562 ± 33% +40.0% 2187 ± 9% sched_debug.cpu#4.curr->pid 42.25 ± 2% +31.4% 55.50 ± 20% sched_debug.cpu#7.cpu_load[1] 42.75 ± 3% +19.3% 51.00 ± 12% sched_debug.cpu#7.cpu_load[2] 43.50 ± 5% +13.2% 49.25 ± 10% sched_debug.cpu#7.cpu_load[3] 441090 ±167% -99.5% 2153 ± 51% sched_debug.cpu#7.ttwu_count 72.75 ± 32% -39.9% 43.75 ± 9% sched_debug.cpu#8.cpu_load[0] ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test: wsm/will-it-scale/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/open1 commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 801101 ± 4% +6.5% 853345 ± 0% will-it-scale.per_process_ops 0.56 ± 10% -11.5% 0.50 ± 0% will-it-scale.scalability 2004 ±100% +233.7% 6688 ±100% latency_stats.sum.wait_woken.inotify_read.__vfs_read.vfs_read.SyS_read.entry_SYSCALL_64_fastpath 40.09 ± 1% -9.6% 36.25 ± 0% time.user_time 357.00 ± 13% +32.4% 472.50 ± 7% slabinfo.mnt_cache.active_objs 357.00 ± 13% +32.4% 472.50 ± 7% slabinfo.mnt_cache.num_objs 90485 ± 2% -7.4% 83796 ± 0% softirqs.RCU 49819 ± 2% -10.7% 44472 ± 1% softirqs.SCHED 30306 ± 2% -91.1% 2685 ± 1% vmstat.system.cs 21928 ± 1% -63.7% 7960 ± 0% vmstat.system.in 41567129 ± 11% -90.4% 3994402 ± 30% cpuidle.C1-NHM.time 3844950 ± 16% -97.0% 115267 ± 3% cpuidle.C1-NHM.usage 16389120 ± 54% -56.8% 7074969 ± 15% cpuidle.C3-NHM.time 1491 ± 49% -91.1% 132.00 ± 17% cpuidle.POLL.usage 1.54 ± 3% -18.8% 1.26 ± 4% perf-profile.cpu-cycles.__call_rcu.call_rcu_sched.__fput.____fput.task_work_run 1.21 ± 5% -18.7% 0.98 ± 2% perf-profile.cpu-cycles.__fd_install.fd_install.do_sys_open.sys_open.entry_SYSCALL_64_fastpath 3.01 ± 4% +11.0% 3.34 ± 3% perf-profile.cpu-cycles.__inode_permission.inode_permission.may_open.path_openat.do_filp_open 1.81 ± 2% +11.1% 2.01 ± 3% perf-profile.cpu-cycles.__memset.get_empty_filp.path_openat.do_filp_open.do_sys_open 1.47 ± 2% -11.4% 1.31 ± 3% perf-profile.cpu-cycles.__slab_alloc.kmem_cache_alloc.get_empty_filp.path_openat.do_filp_open 3.61 ± 2% -10.2% 3.25 ± 1% perf-profile.cpu-cycles.call_rcu_sched.__fput.____fput.task_work_run.prepare_exit_to_usermode 10.03 ± 1% -7.9% 9.24 ± 1% perf-profile.cpu-cycles.do_dentry_open.vfs_open.path_openat.do_filp_open.do_sys_open 1.39 ± 4% -18.2% 1.14 ± 3% perf-profile.cpu-cycles.fd_install.do_sys_open.sys_open.entry_SYSCALL_64_fastpath 2.50 ± 2% +8.9% 2.72 ± 2% perf-profile.cpu-cycles.get_unused_fd_flags.do_sys_open.sys_open.entry_SYSCALL_64_fastpath 3.31 ± 5% +10.4% 3.66 ± 2% perf-profile.cpu-cycles.inode_permission.may_open.path_openat.do_filp_open.do_sys_open 1.15 ± 5% +9.8% 1.26 ± 3% perf-profile.cpu-cycles.kfree.selinux_file_free_security.security_file_free.__fput.____fput 2.14 ± 2% -13.6% 1.85 ± 2% perf-profile.cpu-cycles.kmem_cache_alloc_trace.selinux_file_alloc_security.security_file_alloc.get_empty_filp.path_openat 3.62 ± 2% -8.8% 3.30 ± 1% perf-profile.cpu-cycles.security_file_alloc.get_empty_filp.path_openat.do_filp_open.do_sys_open 1.45 ± 5% +11.9% 1.62 ± 3% perf-profile.cpu-cycles.security_file_free.__fput.____fput.task_work_run.prepare_exit_to_usermode 2.13 ± 6% +15.2% 2.46 ± 4% perf-profile.cpu-cycles.security_inode_permission.__inode_permission.inode_permission.may_open.path_openat 3.30 ± 2% -8.7% 3.01 ± 2% perf-profile.cpu-cycles.selinux_file_alloc_security.security_file_alloc.get_empty_filp.path_openat.do_filp_open 1.18 ± 5% +8.7% 1.29 ± 2% perf-profile.cpu-cycles.selinux_file_free_security.security_file_free.__fput.____fput.task_work_run 3.21 ± 3% +12.8% 3.63 ± 1% perf-profile.cpu-cycles.selinux_inode_permission.security_inode_permission.__inode_permission.inode_permission.link_path_walk 1.83 ± 7% +17.2% 2.15 ± 5% perf-profile.cpu-cycles.selinux_inode_permission.security_inode_permission.__inode_permission.inode_permission.may_open 10.79 ± 1% -9.4% 9.78 ± 1% perf-profile.cpu-cycles.vfs_open.path_openat.do_filp_open.do_sys_open.sys_open 1138 ± 30% +58.8% 1807 ± 9% sched_debug.cfs_rq[11]:/.blocked_load_avg 1199 ± 27% +57.1% 1884 ± 8% sched_debug.cfs_rq[11]:/.tg_load_contrib 3.00 ± 84% +233.3% 10.00 ± 7% sched_debug.cfs_rq[2]:/.nr_spread_over 30468 ± 13% -17.6% 25115 ± 3% sched_debug.cfs_rq[7]:/.avg->runnable_avg_sum 663.25 ± 13% -17.4% 547.75 ± 3% sched_debug.cfs_rq[7]:/.tg_runnable_contrib 3028 ± 59% -57.3% 1291 ± 56% sched_debug.cfs_rq[8]:/.blocked_load_avg 3.25 ± 50% +169.2% 8.75 ± 21% sched_debug.cfs_rq[8]:/.nr_spread_over 3092 ± 57% -56.7% 1340 ± 54% sched_debug.cfs_rq[8]:/.tg_load_contrib 109897 ± 9% -16.2% 92140 ± 3% sched_debug.cpu#0.nr_load_updates 97.75 ± 23% -32.7% 65.75 ± 14% sched_debug.cpu#1.cpu_load[0] 88.00 ± 10% -26.7% 64.50 ± 7% sched_debug.cpu#1.cpu_load[1] 2.00 ± 93% -325.0% -4.50 ±-114% sched_debug.cpu#1.nr_uninterruptible 62.25 ± 8% -12.9% 54.25 ± 13% sched_debug.cpu#11.cpu_load[4] 24550 ± 18% -31.1% 16923 ± 24% sched_debug.cpu#2.ttwu_count 12972 ± 9% -37.0% 8171 ± 30% sched_debug.cpu#2.ttwu_local 1038 ± 41% +92.4% 1998 ± 9% sched_debug.cpu#4.curr->pid 15880 ± 17% +19.8% 19022 ± 4% sched_debug.cpu#4.sched_goidle 1492076 ± 99% -99.4% 8330 ± 50% sched_debug.cpu#6.nr_switches 1492358 ± 99% -99.4% 8627 ± 48% sched_debug.cpu#6.sched_count 614310 ±100% -99.6% 2432 ± 64% sched_debug.cpu#6.sched_goidle 62.75 ± 16% -23.9% 47.75 ± 18% sched_debug.cpu#7.cpu_load[2] 59.50 ± 16% -22.7% 46.00 ± 12% sched_debug.cpu#7.cpu_load[3] 1.00 ±212% +775.0% 8.75 ± 28% sched_debug.cpu#8.nr_uninterruptible ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test: wsm/will-it-scale/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/pthread_mutex2 commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 25290 ± 8% -28.3% 18121 ± 3% softirqs.SCHED 30175 ± 0% -94.6% 1620 ± 5% vmstat.system.cs 21733 ± 0% -66.0% 7399 ± 0% vmstat.system.in 50171812 ± 9% -80.7% 9701068 ± 55% cpuidle.C1-NHM.time 4178650 ± 12% -98.0% 83770 ± 12% cpuidle.C1-NHM.usage 1465 ± 6% -27.9% 1057 ± 17% cpuidle.POLL.usage 60.19 ± 0% -1.5% 59.27 ± 0% turbostat.%Busy 2118 ± 0% -1.6% 2085 ± 0% turbostat.Avg_MHz 17.82 ± 6% +15.7% 20.61 ± 0% turbostat.CPU%c6 0.07 ±133% +1864.3% 1.38 ± 73% perf-profile.cpu-cycles.__libc_start_main 0.07 ±133% +1814.3% 1.34 ± 75% perf-profile.cpu-cycles._start.main.__libc_start_main 1.72 ±110% -100.0% 0.00 ± -1% perf-profile.cpu-cycles.copy_process.part.27._do_fork.sys_clone.entry_SYSCALL_64_fastpath 0.07 ±133% +1814.3% 1.34 ± 75% perf-profile.cpu-cycles.main.__libc_start_main 807.25 ± 17% -26.5% 593.25 ± 24% sched_debug.cfs_rq[3]:/.utilization_load_avg 2069 ± 45% -80.7% 398.50 ± 40% sched_debug.cfs_rq[4]:/.blocked_load_avg 2162 ± 45% -78.0% 475.25 ± 38% sched_debug.cfs_rq[4]:/.tg_load_contrib 419067 ±169% -98.3% 7041 ± 4% sched_debug.cpu#0.sched_goidle 81.50 ± 3% -22.1% 63.50 ± 22% sched_debug.cpu#1.cpu_load[2] 16437 ± 46% -76.4% 3884 ± 57% sched_debug.cpu#10.ttwu_count 80953 ± 18% -24.0% 61501 ± 14% sched_debug.cpu#2.nr_load_updates 68.50 ± 15% -27.7% 49.50 ± 18% sched_debug.cpu#3.cpu_load[0] 67.25 ± 8% -15.2% 57.00 ± 12% sched_debug.cpu#4.cpu_load[3] 65.75 ± 8% -16.7% 54.75 ± 10% sched_debug.cpu#4.cpu_load[4] 39032 ± 14% -38.8% 23893 ± 40% sched_debug.cpu#4.nr_switches 39071 ± 14% -38.8% 23926 ± 40% sched_debug.cpu#4.sched_count 19248 ± 12% -42.4% 11089 ± 26% sched_debug.cpu#4.ttwu_count 9496 ± 6% -67.8% 3057 ± 30% sched_debug.cpu#4.ttwu_local ========================================================================================= tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test: wsm/will-it-scale/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/readseek3 commit: cad8c9e5d6a97898d37b1a8e5cdf838d79ba6e50 62c79204783e188291d880f23d49c02d8c8f498b cad8c9e5d6a97898 62c79204783e188291d880f23d ---------------- -------------------------- %stddev %change %stddev \ | \ 2439605 ± 0% +1.8% 2484631 ± 0% will-it-scale.per_process_ops 1842733 ± 1% +5.8% 1949535 ± 1% will-it-scale.per_thread_ops 0.54 ± 0% -1.8% 0.53 ± 1% will-it-scale.scalability 14706 ± 31% +52.6% 22440 ± 16% will-it-scale.time.involuntary_context_switches 136239 ± 2% +44.4% 196725 ± 3% latency_stats.hits.pipe_wait.pipe_read.__vfs_read.vfs_read.SyS_read.entry_SYSCALL_64_fastpath 7.40 ± 3% -9.7% 6.68 ± 2% perf-profile.cpu-cycles.rw_verify_area.vfs_read.sys_read.entry_SYSCALL_64_fastpath 25530 ± 1% -19.8% 20467 ± 1% softirqs.SCHED 14706 ± 31% +52.6% 22440 ± 16% time.involuntary_context_switches 30169 ± 0% -93.6% 1945 ± 2% vmstat.system.cs 21810 ± 0% -65.8% 7449 ± 0% vmstat.system.in 59.90 ± 0% -1.5% 59.02 ± 0% turbostat.%Busy 2108 ± 0% -1.5% 2076 ± 0% turbostat.Avg_MHz 17.43 ± 0% +17.6% 20.50 ± 0% turbostat.CPU%c6 46658026 ± 4% -77.3% 10588249 ± 14% cpuidle.C1-NHM.time 4461627 ± 0% -97.5% 112631 ± 5% cpuidle.C1-NHM.usage 135274 ± 85% +363.4% 626895 ± 80% cpuidle.C1E-NHM.time 103.25 ± 12% +71.4% 177.00 ± 20% cpuidle.C1E-NHM.usage 38945629 ± 6% -29.3% 27546659 ± 19% cpuidle.POLL.time 1523 ± 10% -26.2% 1124 ± 3% cpuidle.POLL.usage 14514 ± 6% +21.2% 17593 ± 6% sched_debug.cfs_rq[0]:/.tg_load_avg -69524 ±-63% +63.0% -113345 ±-40% sched_debug.cfs_rq[10]:/.spread0 14486 ± 6% +20.9% 17508 ± 6% sched_debug.cfs_rq[10]:/.tg_load_avg 14468 ± 6% +21.0% 17508 ± 6% sched_debug.cfs_rq[11]:/.tg_load_avg 58.25 ± 7% +32.6% 77.25 ± 25% sched_debug.cfs_rq[1]:/.runnable_load_avg 14520 ± 6% +21.1% 17583 ± 6% sched_debug.cfs_rq[1]:/.tg_load_avg 14468 ± 6% +21.4% 17560 ± 6% sched_debug.cfs_rq[2]:/.tg_load_avg 14471 ± 6% +21.3% 17557 ± 6% sched_debug.cfs_rq[3]:/.tg_load_avg 14483 ± 6% +21.1% 17545 ± 6% sched_debug.cfs_rq[4]:/.tg_load_avg 14489 ± 6% +20.8% 17503 ± 6% sched_debug.cfs_rq[5]:/.tg_load_avg 14493 ± 6% +20.7% 17487 ± 6% sched_debug.cfs_rq[6]:/.tg_load_avg 52254 ± 84% -94.8% 2722 ±1818% sched_debug.cfs_rq[7]:/.spread0 14492 ± 6% +20.6% 17481 ± 6% sched_debug.cfs_rq[7]:/.tg_load_avg 14492 ± 6% +20.9% 17523 ± 6% sched_debug.cfs_rq[8]:/.tg_load_avg -34317 ±-276% +228.6% -112775 ±-40% sched_debug.cfs_rq[9]:/.spread0 14486 ± 6% +20.9% 17508 ± 6% sched_debug.cfs_rq[9]:/.tg_load_avg 14454 ± 6% +32.2% 19105 ± 13% sched_debug.cpu#0.ttwu_count 11838 ± 20% -37.9% 7351 ± 30% sched_debug.cpu#1.ttwu_count -6.75 ±-36% -96.3% -0.25 ±-1424% sched_debug.cpu#4.nr_uninterruptible 6151 ± 23% +113.3% 13119 ± 24% sched_debug.cpu#6.nr_switches 6185 ± 23% +112.6% 13147 ± 24% sched_debug.cpu#6.sched_count 1375 ± 29% +242.6% 4711 ± 35% sched_debug.cpu#6.sched_goidle 4277 ± 21% +88.5% 8063 ± 19% sched_debug.cpu#6.ttwu_count 2263 ± 12% -19.2% 1829 ± 0% sched_debug.cpu#9.curr->pid 72929 ± 11% -20.4% 58086 ± 0% sched_debug.cpu#9.nr_load_updates 356400 ±171% -99.6% 1252 ± 11% sched_debug.cpu#9.ttwu_local 0.05 ±100% +430.4% 0.28 ± 98% sched_debug.rt_rq[1]:/.rt_time lkp-a06: Atom Memory: 8G lkp-sb02: Sandy Bridge-EP Memory: 4G nhm4: Nehalem Memory: 4G lituya: Grantley Haswell Memory: 16G wsm: Westmere Memory: 6G lkp-t410: Westmere Memory: 2G nhm-white: Nehalem Memory: 6G To reproduce: git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git cd lkp-tests bin/lkp install job.yaml # job file is attached in this email bin/lkp run job.yaml Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. Thanks, Ying Huang
--- LKP_SERVER: inn LKP_CGI_PORT: 80 LKP_CIFS_PORT: 139 testcase: fsmark default-monitors: wait: activate-monitor kmsg: uptime: iostat: vmstat: numa-numastat: numa-vmstat: numa-meminfo: proc-vmstat: proc-stat: interval: 10 meminfo: slabinfo: interrupts: lock_stat: latency_stats: softirqs: bdi_dev_mapping: diskstats: nfsstat: cpuidle: cpufreq-stats: turbostat: pmeter: sched_debug: interval: 60 cpufreq_governor: performance default-watchdogs: oom-killer: watchdog: commit: ba19772fefe56fca1548d91218c3aeca8f207919 model: Sandy Bridge-EP memory: 4G hdd_partitions: "/dev/disk/by-id/ata-HDT722516DLA380_VDSD1DTCEKYAEJ-part2" swap_partitions: rootfs_partition: "/dev/disk/by-id/ata-HDT722516DLA380_VDSD1DTCEKYAEJ-part1" category: benchmark iterations: 1x nr_threads: 32t disk: 1HDD fs: btrfs fs2: fsmark: filesize: 9B test_size: 400M sync_method: fsyncBeforeClose nr_directories: 16d nr_files_per_directory: 256fpd queue: cyclic testbox: lkp-sb02 tbox_group: lkp-sb02 kconfig: x86_64-rhel enqueue_time: 2015-07-08 12:46:57.152422017 +08:00 user: wfg compiler: gcc-4.9 head_commit: ba19772fefe56fca1548d91218c3aeca8f207919 base_commit: d770e558e21961ad6cfdf0ff7df0eb5d7d4f0754 branch: linux-devel/devel-hourly-2015071021 kernel: "/pkg/linux/x86_64-rhel/gcc-4.9/ba19772fefe56fca1548d91218c3aeca8f207919/vmlinuz-4.2.0-rc1-wl-00882-gba19772" rootfs: debian-x86_64-2015-02-07.cgz result_root: "/result/fsmark/performance-1x-32t-1HDD-btrfs-9B-400M-fsyncBeforeClose-16d-256fpd/lkp-sb02/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/ba19772fefe56fca1548d91218c3aeca8f207919/0" job_file: "/lkp/scheduled/lkp-sb02/cyclic_fsmark-performance-1x-32t-1HDD-btrfs-9B-400M-fsyncBeforeClose-16d-256fpd-x86_64-rhel-CYCLIC_HEAD-ba19772fefe56fca1548d91218c3aeca8f207919-20150708-88967-m5o3fs-0.yaml" dequeue_time: 2015-07-11 12:17:26.250837417 +08:00 nr_cpu: "$(nproc)" max_uptime: 1063.0800000000002 initrd: "/osimage/debian/debian-x86_64-2015-02-07.cgz" bootloader_append: - root=/dev/ram0 - user=wfg - job=/lkp/scheduled/lkp-sb02/cyclic_fsmark-performance-1x-32t-1HDD-btrfs-9B-400M-fsyncBeforeClose-16d-256fpd-x86_64-rhel-CYCLIC_HEAD-ba19772fefe56fca1548d91218c3aeca8f207919-20150708-88967-m5o3fs-0.yaml - ARCH=x86_64 - kconfig=x86_64-rhel - branch=linux-devel/devel-hourly-2015071021 - commit=ba19772fefe56fca1548d91218c3aeca8f207919 - BOOT_IMAGE=/pkg/linux/x86_64-rhel/gcc-4.9/ba19772fefe56fca1548d91218c3aeca8f207919/vmlinuz-4.2.0-rc1-wl-00882-gba19772 - max_uptime=1063 - RESULT_ROOT=/result/fsmark/performance-1x-32t-1HDD-btrfs-9B-400M-fsyncBeforeClose-16d-256fpd/lkp-sb02/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/ba19772fefe56fca1548d91218c3aeca8f207919/0 - LKP_SERVER=inn - |2- earlyprintk=ttyS0,115200 systemd.log_level=err debug apic=debug sysrq_always_enabled rcupdate.rcu_cpu_stall_timeout=100 panic=-1 softlockup_panic=1 nmi_watchdog=panic oops=panic load_ramdisk=2 prompt_ramdisk=0 console=ttyS0,115200 console=tty0 vga=normal rw lkp_initrd: "/lkp/wfg/lkp-x86_64.cgz" modules_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/ba19772fefe56fca1548d91218c3aeca8f207919/modules.cgz" bm_initrd: "/osimage/deps/debian-x86_64-2015-02-07.cgz/lkp.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/run-ipconfig.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/turbostat.cgz,/lkp/benchmarks/turbostat.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/fs.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/fs2.cgz,/lkp/benchmarks/fsmark.cgz" job_state: finished loadavg: 27.66 13.44 5.22 1/170 3028 start_time: '1436588285' end_time: '1436588450' version: "/lkp/wfg/.src-20150711-100505"
echo performance > /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu1/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu2/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu3/cpufreq/scaling_governor mkfs -t btrfs /dev/sda2 mount -t btrfs /dev/sda2 /fs/sda2 ./fs_mark -d /fs/sda2/1 -d /fs/sda2/2 -d /fs/sda2/3 -d /fs/sda2/4 -d /fs/sda2/5 -d /fs/sda2/6 -d /fs/sda2/7 -d /fs/sda2/8 -d /fs/sda2/9 -d /fs/sda2/10 -d /fs/sda2/11 -d /fs/sda2/12 -d /fs/sda2/13 -d /fs/sda2/14 -d /fs/sda2/15 -d /fs/sda2/16 -d /fs/sda2/17 -d /fs/sda2/18 -d /fs/sda2/19 -d /fs/sda2/20 -d /fs/sda2/21 -d /fs/sda2/22 -d /fs/sda2/23 -d /fs/sda2/24 -d /fs/sda2/25 -d /fs/sda2/26 -d /fs/sda2/27 -d /fs/sda2/28 -d /fs/sda2/29 -d /fs/sda2/30 -d /fs/sda2/31 -d /fs/sda2/32 -D 16 -N 256 -n 3200 -L 1 -S 1 -s 9