FYI, we noticed the below changes on git://git.kernel.org/pub/scm/linux/kernel/git/luto/linux.git x86/entry commit 103dd9885d1e96be0c7493ce9a33af0732e35441 ("x86_64, entry: Use sysret to return to userspace when possible")
testbox/testcase/testparams: lkp-sb03/will-it-scale/open2 261ebee6aee55a47 103dd9885d1e96be0c7493ce9a ---------------- -------------------------- %stddev %change %stddev \ | \ 35.96 ± 6% -39.8% 21.66 ± 7% will-it-scale.time.user_time 237495 ± 1% +5.2% 249861 ± 1% will-it-scale.per_thread_ops 739981 ± 0% +7.3% 793911 ± 0% will-it-scale.per_process_ops 127 ± 20% -66.9% 42 ± 30% sched_debug.cfs_rq[14]:/.load 127 ± 20% -66.9% 42 ± 30% sched_debug.cpu#14.load 58 ± 21% -39.3% 35 ± 43% sched_debug.cfs_rq[7]:/.load 58 ± 21% -39.3% 35 ± 43% sched_debug.cpu#7.load 3442 ± 25% -42.5% 1978 ± 20% sched_debug.cpu#28.sched_goidle 3591 ± 36% +128.5% 8204 ± 17% sched_debug.cpu#10.ttwu_local 8241 ± 21% -36.9% 5202 ± 19% sched_debug.cpu#28.nr_switches 5411 ± 18% -39.0% 3301 ± 13% sched_debug.cpu#14.ttwu_local 8800 ± 15% -27.8% 6356 ± 20% sched_debug.cpu#9.ttwu_count 42 ± 22% -33.3% 28 ± 47% sched_debug.cfs_rq[14]:/.runnable_load_avg 67 ± 46% +87.1% 126 ± 31% sched_debug.cpu#12.load 573932 ± 1% +59.5% 915353 ± 1% slabinfo.kmalloc-256.active_objs 8970 ± 1% +59.5% 14305 ± 1% slabinfo.kmalloc-256.num_slabs 8970 ± 1% +59.5% 14305 ± 1% slabinfo.kmalloc-256.active_slabs 574152 ± 1% +59.5% 915575 ± 1% slabinfo.kmalloc-256.num_objs 35.96 ± 6% -39.8% 21.66 ± 7% time.user_time 1358 ± 24% -31.9% 924 ± 34% sched_debug.cpu#31.ttwu_local 2487 ± 14% -32.6% 1677 ± 9% sched_debug.cpu#31.sched_goidle 17643 ± 8% -33.6% 11719 ± 5% sched_debug.cpu#14.nr_switches 8632 ± 11% -34.9% 5617 ± 14% sched_debug.cpu#14.ttwu_count 29 ± 12% +31.9% 38 ± 15% sched_debug.cpu#12.cpu_load[1] 115453 ± 3% +48.3% 171178 ± 2% numa-meminfo.node0.SUnreclaim 28672 ± 2% +49.2% 42775 ± 2% numa-vmstat.node0.nr_slab_unreclaimable 3313 ± 15% +35.0% 4472 ± 13% sched_debug.cpu#11.curr->pid 132090 ± 3% +42.8% 188646 ± 2% numa-meminfo.node0.Slab 200693 ± 1% +42.8% 286535 ± 0% meminfo.SUnreclaim 50431 ± 1% +42.0% 71624 ± 1% proc-vmstat.nr_slab_unreclaimable 26 ± 12% +29.8% 33 ± 12% sched_debug.cpu#12.cpu_load[4] 20434 ± 2% +38.4% 28286 ± 2% numa-vmstat.node1.nr_slab_unreclaimable 82040 ± 2% +36.9% 112273 ± 2% numa-meminfo.node1.SUnreclaim 14623 ± 26% +55.6% 22751 ± 12% sched_debug.cpu#10.nr_switches 235487 ± 1% +36.5% 321407 ± 0% meminfo.Slab 842597 ± 20% -30.2% 588471 ± 26% sched_debug.cfs_rq[27]:/.min_vruntime 6502 ± 25% +54.2% 10028 ± 11% sched_debug.cpu#10.sched_goidle 100197 ± 2% +29.4% 129675 ± 2% numa-meminfo.node1.Slab 237178 ± 1% -23.2% 182260 ± 2% softirqs.RCU 7310 ± 4% -26.2% 5397 ± 6% sched_debug.cpu#14.sched_goidle 6148 ± 12% -22.7% 4751 ± 10% sched_debug.cpu#31.nr_switches 44905 ± 20% -28.5% 32101 ± 25% sched_debug.cfs_rq[27]:/.exec_clock 32 ± 6% +38.3% 44 ± 19% sched_debug.cpu#1.cpu_load[0] 139 ± 15% -25.4% 104 ± 18% sched_debug.cfs_rq[23]:/.tg_load_contrib 3454 ± 14% -17.7% 2844 ± 6% sched_debug.cpu#31.curr->pid 438 ± 8% -11.5% 388 ± 5% sched_debug.cfs_rq[31]:/.tg_runnable_contrib 30 ± 1% +11.6% 33 ± 6% sched_debug.cpu#23.cpu_load[4] 20115 ± 7% -11.3% 17841 ± 6% sched_debug.cfs_rq[31]:/.avg->runnable_avg_sum 35 ± 11% -19.0% 28 ± 9% sched_debug.cpu#21.cpu_load[0] 32 ± 6% +31.2% 42 ± 19% sched_debug.cfs_rq[1]:/.runnable_load_avg 32 ± 11% +27.1% 41 ± 10% sched_debug.cpu#9.cpu_load[2] 30 ± 1% +9.9% 33 ± 6% sched_debug.cpu#23.cpu_load[3] 575 ± 5% +12.4% 647 ± 6% sched_debug.cfs_rq[23]:/.tg_runnable_contrib 26412 ± 5% +12.5% 29710 ± 6% sched_debug.cfs_rq[23]:/.avg->runnable_avg_sum 29 ± 18% +33.6% 38 ± 14% sched_debug.cpu#12.cpu_load[0] 32 ± 5% +32.1% 43 ± 17% sched_debug.cpu#1.cpu_load[1] 33 ± 6% +26.3% 42 ± 16% sched_debug.cpu#1.cpu_load[2] 424232 ± 2% +13.7% 482375 ± 1% numa-meminfo.node0.MemUsed 3558 ± 6% +14.5% 4076 ± 3% sched_debug.cpu#15.curr->pid 1567 ± 0% -9.3% 1422 ± 1% vmstat.system.cs testbox/testcase/testparams: lkp-sb03/will-it-scale/signal1 261ebee6aee55a47 103dd9885d1e96be0c7493ce9a ---------------- -------------------------- fail:runs %reproduction fail:runs | | | 40.59 ± 3% -33.0% 27.21 ± 3% will-it-scale.time.user_time 0.11 ± 0% -10.5% 0.10 ± 0% will-it-scale.scalability 297467 ± 0% +10.2% 327822 ± 0% will-it-scale.per_thread_ops 526508 ± 0% +7.2% 564555 ± 0% will-it-scale.per_process_ops 2527 ± 43% -74.9% 633 ± 20% sched_debug.cpu#24.ttwu_local 3355 ± 33% -66.8% 1113 ± 14% sched_debug.cpu#24.ttwu_count 6776 ± 38% -61.6% 2601 ± 11% sched_debug.cpu#24.nr_switches 47 ± 44% -59.6% 19 ± 12% sched_debug.cfs_rq[10]:/.load 48 ± 40% -60.8% 19 ± 12% sched_debug.cpu#10.load 184 ± 33% -63.1% 68 ± 5% sched_debug.cfs_rq[1]:/.blocked_load_avg 80 ± 36% +97.8% 158 ± 34% sched_debug.cfs_rq[15]:/.blocked_load_avg 98 ± 30% +77.7% 175 ± 31% sched_debug.cfs_rq[15]:/.tg_load_contrib 215 ± 28% -54.3% 98 ± 3% sched_debug.cfs_rq[1]:/.tg_load_contrib 212 ± 32% -60.7% 83 ± 39% sched_debug.cfs_rq[26]:/.blocked_load_avg 238 ± 34% -56.0% 104 ± 35% sched_debug.cfs_rq[26]:/.tg_load_contrib 1175 ± 15% +82.0% 2139 ± 36% sched_debug.cpu#30.sched_goidle 40.59 ± 3% -33.0% 27.21 ± 3% time.user_time 183 ± 23% -32.1% 124 ± 27% sched_debug.cfs_rq[19]:/.tg_load_contrib 3095 ± 11% +81.1% 5606 ± 43% sched_debug.cpu#30.nr_switches 14 ± 5% +39.7% 20 ± 19% sched_debug.cpu#31.cpu_load[0] 1296 ± 11% +28.2% 1661 ± 15% sched_debug.cpu#5.ttwu_count 43 ± 9% +21.5% 52 ± 10% sched_debug.cpu#0.cpu_load[0] 59711 ± 0% -11.8% 52641 ± 7% sched_debug.cfs_rq[5]:/.exec_clock 29 ± 0% +13.8% 33 ± 7% sched_debug.cpu#21.cpu_load[0] 29 ± 6% -12.6% 26 ± 4% sched_debug.cfs_rq[5]:/.runnable_load_avg 30 ± 5% -13.2% 26 ± 6% sched_debug.cpu#5.load 29 ± 1% +12.0% 32 ± 6% sched_debug.cpu#21.cpu_load[2] 29 ± 0% +14.7% 33 ± 6% sched_debug.cpu#21.cpu_load[1] 15 ± 2% +24.6% 19 ± 18% sched_debug.cpu#31.cpu_load[1] 15 ± 2% +19.7% 18 ± 15% sched_debug.cpu#31.cpu_load[2] 30 ± 5% -12.4% 26 ± 6% sched_debug.cfs_rq[5]:/.load 29 ± 0% +13.8% 33 ± 7% sched_debug.cfs_rq[21]:/.runnable_load_avg 29 ± 0% +13.8% 33 ± 7% sched_debug.cfs_rq[21]:/.load 29 ± 0% +13.8% 33 ± 7% sched_debug.cpu#21.load lkp-sb03: Sandy Bridge-EP Memory: 64G will-it-scale.time.user_time 45 ++---------------------------------------------------------------------+ | | 40 ++ .*.. .*.. .*.. | 35 *+.*..*.. .*..* *..*..*. *..*. *..*..*.. ..*..*. *.. .* | *. : : *. *. | 30 ++ : : | 25 ++ : : | O O O O O O: O :O O O O O O O O O O O | 20 ++ : : O | 15 ++ : : | | : : | 10 ++ : : | 5 ++ : : | | : | 0 ++-----------------*---------------------------------------------------+ [*] bisect-good sample [O] bisect-bad sample To reproduce: apt-get install ruby ruby-oj git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git cd lkp-tests bin/setup-local job.yaml # the job file attached in this email bin/run-local job.yaml Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. Thanks, Huang, Ying
--- testcase: will-it-scale default-monitors: wait: pre-test uptime: iostat: vmstat: numa-numastat: numa-vmstat: numa-meminfo: proc-vmstat: proc-stat: meminfo: slabinfo: interrupts: lock_stat: latency_stats: softirqs: bdi_dev_mapping: diskstats: nfsstat: cpuidle: cpufreq-stats: turbostat: pmeter: sched_debug: interval: 10 default_watchdogs: watch-oom: watchdog: cpufreq_governor: commit: 69d668f2356670ab6d59e17c911a222d9546f50b model: Sandy Bridge-EP memory: 64G hdd_partitions: "/dev/disk/by-id/ata-WDC_WD10EARS-00Y5B1_WD-WCAV5F059499-part3" swap_partitions: rootfs_partition: "/dev/disk/by-id/ata-WDC_WD10EARS-00Y5B1_WD-WCAV5F059499-part4" perf-profile: freq: 800 will-it-scale: test: - open2 testbox: lkp-sb03 tbox_group: lkp-sb03 kconfig: x86_64-rhel enqueue_time: 2015-01-31 12:02:48.220308404 +08:00 head_commit: 69d668f2356670ab6d59e17c911a222d9546f50b base_commit: 26bc420b59a38e4e6685a73345a0def461136dce branch: linux-devel/devel-hourly-2015013122 kernel: "/kernel/x86_64-rhel/69d668f2356670ab6d59e17c911a222d9546f50b/vmlinuz-3.19.0-rc6-g69d668f" user: lkp queue: cyclic rootfs: debian-x86_64-2015-01-30.cgz result_root: "/result/lkp-sb03/will-it-scale/open2/debian-x86_64-2015-01-30.cgz/x86_64-rhel/69d668f2356670ab6d59e17c911a222d9546f50b/0" job_file: "/lkp/scheduled/lkp-sb03/cyclic_will-it-scale-open2-x86_64-rhel-HEAD-69d668f2356670ab6d59e17c911a222d9546f50b-0.yaml" dequeue_time: 2015-02-01 00:26:13.233880269 +08:00 nr_cpu: "$(nproc)" job_state: finished loadavg: 22.16 12.46 5.08 1/296 11219 start_time: '1422721619' end_time: '1422721929' version: "/lkp/lkp/.src-20150131-074115"
./runtest.py open2 25 both 1 8 16 24 32
_______________________________________________ LKP mailing list l...@linux.intel.com