I had the same problem, and pretty much the same hardware.
But as Kasper pointed out, https://www.illumos.org/issues/1333#note-21
solved my problem. no more kernel cpu rising...

put this into your /etc/system file

set apix:apic_timer_preferred_mode=0x0

and reboot.


Regards,

Svavar O





Kasper Brink <mailto:k.br...@cs.ru.nl>
23. desember 2011 08:47

I'm not an Illumos expert, but the symptoms you describe sound a lot like
those of https://www.illumos.org/issues/1333 (high sys load, dtrace hang).
Does the workaround in comment #21
(https://www.illumos.org/issues/1333#note-21) solve the problem for you?

Regards,

Kasper


_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss
Chris Bünger <mailto:chris.buen...@googlemail.com>
23. desember 2011 07:55
I experience the same issue. For me, a simple reboot did not help, but
a reboot -p.
I have similar setups with 134 and 151a and only 151a shows this behavior,

Maybe this helps.

Chris


_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss
Mirko Kaffka <mailto:m...@mkaffka.de>
23. desember 2011 00:08
I recently got a new machine and installed oi_151a from scratch
(core i7-2600, intel motherboard DH67BL, 16GB RAM).
After a few days uptime I noticed a constant system load of about 10%
although the desktop was idle and I had not started anything that caused
a permanent load. There was almost no I/O activity, just a few reads
and writes every few seconds. vmstat showed 0-1% user time but
10-13% system time. prstat -v output was far below 1% or 0% user and
system time for all processes.
Over the following days the load increased further. When I took 7 cpu
cores off-line I got about 80% sys load on the remaining core. Where
does it come from?

When I switch from multi user to single user mode the load persists.
When I reboot, everything is fine for a while (0-1% sys load) but the load
slowly starts increasing again. So, I have to reboot the machine about
every 2 days what is very unpleasant.

I tried to analyze the issue using intrstat, lockstat, etc. but have not
got very far.

All following commands were run in single user mode and with only one cpu
core on-line. (I hope it's ok to put the output here?)

~ # vmstat 5
kthr memory page disk faults cpu r b w swap free re mf pi po fr de sr s1 s2 s3 s4 in sy cs us sy id 0 0 0 9070220 2859260 2 4 0 0 0 0 0 5 -1 1 14 397 265 258 0 4 95 0 0 0 10392120 4142932 24 64 0 0 0 0 0 0 0 0 1 508 99 227 0 21 79 0 0 0 10392120 4142960 0 0 0 0 0 0 0 0 0 0 13 511 60 229 0 21 79 0 0 0 10392124 4142964 0 0 0 0 0 0 0 0 0 0 13 509 59 226 0 21 79

~ # ps -ef
     UID   PID  PPID   C    STIME TTY         TIME CMD
    root     0     0   0   Dec 20 ?           0:01 sched
    root     4     0   0   Dec 20 ?           0:00 kcfpoold
    root     6     0   0   Dec 20 ?           2:34 zpool-rpool
    root     1     0   0   Dec 20 ?           0:00 /sbin/init
    root     2     0   0   Dec 20 ?           0:00 pageout
    root     3     0   0   Dec 20 ?           8:54 fsflush
root 10 1 0 Dec 20 ? 0:03 /lib/svc/bin/svc.startd root 12 1 0 Dec 20 ? 0:08 /lib/svc/bin/svc.configd
  netadm    50     1   0   Dec 20 ?           0:00 /lib/inet/ipmgmtd
   dladm    46     1   0   Dec 20 ?           0:00 /sbin/dlmgmtd
    root   167     0   0   Dec 20 ?           1:50 zpool-tank
root 232 1 0 Dec 20 ? 0:00 /usr/lib/sysevent/syseventd
    root  9518    10   0 21:01:56 console     0:00 -bash
    root   262     1   0   Dec 20 ?           0:02 devfsadmd
    root   276     1   0   Dec 20 ?           0:00 /usr/lib/power/powerd
    root 10708  9518   0 21:04:53 console     0:00 ps -ef
    root  3222     1   0   Dec 20 ?           0:00 -bash


~ # intrstat
device | cpu0 %tim cpu1 %tim cpu2 %tim cpu3 %tim -------------+------------------------------------------------------------ e1000g#1 | 1 0.0 0 0.0 0 0.0 0 0.0 ehci#0 | 1 0.0 0 0.0 0 0.0 0 0.0 ehci#1 | 1 0.0 0 0.0 0 0.0 0 0.0 rtls#0 | 1 0.0 0 0.0 0 0.0 0 0.0
(cpu4..7 are all 0.0%)


~ # prstat -v
PID USERNAME USR SYS TRP TFL DFL LCK SLP LAT VCX ICX SCL SIG PROCESS/NLWP
 10711 root     0.0 0.0 0.0 0.0 0.0 0.0 100 0.0   1   0 254   0 prstat/1
  3222 root     0.0 0.0 0.0 0.0 0.0 0.0 100 0.0   0   0   0   0 bash/1
[cut]
Total: 14 processes, 366 lwps, load averages: 0.02, 0.32, 0.67


~ # lockstat -kIW -D20 sleep 30

Profiling interrupt: 2913 events in 30.028 seconds (97 events/sec)

Count indv cuml rcnt     nsec Hottest CPU+PIL        Caller
-------------------------------------------------------------------------------
 2878  99%  99% 0.00      293 cpu[0]                 acpi_cpu_cstate
   12   0%  99% 0.00      224 cpu[0]                 fsflush
   10   0% 100% 0.00      266 cpu[0]                 i86_mwait
[cut]
-------------------------------------------------------------------------------

Is the high count on acpi_cpu_cstate normal?

The hotkernel script from the dtrace toolkit finally froze my system.
After the reboot hotkernel run flawlessly.

How can I further analyze this?

Thanks,
Mirko


_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss
_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss

Reply via email to