are you seeing anyone trying to use the dependency tree in your request logs?

I suggest disabling it (see http://jira.codehaus.org/browse/MRM-665)

- Brett

On 01/03/2011, at 10:17 AM, Michael March wrote:

> In the past  few months our 1.2.1 server (running on Centos 5.5 / JDK
> 6) would, after an hour or two of 'normal' operation, spike the CPU
> load to 10+.
> Recently we upgraded to 1.3.4 and our serveris running a little
> smoother but then after a few hours it will go back to 10+ again.
> Here's a recent output of "top":
> 
> top - 09:08:43 up 1 day, 15:28,  1 user,  load average: 13.52, 12.05, 11.56
> Tasks: 105 total,  10 running,  95 sleeping,   0 stopped,   0 zombie
> Cpu(s): 56.3%us,  6.3%sy,  0.1%ni, 34.5%id,  0.4%wa,  0.0%hi,  2.3%si,  0.2%st
> Mem:   2097152k total,  2035216k used,    61936k free,    51448k buffers
> Swap:   524280k total,      136k used,   524144k free,   766424k cached
>   PID USER      PR  NI  VIRT  RES  SHR S %CPU %MEM    TIME+  COMMAND
> 
> 27233 tomcat    18   0 1241m 906m 9716 S 61.5 44.3 953:33.33 java
> 
> 23841 postgres  15   0  152m  16m  13m R 31.5  0.8  37:44.30 postmaster
> 27269 postgres  15   0  153m  31m  28m S 16.5  1.6 425:21.18 postmaster
> 23783 postgres  15   0  152m  25m  23m S 13.5  1.3  38:24.97 postmaster
> 27268 postgres  15   0  152m  35m  32m S 12.0  1.7 189:07.95 postmaster
> 18653 postgres  16   0  152m  29m  26m R  4.5  1.4 120:31.79 postmaster
> 27668 postgres  16   0  152m  28m  26m R  4.5  1.4 199:01.50 postmaster
>  9560 root      35  19  260m  23m 6844 R  3.0  1.1   0:00.92 yum-updatesd-he
> 23274 postgres  15   0  152m  25m  23m S  1.5  1.3  38:29.17 postmaster
>     1 root      15   0 10348  692  580 S  0.0  0.0   0:00.37 init
>     2 root      RT  -5     0    0    0 S  0.0  0.0   0:00.00 migration/0
>     3 root      34  19     0    0    0 S  0.0  0.0   0:00.00 ksoftirqd/0
>     4 root      RT  -5     0    0    0 S  0.0  0.0   0:00.00 watchdog/0
>     5 root      10  -5     0    0    0 S  0.0  0.0   0:00.00 events/0
>     6 root      10  -5     0    0    0 S  0.0  0.0   0:00.00 khelper
>     7 root      12  -5     0    0    0 S  0.0  0.0   0:00.00 kthread
>     9 root      10  -5     0    0    0 S  0.0  0.0   0:00.00 xenwatch
> 
> 
> ... this is after a few hours of operation and once it hits this high
> load it never comes down.
> 
> Is there anything I should be looking at to fix this?
> 
> 
> --
> <cowmix>

--
Brett Porter
br...@apache.org
http://brettporter.wordpress.com/

Reply via email to