In the past few months our 1.2.1 server (running on Centos 5.5 / JDK 6) would, after an hour or two of 'normal' operation, spike the CPU load to 10+. Recently we upgraded to 1.3.4 and our serveris running a little smoother but then after a few hours it will go back to 10+ again. Here's a recent output of "top":
top - 09:08:43 up 1 day, 15:28, 1 user, load average: 13.52, 12.05, 11.56 Tasks: 105 total, 10 running, 95 sleeping, 0 stopped, 0 zombie Cpu(s): 56.3%us, 6.3%sy, 0.1%ni, 34.5%id, 0.4%wa, 0.0%hi, 2.3%si, 0.2%st Mem: 2097152k total, 2035216k used, 61936k free, 51448k buffers Swap: 524280k total, 136k used, 524144k free, 766424k cached PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 27233 tomcat 18 0 1241m 906m 9716 S 61.5 44.3 953:33.33 java 23841 postgres 15 0 152m 16m 13m R 31.5 0.8 37:44.30 postmaster 27269 postgres 15 0 153m 31m 28m S 16.5 1.6 425:21.18 postmaster 23783 postgres 15 0 152m 25m 23m S 13.5 1.3 38:24.97 postmaster 27268 postgres 15 0 152m 35m 32m S 12.0 1.7 189:07.95 postmaster 18653 postgres 16 0 152m 29m 26m R 4.5 1.4 120:31.79 postmaster 27668 postgres 16 0 152m 28m 26m R 4.5 1.4 199:01.50 postmaster 9560 root 35 19 260m 23m 6844 R 3.0 1.1 0:00.92 yum-updatesd-he 23274 postgres 15 0 152m 25m 23m S 1.5 1.3 38:29.17 postmaster 1 root 15 0 10348 692 580 S 0.0 0.0 0:00.37 init 2 root RT -5 0 0 0 S 0.0 0.0 0:00.00 migration/0 3 root 34 19 0 0 0 S 0.0 0.0 0:00.00 ksoftirqd/0 4 root RT -5 0 0 0 S 0.0 0.0 0:00.00 watchdog/0 5 root 10 -5 0 0 0 S 0.0 0.0 0:00.00 events/0 6 root 10 -5 0 0 0 S 0.0 0.0 0:00.00 khelper 7 root 12 -5 0 0 0 S 0.0 0.0 0:00.00 kthread 9 root 10 -5 0 0 0 S 0.0 0.0 0:00.00 xenwatch ... this is after a few hours of operation and once it hits this high load it never comes down. Is there anything I should be looking at to fix this? -- <cowmix>