We have determined that oom_kill was working. Due to recommendations we changed swappiness to 0, and are running all app servers. By the morning we will know if memory is being cleaned - that seems to be the problem, the system will use memory but never free it. Perhaps thats a bug in any number of things.. We're going to ugprade the pre-dev server to the latest SP of SLES11.
More information from the z/VM level. Still a novice, so I can't initerpret a lot of this. Here’s what’s going on at the CP level right now. [root] bwzld001:~# vmcp indicate load AVGPROC-089% 02 XSTORE-000001/SEC MIGRATE-0000/SEC MDC READS-000000/SEC WRITES-000000/SEC HIT RATIO-000% PAGING-0/SEC STEAL-000% Q0-00001(00000) DORMANT-00010 Q1-00001(00000) E1-00000(00000) Q2-00001(00000) EXPAN-001 E2-00000(00000) Q3-00006(00000) EXPAN-001 E3-00000(00000) PROC 0000-090% IFL PROC 0001-088% IFL LIMITED-00000 [root] bwzld001:~# vmcp indicate user linxd002 exp Userid=LINXD002 Mach=ESA V=V Attached xstore=NONE Iplsys=DEV 0100 Devnum=18 Spool: Reads=0 Writes=0 Owned spaces: Number=1 Owned size=10G Pages: LockedReal=54 LockedLogical=7 Primary space: ID=LINXD002:BASE PRIVATE Defined size=10G Address limit=10G Private spaces: Number=1 Owned size=10G Pages: Main=2601539 Xstore=17444 Dasd=0 WS=2601485 Reserved=0 ResidentLogical=2754 LockedLogical=7 Shared spaces: Number=0 Owned size=0 Pages: Main=0 Xstore=0 Dasd=0 ResidentLogical=0 LockedLogical=0 Private paging: Xstore: Reads=3777 Writes=21554 Migrates=0 Dasd: Reads= 2 Writes= 2 Shared paging: Xstore: Reads= 0 Writes= 0 Migrates=0 Dasd: Reads= 0 Writes= 0 Vconfig: Mode=LINUX CPU 00: Ctime=2 22:42:51 Vtime=0 22:33:50 Ttime=1 00:32:06 Rdr=0 Prt=0 Pch=0 IO=368691 Type=IFL CPUAffinity=ON VtimePrimary=0 22:33:50 TtimePrimary=1 00:32:06 VtimeSecondary=0 00:00:00 TtimeSecondary=0 00:00:00 CPU 01: Ctime=2 22:42:51 Vtime=0 17:46:55 Ttime=0 18:03:40 Rdr=0 Prt=0 Pch=0 IO=424080 Type=IFL CPUAffinity=ON VtimePrimary=0 17:46:55 TtimePrimary=0 18:03:40 VtimeSecondary=0 00:00:00 TtimeSecondary=0 00:00:00 [root] bwzld001:~# vmcp indicate queues exp LINXD001 MP01 Q3 R00 00000000/00000000 .... -.7181 A00 LINXD002 MP01 Q3 R01 00000000/00000000 .... .0474 A01 LINXD002 Q3 R 02601539/02601485 .... .0474 A01 LINXD003 MP01 Q1 PS 00000000/00000000 .I.. 99999 A01 LINXD003 Q1 PS 00148911/00148838 .ID. 99999 A00 LINXDEV Q3 PS 00184038/00183985 ..D. 99999 A00 LINXDEV MP01 Q3 PS 00000000/00000000 .... 99999 A00 LINXD001 Q3 PS 00828196/00828143 ..D. 99999 A00 [root] bwzld001:~# vmcp indicate paging No users in page wait [root] bwzld001:~# vmcp q srm IABIAS : INTENSITY=90%; DURATION=2 LDUBUF : Q1=100% Q2=75% Q3=60% STORBUF: Q1=300% Q2=250% Q3=200% DSPBUF : Q1=32767 Q2=32767 Q3=32767 DISPATCHING MINOR TIMESLICE = 5 MS MAXWSS : LIMIT=9999% ...... : PAGES=999999 XSTORE : 0% [root] bwzld001:~# vmcp q frames All Frames: Configured=3932160 Real=3932160 Usable=3932160 Offline=0 Pageable=3886021 NotInitialized=0 GlobalClearedAvail=64 LocalClearedAvail=64 LocalUnclearedAvail=62 Frames < 2G: GlobalUnclearedAvail=419 Pageable=521368 LogicalFreeStorage=45 RealFreeStorage=4 LockedRS=270 LockedCmd=0 MinidiskCache=0 Nucleus/Prefix=2374 Trace=175 Other=52 Frames > 2G: GlobalUnclearedAvail=1311 Pageable=3364653 LogicalFreeStorage=1394 RealFreeStorage=209 LockedRS=376 LockedCmd=0 MinidiskCache=2012 Other=41240 [root] bwzld001:~# vmcp q alloc map EXTENT EXTENT % ALLOCATION VOLID RDEV START END TOTAL IN USE HIGH USED TYPE ------ ---- ---------- ---------- ------ ------ ------ ---- ------------- VM6RES 9F84 1 20 20 1 1 5% DRCT ACTIVE VM6SPL 9F85 1 10016 1761K 253552 254082 14% SPOOL VM6PG1 9F86 1 10016 1761K 0 0 0% PAGE VM6PG2 9F87 0 0 180 12 12 6% PAGE VM6LX1 9F88 5001 10016 5016 0 0 0% TDISK XSTORE= 2048M online= 2048M XSTORE= 2048M userid= SYSTEM usage= 24% retained= 0M pending= 0M XSTORE MDC min=0M, max=0M, usage=0% XSTORE= 2048M userid= (none) max. attach= 2048M Vmcp q mdc Minidisk cache ON for system Storage MDC min=0M max=128M, usage=0%, bias=1.00 Xstore MDC min=0M max=0M, usage=0%, bias=1.00 On Mon, Jul 26, 2010 at 1:40 PM, Daniel Tate <daniel.t...@gmail.com> wrote: > > FYI, to give our VM as much pagefile as possible, I attached a SAN LUN via > zfcp and used mkswap on it, thinking surely it wouldn't exceed 40GB. > > On Mon, Jul 26, 2010 at 1:18 PM, Daniel Tate <daniel.t...@gmail.com> wrote: >> >> 00: >> 00: CP Q SRM >> 00: IABIAS : INTENSITY=90%; DURATION=2 >> 00: LDUBUF : Q1=100% Q2=75% Q3=60% >> 00: STORBUF: Q1=300% Q2=250% Q3=200% >> 00: DSPBUF : Q1=32767 Q2=32767 Q3=32767 >> 00: DISPATCHING MINOR TIMESLICE = 5 MS >> 00: MAXWSS : LIMIT=9999% >> 00: ...... : PAGES=999999 >> 00: XSTORE : 0% >> 00: >> 00: CP Q ALLOC PAGE >> EXTENT EXTENT TOTAL PAGES HIGH % >> VOLID RDEV START END PAGES IN USE PAGE USED >> ------ ---- ---------- ---------- ------ ------ ------ ---- >> VM6PG1 9F86 1 10016 1761K 0 0 0% >> VM6PG2 9F87 0 0 180 12 12 6% >> ------ ------ ---- >> SUMMARY 1761K 12 1% >> USABLE 1761K 12 1% >> >> >> >> >> On Mon, Jul 26, 2010 at 11:13 AM, Marcy Cortes >> <marcy.d.cor...@wellsfargo.com> wrote: >>> >>> Can you post the results of >>> >>> Q SRM >>> >>> Q ALLOC PAGE >>> >>> >>> >>> >>> Marcy >>> >>> “This message may contain confidential and/or privileged information. If >>> you are not the addressee or authorized to receive this for the addressee, >>> you must not use, copy, disclose, or take any action based on this message >>> or any information herein. If you have received this message in error, >>> please advise the sender immediately by reply e-mail and delete this >>> message. Thank you for your cooperation." >>> >>> >>> >>> ________________________________ >>> >>> From: The IBM z/VM Operating System [mailto:ib...@listserv.uark.edu] On >>> Behalf Of Daniel Tate >>> Sent: Monday, July 26, 2010 8:31 AM >>> To: IBMVM@LISTSERV.UARK.EDU >>> Subject: [IBMVM] Linux on Z/VM running WAS problems - anyone got any tips? >>> >>> >>> I apologize for this not being a "direct" z/VM question. >>> >>> I've posted to the Linux-390 group to get the linux POV.. but exploring all >>> angles here I am attempting to find out if there's anything i can set/do >>> from z/VM that would help the situation.. I'd like it to "finish" the >>> scroll, not sure how to do that except tape down control-C (i'm using >>> c3720). a CP Q ALL is at the bottom of all this mess.. >>> >>> The e-mail sent to Linux-s390 for reference: >>> >>> We're running websphere on a z9 under z/VM 4 systems are live out of 8. >>> it is running apps that consume around 16GB of memory on a Windows machine. >>> on this, we have allocated 10G of real storage (RAM) and around 35GB of >>> Swap. When websphere starts, it consumes all the memory eventually and >>> halts, but not panics, the system. We are running 64-Bit. I'm a z/VM >>> novice so i don't know much to do.. >>> >>> Here is some information from our WAS Admin: >>> "We are running WebSphere 6.1.0.25 with FP EJB3.0,Webservices and Web 2.0 >>> installed. There are two nodes running 14 application servers each. there >>> are currently 32 applications installed but not currently running. No >>> security has been enabled for WebSphere at this time." >>> >>> >>> At this point i see two problems: >>> >>> 1) Why is OOM Kill not functioning properly >>> 2) Why is websphere performance so awful? >>> >>> and have two questions >>> >>> 1) Does anyone have any PRACTICAL experience/tips to optimize SLES11 on >>> z/VM? So far we've been using dated case studies and redbooks that seem to >>> be filled with inaccuracies or outdated information. >>> 2) Is there any way to force a coredump via the cp, like you can with the >>> magic sysrq? >>> >>> All systems are running the same release and patch level: >>> >>> [root] bwzld001:~# lsb_release -a >>> LSB Version: >>> core-2.0-noarch:core-3.2-noarch:core-4.0-noarch:core-2.0-s390x:core-3.2-s390x:core-4.0-s390x:desktop-4.0-noarch:desktop-4.0-s390:desktop-4.0-s390x:graphics-2.0-noarch:graphics-2.0-s390:graphics-2.0-s390x:graphics-3.2-noarch:graphics-3.2-s390:graphics-3.2-s390x:graphics-4.0-noarch:graphics-4.0-s390:graphics-4.0-s390x >>> Distributor ID: SUSE LINUX >>> Description: SUSE Linux Enterprise Server 11 (s390x) >>> Release: 11 >>> Codename: n/a >>> >>> >>> Here is a partial top shortly before system death: >>> >>> top - 08:13:14 up 2 days, 16:08, 2 users, load average: 51.47, 22.20, >>> 10.25 >>> Tasks: 129 total, 4 running, 125 sleeping, 0 stopped, 0 zombie >>> Cpu(s): 16.7%us, 81.5%sy, 0.0%ni, 0.0%id, 0.0%wa, 0.3%hi, 0.3%si, >>> 1.2%st >>> Mem: 10268344k total, 10220568k used, 47776k free, 548k buffers >>> Swap: 35764956k total, 35764956k used, 0k free, 56340k cached >>> >>> PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND >>> >>> 26850 wasadmin 20 0 1506m 253m 2860 S 18 2.5 16:06.28 java >>> 29870 wasadmin 20 0 1497m 279m 2560 S 15 2.8 15:41.13 java >>> 24607 wasadmin 20 0 1502m 223m 2760 S 13 2.2 16:15.14 java >>> 24641 wasadmin 20 0 7229m 1.3g 3172 S 13 13.1 196:35.52 java >>> 26606 wasadmin 20 0 1438m 272m 6212 S 12 2.7 16:02.77 java >>> 27600 wasadmin 20 0 1553m 258m 2920 S 12 2.6 15:46.57 java >>> 24638 wasadmin 20 0 7368m 1.3g 24m S 10 13.7 206:02.05 java >>> 25609 wasadmin 20 0 1528m 219m 2540 S 9 2.2 16:07.33 java >>> 30258 wasadmin 20 0 1515m 249m 2592 S 7 2.5 15:49.79 java >>> 25780 wasadmin 20 0 1604m 277m 2332 S 6 2.8 16:31.41 java >>> 27106 wasadmin 20 0 1458m 273m 2472 S 6 2.7 15:59.13 java >>> 27336 wasadmin 20 0 1528m 238m 2540 S 5 2.4 15:38.82 java >>> 29164 wasadmin 20 0 1527m 224m 2608 S 5 2.2 16:02.56 java >>> 31400 wasadmin 20 0 1509m 259m 2468 S 5 2.6 15:26.38 java >>> 25244 wasadmin 20 0 1509m 290m 2624 S 5 2.9 16:16.07 java >>> 24769 wasadmin 20 0 1409m 259m 2308 S 5 2.6 16:08.12 java >>> 28796 wasadmin 20 0 1338m 263m 3076 S 4 2.6 15:47.72 java >>> 26185 wasadmin 20 0 1493m 274m 2304 S 2 2.7 16:01.97 java >>> 25968 wasadmin 20 0 1427m 257m 2532 S 1 2.6 15:51.50 java >>> 29495 wasadmin 20 0 1466m 259m 2260 S 1 2.6 15:31.82 java >>> 25080 wasadmin 20 0 1445m 236m 2472 S 0 2.4 15:53.19 java >>> 26410 wasadmin 20 0 1475m 271m 2540 S 0 2.7 15:52.48 java >>> 31027 wasadmin 20 0 1413m 238m 2492 S 0 2.4 15:29.78 java >>> 3695 wasadmin 20 0 9968 1352 1352 S 0 0.0 0:00.13 bash >>> 24474 wasadmin 20 0 1468m 205m 2472 S 0 2.0 16:03.63 java >>> 24920 wasadmin 20 0 1522m 263m 2616 S 0 2.6 16:06.29 java >>> 25422 wasadmin 20 0 1584m 229m 2284 S 0 2.3 16:02.18 java >>> 27892 wasadmin 20 0 1414m 263m 2648 S 0 2.6 15:45.96 java >>> 28184 wasadmin 20 0 1523m 241m 2320 S 0 2.4 15:42.21 java >>> 28486 wasadmin 20 0 1450m 231m 2288 S 0 2.3 15:46.53 java >>> 30625 wasadmin 20 0 1477m 251m 3024 S 0 2.5 15:44.80 java >>> >>> ----------------- >>> >>> >>> Here are a few screen grabs from the 3720 Console session: >>> >>> Unless you get a _continuous_flood_ of these messages it means >>> everything is working fine. Allocations from irqs cannot be >>> perfectly reliable and the kernel is designed to handle that. >>> java: page allocation failure. order:0, mode:0x20, alloc_flags:0x7, >>> pflags:0x400 >>> 040 >>> CPU: 1 Not tainted 2.6.27.45-0.1-default #1 >>> Process java (pid: 28831, task: 00000001ab64c638, ksp: 0000000215bbb5e0) >>> 0000000000000000 000000027fbcf7b0 0000000000000002 0000000000000000 >>> 000000027fbcf850 000000027fbcf7c8 000000027fbcf7c8 00000000003b6696 >>> 00000000014a4e88 0000000000000007 0000000000634e00 0000000000000000 >>> 000000000000000d 0000000000000000 000000027fbcf818 000000000000000e >>> 00000000003cdc00 000000000010521a 000000027fbcf7b0 000000027fbcf7f8 >>> Call Trace: >>> ( 0000000000105174> show_trace+0x130/0x134) >>> 000000000019890a> __alloc_pages_internal+0x406/0x55c >>> 00000000001c7056> cache_grow+0x382/0x458 >>> 00000000001c7440> cache_alloc_refill+0x314/0x36c >>> 00000000001c6c12> kmem_cache_alloc+0x82/0x144 >>> 00000000003228f2> __alloc_skb+0x82/0x208 >>> 000000000032378e> dev_alloc_skb+0x36/0x64 >>> 000003e0001a030e> qeth_core_get_next_skb+0x31e/0x704 eth >>> 000003e0000d5f8c> qeth_l3_process_inbound_buffer+0x9c/0x598 eth_l3 >>> 000003e0000d6574> qeth_l3_qdio_input_handler+0xec/0x268 eth_l3 >>> 000003e0000ebc44> qdio_kick_inbound_handler+0xbc/0x178 dio >>> 000003e0000ee58c> __tiqdio_inbound_processing+0x394/0xdf4 dio >>> 000000000013a800> tasklet_action+0x10c/0x1e4 >>> 000000000013b908> __do_softirq+0xe0/0x1c8 >>> 0000000000110252> do_softirq+0xaa/0xb0 >>> 000000000013b772> irq_exit+0xc2/0xcc >>> 00000000002f6586> do_IRQ+0x132/0x1c8 >>> 0000000000114148> io_return+0x0/0x8 >>> 00000000002b850e> _raw_spin_lock_wait+0x86/0xa4 >>> ( 000003e047d6fa00> 0x3e047d6fa00) >>> 000000000019eb9c> shrink_page_list+0x1a0/0x584 >>> 000000000019f184> shrink_inactive_list+0x204/0x5b0 >>> 000000000019f620> shrink_zone+0xf0/0x1d0 >>> 000000000019f882> shrink_zones+0xae/0x184 >>> 00000000001a02be> do_try_to_free_pages+0x96/0x3fc >>> 00000000001a072c> try_to_free_pages+0x74/0x7c >>> 0000000000198730> __alloc_pages_internal+0x22c/0x55c >>> 000000000019b5a2> __do_page_cache_readahead+0x10a/0x2ac >>> 000000000019b7cc> do_page_cache_readahead+0x88/0xa8 >>> 000000000019170e> filemap_fault+0x33a/0x448 >>> 00000000001a55bc> __do_fault+0x78/0x580 >>> 00000000001a962e> handle_mm_fault+0x1e6/0x4c0 >>> 00000000003b9e1e> do_dat_exception+0x29e/0x388 >>> 0000000000113c0c> sysc_return+0x0/0x8 >>> 0000020000214bde> 0x20000214bde >>> Mem-Info: >>> DMA per-cpu: >>> CPU 0: hi: 186, btch: 31 usd: 0 >>> CPU 1: hi: 186, btch: 31 usd: 0 >>> Normal per-cpu: >>> CPU 0: hi: 186, btch: 31 usd: 0 >>> CPU 1: hi: 186, btch: 31 usd: 0 >>> Active:1355277 inactive:1132712 dirty:0 writeback:0 unstable:0 >>> free:9269 slab:17875 mapped:765 pagetables:24402 bounce:0 >>> DMA free:33220kB min:2568kB low:3208kB high:3852kB active:1092112kB >>> inactive:926 >>> 924kB present:2064384kB pages_scanned:21132286 all_unreclaimable? no >>> lowmem_reserveݨ: 0 8064 8064 >>> Normal free:3856kB min:10276kB low:12844kB high:15412kB active:4328996kB >>> inactiv >>> e:3603924kB present:8257536kB pages_scanned:44557906 all_unreclaimable? yes >>> lowmem_reserveݨ: 0 0 0 >>> DMA: 101*4kB 32*8kB 473*16kB 195*32kB 49*64kB 30*128kB 8*256kB 3*512kB >>> 8*1024kB >>> = 33220kB >>> Normal: 0*4kB 0*8kB 1*16kB 0*32kB 0*64kB 0*128kB 1*256kB 1*512kB 3*1024kB = >>> 3856 >>> kB >>> 9283 total pagecache pages >>> 0 pages in swap cache >>> Swap cache stats: add 34513958, delete 34513958, find 6612011/8393146 >>> Free swap = 0kB >>> Total swap = 35764956kB >>> 2621440 pages RAM >>> 54354 pages reserved >>> 22356 pages shared >>> 2538214 pages non-shared >>> The following is only an harmless informational message. >>> Unless you get a _continuous_flood_ of these messages it means >>> everything is working fine. Allocations from irqs cannot be >>> perfectly reliable and the kernel is designed to handle that. >>> java: page allocation failure. order:0, mode:0x20, alloc_flags:0x7, >>> pflags:0x400 >>> 040 >>> CPU: 1 Not tainted 2.6.27.45-0.1-default #1 >>> Process java (pid: 28831, task: 00000001ab64c638, ksp: 0000000215bbb5e0) >>> 0000000000000000 000000027fbcf7b0 0000000000000002 0000000000000000 >>> 000000027fbcf850 000000027fbcf7c8 000000027fbcf7c8 00000000003b6696 >>> 00000000014a5dd3 0000000000000007 0000000000634e00 0000000000000000 >>> 000000000000000d 0000000000000000 000000027fbcf818 000000000000000e >>> 00000000003cdc00 000000000010521a 000000027fbcf7b0 000000027fbcf7f8 >>> Call Trace: >>> ( 0000000000105174> show_trace+0x130/0x134) >>> 000000000019890a> __alloc_pages_internal+0x406/0x55c >>> 00000000001c7056> cache_grow+0x382/0x458 >>> 00000000001c7440> cache_alloc_refill+0x314/0x36c >>> 00000000001c6c12> kmem_cache_alloc+0x82/0x144 >>> 00000000003228f2> __alloc_skb+0x82/0x208 >>> 000000000032378e> dev_alloc_skb+0x36/0x64 >>> 000003e0001a030e> qeth_core_get_next_skb+0x31e/0x704 eth >>> 000003e0000d5f8c> qeth_l3_process_inbound_buffer+0x9c/0x598 eth_l3 >>> 000003e0000d6574> qeth_l3_qdio_input_handler+0xec/0x268 eth_l3 >>> 000003e0000ebc44> qdio_kick_inbound_handler+0xbc/0x178 dio >>> 000003e0000ee58c> __tiqdio_inbound_processing+0x394/0xdf4 dio >>> 000000000013a800> tasklet_action+0x10c/0x1e4 >>> 000000000013b908> __do_softirq+0xe0/0x1c8 >>> 0000000000110252> do_softirq+0xaa/0xb0 >>> 000000000013b772> irq_exit+0xc2/0xcc >>> 00000000002f6586> do_IRQ+0x132/0x1c8 >>> 0000000000114148> io_return+0x0/0x8 >>> 00000000002b850e> _raw_spin_lock_wait+0x86/0xa4 >>> ( 000003e047d6fa00> 0x3e047d6fa00) >>> 000000000019eb9c> shrink_page_list+0x1a0/0x584 >>> 000000000019f184> shrink_inactive_list+0x204/0x5b0 >>> 000000000019f620> shrink_zone+0xf0/0x1d0 >>> 000000000019f882> shrink_zones+0xae/0x184 >>> 00000000001a02be> do_try_to_free_pages+0x96/0x3fc >>> 00000000001a072c> try_to_free_pages+0x74/0x7c >>> 0000000000198730> __alloc_pages_internal+0x22c/0x55c >>> 000000000019b5a2> __do_page_cache_readahead+0x10a/0x2ac >>> 000000000019b7cc> do_page_cache_readahead+0x88/0xa8 >>> 000000000019170e> filemap_fault+0x33a/0x448 >>> 00000000001a55bc> __do_fault+0x78/0x580 >>> 00000000001a962e> handle_mm_fault+0x1e6/0x4c0 >>> 00000000003b9e1e> do_dat_exception+0x29e/0x388 >>> 0000000000113c0c> sysc_return+0x0/0x8 >>> 0000020000214bde> 0x20000214bde >>> Mem-Info: >>> DMA per-cpu: >>> CPU 0: hi: 186, btch: 31 usd: 0 >>> CPU 1: hi: 186, btch: 31 usd: 0 >>> Normal per-cpu: >>> CPU 0: hi: 186, btch: 31 usd: 0 >>> CPU 1: hi: 186, btch: 31 usd: 0 >>> Active:1355277 inactive:1132712 dirty:0 writeback:0 unstable:0 >>> free:9269 slab:17875 mapped:765 pagetables:24402 bounce:0 >>> DMA free:33220kB min:2568kB low:3208kB high:3852kB active:1092112kB >>> inactive:926 >>> 924kB present:2064384kB pages_scanned:21132286 all_unreclaimable? no >>> lowmem_reserveݨ: 0 8064 8064 >>> Normal free:3856kB min:10276kB low:12844kB high:15412kB active:4328996kB >>> inactiv >>> e:3603924kB present:8257536kB pages_scanned:44557906 all_unreclaimable? yes >>> lowmem_reserveݨ: 0 0 0 >>> DMA: 101*4kB 32*8kB 473*16kB 195*32kB 49*64kB 30*128kB 8*256kB 3*512kB >>> 8*1024kB >>> = 33220kB >>> Normal: 0*4kB 0*8kB 1*16kB 0*32kB 0*64kB 0*128kB 1*256kB 1*512kB 3*1024kB = >>> 3856 >>> kB >>> 9283 total pagecache pages >>> 0 pages in swap cache >>> Swap cache stats: add 34513958, delete 34513958, find 6612011/8393146 >>> Free swap = 0kB >>> Total swap = 35764956kB >>> 2621440 pages RAM >>> 54354 pages reserved >>> 22356 pages shared >>> 2538214 pages non-shared >>> __ratelimit: 4 callbacks suppressed >>> The following is only an harmless informational message. >>> Unless you get a _continuous_flood_ of these messages it means >>> everything is working fine. Allocations from irqs cannot be >>> perfectly reliable and the kernel is designed to handle that. >>> java: page allocation failure. order:0, mode:0x20, alloc_flags:0x7, >>> pflags:0x400 >>> 040 >>> CPU: 1 Not tainted 2.6.27.45-0.1-default #1 >>> Process java (pid: 28831, task: 00000001ab64c638, ksp: 0000000215bbb5e0) >>> 0000000000000000 000000027fbcf7b0 0000000000000002 0000000000000000 >>> 000000027fbcf850 000000027fbcf7c8 000000027fbcf7c8 00000000003b6696 >>> >>> etc, etc for HUNDREDS of pages.. perhaps infinite. >>> >>> 00: >>> 00: CP Q ALL >>> 00: STORAGE = 15G CONFIGURED = 15G INC = 64M STANDBY = 0 RESERVED = 0 >>> 00: OSA 039C ATTACHED TO TCPIP 039C DEVTYPE OSA CHPID 01 OSD >>> 00: OSA 039D ATTACHED TO TCPIP 039D DEVTYPE OSA CHPID 01 OSD >>> 00: OSA 039E ATTACHED TO TCPIP 039E DEVTYPE OSA CHPID 01 OSD >>> 00: OSA 03A0 ATTACHED TO DTCVSW2 03A0 DEVTYPE OSA CHPID 01 OSD >>> 00: OSA 03A1 ATTACHED TO DTCVSW2 03A1 DEVTYPE OSA CHPID 01 OSD >>> 00: OSA 03A2 ATTACHED TO DTCVSW2 03A2 DEVTYPE OSA CHPID 01 OSD >>> 00: OSA 03C0 ATTACHED TO DTCVSW1 03C0 DEVTYPE OSA CHPID 02 OSD >>> 00: OSA 03C1 ATTACHED TO DTCVSW1 03C1 DEVTYPE OSA CHPID 02 OSD >>> 00: OSA 03C2 ATTACHED TO DTCVSW1 03C2 DEVTYPE OSA CHPID 02 OSD >>> 00: FCP 5000 ATTACHED TO LINXDEV 5000 CHPID 46 >>> 00: WWPN C05076FAE3000400 >>> 00: FCP 5001 ATTACHED TO LINXD001 5001 CHPID 46 >>> 00: WWPN C05076FAE3000404 >>> 00: FCP 5002 ATTACHED TO LINXD002 5002 CHPID 46 >>> 00: WWPN C05076FAE3000408 >>> 00: FCP 5003 ATTACHED TO LINXD003 5003 CHPID 46 >>> 00: WWPN C05076FAE300040C >>> 00: FCP 5100 ATTACHED TO LINXDEV 5100 CHPID 47 >>> 00: WWPN C05076FAE3000900 >>> 00: FCP 5101 ATTACHED TO LINXD001 5101 CHPID 47 >>> 00: WWPN C05076FAE3000904 >>> 00: FCP 5102 ATTACHED TO LINXD002 5102 CHPID 47 >>> 00: WWPN C05076FAE3000908 >>> 00: FCP 5103 ATTACHED TO LINXD003 5103 CHPID 47 >>> 00: WWPN C05076FAE300090C >>> 00: DASD 9F7D CP SYSTEM VM6LXD 0 >>> 00: DASD 9F7E CP SYSTEM VM6LXE 0 >>> 00: DASD 9F80 CP SYSTEM VM6LX9 2 >>> 00: DASD 9F81 CP SYSTEM VM6LXA 2 >>> 00: DASD 9F82 CP SYSTEM VM6LXB 0 >>> 00: DASD 9F83 CP SYSTEM VM6LXC 0 >>> 00: DASD 9F84 CP OWNED VM6RES 135 >>> 00: DASD 9F85 CP OWNED VM6SPL 0 >>> 00: DASD 9F86 CP OWNED VM6PG1 0 >>> 00: DASD 9F87 CP OWNED VM6PG2 0 >>> 00: DASD 9F88 CP OWNED VM6LX1 4 >>> 00: DASD 9F89 CP SYSTEM VM6LX2 0 >>> 00: DASD 9F8A CP SYSTEM VM6LX3 0 >>> 00: DASD 9F8B CP SYSTEM VM6LX4 0 >>> 00: DASD 9F8C CP SYSTEM VM6LX5 2 >>> 00: DASD 9F8D CP SYSTEM VM6LX6 0 >>> 00: DASD 9F8E CP SYSTEM VM6LX7 0 >>> 00: DASD 9F8F CP SYSTEM VM6LX8 2 >>> 00: DASD 9FC7 CP SYSTEM VM6LX6 0 >>> 00: DASD 9FC8 CP SYSTEM VM6LX5 2 >>> 00: DASD 9FC9 CP SYSTEM VM6LX2 0 >>> 00: DASD 9FCA CP SYSTEM VM6LX4 0 >>> 00: DASD 9FCB CP SYSTEM VM6LX3 0 >>> 00: DASD 9FCE CP SYSTEM VM6LX1 4 >>> 00: DASD 9FCF CP SYSTEM VM6PG2 0 >>> 00: DASD 9FD0 CP SYSTEM VM6PG1 0 >>> 00: DASD 9FD1 CP SYSTEM VM6SPL 0 >>> 00: DASD 9FD2 CP SYSTEM VM6RES 135 >>> >>> >>> >>> >> >