00: 00: CP Q SRM 00: IABIAS : INTENSITY=90%; DURATION=2 00: LDUBUF : Q1=100% Q2=75% Q3=60% 00: STORBUF: Q1=300% Q2=250% Q3=200% 00: DSPBUF : Q1=32767 Q2=32767 Q3=32767 00: DISPATCHING MINOR TIMESLICE = 5 MS 00: MAXWSS : LIMIT=9999% 00: ...... : PAGES=999999 00: XSTORE : 0% 00: 00: CP Q ALLOC PAGE EXTENT EXTENT TOTAL PAGES HIGH % VOLID RDEV START END PAGES IN USE PAGE USED ------ ---- ---------- ---------- ------ ------ ------ ---- VM6PG1 9F86 1 10016 1761K 0 0 0% VM6PG2 9F87 0 0 180 12 12 6% ------ ------ ---- SUMMARY 1761K 12 1% USABLE 1761K 12 1%
On Mon, Jul 26, 2010 at 11:13 AM, Marcy Cortes < marcy.d.cor...@wellsfargo.com> wrote: > Can you post the results of > > Q SRM > > Q ALLOC PAGE > > > > > Marcy > > “This message may contain confidential and/or privileged information. If > you are not the addressee or authorized to receive this for the addressee, > you must not use, copy, disclose, or take any action based on this message > or any information herein. If you have received this message in error, > please advise the sender immediately by reply e-mail and delete this > message. Thank you for your cooperation." > > > > ________________________________ > > From: The IBM z/VM Operating System [mailto:ib...@listserv.uark.edu] On > Behalf Of Daniel Tate > Sent: Monday, July 26, 2010 8:31 AM > To: IBMVM@LISTSERV.UARK.EDU > Subject: [IBMVM] Linux on Z/VM running WAS problems - anyone got any tips? > > > I apologize for this not being a "direct" z/VM question. > > I've posted to the Linux-390 group to get the linux POV.. but exploring all > angles here I am attempting to find out if there's anything i can set/do > from z/VM that would help the situation.. I'd like it to "finish" the > scroll, not sure how to do that except tape down control-C (i'm using > c3720). a CP Q ALL is at the bottom of all this mess.. > > The e-mail sent to Linux-s390 for reference: > > We're running websphere on a z9 under z/VM 4 systems are live out of 8. > it is running apps that consume around 16GB of memory on a Windows machine. > on this, we have allocated 10G of real storage (RAM) and around 35GB of > Swap. When websphere starts, it consumes all the memory eventually and > halts, but not panics, the system. We are running 64-Bit. I'm a z/VM > novice so i don't know much to do.. > > Here is some information from our WAS Admin: > "We are running WebSphere 6.1.0.25 with FP EJB3.0,Webservices and Web 2.0 > installed. There are two nodes running 14 application servers each. there > are currently 32 applications installed but not currently running. No > security has been enabled for WebSphere at this time." > > > At this point i see two problems: > > 1) Why is OOM Kill not functioning properly > 2) Why is websphere performance so awful? > > and have two questions > > 1) Does anyone have any PRACTICAL experience/tips to optimize SLES11 on > z/VM? So far we've been using dated case studies and redbooks that seem to > be filled with inaccuracies or outdated information. > 2) Is there any way to force a coredump via the cp, like you can with the > magic sysrq? > > All systems are running the same release and patch level: > > [root] bwzld001:~# lsb_release -a > LSB Version: > > core-2.0-noarch:core-3.2-noarch:core-4.0-noarch:core-2.0-s390x:core-3.2-s390x:core-4.0-s390x:desktop-4.0-noarch:desktop-4.0-s390:desktop-4.0-s390x:graphics-2.0-noarch:graphics-2.0-s390:graphics-2.0-s390x:graphics-3.2-noarch:graphics-3.2-s390:graphics-3.2-s390x:graphics-4.0-noarch:graphics-4.0-s390:graphics-4.0-s390x > Distributor ID: SUSE LINUX > Description: SUSE Linux Enterprise Server 11 (s390x) > Release: 11 > Codename: n/a > > > Here is a partial top shortly before system death: > > top - 08:13:14 up 2 days, 16:08, 2 users, load average: 51.47, 22.20, > 10.25 > Tasks: 129 total, 4 running, 125 sleeping, 0 stopped, 0 zombie > Cpu(s): 16.7%us, 81.5%sy, 0.0%ni, 0.0%id, 0.0%wa, 0.3%hi, 0.3%si, > 1.2%st > Mem: 10268344k total, 10220568k used, 47776k free, 548k buffers > Swap: 35764956k total, 35764956k used, 0k free, 56340k cached > > PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND > > 26850 wasadmin 20 0 1506m 253m 2860 S 18 2.5 16:06.28 java > 29870 wasadmin 20 0 1497m 279m 2560 S 15 2.8 15:41.13 java > 24607 wasadmin 20 0 1502m 223m 2760 S 13 2.2 16:15.14 java > 24641 wasadmin 20 0 7229m 1.3g 3172 S 13 13.1 196:35.52 java > 26606 wasadmin 20 0 1438m 272m 6212 S 12 2.7 16:02.77 java > 27600 wasadmin 20 0 1553m 258m 2920 S 12 2.6 15:46.57 java > 24638 wasadmin 20 0 7368m 1.3g 24m S 10 13.7 206:02.05 java > 25609 wasadmin 20 0 1528m 219m 2540 S 9 2.2 16:07.33 java > 30258 wasadmin 20 0 1515m 249m 2592 S 7 2.5 15:49.79 java > 25780 wasadmin 20 0 1604m 277m 2332 S 6 2.8 16:31.41 java > 27106 wasadmin 20 0 1458m 273m 2472 S 6 2.7 15:59.13 java > 27336 wasadmin 20 0 1528m 238m 2540 S 5 2.4 15:38.82 java > 29164 wasadmin 20 0 1527m 224m 2608 S 5 2.2 16:02.56 java > 31400 wasadmin 20 0 1509m 259m 2468 S 5 2.6 15:26.38 java > 25244 wasadmin 20 0 1509m 290m 2624 S 5 2.9 16:16.07 java > 24769 wasadmin 20 0 1409m 259m 2308 S 5 2.6 16:08.12 java > 28796 wasadmin 20 0 1338m 263m 3076 S 4 2.6 15:47.72 java > 26185 wasadmin 20 0 1493m 274m 2304 S 2 2.7 16:01.97 java > 25968 wasadmin 20 0 1427m 257m 2532 S 1 2.6 15:51.50 java > 29495 wasadmin 20 0 1466m 259m 2260 S 1 2.6 15:31.82 java > 25080 wasadmin 20 0 1445m 236m 2472 S 0 2.4 15:53.19 java > 26410 wasadmin 20 0 1475m 271m 2540 S 0 2.7 15:52.48 java > 31027 wasadmin 20 0 1413m 238m 2492 S 0 2.4 15:29.78 java > 3695 wasadmin 20 0 9968 1352 1352 S 0 0.0 0:00.13 bash > 24474 wasadmin 20 0 1468m 205m 2472 S 0 2.0 16:03.63 java > 24920 wasadmin 20 0 1522m 263m 2616 S 0 2.6 16:06.29 java > 25422 wasadmin 20 0 1584m 229m 2284 S 0 2.3 16:02.18 java > 27892 wasadmin 20 0 1414m 263m 2648 S 0 2.6 15:45.96 java > 28184 wasadmin 20 0 1523m 241m 2320 S 0 2.4 15:42.21 java > 28486 wasadmin 20 0 1450m 231m 2288 S 0 2.3 15:46.53 java > 30625 wasadmin 20 0 1477m 251m 3024 S 0 2.5 15:44.80 java > > ----------------- > > > Here are a few screen grabs from the 3720 Console session: > > Unless you get a _continuous_flood_ of these messages it means > everything is working fine. Allocations from irqs cannot be > perfectly reliable and the kernel is designed to handle that. > java: page allocation failure. order:0, mode:0x20, alloc_flags:0x7, > pflags:0x400 > 040 > CPU: 1 Not tainted 2.6.27.45-0.1-default #1 > Process java (pid: 28831, task: 00000001ab64c638, ksp: 0000000215bbb5e0) > 0000000000000000 000000027fbcf7b0 0000000000000002 0000000000000000 > 000000027fbcf850 000000027fbcf7c8 000000027fbcf7c8 00000000003b6696 > 00000000014a4e88 0000000000000007 0000000000634e00 0000000000000000 > 000000000000000d 0000000000000000 000000027fbcf818 000000000000000e > 00000000003cdc00 000000000010521a 000000027fbcf7b0 000000027fbcf7f8 > Call Trace: > ( 0000000000105174> show_trace+0x130/0x134) > 000000000019890a> __alloc_pages_internal+0x406/0x55c > 00000000001c7056> cache_grow+0x382/0x458 > 00000000001c7440> cache_alloc_refill+0x314/0x36c > 00000000001c6c12> kmem_cache_alloc+0x82/0x144 > 00000000003228f2> __alloc_skb+0x82/0x208 > 000000000032378e> dev_alloc_skb+0x36/0x64 > 000003e0001a030e> qeth_core_get_next_skb+0x31e/0x704 eth > 000003e0000d5f8c> qeth_l3_process_inbound_buffer+0x9c/0x598 eth_l3 > 000003e0000d6574> qeth_l3_qdio_input_handler+0xec/0x268 eth_l3 > 000003e0000ebc44> qdio_kick_inbound_handler+0xbc/0x178 dio > 000003e0000ee58c> __tiqdio_inbound_processing+0x394/0xdf4 dio > 000000000013a800> tasklet_action+0x10c/0x1e4 > 000000000013b908> __do_softirq+0xe0/0x1c8 > 0000000000110252> do_softirq+0xaa/0xb0 > 000000000013b772> irq_exit+0xc2/0xcc > 00000000002f6586> do_IRQ+0x132/0x1c8 > 0000000000114148> io_return+0x0/0x8 > 00000000002b850e> _raw_spin_lock_wait+0x86/0xa4 > ( 000003e047d6fa00> 0x3e047d6fa00) > 000000000019eb9c> shrink_page_list+0x1a0/0x584 > 000000000019f184> shrink_inactive_list+0x204/0x5b0 > 000000000019f620> shrink_zone+0xf0/0x1d0 > 000000000019f882> shrink_zones+0xae/0x184 > 00000000001a02be> do_try_to_free_pages+0x96/0x3fc > 00000000001a072c> try_to_free_pages+0x74/0x7c > 0000000000198730> __alloc_pages_internal+0x22c/0x55c > 000000000019b5a2> __do_page_cache_readahead+0x10a/0x2ac > 000000000019b7cc> do_page_cache_readahead+0x88/0xa8 > 000000000019170e> filemap_fault+0x33a/0x448 > 00000000001a55bc> __do_fault+0x78/0x580 > 00000000001a962e> handle_mm_fault+0x1e6/0x4c0 > 00000000003b9e1e> do_dat_exception+0x29e/0x388 > 0000000000113c0c> sysc_return+0x0/0x8 > 0000020000214bde> 0x20000214bde > Mem-Info: > DMA per-cpu: > CPU 0: hi: 186, btch: 31 usd: 0 > CPU 1: hi: 186, btch: 31 usd: 0 > Normal per-cpu: > CPU 0: hi: 186, btch: 31 usd: 0 > CPU 1: hi: 186, btch: 31 usd: 0 > Active:1355277 inactive:1132712 dirty:0 writeback:0 unstable:0 > free:9269 slab:17875 mapped:765 pagetables:24402 bounce:0 > DMA free:33220kB min:2568kB low:3208kB high:3852kB active:1092112kB > inactive:926 > 924kB present:2064384kB pages_scanned:21132286 all_unreclaimable? no > lowmem_reserveݨ: 0 8064 8064 > Normal free:3856kB min:10276kB low:12844kB high:15412kB active:4328996kB > inactiv > e:3603924kB present:8257536kB pages_scanned:44557906 all_unreclaimable? yes > lowmem_reserveݨ: 0 0 0 > DMA: 101*4kB 32*8kB 473*16kB 195*32kB 49*64kB 30*128kB 8*256kB 3*512kB > 8*1024kB > = 33220kB > Normal: 0*4kB 0*8kB 1*16kB 0*32kB 0*64kB 0*128kB 1*256kB 1*512kB 3*1024kB = > 3856 > kB > 9283 total pagecache pages > 0 pages in swap cache > Swap cache stats: add 34513958, delete 34513958, find 6612011/8393146 > Free swap = 0kB > Total swap = 35764956kB > 2621440 pages RAM > 54354 pages reserved > 22356 pages shared > 2538214 pages non-shared > The following is only an harmless informational message. > Unless you get a _continuous_flood_ of these messages it means > everything is working fine. Allocations from irqs cannot be > perfectly reliable and the kernel is designed to handle that. > java: page allocation failure. order:0, mode:0x20, alloc_flags:0x7, > pflags:0x400 > 040 > CPU: 1 Not tainted 2.6.27.45-0.1-default #1 > Process java (pid: 28831, task: 00000001ab64c638, ksp: 0000000215bbb5e0) > 0000000000000000 000000027fbcf7b0 0000000000000002 0000000000000000 > 000000027fbcf850 000000027fbcf7c8 000000027fbcf7c8 00000000003b6696 > 00000000014a5dd3 0000000000000007 0000000000634e00 0000000000000000 > 000000000000000d 0000000000000000 000000027fbcf818 000000000000000e > 00000000003cdc00 000000000010521a 000000027fbcf7b0 000000027fbcf7f8 > Call Trace: > ( 0000000000105174> show_trace+0x130/0x134) > 000000000019890a> __alloc_pages_internal+0x406/0x55c > 00000000001c7056> cache_grow+0x382/0x458 > 00000000001c7440> cache_alloc_refill+0x314/0x36c > 00000000001c6c12> kmem_cache_alloc+0x82/0x144 > 00000000003228f2> __alloc_skb+0x82/0x208 > 000000000032378e> dev_alloc_skb+0x36/0x64 > 000003e0001a030e> qeth_core_get_next_skb+0x31e/0x704 eth > 000003e0000d5f8c> qeth_l3_process_inbound_buffer+0x9c/0x598 eth_l3 > 000003e0000d6574> qeth_l3_qdio_input_handler+0xec/0x268 eth_l3 > 000003e0000ebc44> qdio_kick_inbound_handler+0xbc/0x178 dio > 000003e0000ee58c> __tiqdio_inbound_processing+0x394/0xdf4 dio > 000000000013a800> tasklet_action+0x10c/0x1e4 > 000000000013b908> __do_softirq+0xe0/0x1c8 > 0000000000110252> do_softirq+0xaa/0xb0 > 000000000013b772> irq_exit+0xc2/0xcc > 00000000002f6586> do_IRQ+0x132/0x1c8 > 0000000000114148> io_return+0x0/0x8 > 00000000002b850e> _raw_spin_lock_wait+0x86/0xa4 > ( 000003e047d6fa00> 0x3e047d6fa00) > 000000000019eb9c> shrink_page_list+0x1a0/0x584 > 000000000019f184> shrink_inactive_list+0x204/0x5b0 > 000000000019f620> shrink_zone+0xf0/0x1d0 > 000000000019f882> shrink_zones+0xae/0x184 > 00000000001a02be> do_try_to_free_pages+0x96/0x3fc > 00000000001a072c> try_to_free_pages+0x74/0x7c > 0000000000198730> __alloc_pages_internal+0x22c/0x55c > 000000000019b5a2> __do_page_cache_readahead+0x10a/0x2ac > 000000000019b7cc> do_page_cache_readahead+0x88/0xa8 > 000000000019170e> filemap_fault+0x33a/0x448 > 00000000001a55bc> __do_fault+0x78/0x580 > 00000000001a962e> handle_mm_fault+0x1e6/0x4c0 > 00000000003b9e1e> do_dat_exception+0x29e/0x388 > 0000000000113c0c> sysc_return+0x0/0x8 > 0000020000214bde> 0x20000214bde > Mem-Info: > DMA per-cpu: > CPU 0: hi: 186, btch: 31 usd: 0 > CPU 1: hi: 186, btch: 31 usd: 0 > Normal per-cpu: > CPU 0: hi: 186, btch: 31 usd: 0 > CPU 1: hi: 186, btch: 31 usd: 0 > Active:1355277 inactive:1132712 dirty:0 writeback:0 unstable:0 > free:9269 slab:17875 mapped:765 pagetables:24402 bounce:0 > DMA free:33220kB min:2568kB low:3208kB high:3852kB active:1092112kB > inactive:926 > 924kB present:2064384kB pages_scanned:21132286 all_unreclaimable? no > lowmem_reserveݨ: 0 8064 8064 > Normal free:3856kB min:10276kB low:12844kB high:15412kB active:4328996kB > inactiv > e:3603924kB present:8257536kB pages_scanned:44557906 all_unreclaimable? yes > lowmem_reserveݨ: 0 0 0 > DMA: 101*4kB 32*8kB 473*16kB 195*32kB 49*64kB 30*128kB 8*256kB 3*512kB > 8*1024kB > = 33220kB > Normal: 0*4kB 0*8kB 1*16kB 0*32kB 0*64kB 0*128kB 1*256kB 1*512kB 3*1024kB = > 3856 > kB > 9283 total pagecache pages > 0 pages in swap cache > Swap cache stats: add 34513958, delete 34513958, find 6612011/8393146 > Free swap = 0kB > Total swap = 35764956kB > 2621440 pages RAM > 54354 pages reserved > 22356 pages shared > 2538214 pages non-shared > __ratelimit: 4 callbacks suppressed > The following is only an harmless informational message. > Unless you get a _continuous_flood_ of these messages it means > everything is working fine. Allocations from irqs cannot be > perfectly reliable and the kernel is designed to handle that. > java: page allocation failure. order:0, mode:0x20, alloc_flags:0x7, > pflags:0x400 > 040 > CPU: 1 Not tainted 2.6.27.45-0.1-default #1 > Process java (pid: 28831, task: 00000001ab64c638, ksp: 0000000215bbb5e0) > 0000000000000000 000000027fbcf7b0 0000000000000002 0000000000000000 > 000000027fbcf850 000000027fbcf7c8 000000027fbcf7c8 00000000003b6696 > > etc, etc for HUNDREDS of pages.. perhaps infinite. > > 00: > 00: CP Q ALL > 00: STORAGE = 15G CONFIGURED = 15G INC = 64M STANDBY = 0 RESERVED = 0 > 00: OSA 039C ATTACHED TO TCPIP 039C DEVTYPE OSA CHPID 01 OSD > 00: OSA 039D ATTACHED TO TCPIP 039D DEVTYPE OSA CHPID 01 OSD > 00: OSA 039E ATTACHED TO TCPIP 039E DEVTYPE OSA CHPID 01 OSD > 00: OSA 03A0 ATTACHED TO DTCVSW2 03A0 DEVTYPE OSA CHPID 01 OSD > 00: OSA 03A1 ATTACHED TO DTCVSW2 03A1 DEVTYPE OSA CHPID 01 OSD > 00: OSA 03A2 ATTACHED TO DTCVSW2 03A2 DEVTYPE OSA CHPID 01 OSD > 00: OSA 03C0 ATTACHED TO DTCVSW1 03C0 DEVTYPE OSA CHPID 02 OSD > 00: OSA 03C1 ATTACHED TO DTCVSW1 03C1 DEVTYPE OSA CHPID 02 OSD > 00: OSA 03C2 ATTACHED TO DTCVSW1 03C2 DEVTYPE OSA CHPID 02 OSD > 00: FCP 5000 ATTACHED TO LINXDEV 5000 CHPID 46 > 00: WWPN C05076FAE3000400 > 00: FCP 5001 ATTACHED TO LINXD001 5001 CHPID 46 > 00: WWPN C05076FAE3000404 > 00: FCP 5002 ATTACHED TO LINXD002 5002 CHPID 46 > 00: WWPN C05076FAE3000408 > 00: FCP 5003 ATTACHED TO LINXD003 5003 CHPID 46 > 00: WWPN C05076FAE300040C > 00: FCP 5100 ATTACHED TO LINXDEV 5100 CHPID 47 > 00: WWPN C05076FAE3000900 > 00: FCP 5101 ATTACHED TO LINXD001 5101 CHPID 47 > 00: WWPN C05076FAE3000904 > 00: FCP 5102 ATTACHED TO LINXD002 5102 CHPID 47 > 00: WWPN C05076FAE3000908 > 00: FCP 5103 ATTACHED TO LINXD003 5103 CHPID 47 > 00: WWPN C05076FAE300090C > 00: DASD 9F7D CP SYSTEM VM6LXD 0 > 00: DASD 9F7E CP SYSTEM VM6LXE 0 > 00: DASD 9F80 CP SYSTEM VM6LX9 2 > 00: DASD 9F81 CP SYSTEM VM6LXA 2 > 00: DASD 9F82 CP SYSTEM VM6LXB 0 > 00: DASD 9F83 CP SYSTEM VM6LXC 0 > 00: DASD 9F84 CP OWNED VM6RES 135 > 00: DASD 9F85 CP OWNED VM6SPL 0 > 00: DASD 9F86 CP OWNED VM6PG1 0 > 00: DASD 9F87 CP OWNED VM6PG2 0 > 00: DASD 9F88 CP OWNED VM6LX1 4 > 00: DASD 9F89 CP SYSTEM VM6LX2 0 > 00: DASD 9F8A CP SYSTEM VM6LX3 0 > 00: DASD 9F8B CP SYSTEM VM6LX4 0 > 00: DASD 9F8C CP SYSTEM VM6LX5 2 > 00: DASD 9F8D CP SYSTEM VM6LX6 0 > 00: DASD 9F8E CP SYSTEM VM6LX7 0 > 00: DASD 9F8F CP SYSTEM VM6LX8 2 > 00: DASD 9FC7 CP SYSTEM VM6LX6 0 > 00: DASD 9FC8 CP SYSTEM VM6LX5 2 > 00: DASD 9FC9 CP SYSTEM VM6LX2 0 > 00: DASD 9FCA CP SYSTEM VM6LX4 0 > 00: DASD 9FCB CP SYSTEM VM6LX3 0 > 00: DASD 9FCE CP SYSTEM VM6LX1 4 > 00: DASD 9FCF CP SYSTEM VM6PG2 0 > 00: DASD 9FD0 CP SYSTEM VM6PG1 0 > 00: DASD 9FD1 CP SYSTEM VM6SPL 0 > 00: DASD 9FD2 CP SYSTEM VM6RES 135 > > > > >