Gregory:

Do you have a z114 with 10 IFLs? That is the maximum number of IFLs
available on a z114 (2818-M10) and would be unusual. Is this a single z/VM
LPAR? How much memory is on the z114 (and in this LPAR)? Also, what was the
specific MT/Model for the P7 box?

If you were to compare a 12-core Power 730 (8231-E2C) to a 10 IFL z114 the
Power system has 1.4 to 2.0 times the capacity of the z114.

Jim Elliott
Senior IT Consultant - GlassHouse Systems Inc.


On Tue, Nov 3, 2020 at 8:47 AM Grzegorz Powiedziuk <gpowiedz...@gmail.com>
wrote:

> Hi, I could use some ideas. We moved a huge db2 from old p7 aix to rhel7 on
> Z and we are having big performance issues.
> Same memory, CPU number is down from 12 to 10.  Although they had
> multithreading ON so they saw more "cpus" We have faster disks (moved to
> flash), faster FCP cards and faster network adapters.
> We are running on z114 and at this point that is practically the only VM
> running with IFLs on this box.
>
> It seems that when "jobs" run on their own, they finish faster than what
> they were getting on AIX.
> But problems start if there is more than we can chew. So either few jobs
> running at the same time or some reorgs running in the database.
>
> Load average goes to 150-200, cpus are at 100%  (kernel time can go to
> 20-30% ) but no iowaits.
> Plenty of memory available.
> At this point everything becomes extremely slow, people are starting having
> problems with connecting to db2 (annd sshing), basically it becomes a
> nightmare
>
> This db2 is massive (30+TB) and it is a multinode configuration (17 nodes
> running on the same host). We moved it like this 1:1 from that old AIX.
>
> DB2 is running on the ext4 filesystem (Actually a huge number of
> filesystems- each NODE is a separate logical volume). Separate for logs,
> data.
>
> If this continues like this, we will add 2 cpus but I have a feeling that
> it will not make much difference.
>
> I know that we end up with a massive number of processes and a massive
> number of file descriptors (lsof sice it shows also threads now, is
> practically useless - it would run for way too long - 10-30 minutes
> probably) .
>
> A snapshot from just now:
>
> top - 08:37:50 up 11 days, 12:04, 28 users,  load average: 188.29, 151.07,
> 133.54
> Tasks: 1843 total,  11 running, 1832 sleeping,   0 stopped,   0 zombie
> %Cpu0  : 76.3 us, 16.6 sy,  0.0 ni,  0.0 id,  0.0 wa,  1.0 hi,  3.2 si,
>  2.9 st
> %Cpu1  : 66.1 us, 31.3 sy,  0.0 ni,  0.0 id,  0.0 wa,  0.6 hi,  1.3 si,
>  0.6 st
> %Cpu2  : 66.9 us, 31.2 sy,  0.0 ni,  0.0 id,  0.0 wa,  0.3 hi,  1.3 si,
>  0.3 st
> %Cpu3  : 74.7 us, 23.4 sy,  0.0 ni,  0.0 id,  0.0 wa,  0.3 hi,  1.3 si,
>  0.3 st
> %Cpu4  : 86.7 us, 10.7 sy,  0.0 ni,  0.0 id,  0.0 wa,  0.6 hi,  1.3 si,
>  0.6 st
> %Cpu5  : 83.8 us, 13.6 sy,  0.0 ni,  0.0 id,  0.0 wa,  0.6 hi,  1.6 si,
>  0.3 st
> %Cpu6  : 81.6 us, 15.2 sy,  0.0 ni,  0.0 id,  0.0 wa,  0.6 hi,  1.9 si,
>  0.6 st
> %Cpu7  : 70.6 us, 26.2 sy,  0.0 ni,  0.0 id,  0.0 wa,  0.6 hi,  1.9 si,
>  0.6 st
> %Cpu8  : 70.5 us, 26.6 sy,  0.0 ni,  0.0 id,  0.0 wa,  0.6 hi,  1.6 si,
>  0.6 st
> %Cpu9  : 84.1 us, 13.6 sy,  0.0 ni,  0.0 id,  0.0 wa,  0.3 hi,  1.3 si,
>  0.6 st
> KiB Mem : 15424256+total,  1069280 free, 18452168 used, 13472112+buff/cache
> KiB Swap: 52305904 total, 51231216 free,  1074688 used. 17399028 avail Mem
>
> Where  can I look for potential relief? Everyone was hoping for a better
> performance not worse.I am hoping that there is something we can tweak to
> make this better.
> I will appreciate any ideas!
> thanks
> Gregory
>
> ----------------------------------------------------------------------
> For LINUX-390 subscribe / signoff / archive access instructions,
> send email to lists...@vm.marist.edu with the message: INFO LINUX-390 or
> visit
> http://www2.marist.edu/htbin/wlvindex?LINUX-390
>

----------------------------------------------------------------------
For LINUX-390 subscribe / signoff / archive access instructions,
send email to lists...@vm.marist.edu with the message: INFO LINUX-390 or visit
http://www2.marist.edu/htbin/wlvindex?LINUX-390

Reply via email to