Re: [gpfsug-discuss] Request for folks using encryption on SKLM, run a word count

2020-09-11 Thread J. Eric Wonderley
We have spectrum archive with encryption on disk and tape. We get maybe a 100 or so messages like this daily. It would be nice if message had some information about which client is the issue. We have had client certs expire in the past. The root cause of the outage was a network outage...iirc

Re: [gpfsug-discuss] gpfs filesets question

2020-04-16 Thread J. Eric Wonderley
> > Fred > __ > Fred Stock | IBM Pittsburgh Lab | 720-430-8821 > sto...@us.ibm.com > > > > ----- Original message - > From: "J. Eric Wonderley" > Sent by: gpfsug-discuss-boun...@spectrumscale.org > To: gpfsug ma

[gpfsug-discuss] gpfs filesets question

2020-04-16 Thread J. Eric Wonderley
I have filesets setup in a filesystem...looks like: [root@cl005 ~]# mmlsfileset home -L Filesets in file system 'home': NameId RootInode ParentId Created InodeSpace MaxInodesAllocInodes Comment root 0 3

Re: [gpfsug-discuss] GPFS v5: Blocksizes and subblocks

2019-03-27 Thread J. Eric Wonderley
mmlspool might suggest there's only 1 system pool per cluster. We have 2 clusters and it has id=0 on both. One of our clusters has 2 filesystems that have same id for two different dataonly pools: [root@cl001 ~]# mmlspool home all NameId system 0 fc_8T65537 fc_ss

Re: [gpfsug-discuss] Multihomed nodes and failover networks

2018-10-26 Thread J. Eric Wonderley
Multihoming is accomplished by using subnets...see mmchconfig. Failover networks on the other hand are not allowed. Bad network behavior is dealt with by expelling nodes. You must have decent/supported network gear...we have learned that lesson the hard way On Fri, Oct 26, 2018 at 10:37 AM Luka

Re: [gpfsug-discuss] Mixing RDMA Client Fabrics for a single NSD Cluster

2018-07-19 Thread J. Eric Wonderley
Hi Stephan: I think every node in C1 and in C2 have to see every node in the server cluster NSD-[AD]. We have a 10 node server cluster where 2 nodes do nothing but server out nfs. Since these two are apart of the server cluster...client clusters wanting to mount the server cluster via gpfs need

Re: [gpfsug-discuss] GPFS Encryption

2018-03-26 Thread J. Eric Wonderley
Hi Gareth: We have the spectrum archive product with encryption. It encrypts data on disk and tape...but not metadata. We originally had hoped to write small files with metadata...that does not happen with encryption. My guess is that the system pool(where metadata lives) cannot be encrypted.

Re: [gpfsug-discuss] more than one mlx connectx-4 adapter in same host

2017-12-20 Thread J. Eric Wonderley
Ethernet based because of the mention of Juniper. > > Are you attempting to do RoCE or just plain TCP/IP? > > > On December 20, 2017 at 14:40:48 EST, J. Eric Wonderley < > eric.wonder...@vt.edu> wrote: > > Hello: > > Does anyone have this type of config? > >

[gpfsug-discuss] more than one mlx connectx-4 adapter in same host

2017-12-20 Thread J. Eric Wonderley
Hello: Does anyone have this type of config? The host configuration looks sane but we seem to observe link-down on all mlx adapters no matter what we do. Big picture is that we are attempting to do mc(multichassis)-lags to a core switch. I'm somewhat fearful as to how this is implemented in the

[gpfsug-discuss] mm'add|del'node with ccr enabled

2017-12-08 Thread J. Eric Wonderley
Hello: If I recall correctly this does not work...correct? I think the last time I attempted this was gpfs version <=4.1. I think I attempted to add a quorum node. The process was that I remember doing was mmshutdown -a, mmchcluster --ccr-disable, mmaddnode yadayada, mmchcluster --ccr-enable, m

[gpfsug-discuss] sas avago/lsi hba reseller recommendation

2017-08-28 Thread J. Eric Wonderley
We have several avago/lsi 9305-16e that I believe came from Advanced HPC. Can someone recommend a another reseller of these hbas or a contact with Advance HPC? ___ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/

Re: [gpfsug-discuss] mmsetquota produces error

2017-08-17 Thread J. Eric Wonderley
gt; > NameId RootInode ParentId > Created InodeSpace MaxInodesAllocInodes > Comment > > setquotafoo2518408295 0 Thu Aug 17 > 15:17:18 201700 0 > >

Re: [gpfsug-discuss] mmsetquota produces error

2017-08-17 Thread J. Eric Wonderley
, 2017 at 9:43 AM, Edward Wahl wrote: > On Fri, 4 Aug 2017 01:02:22 -0400 > "J. Eric Wonderley" wrote: > > > 4.2.2.3 > > > > I want to think maybe this started after expanding inode space > > What does 'mmlsfileset home nathanfootest -L&#

Re: [gpfsug-discuss] data integrity documentation

2017-08-04 Thread J. Eric Wonderley
i actually hit this assert and turned it in to support on this version: Build branch "4.2.2.3 efix6 (987197)". i was told do to exactly what sven mentioned. i thought it strange that i did NOT hit the assert in a no pass but hit it in a yes pass. On Thu, Aug 3, 2017 at 9:06 AM, Sven Oehme wrote

Re: [gpfsug-discuss] mmsetquota produces error

2017-08-03 Thread J. Eric Wonderley
4.2.2.3 I want to think maybe this started after expanding inode space On Thu, Aug 3, 2017 at 9:11 AM, James Davis wrote: > Hey, > > Hmm, your invocation looks valid to me. What's your GPFS level? > > Cheers, > > Jamie > > > ----- Original message - >

[gpfsug-discuss] mmsetquota produces error

2017-08-02 Thread J. Eric Wonderley
for one of our home filesystem we get: mmsetquota home:nathanfootest --block 10T:10T --files 10M:10M tssetquota: Could not get id of fileset 'nathanfootest' error (22): 'Invalid argument'. mmedquota -j home:nathanfootest does work however ___ gpfsug-d

Re: [gpfsug-discuss] data integrity documentation

2017-08-02 Thread J. Eric Wonderley
No guarantee...unless you are using ess/gss solution. Crappy network will get you loads of expels and occasional fscks. Which I guess beats data loss and recovery from backup. YOu probably have a network issue...they can be subtle. Gpfs is a very extremely thorough network tester. Eric On We

Re: [gpfsug-discuss] Quota and hardlimit enforcement

2017-07-31 Thread J. Eric Wonderley
Hi Renar: What does 'mmlsquota -j fileset filesystem' report? I did not think you would get a grace period of none unless the hardlimit=softlimit. On Mon, Jul 31, 2017 at 1:44 PM, Grunenberg, Renar < renar.grunenb...@huk-coburg.de> wrote: > Hallo All, > we are on Version 4.2.3.2 and see some mi

[gpfsug-discuss] gui related connection fail in gpfs logs

2017-06-20 Thread J. Eric Wonderley
These type messages repeat often in our logs: 017-06-20_09:25:13.676-0400: [E] An%20attempt%20to%20send%20notification%20to%20the%20GUI%20subsystem%20failed%2E%20response%3Dcurl%3A%20%287%29%20Failed%20connect%20to%20arproto2%2Ear%2Enis%2Eisb%2Einternal%3A443%3B%20Connection%20refused%20rc%3D7 rc=

Re: [gpfsug-discuss] help with multi-cluster setup: Network is unreachable

2017-05-08 Thread J. Eric Wonderley
Hi Jamie: I think typically you want to keep the clients ahead of the server in version. I would advance the version of you client nodes. New clients can communicate with older versions of server nsds. Vice versa...no so much. ___ gpfsug-discuss maili

Re: [gpfsug-discuss] GPFS Network Configuration - 1 Daemon Network , 1 Admin Network

2017-04-10 Thread J. Eric Wonderley
1) You want more that one quorum node on your server cluster. The non-quorum node does need a daemon network interface exposed to the client cluster as does the quorum nodes. 2) No. Admin network is for intra cluster communications...not inter cluster(between clusters). Daemon interface(port

Re: [gpfsug-discuss] snapshots & tiering in a busy filesystem

2017-03-22 Thread J. Eric Wonderley
apshot gpfs | wc -l > > 6916 > > > > *From:* gpfsug-discuss-boun...@spectrumscale.org [mailto:gpfsug-discuss- > boun...@spectrumscale.org] *On Behalf Of *J. Eric Wonderley > *Sent:* 20 March 2017 14:03 > *To:* gpfsug main discussion list > *Subject:* [gpfsug-discuss] snapshot

[gpfsug-discuss] snapshots & tiering in a busy filesystem

2017-03-20 Thread J. Eric Wonderley
I found this link and it didn't give me much hope for doing snapshots & backup in a home(busy) filesystem: http://www.spectrumscale.org/pipermail/gpfsug-discuss/2013- February/000200.html I realize this is dated and I wondered if qos etc have made is a tolerable thing to do now. Gpfs I think was

Re: [gpfsug-discuss] Fw: mmbackup examples using policy

2017-02-15 Thread J. Eric Wonderley
al Parallel File > System Dev. > Pittsburgh, PA (412) 667-6993 Tie-Line 989-6993 >sber...@us.ibm.com > Every once in a while, it is a good idea to call out, "Computer, end > program!" just to check. --David Noelle > All Your Bas

[gpfsug-discuss] mmbackup examples using policy

2017-02-13 Thread J. Eric Wonderley
Anyone have any examples of this? I have a filesystem that has 2 pools and several filesets and would like daily progressive incremental backups of its contents. I found some stuff here(nothing real close to what I wanted however): /usr/lpp/mmfs/samples/ilm I have the tsm client installed on the

Re: [gpfsug-discuss] proper gpfs shutdown when node disappears

2017-02-03 Thread J. Eric Wonderley
Well we got it into the down state using mmsdrrestore -p to recover stuff into /var/mmfs/gen to cl004. Anyhow we ended up unknown for cl004 when it powered off. Short of removing node, unknown is the state you get. Unknown seems stable for a hopefully short outage of cl004. Thanks On Thu, Feb

[gpfsug-discuss] proper gpfs shutdown when node disappears

2017-02-02 Thread J. Eric Wonderley
Is there a way to accomplish this so the rest of cluster knows its down? My state now: [root@cl001 ~]# mmgetstate -aL cl004.cl.arc.internal: mmremote: determineMode: Missing file /var/mmfs/gen/mmsdrfs. cl004.cl.arc.internal: mmremote: This node does not belong to a GPFS cluster. mmdsh: cl004.cl.

Re: [gpfsug-discuss] Path to NSD lost when host_sas_address changed on port

2017-01-20 Thread J. Eric Wonderley
Maybe multipath is not seeing all of the wwns? multipath -v3 | grep ^51855 look ok? For some unknown reason multipath does not see our sandisk array...we have to add them to the end of /etc/multipath/wwids file On Fri, Jan 20, 2017 at 10:32 AM, David D. Johnson wrote: > We have most of our GP

[gpfsug-discuss] rmda errors scatter thru gpfs logs

2017-01-17 Thread J. Eric Wonderley
I have messages like these frequent my logs: Tue Jan 17 11:25:49.731 2017: [E] VERBS RDMA rdma write error IBV_WC_REM_ACCESS_ERR to 10.51.10.5 (cl005) on mlx5_0 port 1 fabnum 0 vendor_err 136 Tue Jan 17 11:25:49.732 2017: [E] VERBS RDMA closed connection to 10.51.10.5 (cl005) on mlx5_0 port 1 fabnu

Re: [gpfsug-discuss] Authorized Key Messages

2017-01-13 Thread J. Eric Wonderley
Our intent was to have ccr turned off since all nodes are quorum in the server cluster: Considering this: [root@cl001 ~]# mmfsadm dump config | grep -i ccr ! ccrEnabled 0 ccrMaxChallengeCheckRetries 4 ccr : 0 (cluster configuration repository) ccr : 1 (cluster c

Re: [gpfsug-discuss] replication and no failure groups

2017-01-09 Thread J. Eric Wonderley
> Fax: +972-3-916-5672 <+972%203-916-5672> > Mobile: +972-52-8395593 <+972%2052-839-5593> > e-mail: y...@il.ibm.com > *IBM Israel* <http://www.ibm.com/il/he/> > > > > > > > > From:"J. Eric Wonderley" > To:gpfsug main di

Re: [gpfsug-discuss] replication and no failure groups

2017-01-09 Thread J. Eric Wonderley
Hi Yaron: This is the filesystem: [root@cl005 net]# mmlsdisk work disk driver sector failure holds holdsstorage name type size group metadata data status availability pool -- --- - --

Re: [gpfsug-discuss] nsd not adding with one quorum node down?

2017-01-05 Thread J. Eric Wonderley
..@spectrumscale.org [mailto:gpfsug-discuss- > boun...@spectrumscale.org] *On Behalf Of *J. Eric Wonderley > *Sent:* Thursday, January 05, 2017 2:01 PM > *To:* gpfsug main discussion list > *Subject:* [gpfsug-discuss] nsd not adding with one quorum node down? > > > > I have one quorum node do

[gpfsug-discuss] nsd not adding with one quorum node down?

2017-01-05 Thread J. Eric Wonderley
I have one quorum node down and attempting to add a nsd to a fs: [root@cl005 ~]# mmadddisk home -F add_1_flh_home -v no |& tee /root/adddisk_flh_home.out Verifying file system configuration information ... The following disks of home will be formatted on node cl003: r10f1e5: size 1879610 MB Ex

[gpfsug-discuss] strange mmchnsd error?

2017-01-04 Thread J. Eric Wonderley
[root@cl001 ~]# cat chnsd_home_flh %nsd: nsd=r10f1e5 servers=cl008,cl001,cl002,cl003,cl004,cl005,cl006,cl007 %nsd: nsd=r10f6e5 servers=cl007,cl008,cl001,cl002,cl003,cl004,cl005,cl006 %nsd: nsd=r10f1e6 servers=cl006,cl007,cl008,cl001,cl002,cl003,cl004,cl005 %nsd: nsd=r10f6e6 servers=cl005,cl006,cl00

Re: [gpfsug-discuss] Quotas on Multiple Filesets

2016-12-02 Thread J. Eric Wonderley
Hi Michael: I was about to ask a similar question about nested filesets. I have this setup: [root@cl001 ~]# mmlsfileset home Filesets in file system 'home': Name StatusPath root Linked/gpfs/home groupLinked/gpfs/home/group pr

Re: [gpfsug-discuss] rpldisk vs deldisk & adddisk

2016-12-02 Thread J. Eric Wonderley
disks were removed. > Can do multiple that why and us the entire cluster to move data if you > want. > > On 12/1/16 1:10 PM, J. Eric Wonderley wrote: > > I have a few misconfigured disk groups and I have a few same size > correctly configured disk groups. > > Is there an

[gpfsug-discuss] rpldisk vs deldisk & adddisk

2016-12-01 Thread J. Eric Wonderley
I have a few misconfigured disk groups and I have a few same size correctly configured disk groups. Is there any (dis)advantage to running mmrpldisk over mmdeldisk and mmadddisk? Everytime I have ever run mmdeldisk...it been somewhat painful(even with qos) process. ___

Re: [gpfsug-discuss] wanted...gpfs policy that places larger files onto a pool based on size

2016-10-31 Thread J. Eric Wonderley
ot use the size >>> of the file to determine the placement of the file in a GPFS Storage Pool. >>> This is because GPFS has no idea what the file size will be when the file >>> is open()’d for writing. >>> >>> >>> >>> Hope that helps! >

[gpfsug-discuss] wanted...gpfs policy that places larger files onto a pool based on size

2016-10-31 Thread J. Eric Wonderley
I wanted to do something like this... [root@cl001 ~]# cat /opt/gpfs/home.ply /*Failsafe migration of old small files back to spinning media pool(fc_8T) */ RULE 'theshold' MIGRATE FROM POOL 'system' THRESHOLD(90,70) WEIGHT(ACCESS_TIME) TO POOL 'fc_8T' /*Write files larger than 16MB to pool called "

[gpfsug-discuss] migrate policy vs restripe

2016-10-04 Thread J. Eric Wonderley
We have the need to move data from one set of spindles to another. Are there any performance or availability considerations when choosing to do either a migration policy or a restripe to make this move? I did discover that a restripe only works within the same pool...even though you setup two poo

Re: [gpfsug-discuss] Blocksize

2016-09-22 Thread J. Eric Wonderley
It defaults to 4k: mmlsfs testbs8M -i flagvaluedescription --- --- -i 4096 Inode size in bytes I think you can make as small as 512b. Gpfs will store