[gpfsug-discuss] Strategies for keeping GPFS copy for Disaster Recovery

2021-04-07 Thread David Johnson
We plan to use rsync to keep a DR copy of our filesystem. The production filesystem contains hundreds of dependent fillets and a much smaller number of independent filesets. A question for any of you folks out there with a similar situation: do you synchronize filesets in parallel on the DR copy?

[gpfsug-discuss] internal details on GPFS inode expansion

2020-11-30 Thread David Johnson
When GPFS needs to add inodes to the filesystem, it seems to pre-create about 4 million of them. Judging by the logs, it seems it only takes a few (13 maybe) seconds to do this. However we are suspecting that this might only be to request the additional inodes and that there is some background a

[gpfsug-discuss] mmremotecluster access from SS 5.0.x to 4.2.3-x refuses id_rsa.pub

2020-06-11 Thread David Johnson
I'm trying to access an old GPFS filesystem from a new cluster. It is good up to the point of adding the SSL keys of the old cluster on the new one. I get from mmremotecluster add command: File _id_rsa.pub does not contain a nist sp 800-131a compliance key Is there any way to override this

Re: [gpfsug-discuss] GPFS 5 and supported rhel OS

2020-02-20 Thread David Johnson
Instead of keeping whole legacy systems around, could they achieve the same with a container built from the legacy software? > On Feb 20, 2020, at 11:59 AM, Skylar Thompson wrote: > > On Thu, Feb 20, 2020 at 04:29:40PM +, Ken Atkinson wrote: >> Fred, >> It may be that some HPC users "have t

Re: [gpfsug-discuss] question about spectrum scale 5.0.3 installer

2019-10-23 Thread David Johnson
By the way, we have been dealing with adding and deleting nodes manually since GPFS 3.4, back in 2009. At what point is the spectrumscale command line utility more trouble than it’s worth? > On Oct 23, 2019, at 11:19 AM, David Johnson wrote: > > I built a test cluster a month

[gpfsug-discuss] question about spectrum scale 5.0.3 installer

2019-10-23 Thread David Johnson
I built a test cluster a month ago on 14 nodes. Today I want to install two more NSD nodes. When I tried to run the installer, it looks like it is going back and fiddling with the nodes that were installed earlier, and are up and running with the filesystem mounted. I ended up having to abort

[gpfsug-discuss] CIFS protocol access does not honor secondary groups

2019-10-02 Thread David Johnson
After converting from clustered CIFS to CES protocols, we’ve noticed that SMB users can’t access files owned by groups that they are members of, unless that group happens to be their primary group. Have read the smb.conf man page, and don’t see anything obvious that would control this… What migh

Re: [gpfsug-discuss] Building GPFS filesystem system data pool on shared nothing NVMe drives

2019-07-30 Thread David Johnson
n...@spectrumscale.org > <mailto:gpfsug-discuss-boun...@spectrumscale.org> > <mailto:gpfsug-discuss-boun...@spectrumscale.org>> On Behalf Of David Johnson > Sent: Tuesday, July 30, 2019 7:46 AM > To: gpfsug main discussion list <mailto:gpfsug-discuss@spectrumscale.org&g

Re: [gpfsug-discuss] Building GPFS filesystem system data pool on shared nothing NVMe drives

2019-07-30 Thread David Johnson
ps://www.ibm.com/support/knowledgecenter/STXKQY_ECE_5.0.3/com.ibm.spectrum.scale.ece.v5r03.doc/b1lece_intro.htm > > <https://www.ibm.com/support/knowledgecenter/STXKQY_ECE_5.0.3/com.ibm.spectrum.scale.ece.v5r03.doc/b1lece_intro.htm> > -- > Cheers > > El 29 jul 2019, a las 19:06, David

[gpfsug-discuss] Building GPFS filesystem system data pool on shared nothing NVMe drives

2019-07-29 Thread David Johnson
We are planning a 5.0.x upgrade onto new hardware to make use of the new 5.x GPFS features. The goal is to use up to four NSD nodes for metadata, each one with 6 NVMe drives (to be determined whether we use Intel VROC for raid 5 or raid 1, or just straight disks). So questions — Has anyone do

[gpfsug-discuss] Spectrum Scale Standard 4.2.3-13 download broken

2019-03-21 Thread David Johnson
I tried twice to download the latest PTF, but the md5sum did not match and the package will not install. Succeeded with the Protocols version. There is no link on the web page to report problems, so I'm posting here hoping someone can get it fixed. -- ddj __

[gpfsug-discuss] Rebalancing with mmrestripefs -P

2018-08-20 Thread David Johnson
I have one storage pool that was recently doubled, and another pool migrated there using mmapplypolicy. The new half is only 50% full, and the old half is 94% full. Disks in storage pool: cit_10tb (Maximum disk size allowed is 516 TB) d05_george_23 50.49T 23 No Yes

[gpfsug-discuss] recommendations for gpfs 5.x GUI and perf/health monitoring collector nodes

2018-05-31 Thread David Johnson
We are planning to bring up the new ZIMon tools on our 450+ node cluster, and need to purchase new nodes to run the collector federation and GUI function on. What would you choose as a platform for this? — memory size? — local disk space — SSD? shared? — net attach — 10Gig? 25Gig? IB? — CPU

Re: [gpfsug-discuss] Singularity + GPFS

2018-04-26 Thread David Johnson
Regarding MPI-IO, how do you mean “building the applications against GPFS”? We try to advise our users about things to avoid, but we have some poster-ready “chaos monkeys” as well, who resist guidance. What apps do your users favor? Molpro is one of our heaviest apps right now. Thanks, — ddj >

Re: [gpfsug-discuss] Converting a dependent fileset to independent

2018-04-25 Thread David Johnson
Yes, independent snapshotting would be an issue. However at the moment we have 570 dependent filesets in our main filesystem, which is not all that far from the limit of 1000 independent filesets per filesystem. There was a thread concerning fileset issues back in February, wondering if the l

Re: [gpfsug-discuss] mmrestripefs "No space left on device"

2017-11-02 Thread David Johnson
One thing that may be relevant is if you have snapshots, depending on your release level, inodes in the snapshot may considered immutable, and will not be migrated. Once the snapshots have been deleted, the inodes are freed up and you won’t see the (somewhat misleading) message about no space.

Re: [gpfsug-discuss] Permissions issue in GPFS 4.2.3-4?

2017-08-30 Thread David Johnson
We ran into this back in mid February. Never really got a satisfactory answer how it got this way, the thought was that a bunch of nodes were expelled during an mmchconfig, and the files ended up with the wrong permissions. — ddj > On Aug 30, 2017, at 3:26 PM, Buterbaugh, Kevin L > wrote:

Re: [gpfsug-discuss] mmsysmon.py revisited

2017-07-19 Thread David Johnson
r/mmfs/mmsysmon/ > > mmsysmonitor.conf. > > > > It’s a huge problem. I don’t understand why it hasn’t been given > > much credit by dev or support. > > > > ~jonathon > > > > > > On 7/18/17, 11:21 AM, "gpfsug-discuss-boun...@spectrumsc

[gpfsug-discuss] mmsysmon.py revisited

2017-07-18 Thread David Johnson
We also noticed a fair amount of CPU time accumulated by mmsysmon.py on our diskless compute nodes. I read the earlier query, where it was answered: > ces == Cluster Export Services, mmsysmon.py comes from mmcesmon. It is used > for managing export services of GPFS. If it is killed, your nfs/sm