Re: [lustre-discuss] LNET ports and connections
LNet is a peer-to-peer protocol, it has no concept of client and server. If one host needs to send a message to another but doesn't already have a connection, it creates a new connection. I don't yet know enough specifics of the lustre protocol to be certain of the circumstances when a lustre server will need to initiate a message to a client, but I imagine that recalling a lock might be one. I think you should assume that any LNet node might receive a connection from any other LNet node (for which they share an LNet network), and that the connection could come from any port between 512 and 1023 (LNET_ACCEPTOR_MIN_PORT to LNET_ACCEPTOR_MAX_PORT). NeilBrown On Mon, Feb 17 2020, Degremont, Aurelien wrote: > Hi all, > > From what I've understood so far, LNET listens on port 988 by default and > peers connect to it using 1021-1023 TCP ports as source ports. > At Lustre level, servers listen on 988 and clients connect to them using the > same source ports 1021-1023. > So only accepting connections to port 988 on server side sounded pretty safe > to me. However, I've seen connections from 1021-1023 to 988, from server > hosts to client hosts sometimes. > I can't understand what mechanism could trigger these connections. Did I miss > something? > > Thanks > > Aurélien > > ___ > lustre-discuss mailing list > lustre-discuss@lists.lustre.org > http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org signature.asc Description: PGP signature ___ lustre-discuss mailing list lustre-discuss@lists.lustre.org http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
[lustre-discuss] enable quota enforcement on the fly?
We recently noticed we apparently did not enable group quota enforcement early last year during the most recent rebuild of our Lustre filesystem. Is it possible to do so on the fly, or is it better/required for the filesystem to be quiesced first? We are using Lustre 2.10.7 with ZFS 0.7.5 (project quotas are not needed). [loforbes@mds01 ~]$ lctl get_param osd-*.*.quota_slave.info osd-zfs.lustre2-MDT.quota_slave.info= target name:lustre2-MDT pool ID:0 type: md quota enabled: none conn to master: setup space acct: ug user uptodate: glb[0],slv[0],reint[0] group uptodate: glb[0],slv[0],reint[0] project uptodate: glb[0],slv[0],reint[0] Thank you! -- Regards, -liam -There are uncountably more irrational fears than rational ones. -P. Dolan Liam Forbes lofor...@alaska.edu ph: 907-450-8618 fax: 907-450-8601 UAF Research Computing Systems Senior HPC Engineer CISSP ___ lustre-discuss mailing list lustre-discuss@lists.lustre.org http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
[lustre-discuss] LNET ports and connections
Hi all, From what I've understood so far, LNET listens on port 988 by default and peers connect to it using 1021-1023 TCP ports as source ports. At Lustre level, servers listen on 988 and clients connect to them using the same source ports 1021-1023. So only accepting connections to port 988 on server side sounded pretty safe to me. However, I've seen connections from 1021-1023 to 988, from server hosts to client hosts sometimes. I can't understand what mechanism could trigger these connections. Did I miss something? Thanks Aurélien ___ lustre-discuss mailing list lustre-discuss@lists.lustre.org http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
Re: [lustre-discuss] Jobstats harvesting
On Mon., 17 Feb. 2020, 18:06 Andreas Dilger, wrote: > You don't mention which Lustre release you are using, but newer > releases allow "complex JobIDs" that can contain both the SLURMJobID > as well as other constant strings (e.g. cluster name), hostname, UID, GID, > and process name. > Yeah, i twigged that once I'd sent the mail: we're still 2.10.8 in production, so having the option of the more complex jobid string is another reason for upgrading Related, ive found the DDN fork of collectd, and i see the lustre2.c plugin is GPL2 but are there any plans to get it merged upstream? Andrew (Also who's mad enough to be running mythtv on lustre judging from the examples?) > ___ lustre-discuss mailing list lustre-discuss@lists.lustre.org http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
[lustre-discuss] Speed of deleting small files on OST vs DoM
Hi! Is there a good reason why deleting lots of small files (io500 test md_easy/hard_delete) with the files on OSTs are up to two times faster then when using DoM with the whole file(s) on the MDT? Using server/client 2.13.0 DoM up to 64k, test files < 4k I can see that the actual data deletion with the data on OST is asynchronous, but I see no reason for it to be almost two times faster. Both MDT's and OSTs are SSDs. The situation is basically the same for a single task and for multiple clients with multiple tasks/client. -- Ake Sandgren, HPC2N, Umea University, S-90187 Umea, Sweden Internet: a...@hpc2n.umu.se Phone: +46 90 7866134 Fax: +46 90-580 14 Mobile: +46 70 7716134 WWW: http://www.hpc2n.umu.se ___ lustre-discuss mailing list lustre-discuss@lists.lustre.org http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org