[slurm-users] SLURM configuration help

2024-04-04 Thread Alison Peterson via slurm-users
I am writing to seek assistance with a critical issue on our single-node system managed by Slurm. Our jobs are queued and marked as awaiting resources, but they are not starting despite seeming availability. I'm new with SLURM and my only experience was a class on installing it so I have no experie

[slurm-users] SLURM configuration for LDAP users

2024-02-03 Thread Richard Chang via slurm-users
Hi, I am a little new to this, so please pardon my ignorance. I have configured slurm in my cluster and it works fine with local users. But I am not able to get it working with LDAP/SSSD authentication. User logins using ssh are working fine. An LDAP user can login to the login, slurmctld an

[slurm-users] Slurm Configuration assistance: Unable to use srun after installation (slurm on fedora 33)

2021-04-18 Thread Johnsy K. John
Hello SchedMD team, I would like to use your slurm workload manager for learning purposes. And I tried installing the the software (downloaded from: https://www.schedmd.com/downloads.php ) and followed the steps as mentioned in: https://slurm.schedmd.com/download.html https://slurm.schedmd.com/qu

Re: [slurm-users] Slurm configuration, Weight Parameter

2019-12-05 Thread Sarlo, Jeffrey S
We have weights and priority/multifactor. Jeff From: Sistemas NLHPC [mailto:siste...@nlhpc.cl] Sent: Thursday, December 05, 2019 12:01 PM To: Sarlo, Jeffrey S; Slurm User Community List Subject: Re: [slurm-users] Slurm configuration, Weight Parameter Thanks Jeff ! We upgrade slurm to 18.08.4

Re: [slurm-users] Slurm configuration, Weight Parameter

2019-12-05 Thread Sistemas NLHPC
*On > Behalf Of *Sistemas NLHPC > *Sent:* Tuesday, December 03, 2019 12:33 PM > *To:* Slurm User Community List > *Subject:* Re: [slurm-users] Slurm configuration, Weight Parameter > > > > Hi Renfro > > > > I am testing this configuration, test configuration an

Re: [slurm-users] Slurm configuration, Weight Parameter

2019-12-03 Thread Sistemas NLHPC
Hi Renfro I am testing this configuration, test configuration and as clean as possible: NodeName=devcn050 RealMemory=3007 Features=3007MB Weight=200 State=idle Sockets=2 CoresPerSocket=1 NodeName=devcn002 RealMemory=3007 Features=3007MB Weight=1 State=idle Sockets=2 CoresPerSocket=1 NodeNam

Re: [slurm-users] Slurm configuration, Weight Parameter

2019-11-30 Thread Renfro, Michael
We’ve been using that weighting scheme for a year or so, and it works as expected. Not sure how Slurm would react to multiple NodeName=DEFAULT lines like you have, but here’s our node settings and a subset of our partition settings. In our environment, we’d often have lots of idle cores on GPU

Re: [slurm-users] Slurm configuration, Weight Parameter

2019-11-29 Thread Sistemas NLHPC
Hi All, Thanks all for your posts Reading the documentation of Slurm and other sites like Niflheim https://wiki.fysik.dtu.dk/niflheim/Slurm_configuration#node-weight (Ole Holm Nielsen) the parameter "Weight" is to assign a value to the nodes, with this you can have priority in the nodes. But I ha

Re: [slurm-users] Slurm configuration, Weight Parameter

2019-11-23 Thread Chris Samuel
On 23/11/19 9:14 am, Chris Samuel wrote: My gut instinct (and I've never tried this) is to make the 3GB nodes be in a separate partition that is guarded by AllowQos=3GB and have a QOS called "3GB" that uses MinTRESPerJob to require jobs to ask for more than 2GB of RAM to be allowed into the QO

Re: [slurm-users] Slurm configuration, Weight Parameter

2019-11-23 Thread Chris Samuel
On 21/11/19 7:25 am, Sistemas NLHPC wrote: Currently we have two types of nodes, one with 3GB and another with 2GB of RAM, it is required that in nodes of 3 GB it is not allowed to execute tasks with less than 2GB, to avoid underutilization of resources. My gut instinct (and I've never tried

Re: [slurm-users] Slurm configuration, Weight Parameter

2019-11-22 Thread Goetz, Patrick G
Can't you just set the usage priority to be higher for the 2GB machines? This way, if the requested memory is less than 2GB those machines will be used first, and larger jobs skip to the higher memory machines. On 11/21/19 9:44 AM, Jim Prewett wrote: > > Hi Sistemas, > > I could be mista

Re: [slurm-users] Slurm configuration, Weight Parameter

2019-11-21 Thread Jim Prewett
Hi Sistemas, I could be mistaken, but I don't think there is a way to require jobs on the 3GB nodes to request more than 2GB! https://slurm.schedmd.com/slurm.conf.html states this: "Note that if a job allocation request can not be satisfied using the nodes with the lowest weight, the set o

[slurm-users] Slurm configuration, Weight Parameter

2019-11-21 Thread Sistemas NLHPC
Hi all, Currently we have two types of nodes, one with 3GB and another with 2GB of RAM, it is required that in nodes of 3 GB it is not allowed to execute tasks with less than 2GB, to avoid underutilization of resources. This, because we have nodes that can fulfill the condition of executing tasks

Re: [slurm-users] Slurm configuration

2019-10-29 Thread Sistemas NLHPC
Hi Daniel I have tried this configuration but it has not given me results. Is there any other option to be able to do this, or should something else be configured to use the weight parameter? Thanks in advance. Regards, El lun., 5 ago. 2019 a las 5:35, Daniel Letai () escribió: > Hi. > > > On

Re: [slurm-users] Slurm configuration

2019-08-05 Thread Daniel Letai
Hi. On 8/3/19 12:37 AM, Sistemas NLHPC wrote: Hi all, Currently we have two types of nodes, one with 192GB and another with 768GB of RAM, it is required that in nodes of 768 GB it is not allowed to execute tasks

Re: [slurm-users] Slurm configuration

2019-08-04 Thread Loris Bennett
Hi NLHPC employee, Sistemas NLHPC writes: > Hi all, > > Currently we have two types of nodes, one with 192GB and another with > 768GB of RAM, it is required that in nodes of 768 GB it is not allowed > to execute tasks with less than 192GB, to avoid underutilization of > resources. > > This, beca

Re: [slurm-users] Slurm configuration

2019-08-03 Thread Henkel, Andreas
Hi, Have you checked documentation of MinMemory in Slurm.conf for node definition? Best, Andreas > Am 02.08.2019 um 23:53 schrieb Sistemas NLHPC : > > Hi all, > > Currently we have two types of nodes, one with 192GB and another with 768GB > of RAM, it is required that in nodes of 768 GB it is

[slurm-users] Slurm configuration

2019-08-02 Thread Sistemas NLHPC
Hi all, Currently we have two types of nodes, one with 192GB and another with 768GB of RAM, it is required that in nodes of 768 GB it is not allowed to execute tasks with less than 192GB, to avoid underutilization of resources. This, because we have nodes that can fulfill the condition of executi

Re: [slurm-users] slurm configuration for reprise license mangagement - cloud version

2019-06-25 Thread pim schravendijk
Thanks, Ahmet, in the settings it mentions servername and servertype so it seemed at first glance that some relatively elaborate communication with a license server was involved. I'll try if with some scripting around the options you mentioned I can find a suitable solution. Greetings, Pim On Tue

Re: [slurm-users] slurm configuration for reprise license mangagement - cloud version

2019-06-25 Thread mercan
Hi; As far as I know, the slurm is not able to work (communicate) with reprise license manager or any other license manager. Slurm just sums the used licenses according to the -L parameter of the jobs, and subtracts this sum from the total license count which given by using "sacctmgr add/modi

[slurm-users] slurm configuration for reprise license mangagement - cloud version

2019-06-25 Thread pim schravendijk
Hi all, We use a piece of software that works with reprise license management via the cloud http://www.reprisesoftware.com./products/cloud-licensing.php According to https://slurm.schedmd.com/licenses.html slurm is able to work with reprise license manager, though the examples at schedmd seem to

Re: [slurm-users] Slurm configuration on multi computers with ldap and dedicated resources

2019-02-11 Thread Renfro, Michael
I’m assuming you have LDAP and Slurm already working on all your nodes, and want to restrict access to two of the nodes based off of Unix group membership, while letting all users access the rest of the nodes. If that’s the case, you should be able to put the two towers into a separate partitio

Re: [slurm-users] Slurm configuration on multi computers with ldap and dedicated resources

2019-02-11 Thread Jean-Sébastien Lerat
Dear, Can someone help me to configure several machines under Slurm with different rights per user group? Or can someone redirect me to a tuoriel that explains these different points? Regards, Jean-Sébastien Le lun. 28 janv. 2019 à 17:52, Jean-Sébastien Lerat a écrit : > Hi, > > I have two to

[slurm-users] Slurm configuration on multi computers with ldap and dedicated resources

2019-01-28 Thread Jean-Sébastien Lerat
Hi, I have two towers (computers), few Odroid XU4 (rapsberry like) and an additional computer which will be the controller. I am planning to use spack (https://spack.io/) on a NFS share from the controller such that all other computers (odroid included) can use theses modules. I wish that users c