But won't that first process be able to use 100% of a core? What if enough users do this such that every core is at 100% utilization? Or, what if the application is MPI + OpenMP? In that case, that one process on the login node could spawn multiple threads that use the remaining cores on the login node.

Prentice

On 4/26/21 2:01 AM, Marcus Wagner wrote:
Hi,

we also have a wrapper script, together with a number of "MPI-Backends".
If mpiexec is called on the login nodes, only the first process is started on the login node, the rest runs on the MPI backends.

Best
Marcus

Am 25.04.2021 um 09:46 schrieb Patrick Begou:
Hi,

I also saw a cluster setup where mpirun or mpiexec commands were
replaced by a shell script just saying "please use srun or sbatch...".

Patrick

Le 24/04/2021 à 10:03, Ole Holm Nielsen a écrit :
On 24-04-2021 04:37, Cristóbal Navarro wrote:
Hi Community,
I have a set of users still not so familiar with slurm, and yesterday
they bypassed srun/sbatch and just ran their CPU program directly on
the head/login node thinking it would still run on the compute node.
I am aware that I will need to teach them some basic usage, but in
the meanwhile, how have you solved this type of user-behavior
problem? Is there a preffered way to restrict the master/login
resources, or actions,  to the regular users ?

We restrict user limits in /etc/security/limits.conf so users can't
run very long or very big tasks on the login nodes:

# Normal user limits
*               hard    cpu             20
*               hard    rss             50000000
*               hard    data            50000000
*               soft    stack           40000000
*               hard    stack           50000000
*               hard    nproc           250

/Ole





Reply via email to