Hi Manuel Yes, have seen the same version-dependent behaviour, although perhaps have not diagnosed it is thoroughly as you. We principally use Intel MPI and thought it mainly affected that. I did report a bug (3032) but unfortunately (to my chagrin) we don’t have a support contract at the moment.
Sorry I can’t offer anything other than that. ~~ Ade From: Manuel Rodríguez Pascual [mailto:manuel.rodriguez.pasc...@gmail.com] Sent: 21 October 2016 14:54 To: slurm-dev <slurm-dev@schedmd.com> Subject: [slurm-dev] Wrong behaviour of "--tasks-per-node" flag Hi all, I am having the weirdest error ever. I am pretty sure this is a bug. I have reproduced the error in latest slurm commit (slurm 17.02.0-0pre2, commit 406d3fe429ef6b694f30e19f69acf989e65d7509 ) and in slurm 16.05.5 branch. It does NOT happen in slurm 15.08.12 . My cluster is composed by 8 nodes, each with 2 sockets, each with 8 cores. Slurm.conf content is SchedulerType=sched/backfill SchedulerPort=7321 SelectType=select/linear #DEDICATED NODES NodeName=acme[11-14,21-24] CPUs=16 Sockets=2 CoresPerSocket=8 ThreadsPerCore=1 State=UNKNOWN I am running a simple hello World parallel code. It is submitted as "sbatch --ntasks=X --tasks-per-node=Y myScript.sh ". The problem is that, depending on the values of X and Y, Slurm performs a wrong opperation and returns an error. " sbatch --ntasks=8 --tasks-per-node=2 myScript.sh srun: Warning: can't honor --ntasks-per-node set to 2 which doesn't match the requested tasks 4 with the number of requested nodes 4. Ignoring --ntasks-per-node. " Note that I did not request 4 but 8 tasks, and I did not request any number of nodes. Same happens with " sbatch --ntasks=16 --tasks-per-node=2 myScript.sh srun: Warning: can't honor --ntasks-per-node set to 2 which doesn't match the requested tasks 8 with the number of requested nodes 8. Ignoring --ntasks-per-node. " and " sbatch --ntasks=32 --tasks-per-node=4 myScript.sh srun: Warning: can't honor --ntasks-per-node set to 4 which doesn't match the requested tasks 8 with the number of requested nodes 8. Ignoring --ntasks-per-node. " All the rest of configurations work correctly and do not return any error. In particular, I have tried the following combinations with no problem: (ntasks, tasks-per-node) (1,1) (2,1), (2,2) (4,1), (4,2), (4,4) (8,1), (4,4), (8,8) (16,4), (16,8), (16,16) (32,8), (32,16) (64,8), (64, 16) (128, 16) As said, this does not happen when executing the very same commands and scripts with slurm 15.08.12. So, have you had any similar experiences? Is this a bug, a desired behaviour, or am I doing something wrong? Thanks for your help. Best regards, Manuel ________________________________ [HPC Wales - www.hpcwales.co.uk] <http://www.hpcwales.co.uk> ________________________________ The contents of this email and any files transmitted with it are confidential and intended solely for the named addressee only. Unless you are the named addressee (or authorised to receive this on their behalf) you may not copy it or use it, or disclose it to anyone else. If you have received this email in error, please notify the sender by email or telephone. All emails sent by High Performance Computing Wales have been checked using an Anti-Virus system. We would advise you to run your own virus check before opening any attachments received as we will not in any event accept any liability whatsoever, once an email and/or attachment is received. High Performance Computing Wales is a private limited company incorporated in Wales on 8 March 2010 as company number 07181701. Our registered office is at Finance Office, Bangor University, Cae Derwen, College Road, Bangor, Gwynedd. LL57 2DG. UK. High Performance Computing Wales is part funded by the European Regional Development Fund through the Welsh Government.