[slurm-dev] Re: slurm 2.6 partition node specification problem

2013-07-10 Thread John Thiltges
Hi Eva, I wasn't able to reproduce the problem with a quick test. You have config lines similar to these? NodeName=gpu-1-[4-17],gpu-2-[4,6-16],gpu-3-9 ... PartitionName=... Nodes=gpu-1-[4-17],gpu-2-[4,6-16],gpu-3-9 Regards, John On 2013-07-10 19:20, Eva Hocks wrote: > > > > > Thanks

[slurm-dev] Re: Slurm task assignment by actual physical core

2013-07-10 Thread Danny Auble
The llnl documentation is very out of date. You should be consulting http://slurm.schedmd.com. It might not be that different in this case but you should probably switch in any case. Teddy wrote: >Hi Daniel and Moe, > >Thanks for your reply. > > >Some of users of ours said the srun with --dis

[slurm-dev] Re: Slurm task assignment by actual physical core

2013-07-10 Thread Teddy
Hi Daniel and Moe, Thanks for your reply. Some of users of ours said the srun with --distribution somehow not working as expected. When i read the tutorial carefully ( https://computing.llnl.gov/linux/slurm/mc_support.html), i found this: "It is important to note that many of these flags are o

[slurm-dev] Re: slurm 2.6 partition node specification problem

2013-07-10 Thread Eva Hocks
Thanks, John but this is what I have in the partition file: nodes=gpu-1-[4-17],gpu-2-[4,6-16],gpu-3-9 slurm gets confused when it can't look up gpu-2-4 and then splits the gpu-2-[4,6-16] into gpu-[2]-[4] (failed lookup) and 6-16] (which is actually no node name at all but a wrong parsing af

[slurm-dev] Re: slurm 2.6 partition node specification problem

2013-07-10 Thread John Thiltges
On 07/10/2013 06:16 PM, Eva Hocks wrote: > The entry in partiton.conf: > PartitionName=CLUSTER Default=yes State=UP > nodes=gpu-[1]-[4-17],gpu-[2]-[4,6-16],gpu-[3]-[9] > > > causes slurmctl to crash: > > 2013-07-10T16:03:22.923] error: find_node_record: lookup failure for > gpu-[2]-[4] > [2013-0

[slurm-dev] slurm 2.6 partition node specification problem

2013-07-10 Thread Eva Hocks
The entry in partiton.conf: PartitionName=CLUSTER Default=yes State=UP nodes=gpu-[1]-[4-17],gpu-[2]-[4,6-16],gpu-[3]-[9] causes slurmctl to crash: 2013-07-10T16:03:22.923] error: find_node_record: lookup failure for gpu-[2]-[4] [2013-07-10T16:03:22.923] error: node_name2bitmap: invalid no

[slurm-dev] slurm-dev Re: slurm integration with FlexLM license manager

2013-07-10 Thread Bill . Brophy
Developers at Bull have been discussing a SLURM license management project with SchedMD.  The first phase of this project would provide initial support of licenses for accounting purposes within SLURM.   This would involve the introduction of new license tables in the accounting information.  Th

[slurm-dev] Slurm version 2.6 is now available

2013-07-10 Thread Moe Jette
We are pleased to announce the availability of Slurm version 2.6. Changes from version 2.5 are extensive and highlights are listed below. Please see the RELEASE_NOTES file in the Slurm distribution for more details. Note the Slurm documentation at schedmd.com has been updated to version 2.6. Down

[slurm-dev] Re: Grant user priority on X cores?

2013-07-10 Thread Lennart Karlsson
On 07/09/2013 08:38 PM, Neil Van Lysel wrote: > Is it possible to grant a user priority on X cores? For example, we have > a small 768 core SLURM cluster, and we would like to give user A > priority on only 512 cores. I am currently using QOS to give specific > users priority on all cores, but I d