[slurm-users] Re: GPU shards not exclusive

2024-02-29 Thread Reed Dier via slurm-users
Hi Will, I appreciate your corroboration. After we upgraded to 23.02.$latest, it seemed to make it easier to reproduce than before. However, the issue appears to have subsided, and the only change I can potentially attribute it to was after turning on > SlurmctldParameters=rl_enable in

[slurm-users] Re: [EXTERN] Re: sbatch and cgroup v2

2024-02-29 Thread Dietmar Rieder via slurm-users
Hi list, I finally got it working, I completely overlooked that I set Oversubscribe=EXCLUSIVE for the partition that I used to test, stupid me. sorry for the noise and thanks again for your answers Best Dietmar On 2/29/24 13:19, Dietmar Rieder via slurm-users wrote: Hi Josef, hi

[slurm-users] Re: [EXTERN] Re: sbatch and cgroup v2

2024-02-29 Thread Dietmar Rieder via slurm-users
Hi Josef, hi list, I now rebuild the rpms from OpenHPC but using the original sources form version 23.11.4. The configure command that is genereated from the spec is the following: ./configure --build=x86_64-redhat-linux-gnu \ --host=x86_64-redhat-linux-gnu \ --program-prefix= \

[slurm-users] Re: slurmdbd error - Symbol `slurm_conf' has different size in shared object

2024-02-29 Thread Miriam Olmi via slurm-users
Dear Josef, thanks a lot again for your help. Unfortunately I cannot solve this problem. According to the Slurm documentation (https://slurm.schedmd.com/quickstart_admin.html#upgrade) I have to upgrade only slurmdbd at the very beginning and the cluster should be able to work even with

[slurm-users] Re: Can't schedule on cloud node: State=IDLE+CLOUD+POWERED_DOWN+NOT_RESPONDING

2024-02-29 Thread Xaver Stiensmeier via slurm-users
I am wondering why my question (below) didn't catch anyone's attention. Just for me as a feedback. Is it unclear where my problem lies or is it clear, but no solution is known? I looked through the documentation and now searched the Slurm repository, but am still unable to clearly identify how to