Re: [slurm-users] slurm & rstudio

2020-07-20 Thread Benjamin Redling
Hi Kush, have you tried searching for (parts of) that error message? Is R Studio Pro aware of the following changes? https://lists.schedmd.com/pipermail/slurm-users/2018-May/001296.html Regards, Benjamin Am 20. Juli 2020 16:50:08 MESZ schrieb "Sidhu, Khushwant" : >Hi, > >I'm trying to use rstu

Re: [slurm-users] SLURM on a large shared memory node

2020-12-03 Thread Benjamin Redling
Hello Benson, On 24/11/2020 14.20, Benson Muite wrote: Am setting up SLURM on a single shared memory machine. Found the following blog post: http://rolk.github.io/2015/04/20/slurm-cluster sorry, but that is only a random, outdated blog post from 2015. Even the Debian 9 stretch provided Slurm

Re: [slurm-users] Assigning two "cores" when I'm only request one.

2021-07-13 Thread Benjamin Redling
On 12/07/2021 21.16, Luis R. Torres wrote: I'm trying to run one task on one "core", however, when I test the affinity, the system gives me "two"; I'm assuming the two are threads since the system is a dual socket system. [...] AFAI understand, you reason that the no. of hw threads depends on

Re: [slurm-users] Having errors trying to run a packed jobs script

2017-11-07 Thread Benjamin Redling
Hello Marius, Am 07.11.2017 um 10:12 schrieb Marius Cetateanu: > I have a very small cluster(if it even could be called a cluster) with only > one node for the moment; the node is a dual Xeon with 14 cores/socket, > hyper-threaded and 256GB of memory, running CentOS 7.3. Bigger than a small clust

Re: [slurm-users] [slurm-dev] Re: Installing SLURM locally on Ubuntu 16.04

2017-11-08 Thread Benjamin Redling
On 11/8/17 3:01 PM, Douglas Jacobsen wrote: Also please make sure you have the slurm-munge package installed (at least for the RPMs this is the name of the package, I'm unsure if that packaging layout was conserved for Debian) nope, it's just "munge" Regards, Benjamin -- FSU Jena | JULIELab.de

Re: [slurm-users] [slurm-dev] Re: Installing SLURM locally on Ubuntu 16.04

2017-11-13 Thread Benjamin Redling
On 11/12/17 4:52 PM, Gennaro Oliva wrote: On Sun, Nov 12, 2017 at 10:03:18AM -0500, Will L wrote: I just tried `sudo apt-get remove --purge munge`, etc., and munge itself this should have uninstalled slurm-wlm also, did you reinstalled it with apt? seems to be working fine. But I still g

Re: [slurm-users] Priority wait

2017-11-13 Thread Benjamin Redling
Hi Roy, On 11/13/17 2:37 PM, Roe Zohar wrote: [...] I sent 3000 jobs with feature Optimus and part are running while part are pendind. Which is ok. But I have sent 1000 jobs to Megatron and they are all in pending stating they wait because of priority. Whay os that? B.t.w if I change their pr

Re: [slurm-users] slurm conf with single machine with multi cores.

2017-11-29 Thread Benjamin Redling
On 11/29/17 4:32 PM, david vilanova wrote: Hi, I have updated the slurm.conf as follows: SelectType=select/cons_res SelectTypeParameters=CR_CPU NodeName=linuxcluster CPUs=2 PartitionName=testq Nodes=linuxcluster Default=YES MaxTime=INFINITE State=UP Still get testq node in down status ??? Any

Re: [slurm-users] slurm conf with single machine with multi cores.

2017-12-02 Thread Benjamin Redling
Am 30.11.2017 um 09:31 schrieb david vilanova: > Here below my slurm.conf file: > NodeName=linuxcluster CPUs=12 > PartitionName=testq Nodes=linuxclusterDefault=YES MaxTime=INFINITE State=UP missing white space! Nodes=linuxclusterDefault=Yes is not a node node, > [2017-11-30T09:24:28.430] layouts

Re: [slurm-users] Remote submission hosts and security

2017-12-05 Thread Benjamin Redling
Am 05.12.2017 um 22:27 schrieb Jeff White: > I have a need to allow a server which is outside of my cluster access to > submit jobs to the cluster.  I can do that easily enough by handing my > Slurm RPMs, config, and munge key to the owner of that server and > opening access in my firewall.  Howeve

Re: [slurm-users] Multithreads config

2018-02-16 Thread Benjamin Redling
Am 16.02.2018 um 15:28 schrieb david martin: > *I have a single physical server with :* > * *63 cpus (each cpu has 16 cores) * > * *480Gb total memory* > > *NodeNAME= Sockets=1 CoresPerSocket=16 ThreadsPerCore=1 Procs=63 > REALMEMORY=48*** > *This configuration will not work. What is s

Re: [slurm-users] slurm jobs are pending but resources are available

2018-04-17 Thread Benjamin Redling
Hello, Am 16.04.2018 um 18:50 schrieb Michael Di Domenico: > On Mon, Apr 16, 2018 at 6:35 AM, wrote: > perhaps i missed something in the email, but it sounds like you have > 56 cores, you have two running jobs that consume 52 cores, leaving you > four free. No. From the original mail: <--- %

Re: [slurm-users] Include some cores of the head node to a partition

2018-04-22 Thread Benjamin Redling
Hello Mahmood, Am 22.04.2018 um 04:55 schrieb Mahmood Naderan: > I think that will limit other nodes to 20 too. Isn't that? you can declare less CPUs than phys. available. I do that for our cluster; it is working robust for ages. > Currently computes have 32 cores per node and I want all 32 core

Re: [slurm-users] Controller / backup controller q's

2018-05-25 Thread Benjamin Redling
Am 24.05.2018 um 17:43 schrieb Will Dennis: > 3)  What are the steps to replace a primary controller, given that a > backup controller exists? (Hopefully this is already documented > somewhere that I haven’t found yet) Why not drive such a small cluster with a single primary controller in a mig

Re: [slurm-users] Help buidling slurm on OS X High Sierra

2018-06-24 Thread Benjamin Redling
On 24/06/18 01:27, George Hartzell wrote: > I'm trying to build Slurm on a Mac running OS X Sierra (via Spack) [...] > There are enough mentions of darwin in the src tree that it seems like > it should work (or has worked). > > Am I chasing something hopeless? Maybe: https://slurm.schedmd.com/p

Re: [slurm-users] Help buidling slurm on OS X High Sierra

2018-06-27 Thread Benjamin Redling
On 24/06/18 22:04, Pär Lindfors wrote: > On 06/24/2018 01:55 PM, Benjamin Redling wrote: >> https://slurm.schedmd.com/platforms.html > Does not seem to have been updated in a while. Solaris support was > removed recently, probably in 17.11. True. Apart form the things gettin

Re: [slurm-users] Resource sharing between different clusters

2018-10-19 Thread Benjamin Redling
On 18/10/2018 18:16, Cao, Lei wrote: > I am pretty new to slurm so please bear with me. I have the following > scenario and I wonder if slurm currently supports this in someway. > > Let's say I have 3 clusters. Cluster1 and cluster2 run their own > slurmctld and slurmds(this is a hard re

Re: [slurm-users] How to partition nodes into smaller units

2019-02-09 Thread Benjamin Redling
Hello, On 05.02.19 16:46, Ansgar Esztermann-Kirchner wrote: > [...]-- we'd like to have two "half nodes", where > jobs will be able to use one of the two GPUs, plus (at most) half of > the CPUs. With SGE, we've put two queues on the nodes, but this > effectively prevents certain maintenance jobs f

Re: [slurm-users] How should I do so that jobs are allocated to the thread and not to the core ?

2019-05-02 Thread Benjamin Redling
Have you Seen the slurm FAQ? You may want to search on that site for "Hyperthreading" (Sorry for the TOFU. vacation, mobile) Am 30. April 2019 18:07:03 MESZ schrieb Jean-mathieu CHANTREIN : >Hello, > >Most jobs of my users are single-thread. I have multithreaded >processors. The jobs seem to re

Re: [slurm-users] [Long] Why are tasks started on a 30 second clock?

2019-07-25 Thread Benjamin Redling
On 25.07.19 20:11, Kirill Katsnelson wrote: On Thu, Jul 25, 2019 at 8:16 AM Mark Hahn > wrote: how about a timeout from elsewhere?  for instance, when I see a 30s delay, I normally at least check DNS, which can introduce such quantized delays. Thanks

Re: [slurm-users] OverSubscribe=FORCE:1overloads nodes?

2019-09-08 Thread Benjamin Redling
Hello Menglong, which selection plugin and in case of of cons_res what consumable resources have you configured? Maybe review: https://slurm.schedmd.com/cons_res_share.html Regards, Benjamin Am 9. September 2019 03:38:13 MESZ schrieb "hu...@sugon.com" : >Dear there, >I have two jobs in my clust

[slurm-users] Archived docs show 19.05 news

2019-10-24 Thread Benjamin Redling
Hello everybody, confusing: https://slurm.schedmd.com/archive/slurm-18.08.8/news.html " RELEASE NOTES FOR SLURM VERSION 19.05 28 May 2019 ... " Bug-tracking is only via commercial support? Regards, Benjamin

Re: [slurm-users] RHEL8 support

2019-10-28 Thread Benjamin Redling
On 28/10/2019 08.26, Bjørn-Helge Mevik wrote: > Taras Shapovalov writes: > >> Do I understand correctly that Slurm19 is not compatible with rhel8? It is >> not in the list https://slurm.schedmd.com/platforms.html > > It says > > "RedHat Enterprise Linux 7 (RHEL7), CentOS 7, Scientific Linux 7 (

[slurm-users] ProEpiLogInterfacePlugin -> PerilogueInterfacePlugin (E.A. Schneider @ CMU'76?)

2020-02-21 Thread Benjamin Redling
Hello everybody, only yesterday I had time to review: https://slurm.schedmd.com/SLUG19/Slurm_20.02_and_Beyond.pdf " If you have a good name for this plugin type, I haven't found a good name - "ProEpiLogInterfacePlugin" is a bit unwieldy " So, I searched for a hypernym of "prologue" and "epilog" a