CLASSIFICATION: UNCLASSIFIED
I've got a cluster with about 39 nodes, with 8 to 12 cores each. When I submit
a job array job of say 15k, about 300 of those jobs start up across the
cluster, but once those jobs complete, I only see one node's worth of jobs (say
8) going at a time from then on an
Tony
-Original Message-
From: Glover, Anthony E CTR USARMY RDECOM (US)
[mailto:anthony.e.glover@mail.mil]
Sent: Thursday, June 29, 2017 8:39 AM
To: slurm-dev
Subject: [Non-DoD Source] [slurm-dev] Job Array not using all nodes
(UNCLASSIFIED)
CLASSIFICATION: UNCLASSIFIED
I've go
s Min" and
"Nodes Max" are both set to 1 even though I had
#SBATCH -N 1-40
In my script. Maybe the Nodes Max is preventing my distribution? Anyway to get
around that?
Thanks,
Tony
-Original Message-
From: Glover, Anthony E CTR USARMY RDECOM (US)
[mailto:anthony.e.glover
CLASSIFICATION: UNCLASSIFIED
Got a general question, but one that might be specifically addressed by Slurm -
don't know.
We have a multi-process, distributed simulation that runs as a single job and
generates a significant amount of data. At the end of that run, we would like
to be able to po
but set a dependency on the post-processing job so that it can't start until
the first job has finished successfully. We've had users who manage fairly
complicated analysis pipelines entirely with job dependencies.
Regards,
Pete
On 7/19/17, 10:07 AM, "Glover, Anthony E CTR US