What is your SLURM_TASKS_PER_NODE?

On Feb 12, 2014, at 6:58 AM, Adrian Reber <adr...@lisas.de> wrote:

> No, the system has only a few MOAB_* variables and many SLURM_*
> variables:
> 
> $BASH                         $IFS                          $SECONDS          
>             $SLURM_PTY_PORT
> $BASHOPTS                     $LINENO                       $SHELL            
>             $SLURM_PTY_WIN_COL
> $BASHPID                      $LINES                        $SHELLOPTS        
>             $SLURM_PTY_WIN_ROW
> $BASH_ALIASES                 $MACHTYPE                     $SHLVL            
>             $SLURM_SRUN_COMM_HOST
> $BASH_ARGC                    $MAILCHECK                    $SLURMD_NODENAME  
>             $SLURM_SRUN_COMM_PORT
> $BASH_ARGV                    $MOAB_CLASS                   
> $SLURM_CHECKPOINT_IMAGE_DIR   $SLURM_STEPID
> $BASH_CMDS                    $MOAB_GROUP                   $SLURM_CONF       
>             $SLURM_STEP_ID
> $BASH_COMMAND                 $MOAB_JOBID                   
> $SLURM_CPUS_ON_NODE           $SLURM_STEP_LAUNCHER_PORT
> $BASH_LINENO                  $MOAB_NODECOUNT               
> $SLURM_DISTRIBUTION           $SLURM_STEP_NODELIST
> $BASH_SOURCE                  $MOAB_PARTITION               $SLURM_GTIDS      
>             $SLURM_STEP_NUM_NODES
> $BASH_SUBSHELL                $MOAB_PROCCOUNT               $SLURM_JOBID      
>             $SLURM_STEP_NUM_TASKS
> $BASH_VERSINFO                $MOAB_SUBMITDIR               
> $SLURM_JOB_CPUS_PER_NODE      $SLURM_STEP_TASKS_PER_NODE
> $BASH_VERSION                 $MOAB_USER                    $SLURM_JOB_ID     
>             $SLURM_SUBMIT_DIR
> $COLUMNS                      $OPTERR                       
> $SLURM_JOB_NODELIST           $SLURM_SUBMIT_HOST
> $COMP_WORDBREAKS              $OPTIND                       
> $SLURM_JOB_NUM_NODES          $SLURM_TASKS_PER_NODE
> $DIRSTACK                     $OSTYPE                       
> $SLURM_LAUNCH_NODE_IPADDR     $SLURM_TASK_PID
> $EUID                         $PATH                         $SLURM_LOCALID    
>             $SLURM_TOPOLOGY_ADDR
> $GROUPS                       $POSIXLY_CORRECT              $SLURM_NNODES     
>             $SLURM_TOPOLOGY_ADDR_PATTERN
> $HISTCMD                      $PPID                         $SLURM_NODEID     
>             $SRUN_DEBUG
> $HISTFILE                     $PS1                          $SLURM_NODELIST   
>             $TERM
> $HISTFILESIZE                 $PS2                          $SLURM_NPROCS     
>             $TMPDIR
> $HISTSIZE                     $PS4                          $SLURM_NTASKS     
>             $UID
> $HOSTNAME                     $PWD                          
> $SLURM_PRIO_PROCESS           $_
> $HOSTTYPE                     $RANDOM                       $SLURM_PROCID     
>             
> 
> 
> 
> On Wed, Feb 12, 2014 at 06:12:45AM -0800, Ralph Castain wrote:
>> Seems rather odd - since this is managed by Moab, you shouldn't be seeing 
>> SLURM envars at all. What you should see are PBS_* envars, including a 
>> PBS_NODEFILE that actually contains the allocation.
>> 
>> 
>> On Feb 12, 2014, at 4:42 AM, Adrian Reber <adr...@lisas.de> wrote:
>> 
>>> I tried the nightly snapshot (openmpi-1.7.5a1r30692.tar.gz) on a system
>>> with slurm and moab. I requested an interactive session using:
>>> 
>>> msub -I -l nodes=3:ppn=8
>>> 
>>> and started a simple test case which fails:
>>> 
>>> $ mpirun -np 2 ./mpi-test 1
>>> --------------------------------------------------------------------------
>>> There are not enough slots available in the system to satisfy the 2 slots 
>>> that were requested by the application:
>>> ./mpi-test
>>> 
>>> Either request fewer slots for your application, or make more slots 
>>> available
>>> for use.
>>> --------------------------------------------------------------------------
>>> srun: error: xxxx108: task 1: Exited with exit code 1
>>> srun: Terminating job step 131823.4
>>> srun: error: xxxx107: task 0: Exited with exit code 1
>>> srun: Job step aborted
>>> slurmd[xxxx108]: *** STEP 131823.4 KILLED AT 2014-02-12T13:30:32 WITH 
>>> SIGNAL 9 ***
>>> 
>>> 
>>> requesting only one core works:
>>> 
>>> $ mpirun  ./mpi-test 1
>>> 4.4.7 20120313 (Red Hat 4.4.7-4):Process 0 on xxxx106 out of 1: 0.000000
>>> 4.4.7 20120313 (Red Hat 4.4.7-4):Process 0 on xxxx106 out of 1: 0.000000
>>> 
>>> 
>>> using openmpi-1.6.5 works with multiple cores:
>>> 
>>> $ mpirun -np 24 ./mpi-test 2
>>> 4.4.7 20120313 (Red Hat 4.4.7-4):Process 0 on xxxx106 out of 24: 0.000000
>>> 4.4.7 20120313 (Red Hat 4.4.7-4):Process 12 on xxxx106 out of 24: 12.000000
>>> 4.4.7 20120313 (Red Hat 4.4.7-4):Process 11 on xxxx108 out of 24: 11.000000
>>> 4.4.7 20120313 (Red Hat 4.4.7-4):Process 18 on xxxx106 out of 24: 18.000000
>>> 
>>> $ echo $SLURM_JOB_CPUS_PER_NODE 
>>> 8(x3)
>>> 
>>> I never used slurm before so this could also be a user error on my side.
>>> But as 1.6.5 works it seems something has changed and wanted to let
>>> you know in case it was not intentionally.
>>> 
>>>             Adrian
>>> _______________________________________________
>>> devel mailing list
>>> de...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> 
>> _______________________________________________
>> devel mailing list
>> de...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
> 
>               Adrian
> 
> -- 
> Adrian Reber <adr...@lisas.de>            http://lisas.de/~adrian/
> "Let us all bask in television's warm glowing warming glow." -- Homer Simpson
> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel

Reply via email to