Look in the log file, by default it's written to the current directory if you didn't use -o in sbatch.
On 5 Jan 2014 11:09, ian <ian.h.0...@gmail.com> wrote: Hello, I have problem with sbatch. I run the following command. Why can't I see results of control node and compute nodes at standard output? Thank you. miao@SLURM0:~$ cat my.script #!/bin/sh #SBATCH --time=30 /bin/hostname srun -l /bin/hostname srun -l /bin/pwd miao@SLURM0:~$ sbatch –n2 my.script Submitted batch job 74 miao@SLURM0:~$ miao@SLURM0:~$ scontrol show config Configuration data as of 2014-01-04T18:47:22 AccountingStorageBackupHost = (null) AccountingStorageEnforce = none AccountingStorageHost = localhost AccountingStorageLoc = /var/log/slurm_jobacct.log AccountingStoragePort = 0 AccountingStorageType = accounting_storage/none AccountingStorageUser = root AccountingStoreJobComment = YES AuthType = auth/munge BackupAddr = (null) BackupController = (null) BatchStartTimeout = 10 sec BOOT_TIME = 2014-01-04T16:53:52 CacheGroups = 0 CheckpointType = checkpoint/none ClusterName = slurm CompleteWait = 0 sec ControlAddr = SLURM0 ControlMachine = SLURM0 CryptoType = crypto/munge DebugFlags = (null) DefMemPerNode = UNLIMITED DisableRootJobs = NO EnforcePartLimits = NO Epilog = (null) EpilogMsgTime = 2000 usec EpilogSlurmctld = (null) FastSchedule = 1 FirstJobId = 1 GetEnvTimeout = 2 sec GresTypes = (null) GroupUpdateForce = 0 GroupUpdateTime = 600 sec HASH_VAL = Match HealthCheckInterval = 0 sec HealthCheckProgram = (null) InactiveLimit = 0 sec JobAcctGatherFrequency = 30 sec JobAcctGatherType = jobacct_gather/none JobCheckpointDir = /var/slurm/checkpoint JobCompHost = localhost JobCompLoc = /var/log/slurm_jobcomp.log JobCompPort = 0 JobCompType = jobcomp/none JobCompUser = root JobCredentialPrivateKey = (null) JobCredentialPublicCertificate = (null) JobFileAppend = 0 JobRequeue = 1 JobSubmitPlugins = (null) KillOnBadExit = 0 KillWait = 30 sec Licenses = (null) MailProg = /usr/bin/mail MaxJobCount = 10000 MaxJobId = 4294901760 MaxMemPerNode = UNLIMITED MaxStepCount = 40000 MaxTasksPerNode = 128 MessageTimeout = 10 sec MinJobAge = 300 sec MpiDefault = none MpiParams = (null) NEXT_JOB_ID = 79 OverTimeLimit = 0 min PluginDir = /usr/lib/slurm PlugStackConfig = /etc/slurm-llnl/plugstack.conf PreemptMode = OFF PreemptType = preempt/none PriorityType = priority/basic PrivateData = none ProctrackType = proctrack/pgid Prolog = (null) PrologSlurmctld = (null) PropagatePrioProcess = 0 PropagateResourceLimits = ALL PropagateResourceLimitsExcept = (null) ResumeProgram = (null) ResumeRate = 300 nodes/min ResumeTimeout = 60 sec ResvOverRun = 0 min ReturnToService = 0 SallocDefaultCommand = (null) SchedulerParameters = (null) SchedulerPort = 7321 SchedulerRootFilter = 1 SchedulerTimeSlice = 30 sec SchedulerType = sched/backfill SelectType = select/linear SlurmUser = root(0) SlurmctldDebug = 6 SlurmctldLogFile = /var/log/slurm-llnl/slurmctld.log SlurmSchedLogFile = (null) SlurmctldPort = 6817 SlurmctldTimeout = 300 sec SlurmdDebug = 5 SlurmdLogFile = /var/log/slurm-llnl/slurmd.log SlurmdPidFile = /var/run/slurmd.pid SlurmdPort = 6818 SlurmdSpoolDir = /tmp/slurmd SlurmdTimeout = 300 sec SlurmdUser = root(0) SlurmSchedLogLevel = 0 SlurmctldPidFile = /var/run/slurmctld.pid SLURM_CONF = /etc/slurm-llnl/slurm.conf SLURM_VERSION = 2.3.2 SrunEpilog = (null) SrunProlog = (null) StateSaveLocation = /tmp SuspendExcNodes = (null) SuspendExcParts = (null) SuspendProgram = (null) SuspendRate = 60 nodes/min SuspendTime = NONE SuspendTimeout = 30 sec SwitchType = switch/none TaskEpilog = (null) TaskPlugin = task/none TaskPluginParam = (null type) TaskProlog = (null) TmpFS = /tmp TopologyPlugin = topology/none TrackWCKey = 0 TreeWidth = 50 UsePam = 0 UnkillableStepProgram = (null) UnkillableStepTimeout = 60 sec VSizeFactor = 0 percent WaitTime = 0 sec Slurmctld(primary/backup) at SLURM0/(NULL) are UP/DOWN miao@SLURM0:~$ sinfo PARTITION AVAIL TIMELIMIT NODES STATE NODELIST debug* up infinite 2 idle SLURM[1-2] miao@SLURM0:~$ cat /var/log/slurm-llnl/slurmctld.log [2014-01-04T16:59:15] debug2: Processing RPC: REQUEST_SUBMIT_BATCH_JOB from uid=1000 [2014-01-04T16:59:15] debug2: found 2 usable nodes from config containing SLURM[1-2] [2014-01-04T16:59:15] debug2: sched: JobId=74 allocated resources: NodeList=(null) [2014-01-04T16:59:15] _slurm_rpc_submit_batch_job JobId=74 usec=456 [2014-01-04T16:59:15] debug: sched: Running job scheduler [2014-01-04T16:59:15] debug2: found 2 usable nodes from config containing SLURM[1-2] [2014-01-04T16:59:15] sched: Allocate JobId=74 NodeList=SLURM1 #CPUs=1 [2014-01-04T16:59:15] debug2: Spawning RPC agent for msg_type 4005 [2014-01-04T16:59:15] debug2: got 1 threads to send out [2014-01-04T16:59:15] debug2: Tree head got back 0 looking for 1 [2014-01-04T16:59:15] debug2: Tree head got back 1 [2014-01-04T16:59:15] debug2: Tree head got them all [2014-01-04T16:59:15] debug2: Processing RPC: REQUEST_JOB_ALLOCATION_INFO_LITE from uid=1000 [2014-01-04T16:59:15] debug: _slurm_rpc_job_alloc_info_lite JobId=74 NodeList=SLURM1 usec=73 [2014-01-04T16:59:15] debug2: Processing RPC: REQUEST_JOB_STEP_CREATE from uid=1000 [2014-01-04T16:59:15] debug: Configuration for job 74 complete [2014-01-04T16:59:15] debug: laying out the 1 tasks on 1 hosts SLURM1 dist 1 [2014-01-04T16:59:15] sched: _slurm_rpc_job_step_create: StepId=74.0 SLURM1 usec=490 [2014-01-04T16:59:15] debug: Processing RPC: REQUEST_STEP_COMPLETE for 74.0 nodes 0-0 rc=0 uid=0 [2014-01-04T16:59:15] sched: _slurm_rpc_step_complete StepId=74.0 usec=106 [2014-01-04T16:59:15] debug2: Processing RPC: REQUEST_JOB_ALLOCATION_INFO_LITE from uid=1000 [2014-01-04T16:59:15] debug: _slurm_rpc_job_alloc_info_lite JobId=74 NodeList=SLURM1 usec=307 [2014-01-04T16:59:15] debug2: Processing RPC: REQUEST_JOB_STEP_CREATE from uid=1000 [2014-01-04T16:59:15] debug: laying out the 1 tasks on 1 hosts SLURM1 dist 1 [2014-01-04T16:59:15] sched: _slurm_rpc_job_step_create: StepId=74.1 SLURM1 usec=592 [2014-01-04T16:59:15] debug: Processing RPC: REQUEST_STEP_COMPLETE for 74.1 nodes 0-0 rc=0 uid=0 [2014-01-04T16:59:15] sched: _slurm_rpc_step_complete StepId=74.1 usec=104 [2014-01-04T16:59:15] debug2: node_did_resp SLURM1 [2014-01-04T16:59:15] debug2: Processing RPC: REQUEST_COMPLETE_BATCH_SCRIPT from uid=0 JobId=74 [2014-01-04T16:59:15] completing job 74 [2014-01-04T16:59:15] debug2: Spawning RPC agent for msg_type 6011 [2014-01-04T16:59:15] sched: job_complete for JobId=74 successful [2014-01-04T16:59:15] debug2: _slurm_rpc_complete_batch_script JobId=74 usec=224 [2014-01-04T16:59:15] debug2: got 1 threads to send out [2014-01-04T16:59:15] debug2: Tree head got back 0 looking for 1 [2014-01-04T16:59:15] debug2: Tree head got back 1 [2014-01-04T16:59:15] debug2: Tree head got them all [2014-01-04T16:59:16] debug2: node_did_resp SLURM1 miao@SLURM0:~$ srun -n2 -l hostname 1: SLURM2 0: SLURM1 miao@SLURM0:~$ Thank you very much! Best regards Ian Malcolm ________________________________ This email and any files transmitted with it are confidential and are intended solely for the use of the individual or entity to whom they are addressed. If you are not the original recipient or the person responsible for delivering the email to the intended recipient, be advised that you have received this email in error, and that any use, dissemination, forwarding, printing, or copying of this email is strictly prohibited. If you received this email in error, please immediately notify the sender and delete the original.