Hi Dmitry,

I can see the confusion, my last mail was poorly worded.  When I ran the plugin 
in the S_CTX_REMOTE context I was able to retrieve some information.  I wasn't 
able to retrieve any data when running in the S_CTX_ALLOCATOR context.

Again, apologies for the confusion.

---
Sam Gallop

From: Dmitry Chirikov [mailto:dmi...@chirikov.ru]
Sent: 30 January 2017 22:10
To: slurm-dev <slurm-dev@schedmd.com>
Subject: [slurm-dev] RE: SPANK and job variables/options

Hi Sam,

not sure I get you right how you achieve that. My code is here
#include <sys/types.h>
#include <stdio.h>
#include <stdlib.h>
#include <unistd.h>
#include <string.h>
#include <sys/resource.h>
#include <slurm/spank.h>

SPANK_PLUGIN(dc_test, 1);

int slurm_spank_init (spank_t sp, int ac, char **av) {
    int i, argc;
    char function[50];
    char **argv;
    char **env;
    if (spank_context () != S_CTX_ALLOCATOR) {
       return (0);
    }
    sprintf(function, "slurm_spank_init");
    if (spank_get_item(sp, S_JOB_ARGV, &argc, &argv) == ESPANK_SUCCESS) {
        for (i = 0; i < argc; i++) {
            slurm_info( "%s (ARGV): %s\n", function, argv[i]);
        }
    } else {
        slurm_error("Unable to retrieve S_JOB_ARGV");
    }
    return (0);
}

Then submit job as user:

$ sbatch test.sh
sbatch: error: Unable to retrieve S_JOB_ARGV
Submitted batch job 766

So, no cigar yet.

Kind regards,
Dmitry

Kind regards,
Dmitry Chirikov

On Mon, Jan 30, 2017 at 6:59 PM, Sam Gallop (NBI) 
<sam.gal...@nbi.ac.uk<mailto:sam.gal...@nbi.ac.uk>> wrote:
Hi,

By changing the entry point to slurm_spank_init I was able to retrieve the 
following (see below), but this was only when the context was S_CTX_ALLOCATOR.  
When the context is != S_CTX_ALLOCATOR I got nothing.

slurm_spank_init (ARGV): sleep
slurm_spank_init (ARGV): 30
slurm_spank_init (ENV): SLURM_CHECKPOINT_IMAGE_DIR=/root
slurm_spank_init (ENV): SLURM_NODELIST=centos1
slurm_spank_init (ENV): SLURM_JOB_NAME=slurm1.sh
slurm_spank_init (ENV): SLURMD_NODENAME=centos1
slurm_spank_init (ENV): SLURM_TOPOLOGY_ADDR=centos1
slurm_spank_init (ENV): SLURM_PRIO_PROCESS=0
slurm_spank_init (ENV): SLURM_TOPOLOGY_ADDR_PATTERN=node
slurm_spank_init (ENV): SLURM_NNODES=1
slurm_spank_init (ENV): SLURM_JOBID=268
slurm_spank_init (ENV): SLURM_TASKS_PER_NODE=3
slurm_spank_init (ENV): SLURM_JOB_ID=268
slurm_spank_init (ENV): SLURM_CPUS_PER_TASK=1
slurm_spank_init (ENV): SLURM_JOB_USER=root
slurm_spank_init (ENV): SLURM_JOB_UID=0
slurm_spank_init (ENV): SLURM_NODEID=0
…
slurm_spank_init (ENV): SLURM_SUBMIT_DIR=/root
slurm_spank_init (ENV): SLURM_TASK_PID=2470
slurm_spank_init (ENV): SLURM_CPUS_ON_NODE=3
slurm_spank_init (ENV): SLURM_PROCID=0
slurm_spank_init (ENV): SLURM_JOB_NODELIST=centos1
slurm_spank_init (ENV): SLURM_LOCALID=0
slurm_spank_init (ENV): SLURM_JOB_CPUS_PER_NODE=3
slurm_spank_init (ENV): SLURM_CLUSTER_NAME=cluster
slurm_spank_init (ENV): SLURM_GTIDS=0
slurm_spank_init (ENV): SLURM_SUBMIT_HOST=centos1
slurm_spank_init (ENV): SLURM_JOB_PARTITION=debug
slurm_spank_init (ENV): SLURM_JOB_NUM_NODES=1
slurm_spank_init (ENV): SLURM_MEM_PER_NODE=1024
slurm_spank_init (ENV): SLURM_RLIMIT_CPU=18446744073709551615
slurm_spank_init (ENV): SLURM_RLIMIT_FSIZE=18446744073709551615
slurm_spank_init (ENV): SLURM_RLIMIT_DATA=18446744073709551615
slurm_spank_init (ENV): SLURM_RLIMIT_STACK=10485760
slurm_spank_init (ENV): SLURM_RLIMIT_CORE=0
slurm_spank_init (ENV): SLURM_RLIMIT_RSS=1073741824
slurm_spank_init (ENV): SLURM_RLIMIT_NPROC=15188
slurm_spank_init (ENV): SLURM_RLIMIT_NOFILE=1024
slurm_spank_init (ENV): SLURM_RLIMIT_MEMLOCK=65536
slurm_spank_init (ENV): SLURM_RLIMIT_AS=18446744073709551615
slurm_spank_init (ENV): SRUN_DEBUG=3
slurm_spank_init (ENV): SLURM_UMASK=0022
slurm_spank_init (ENV): SLURM_NTASKS=3
slurm_spank_init (ENV): SLURM_NPROCS=3
slurm_spank_init (ENV): SLURM_DISTRIBUTION=block
slurm_spank_init (ENV): SLURM_CPU_BIND_VERBOSE=quiet
slurm_spank_init (ENV): SLURM_CPU_BIND_TYPE=cores
slurm_spank_init (ENV): SLURM_CPU_BIND_LIST=
slurm_spank_init (ENV): SLURM_CPU_BIND=quiet,cores
slurm_spank_init (ENV): SLURM_STEP_ID=0
slurm_spank_init (ENV): SLURM_STEPID=0
slurm_spank_init (ENV): SLURM_SRUN_COMM_PORT=42199
slurm_spank_init (ENV): SLURM_STEP_NODELIST=centos1
slurm_spank_init (ENV): SLURM_STEP_NUM_NODES=1
slurm_spank_init (ENV): SLURM_STEP_NUM_TASKS=3
slurm_spank_init (ENV): SLURM_STEP_TASKS_PER_NODE=3
slurm_spank_init (ENV): SLURM_STEP_LAUNCHER_PORT=42199
slurm_spank_init (ENV): SLURM_SRUN_COMM_HOST=…

---
Sam Gallop

From: Dmitry Chirikov [mailto:dmi...@chirikov.ru<mailto:dmi...@chirikov.ru>]
Sent: 30 January 2017 17:25
To: slurm-dev <slurm-dev@schedmd.com<mailto:slurm-dev@schedmd.com>>
Subject: [slurm-dev] RE: SPANK and job variables/options

Hi Sam,

thanks for sharing. If i get it right you are enumerating variables after 
allocation (slurm_spank_task_init) on the running node, but I am really curious 
if it even possible to get those vars (or some vars) before nodes will be 
allocated:

int slurm_spank_init (spank_t sp, int ac, char **av) {
    if (spank_context () != S_CTX_ALLOCATOR) {
       return (0);
    }
...
}

I believe at this step srun command line options and #SBATCH instructions 
should be parsed already or, at least, available.

Kind regards,
Dmitry

Kind regards,
Dmitry Chirikov

On Mon, Jan 30, 2017 at 4:41 PM, Sam Gallop (NBI) 
<sam.gal...@nbi.ac.uk<mailto:sam.gal...@nbi.ac.uk>> wrote:
Hi Dmitry,

I just recently posted this for similar query about SPANK plugins (see below 
for the example code).  It helps dump some of the S_JOB_ARGV or S_JOB_ENV 
variables you are interested in.  It's very basic and there is lots that can be 
done with SPANK.  Example output looks like this …
<snip>
slurm_spank_task_init (ARGV): /bin/sleep
slurm_spank_task_init (ARGV): 30
…
slurm_spank_task_init (ENV): SLURM_JOB_NAME=slurm1.sh
slurm_spank_task_init (ENV): SLURMD_NODENAME=centos1
slurm_spank_task_init (ENV): SLURM_JOBID=266
slurm_spank_task_init (ENV): SLURM_JOB_PARTITION=debug
slurm_spank_task_init (ENV): SLURM_JOB_NUM_NODES=1
slurm_spank_task_init (ENV): SLURM_MEM_PER_NODE=1024
<snip>

# cat userenv.c
/*
*   To compile:
*    gcc -shared -o userenv.so userenv.c
*
*/
#include <sys/types.h>
#include <stdio.h>
#include <stdlib.h>
#include <unistd.h>
#include <string.h>
#include <libgen.h>
#include <signal.h>
#include <sys/resource.h>

#include <slurm/spank.h>
#include <slurm/slurm.h>

SPANK_PLUGIN(userenv, 1);

static int _userenv_opt_process (int val,
                                const char *optarg,
                                int remote);
void get_env(spank_t sp, char *function);

struct spank_option spank_options[] =
{
        {
                "userenv",
                "",
                "helper function",
                0,
                0,
                (spank_opt_cb_f) _userenv_opt_process
        },
        SPANK_OPTIONS_TABLE_END
};

int slurm_spank_task_init(spank_t sp, int ac, char **av)
{
        FILE *f;
        char filename[256];
        char function[50];
        int i, argc;
        char **argv;
        char **env;
        uint32_t s_id;

        sprintf(function, "slurm_spank_task_init");

        if (spank_get_item(sp, S_JOB_ID, &s_id) != ESPANK_SUCCESS) {
                slurm_info("userenv: Failed to get id");
                return (0);
        }

        sprintf(filename, "/tmp/%s_%d_%d", function, spank_context(), s_id);
        f = fopen(filename, "w");

        if (spank_get_item(sp, S_JOB_ARGV, &argc, &argv) == ESPANK_SUCCESS) {
                for (i = 0; i < argc; i++) {
                        fprintf(f, "%s (ARGV): %s\n", function, argv[i]);
                }
        }

        if (spank_get_item(sp, S_JOB_ENV, &env) == ESPANK_SUCCESS) {
                        while (*env != NULL) {
                                fprintf(f, "%s (ENV): %s\n", function, *env);
                                ++env;
                        }
        }

        fclose(f);
}

static int _userenv_opt_process(int val, const char *optarg, int remote)
{
        slurm_info("userenv: userenv plugin has been registered");
        return (0);
}

---
Sam Gallop

[Description: Macintosh HD:Users:fretter:Documents:SugarSync Shared 
Folders:NBI:pics for hpc wiki:CiSlogo578x293.png]

Have you tried looking through our Documentation 
Portal<http://docs.cis.nbi.ac.uk/display/CIS/CiSSupport+Home>
Our documentation isn’t all text, check out our Video 
Tutorials<http://docs.cis.nbi.ac.uk/display/CIS/Video+tutorials>
Have you tried looking through CiS Service Desk<http://support.cis.nbi.ac.uk/>
Keep up to date with availability at CiS Service 
Status<http://status.cis.nbi.ac.uk/>
More information on our HPC, Linux, Storage at HPC 
Support<http://docs.cis.nbi.ac.uk/display/CIS/CiSSupport+Home> Site


To speak to us about technical issues feel free to call the Computing 
infrastructure for Science team on group phone extension 2003.
If your request is urgent, please contact the NBIP Computing Helpdesk at 
computing.helpd...@nbi.ac.uk<mailto:computing.helpd...@nbi.ac.uk> or call phone 
extension 1234.

From: Dmitry Chirikov [mailto:dmi...@chirikov.ru<mailto:dmi...@chirikov.ru>]
Sent: 26 January 2017 12:04
To: slurm-dev <slurm-dev@schedmd.com<mailto:slurm-dev@schedmd.com>>
Subject: [slurm-dev] SPANK and job variables/options

Hi all,

Playing with SPANK I faced up with an issue
Seems I can't get any useful bits about job which user is about to submit.

I am trying to add a hook to allocator context in slurm_spank_init or 
slurm_spank_init_post_opt functions.

Idea was to check some option of the committed job file (or srun options) and 
make a decision about should it be accepted to be queued or not:

But seems like it is not possible to get very basic variables like S_JOB_ARGV 
or S_JOB_ENV

If i get it right this check limits the variables spank able to use to, 
literally, slurm version.
https://github.com/SchedMD/slurm/blob/d7770b9b68d992f4f38417d5ef7204895954eff8/src/common/plugstack.c#L1765

Prolog facilies were looking promised, but they cannot drop the job - it can be 
re-queued only.

So the question is, how to do it properly, when variables are available to use, 
and it is not too late to deny user to submit a job?

Thank you in advance.

Kind regards,
Dmitry Chirikov


Reply via email to