> $ qconf -sc | grep gpu
> gpu                 gpu        INT       <=      YES         YES 0        0
>
> Should be:
>
> # qconf -sc | grep gpu
> gpu                 gpu        INT       <=      YES         JOB 0        0
>
> I will poke at this on Monday and hopefully not mess things up too badly...
> :-)
>
> Rajil, does this help you as well?  I think this answers the same
> question(s) that I had.
>
> -Dj

Yes, this seems to be working. I also set up a load_sensor script
using the script mentioned here
http://serverfault.com/questions/322073/howto-set-up-sge-for-cuda-devices
something like this
#qconf -sconf compute-4-0
compute-4-0.local: mailer /bin/mail
xterm /usr/bin/X11/xterm
execd_spool_dir /opt/gridengine/default/spool
load_sensor /share/apps/load_sensor.sh

The third submitted job is now in 'qw' as i wanted.
#qstat
425 0.50500 j1 rajil r 12/05/2015 12:36:32 gpu.q@compute-4-0.local 8
426 0.50500 j2 rajil r 12/05/2015 12:37:17 gpu.q@compute-4-1.local 8
427 0.50500 j3 rajil qw 12/05/2015 12:39:18 8

Also i can see the gpu values to be 0 when they are being used
#qhost -F gpu
compute-4-0 linux-x64 32 1.73 126.1G 26.2G 1000.0M 0.0
      Host Resource(s): hc:gpu=0.000000
compute-4-1 linux-x64 32 1.81 126.1G 24.4G 1000.0M 0.0
      Host Resource(s): hc:gpu=0.000000

On the compute nodes i had to specify the gpu value
#qconf -se compute-4-0
hostname compute-4-0.local
load_scaling NONE
complex_values gpu=1

But leave it the complex type empty in the queue
#qconf -sq gpu.q
complex_values NONE

The queue submission script is like
#!/bin/csh
#$ -V #$ -S /bin/csh
#$ -N j1
#$ -q gpu.q
#$ -l gpu=1
#$ -m beas
#$ -j y -o /home/rajil/tmp/tst/j1.qlog
#$ -pe mpi 8
abaqus python /share/apps/abaqus/6.14-2/../abaJobHandler.py j1
/home/rajil/tmp/tst j1.fs.133848 0 j1.com model.inp

-Rajil
_______________________________________________
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users

Reply via email to