qconf -se ubuntu-node2 : hostname ubuntu-node2 load_scaling NONE complex_values NONE load_values arch=lx26-amd64,num_proc=16,mem_total=48201.960938M, \
swap_total=95746.996094M,virtual_total=143948.957031M, \ load_avg=3.740000,load_short=4.000000, \ load_medium=3.740000,load_long=2.360000, \ mem_free=47376.683594M,swap_free=95746.996094M, \ virtual_free=143123.679688M,mem_used=825.277344M, \ swap_used=0.000000M,virtual_used=825.277344M, \ cpu=25.000000,m_topology=NONE,m_topology_inuse=NONE, \ m_socket=0,m_core=0,np_load_avg=0.233750, \ np_load_short=0.250000,np_load_medium=0.233750, \ np_load_long=0.147500 processors 16 user_lists NONE xuser_lists NONE projects NONE xprojects NONE usage_scaling NONE report_variables NONE El jue., 6 dic. 2018 a las 11:17, Dimar Jaime González Soto (< dimar.gonzalez.s...@gmail.com>) escribió: > qhost : > > HOSTNAME ARCH NCPU LOAD MEMTOT MEMUSE SWAPTO > SWAPUS > > ------------------------------------------------------------------------------- > global - - - - - - > - > ubuntu-frontend lx26-amd64 16 4.13 31.4G 1.2G 0.0 > 0.0 > ubuntu-node11 lx26-amd64 16 4.55 47.1G 397.5M 93.5G > 0.0 > ubuntu-node12 lx26-amd64 16 3.64 47.1G 1.0G 93.5G > 0.0 > ubuntu-node13 lx26-amd64 16 4.54 47.1G 399.9M 93.5G > 0.0 > ubuntu-node2 lx26-amd64 16 3.67 47.1G 818.5M 93.5G > 0.0 > > El jue., 6 dic. 2018 a las 11:13, Reuti (<re...@staff.uni-marburg.de>) > escribió: > >> >> > Am 06.12.2018 um 15:07 schrieb Dimar Jaime González Soto < >> dimar.gonzalez.s...@gmail.com>: >> > >> > qalter -w p doesn't shows anything, qstat shows 16 processes and not >> 60: >> > >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node2 1 1 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node12 1 2 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node13 1 3 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node11 1 4 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node11 1 5 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node13 1 6 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node12 1 7 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node2 1 8 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node2 1 9 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node12 1 10 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node13 1 11 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node11 1 12 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node11 1 13 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node13 1 14 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node12 1 15 >> > 250 0.50000 OMA cbuach r 12/06/2018 11:04:15 >> main.q@ubuntu-node2 1 16 >> > 250 0.50000 OMA cbuach qw 12/06/2018 11:04:02 >> 1 17-60:1 >> >> Aha, so they are running already on remote nodes – fine. As the setting >> in the queue configuration is per host, this should work and provide more >> processes per node instead of four. >> >> Is there a setting for the exechosts: >> >> qconf -se ubuntu-node2 >> >> limiting the slots to 4 in complex_values? Can you please also provide >> the `qhost` output. >> >> -- Reuti >> >> >> >> > >> > El jue., 6 dic. 2018 a las 10:59, Reuti (<re...@staff.uni-marburg.de>) >> escribió: >> > >> > > Am 06.12.2018 um 09:47 schrieb Hay, William <w....@ucl.ac.uk>: >> > > >> > > On Wed, Dec 05, 2018 at 03:29:23PM -0300, Dimar Jaime Gonz??lez Soto >> wrote: >> > >> the app site is https://omabrowser.org/standalone/ I tried to >> make a >> > >> parallel environment but it didn't work. >> > > The website indicates that an array job should work for this. >> > > Has the load average spiked to the point where np_load_avg>=1.75? >> > >> > Yes, I noticed this too. Hence we need no parallel environement at all, >> as OMA will just start several serial jobs as long as slots are available >> AFAICS. >> > >> > What does `qstat` show for a running job. There should be a line for >> each executing task while the waiting once are abbreviated in one line. >> > >> > -- Reuti >> > >> > >> > > >> > > I would try running qalter -w p against the job id to see what it >> says. >> > > >> > > William >> > > >> > > >> > > >> > >> >> > >>> Am 05.12.2018 um 19:10 schrieb Dimar Jaime Gonzalez Soto >> > >> <dimar.gonzalez.s...@gmail.com>: >> > >>> >> > >>> Hi everyone I'm trying to run OMA standalone on a grid engine setup >> > >> with this line: >> > >>> >> > >>> qsub -v NR_PROCESSES=60 -b y -j y -t 1-60 -cwd >> /usr/local/OMA/bin/OMA >> > >>> >> > >>> it works but only execute 4 processes per node, there are 4 nodes >> > >> with 16 logical threads. My main.q configuration is: >> > >>> >> > >>> qname main.q >> > >>> hostlist @allhosts >> > >>> seq_no 0 >> > >>> load_thresholds np_load_avg=1.75 >> > >>> suspend_thresholds NONE >> > >>> nsuspend 1 >> > >>> suspend_interval 00:05:00 >> > >>> priority 0 >> > >>> min_cpu_interval 00:05:00 >> > >>> processors UNDIFINED >> > >>> qtype BATCH INTERACTIVE >> > >>> ckpt_list NONE >> > >>> pe_list make >> > >>> rerun FALSE >> > >>> slots 16 >> > >> > >> > >> > -- >> > Atte. >> > >> > Dimar González Soto >> > Ingeniero Civil en Informática >> > Universidad Austral de Chile >> > >> > >> >> > > -- > Atte. > > Dimar González Soto > Ingeniero Civil en Informática > Universidad Austral de Chile > > > -- Atte. Dimar González Soto Ingeniero Civil en Informática Universidad Austral de Chile
_______________________________________________ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users