Hello, i have a strange effect, where i am not sure whether it is "only" a misconfiguration or a bug.
First: I run son of gridengine 8.1.9-1.el6.x86_64 (i installed the rhel rpm on an opensuse 13.1 machine. This should not matter in this case, and it is reported to be able to run on opensuse). mpirun and mpiexec are from openmpi-1.10.3 (no other mpi was installed, neither on master, nor on slaves). The installation was made with: ./configure --prefix=`pwd`/build --disable-dlopen --disable-mca-dso --with-orte --with-sge --with-x --enable-mpi-thread-multiple --enable-orterun-prefix-by-default --enable-mpirun-prefix-by-default --enable-orte-static-ports --enable-mpi-cxx --enable-mpi-cxx-seek --enable-oshmem --enable-java --enable-mpi-java make make install I attached the outputs of 'qconf -ap all.q' , 'qconf -sconf' and 'qconf -sp orte' as textfiles. Now my problem: I asked for 20 cores and if i run qstat -u '*' it shows that this job is being run in slave07 using 20 cores but is not true! if i run qstat -f -u '*' i see that this job is only using 3 cores in salve07 and there are 17 cores in other nodes allocated to this job which are in fact unused! Or other example: My job took say 6 cpus on slave07 and 14 on slave06 but nothing was running on 06 and therefore a waste of ressource on 06 and overload on 07 becomes highly possible (the numbers are made up). If i ran 1 Cpus in many independent jobs that would not be an issue, but imagine i now request 60 cpus on slave07, that would seriously overload the node in many cases. Or other example: if i ask for say 50 CPUs, the job will start on one node, e.g, slave01, but only reserving say 15 CPUs out of 64 and reserve the rest on many other nodes (obviously wasting space doing nothing). This has the bad consequence of allocating many more CPUs than available when many jobs are running, imagine you have 10 jobs like this one... some nodes will run maybe 3 even if they only have 24 CPUs... I hope that i have made clear what the issue is. I also see that the `qstat` and `qstat -f` are in disagreement. The latter is correct, i checked the processes running on the nodes. Did somebody already encounter such a problem? Does somebody have an idea where to look into or what to test? With kind regards, ulrich
HOSTNAME ARCH NCPU NSOC NCOR NTHR LOAD MEMTOT MEMUSE SWAPTO SWAPUS ---------------------------------------------------------------------------------------------- global - - - - - - - - - - slave01 lx-amd64 64 4 64 64 0.01 504.9G 2.3G 10.0G 0.0 slave02 lx-amd64 64 4 64 64 123.0 504.8G 171.7G 10.0G 0.0 slave03 lx-amd64 64 4 64 64 0.01 504.9G 2.0G 10.0G 0.0 slave04 lx-amd64 64 4 64 64 0.01 504.9G 2.0G 10.0G 0.0 slave05 lx-amd64 64 4 64 64 2.01 504.9G 40.2G 10.0G 0.0 slave06 lx-amd64 40 4 40 40 0.01 314.8G 1.1G 32.0G 0.0 slave07 lx-amd64 24 2 24 24 166.9 188.8G 188.0G 30.0G 28.7G slave08 lx-amd64 24 2 24 24 0.01 188.8G 660.5M 30.0G 7.5M
#global: execd_spool_dir /opt/sge/default/spool mailer /bin/mail xterm /usr/bin/xterm load_sensor none prolog none epilog none shell_start_mode posix_compliant login_shells sh,bash,ksh,csh,tcsh min_uid 100 min_gid 100 user_lists none xuser_lists none projects none xprojects none enforce_project false enforce_user auto load_report_time 00:00:40 max_unheard 00:05:00 reschedule_unknown 00:00:00 loglevel log_warning administrator_mail hil...@mpia-hd.mpg.de set_token_cmd none pag_cmd none token_extend_time none shepherd_cmd none qmaster_params none execd_params none reporting_params accounting=true reporting=false \ flush_time=00:00:15 joblog=false sharelog=00:00:00 finished_jobs 100 gid_range 20000-20100 qlogin_command builtin qlogin_daemon builtin rlogin_command builtin rlogin_daemon builtin rsh_command builtin rsh_daemon builtin max_aj_instances 2000 max_aj_tasks 75000 max_u_jobs 0 max_jobs 0 max_advance_reservations 0 auto_user_oticket 0 auto_user_fshare 0 auto_user_default_project none auto_user_delete_time 86400 delegated_file_staging false reprioritize 0 jsv_url none jsv_allowed_mod ac,h,i,e,o,j,M,N,p,w
pe_name orte slots 408 user_lists NONE xuser_lists NONE start_proc_args NONE stop_proc_args NONE allocation_rule $round_robin control_slaves TRUE job_is_first_task FALSE urgency_slots min accounting_summary FALSE qsort_args NONE
qname all.q hostlist @allhosts seq_no 0 load_thresholds np_load_avg=8.75 suspend_thresholds NONE nsuspend 1 suspend_interval 00:05:00 priority 0 min_cpu_interval 00:05:00 processors UNDEFINED qtype BATCH INTERACTIVE ckpt_list NONE pe_list make smp mpi orte rerun FALSE slots 1,[slave01=64],[slave02=64],[slave03=64],[slave04=64], \ [slave05=64],[slave06=40],[slave07=24],[slave08=24] tmpdir /tmp shell /bin/sh prolog NONE epilog NONE shell_start_mode posix_compliant starter_method NONE suspend_method NONE resume_method NONE terminate_method NONE notify 00:00:60 owner_list NONE user_lists NONE xuser_lists NONE subordinate_list NONE complex_values NONE projects NONE xprojects NONE calendar NONE initial_state default s_rt INFINITY h_rt INFINITY s_cpu INFINITY h_cpu INFINITY s_fsize INFINITY h_fsize INFINITY s_data INFINITY h_data INFINITY s_stack INFINITY h_stack INFINITY s_core INFINITY h_core INFINITY s_rss INFINITY h_rss INFINITY s_vmem INFINITY h_vmem INFINITY
_______________________________________________ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users