hi not sure what is the issues here, see in-line On 5/11/2012 8:11 AM, iqtcub wrote:
Hi,Following up with the thread with the same subject ( http://thread.gmane.org/gmane.comp.clustering.opengridengine.user/894/ ).We're using sge 6.2u5, our setup is 2 machines(its a testing cluster) with 2 cores each machine.-qsub -q v20z.q -pe smp 1 script.sub -wait until the job runs -qsub -q v20z.q -pe smp 1 script.sub Each job enters a different node.
this is not ok for you?
However if we do: -for i in 1 2; do qsub -q v20z.q -pe smp 1 script.sub; done Then both jobs enter into the same node.
this is ok for you?
Our scheduling conf is as follows: ---------------------------- algorithm default schedule_interval 0:0:15 maxujobs 0 queue_sort_method load job_load_adjustments NONE load_adjustment_decay_time 0:7:30 load_formula slots schedd_job_info true flush_submit_sec 0 flush_finish_sec 0 params MONITOR=1 reprioritize_interval 0:0:0 halftime 168 usage_weight_list cpu=1.000000,mem=0.000000,io=0.000000 compensation_factor 2.000000 weight_user 0.250000 weight_project 0.250000 weight_department 0.250000 weight_job 0.250000 weight_tickets_functional 0 weight_tickets_share 1000000 share_override_tickets TRUE share_functional_shares FALSE max_functional_jobs_to_schedule 200 report_pjob_tickets TRUE max_pending_tasks_per_job 50 halflife_decay_list none policy_hierarchy OS weight_ticket 0.890000 weight_waiting_time 0.000000 weight_deadline 3600000.000000 weight_urgency 0.100000 weight_priority 0.010000 max_reservation 50 default_duration 9999:00:00 -------------------------------------------- The smp PE config is: pe_name smp slots 999 user_lists NONE xuser_lists NONE start_proc_args /bin/true stop_proc_args /bin/true allocation_rule $pe_slots control_slaves FALSE job_is_first_task TRUE urgency_slots min accounting_summary TRUE ------------------------- The config on both nodes are like this: hostname v20z-03 load_scaling NONE complex_values mem_free=7891.796875M,slots=2load_values arch=lx24-amd64,num_proc=2,mem_total=7935.984375M, \ swap_total=4095.992188M,virtual_total=12031.976562M, \ h_fsize=9.7G,load_avg=0.180000,load_short=0.080000, \load_medium=0.180000,load_long=0.090000, \ mem_free=7830.246094M,swap_free=4095.992188M, \ virtual_free=11926.238281M,mem_used=105.738281M, \ swap_used=0.000000M,virtual_used=105.738281M, \cpu=0.000000,m_topology=SCSC,m_topology_inuse=SCSC, \m_socket=2,m_core=2,np_load_avg=0.090000, \ np_load_short=0.040000,np_load_medium=0.090000, \ np_load_long=0.045000 processors 2 user_lists NONE xuser_lists NONE projects NONE xprojects NONE usage_scaling cpu=12.300000 report_variables NONE ------------------------------- The queue config: qname v20z.q hostlist @v20z seq_no 0 load_thresholds np_load_avg=1.75 suspend_thresholds NONE nsuspend 1 suspend_interval 00:01:00 priority 0 min_cpu_interval 00:01:00 processors UNDEFINED qtype BATCH INTERACTIVE ckpt_list BLCR pe_list make smp rerun FALSE slots 2 tmpdir /scratch shell /bin/csh prolog NONE epilog NONE shell_start_mode posix_compliant starter_method NONE suspend_method NONE resume_method NONE terminate_method NONE notify 00:00:60 owner_list NONE user_lists NONE xuser_lists NONE subordinate_list NONE complex_values split=2 projects NONE xprojects NONE calendar NONE initial_state default s_rt INFINITY h_rt INFINITY s_cpu INFINITY h_cpu INFINITY s_fsize INFINITY h_fsize INFINITY s_data INFINITY h_data INFINITY s_stack INFINITY h_stack INFINITY s_core INFINITY h_core INFINITY s_rss INFINITY h_rss INFINITY s_vmem INFINITY h_vmem INFINITY -----------------------------------From what i understood, its possible that this method is broken, am i right?I've also tried the scheduler configuration in the following links, with the same result:http://article.gmane.org/gmane.comp.clustering.opengridengine.user/1037 http://wiki.gridengine.info/wiki/index.php/StephansBlog Thanks in advance! _______________________________________________ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users
--
<<attachment: laotsao.vcf>>
_______________________________________________ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users