Hi, Am 07.12.2012 um 11:16 schrieb Arnau Bria:
> I've configured our cluster in the way that slots slots are consumable by default in `qconf -sc`. > /memory are consumable > resources. Our nodes have their limits and there are some default > resources requirements at job submission. All this conf should avoid > memory/processor oversubscription (at least, from what I've read). > something like http://jeetworks.org/node/93 ... is this the recommended > way for avoiding over-subscription? Yes. OTOH you can use an RQS. But instead of the intermediate files the author provided I would prefer the command similar to: $ for node in node{01..24}; do qconf -mattr exechost complex_values slots=8 $node; done (You need Bash 4 for leading zeros here.) NB: default requests for consumables I would put in the complex definition. > I've also configured core-binding, and the default for each job is 1 > slot. > > But with this conf I have some questions: > > 1.-) when submitting a job specifying more than 1 job slot (-l slots=2 > -binding linear:2), OGS fails and suggest to use parallel environment. Yep. > I've read > somewhere that this is OGS design, so I need a pe. I've not found a > clear doc about pe (yes, how to create and manage, but not a complete > definition of each parameter and its implications) so, anyone could > share some doc about it? The definitions of the entries are in the man page. What is unclear therein? > what is the minimun conf I need for allowing > slots requirement al job submission? somethign like: > > $ qconf -sp smp > pe_name smp > slots 1024 > user_lists NONE > xuser_lists NONE > start_proc_args NONE > stop_proc_args NONE > allocation_rule $pe_slots With this setting you are limited to one host, but this is correct for smp. > control_slaves FALSE For tightly integrated jobs the above needs to be set to TRUE in case you want to issue a local startup of a second process (or going to a second node). For forks and threads instead it doesn't matter. > job_is_first_task TRUE > urgency_slots min > accounting_summary FALSE > > > is enough? (this is what I'm using and works fine) Yes. > 2.-) I've still not configured share based priorities, but when done, > if a user request more slots/memory than the default, but it does not > use them, is this request taken into account for share calculation? > I mean, user A request 8GB and 1 cpu and uses 8GB and 3600 sec of cpu, h_cpu or h_rt? > and user B requests 16GB and 2 cpus, but uses 8GB and 3600 of cpu. Are > both users priorities recalculated by resource usage or by resource > requests? Depends of the setting: ACCT_RESERVED_USAGE and SHARETREE_RESERVED_USAGE in SGE's configuration (`man sge_conf`). -- Reuti > TIA, > Arnau > _______________________________________________ > users mailing list > [email protected] > https://gridengine.org/mailman/listinfo/users _______________________________________________ users mailing list [email protected] https://gridengine.org/mailman/listinfo/users
