I think it was a silly mistake. "placement-strategy" was not enabled. We have enabled it now and testing it out.
Thanks On Mon, Oct 24, 2016 at 7:35 PM, Ulrich Windl < ulrich.wi...@rz.uni-regensburg.de> wrote: > >>> Nikhil Utane <nikhil.subscri...@gmail.com> schrieb am 24.10.2016 um > 13:22 > in > Nachricht > <cagnwmjxmlynjjsjiprmrkhcapb3rnqx9xuwucqmwl5qa8ww...@mail.gmail.com>: > > I had set resource utilization to 1. Even then it scheduled 2 resources. > > Doesn't it honor utilization resources if it doesn't find a free node? > > Show us the config and the logs, please! > > > > > > -Nikhil > > > > On Mon, Oct 24, 2016 at 4:43 PM, Vladislav Bogdanov < > bub...@hoster-ok.com> > > wrote: > > > >> 24.10.2016 14:04, Nikhil Utane wrote: > >> > >>> That is what happened here :(. > >>> When 2 nodes went down, two resources got scheduled on single node. > >>> Isn't there any way to stop this from happening. Colocation constraint > >>> is not helping. > >>> > >> > >> If it is ok to have some instances not running in such outage cases, you > >> can limit them to 1-per-node with utilization attributes (as was > suggested > >> earlier). Then, when nodes return, resource instances will return with > (and > >> on!) them. > >> > >> > >> > >>> -Regards > >>> Nikhil > >>> > >>> On Sat, Oct 22, 2016 at 12:57 AM, Vladislav Bogdanov > >>> <bub...@hoster-ok.com <mailto:bub...@hoster-ok.com>> wrote: > >>> > >>> 21.10.2016 19:34, Andrei Borzenkov wrote: > >>> > >>> 14.10.2016 10:39, Vladislav Bogdanov пишет: > >>> > >>> > >>> use of utilization (balanced strategy) has one caveat: > >>> resources are > >>> not moved just because of utilization of one node is less, > >>> when nodes > >>> have the same allocation score for the resource. So, after > the > >>> simultaneus outage of two nodes in a 5-node cluster, it may > >>> appear > >>> that one node runs two resources and two recovered nodes > run > >>> nothing. > >>> > >>> > >>> I call this a feature. Every resource move potentially means > >>> service > >>> outage, so it should not happen without explicit action. > >>> > >>> > >>> In a case I describe that moves could be easily prevented by using > >>> stickiness (it increases allocation score on a current node). > >>> The issue is that it is impossible to "re-balance" resources in > >>> time-frames when stickiness is zero (over-night maintenance > window). > >>> > >>> > >>> > >>> Original 'utilization' strategy only limits resource > >>> placement, it is > >>> not considered when choosing a node for a resource. > >>> > >>> > >>> > >>> _______________________________________________ > >>> Users mailing list: Users@clusterlabs.org > >>> <mailto:Users@clusterlabs.org> > >>> http://clusterlabs.org/mailman/listinfo/users > >>> <http://clusterlabs.org/mailman/listinfo/users> > >>> > >>> Project Home: http://www.clusterlabs.org > >>> Getting started: > >>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > >>> <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf> > >>> Bugs: http://bugs.clusterlabs.org > >>> > >>> > >>> > >>> _______________________________________________ > >>> Users mailing list: Users@clusterlabs.org <mailto: > >>> Users@clusterlabs.org> > >>> http://clusterlabs.org/mailman/listinfo/users > >>> <http://clusterlabs.org/mailman/listinfo/users> > >>> > >>> Project Home: http://www.clusterlabs.org > >>> Getting started: > >>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > >>> <http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf> > >>> Bugs: http://bugs.clusterlabs.org > >>> > >>> > >>> > >>> > >>> _______________________________________________ > >>> Users mailing list: Users@clusterlabs.org > >>> http://clusterlabs.org/mailman/listinfo/users > >>> > >>> Project Home: http://www.clusterlabs.org > >>> Getting started: http://www.clusterlabs.org/ > doc/Cluster_from_Scratch.pdf > >>> Bugs: http://bugs.clusterlabs.org > >>> > >>> > >> > >> _______________________________________________ > >> Users mailing list: Users@clusterlabs.org > >> http://clusterlabs.org/mailman/listinfo/users > >> > >> Project Home: http://www.clusterlabs.org > >> Getting started: http://www.clusterlabs.org/ > doc/Cluster_from_Scratch.pdf > >> Bugs: http://bugs.clusterlabs.org > >> > > > > > _______________________________________________ > Users mailing list: Users@clusterlabs.org > http://clusterlabs.org/mailman/listinfo/users > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org >
_______________________________________________ Users mailing list: Users@clusterlabs.org http://clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org