Maybe we could add a user parameter to specify a cluster name that is used to make the paths unique.
On Thu, Nov 19, 2015, 11:24 Till Rohrmann <trohrm...@apache.org> wrote: > I agree that this would make the configuration easier. However, it entails > also that the user has to retrieve the randomized path from the logs if he > wants to restart jobs after the cluster has crashed or intentionally > restarted. Furthermore, the system won't be able to clean up old checkpoint > and job handles in case that the cluster stop was intentional. > > Thus, the question is how do we define the behaviour in order to retrieve > handles and to clean up old handles so that ZooKeeper won't be cluttered > with old handles? > > There are basically two modes: > > 1. Keep state handles when shutting down the cluster. Provide a mean to > define a fixed path when starting the cluster and also a mean to purge old > state handles. Furthermore, add a shutdown mode where the handles under the > current path are directly removed. This mode would guarantee to always have > the state handles available if not explicitly told differently. However, > the downside is that ZooKeeper will be cluttered most certainly. > > 2. Remove the state handles when shutting down the cluster. Provide a > shutdown mode where we keep the state handles. This will keep ZooKeeper > clean but will give you also the possibility to keep a checkpoint around if > necessary. However, the user is more likely to lose his state when shutting > down the cluster. > > On Thu, Nov 19, 2015 at 10:55 AM, Robert Metzger <rmetz...@apache.org> > wrote: > >> I agree with Aljoscha. Many companies install Flink (and its config) in a >> central directory and users share that installation. >> >> On Thu, Nov 19, 2015 at 10:45 AM, Aljoscha Krettek <aljos...@apache.org> >> wrote: >> >>> I think we should find a way to randomize the paths where the HA stuff >>> stores data. If users don’t realize that they store data in the same paths >>> this could lead to problems. >>> >>> > On 19 Nov 2015, at 08:50, Till Rohrmann <trohrm...@apache.org> wrote: >>> > >>> > Hi Gwenhaël, >>> > >>> > good to hear that you could resolve the problem. >>> > >>> > When you run multiple HA flink jobs in the same cluster, then you >>> don’t have to adjust the configuration of Flink. It should work out of the >>> box. >>> > >>> > However, if you run multiple HA Flink cluster, then you have to set >>> for each cluster a distinct ZooKeeper root path via the option >>> recovery.zookeeper.path.root in the Flink configuraiton. This is necessary >>> because otherwise all JobManagers (the ones of the different clusters) will >>> compete for a single leadership. Furthermore, all TaskManagers will only >>> see the one and only leader and connect to it. The reason is that the >>> TaskManagers will look up their leader at a ZNode below the ZooKeeper root >>> path. >>> > >>> > If you have other questions then don’t hesitate asking me. >>> > >>> > Cheers, >>> > Till >>> > >>> > >>> > On Wed, Nov 18, 2015 at 6:37 PM, Gwenhael Pasquiers < >>> gwenhael.pasqui...@ericsson.com> wrote: >>> > Nevermind, >>> > >>> > >>> > >>> > Looking at the logs I saw that it was having issues trying to connect >>> to ZK. >>> > >>> > To make I short is had the wrong port. >>> > >>> > >>> > >>> > It is now starting. >>> > >>> > >>> > >>> > Tomorrow I’ll try to kill some JobManagers *evil*. >>> > >>> > >>> > >>> > Another question : if I have multiple HA flink jobs, are there some >>> points to check in order to be sure that they won’t collide on hdfs or ZK ? >>> > >>> > >>> > >>> > B.R. >>> > >>> > >>> > >>> > Gwenhaël PASQUIERS >>> > >>> > >>> > >>> > From: Till Rohrmann [mailto:till.rohrm...@gmail.com] >>> > Sent: mercredi 18 novembre 2015 18:01 >>> > To: user@flink.apache.org >>> > Subject: Re: YARN High Availability >>> > >>> > >>> > >>> > Hi Gwenhaël, >>> > >>> > >>> > >>> > do you have access to the yarn logs? >>> > >>> > >>> > >>> > Cheers, >>> > >>> > Till >>> > >>> > >>> > >>> > On Wed, Nov 18, 2015 at 5:55 PM, Gwenhael Pasquiers < >>> gwenhael.pasqui...@ericsson.com> wrote: >>> > >>> > Hello, >>> > >>> > >>> > >>> > We’re trying to set up high availability using an existing zookeeper >>> quorum already running in our Cloudera cluster. >>> > >>> > >>> > >>> > So, as per the doc we’ve changed the max attempt in yarn’s config as >>> well as the flink.yaml. >>> > >>> > >>> > >>> > recovery.mode: zookeeper >>> > >>> > recovery.zookeeper.quorum: host1:3181,host2:3181,host3:3181 >>> > >>> > state.backend: filesystem >>> > >>> > state.backend.fs.checkpointdir: hdfs:///flink/checkpoints >>> > >>> > recovery.zookeeper.storageDir: hdfs:///flink/recovery/ >>> > >>> > yarn.application-attempts: 1000 >>> > >>> > >>> > >>> > Everything is ok as long as recovery.mode is commented. >>> > >>> > As soon as I uncomment recovery.mode the deployment on yarn is stuck >>> on : >>> > >>> > >>> > >>> > “Deploying cluster, current state ACCEPTED”. >>> > >>> > “Deployment took more than 60 seconds….” >>> > >>> > Every second. >>> > >>> > >>> > >>> > And I have more than enough resources available on my yarn cluster. >>> > >>> > >>> > >>> > Do you have any idea of what could cause this, and/or what logs I >>> should look for in order to understand ? >>> > >>> > >>> > >>> > B.R. >>> > >>> > >>> > >>> > Gwenhaël PASQUIERS >>> > >>> > >>> > >>> > >>> >>> >> >