Hi, Esteban.
Thanks for your reply. Thus, QJM automatic failover option is a cut-edge
thing. Am I right?
I think that it's a good idea to have truly equal NNs doing their work
in parallel, as Konstantin Shvachko mentioned.
On 07/02/2015 04:49 PM, Esteban Gutierrez wrote:
Hi Dmitry,
Have you looked into the QJM automatic failover mode using the
ZKFailoverController?
https://hadoop.apache.org/docs/r2.6.0/hadoop-project-dist/hadoop-hdfs/HDFSHighAvailabilityWithQJM.html#Automatic_Failover
This is the most commonly used HA mode in production environments. Also
there is some recent work that will be in Hadoop 3 that will allow to have
more than 1 stand-by NNs: https://issues.apache.org/jira/browse/HDFS-6440
cheers,
esteban.
--
Cloudera, Inc.
On Thu, Jul 2, 2015 at 7:42 AM, Dmitry Salychev <darkness....@gmail.com>
wrote:
Sure, I did. It's actually not what I'm looking for. I don't want to spend
time to make dead NN alive by my hands. There should be a solution for
NN-SPOF problem.
On 07/02/2015 04:36 PM, Vinayakumar B wrote:
Hi..
Did you look at the HDFS Namenode high availability?
-Vinay
On Jul 2, 2015 11:50 AM, "Dmitry Salychev" <darkness....@gmail.com>
wrote:
Hello, HDFS Developers.
I know that NN is a single point of failure of an entire HDFS cluster. If
it fails, the cluster will be unavailable no matter how many DN there. I
know that there is an initiative <
http://www.wandisco.com/system/files/documentation/Meetup-ConsensusReplication.pdf
which introduces ConsensusNode (as I can see it looks like distributed
NN)
and related issues (HDFS-6469 <
https://issues.apache.org/jira/browse/HDFS-6469>, HADOOP-10641 <
https://issues.apache.org/jira/browse/HADOOP-10641> and HDFS-7007 <
https://issues.apache.org/jira/browse/HDFS-7007>). So, I'd like to ask.
Has this NN-SPOF problem been solved? If it hasn't, can you show me an
entry point where I can help to solve it?
Thanks for your time.