[
https://issues.apache.org/jira/browse/AMBARI-8244?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14385052#comment-14385052
]
Yusaku Sako commented on AMBARI-8244:
-------------------------------------
The above Hadoop QA failure is unrelated to this patch.
When I tested AMBARI-8244.6.patch end-to-end on a real cluster, I've found the
following:
* Cluster installation and service checks are fine.
* Post-install, I was able to move the NameNode on c6401 to c6402 and things
were ok after the move.
* When enabled NameNode HA, things seemed to have gone smoothly.
dfs.namenode.rpc-address is still pointing to c6402, but it worked fine after
NameNode on c6402 was shut down; *hadoop fs* commands to read/write on HDFS
were running fine.
* HDFS service check failed because it ran "hadoop dfsadmin -safemode get" with
"-fs" parameter that specifies a specific NameNode based on
dfs.namenode.rpc-address (this is due to changes in the patch); if that
NameNode is down, then the command would obviously fail. So we need to remove
this property when NameNode HA is enabled.
* However, that might not be the full story. *hadoop dfsadmin -safemode get*,
even when triggered from the command line caused issues; it still tried to
access the NameNode that is down and not the other one even without the "-fs"
parameter. I've hand edited hdfs-site.xml to remove dfs.namenode.rpc-address,
but "hadoop dfsadmin -safemode get" still tries to connect to the NameNode that
is down.
{code}
safemode: Call From c6401.ambari.apache.org/192.168.64.101 to
c6402.ambari.apache.org:8020 failed on connection exception:
java.net.ConnectException: Connection refused; For more details see:
http://wiki.apache.org/hadoop/ConnectionRefused
{code}
What's confusing is that the "
> Ambari HDP 2.0.6+ stacks do not work with fs.defaultFS not being hdfs
> ---------------------------------------------------------------------
>
> Key: AMBARI-8244
> URL: https://issues.apache.org/jira/browse/AMBARI-8244
> Project: Ambari
> Issue Type: Bug
> Components: stacks
> Affects Versions: 2.0.0
> Reporter: Ivan Mitic
> Assignee: Ivan Mitic
> Labels: HDP
> Fix For: 2.1.0
>
> Attachments: AMBARI-8244.2.patch, AMBARI-8244.3.patch,
> AMBARI-8244.4.patch, AMBARI-8244.5.patch, AMBARI-8244.6.patch,
> AMBARI-8244.patch
>
>
> Right now changing the default file system does not work with the HDP 2.0.6+
> stacks. Given that it might be common to run HDP against some other file
> system in the cloud, adding support for this will be super useful. One
> alternative is to consider a separate stack definition for other file
> systems, however, given that I noticed just 2 minor bugs needed to support
> this, I would rather extend on the existing code.
> Bugs:
> - One issue is in Nagios install scripts, where it is assumed that
> fs.defaultFS has the namenode port number.
> - Another issue is in HDFS install scripts, where {{hadoop dfsadmin}}
> command only works when hdfs is the default file system.
> Fix for both places is to extract the namenode address/port from
> {{dfs.namenode.rpc-address}} if one is defined and use it instead of relying
> on {{fs.defaultFS}}.
> Haven't included any tests yet (my first Ambari patch, not sure what is
> appropriate, so please comment).
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)