Repository: ambari Updated Branches: refs/heads/trunk ef9a456a7 -> cbcadd2c3
AMBARI-11704. Mahout service check fails due to bad folder permission for /user/ambari-qa Project: http://git-wip-us.apache.org/repos/asf/ambari/repo Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/cbcadd2c Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/cbcadd2c Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/cbcadd2c Branch: refs/heads/trunk Commit: cbcadd2c3e1006eece7ff7f45122cf606cc2160b Parents: ef9a456 Author: Sumit Mohanty <smoha...@hortonworks.com> Authored: Thu Jun 4 19:35:18 2015 -0700 Committer: Sumit Mohanty <smoha...@hortonworks.com> Committed: Thu Jun 4 19:35:18 2015 -0700 ---------------------------------------------------------------------- .../2.1.0.2.0/package/scripts/hdfs_namenode.py | 5 ++++- .../python/stacks/2.0.6/HDFS/test_namenode.py | 18 +++++++++--------- 2 files changed, 13 insertions(+), 10 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/ambari/blob/cbcadd2c/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_namenode.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_namenode.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_namenode.py index eec864c..d26d145 100644 --- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_namenode.py +++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_namenode.py @@ -90,7 +90,10 @@ def namenode(action=None, do_format=True, rolling_restart=False, env=None): user = params.hdfs_user) is_namenode_safe_mode_off = format("hadoop dfsadmin -fs {namenode_address} -safemode get | grep 'Safe mode is OFF'") - is_active_namenode_cmd = as_user(format("hdfs --config {hadoop_conf_dir} haadmin -getServiceState {namenode_id} | grep active"), params.hdfs_user, env={'PATH':params.hadoop_bin_dir}) + if params.dfs_ha_enabled: + is_active_namenode_cmd = as_user(format("hdfs --config {hadoop_conf_dir} haadmin -getServiceState {namenode_id} | grep active"), params.hdfs_user, env={'PATH':params.hadoop_bin_dir}) + else: + is_active_namenode_cmd = None # During normal operations, if HA is enabled and it is in standby, then stay in current state, otherwise, leave safemode. # During Rolling Upgrade, both namenodes must leave safemode. http://git-wip-us.apache.org/repos/asf/ambari/blob/cbcadd2c/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py index ddab048..b920c17 100644 --- a/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py +++ b/ambari-server/src/test/python/stacks/2.0.6/HDFS/test_namenode.py @@ -101,7 +101,7 @@ class TestNamenode(RMFTestCase): ) self.assertResourceCalled('HdfsResource', '/tmp', security_enabled = False, - only_if="ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState None | grep active'", + only_if=None, keytab = UnknownConfigurationMock(), hadoop_bin_dir = '/usr/bin', default_fs = 'wasb://a...@c6401.ambari.apache.org', @@ -117,7 +117,7 @@ class TestNamenode(RMFTestCase): ) self.assertResourceCalled('HdfsResource', '/user/ambari-qa', security_enabled = False, - only_if="ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState None | grep active'", + only_if=None, keytab = UnknownConfigurationMock(), hadoop_bin_dir = '/usr/bin', default_fs = 'wasb://a...@c6401.ambari.apache.org', @@ -133,7 +133,7 @@ class TestNamenode(RMFTestCase): ) self.assertResourceCalled('HdfsResource', None, security_enabled = False, - only_if="ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState None | grep active'", + only_if=None, keytab = UnknownConfigurationMock(), hadoop_bin_dir = '/usr/bin', default_fs = 'wasb://a...@c6401.ambari.apache.org', @@ -217,7 +217,7 @@ class TestNamenode(RMFTestCase): ) self.assertResourceCalled('HdfsResource', '/tmp', security_enabled = False, - only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState None | grep active'", + only_if = None, keytab = UnknownConfigurationMock(), hadoop_bin_dir = '/usr/bin', default_fs = 'hdfs://c6401.ambari.apache.org:8020', @@ -233,7 +233,7 @@ class TestNamenode(RMFTestCase): ) self.assertResourceCalled('HdfsResource', '/user/ambari-qa', security_enabled = False, - only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState None | grep active'", + only_if = None, keytab = UnknownConfigurationMock(), hadoop_bin_dir = '/usr/bin', default_fs = 'hdfs://c6401.ambari.apache.org:8020', @@ -249,7 +249,7 @@ class TestNamenode(RMFTestCase): ) self.assertResourceCalled('HdfsResource', None, security_enabled = False, - only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState None | grep active'", + only_if = None, keytab = UnknownConfigurationMock(), hadoop_bin_dir = '/usr/bin', default_fs = 'hdfs://c6401.ambari.apache.org:8020', @@ -365,7 +365,7 @@ class TestNamenode(RMFTestCase): type = 'directory', action = ['create_on_execute'], hdfs_site=self.getConfig()['configurations']['hdfs-site'], principal_name='hdfs', default_fs='hdfs://c6401.ambari.apache.org:8020', mode = 0777, - only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState None | grep active'" + only_if = None ) self.assertResourceCalled('HdfsResource', '/user/ambari-qa', security_enabled = True, @@ -378,11 +378,11 @@ class TestNamenode(RMFTestCase): type = 'directory', action = ['create_on_execute'], hdfs_site=self.getConfig()['configurations']['hdfs-site'], principal_name='hdfs', default_fs='hdfs://c6401.ambari.apache.org:8020', mode = 0770, - only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState None | grep active'" + only_if = None ) self.assertResourceCalled('HdfsResource', None, security_enabled = True, - only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState None | grep active'", + only_if = None, keytab = '/etc/security/keytabs/hdfs.headless.keytab', hadoop_bin_dir = '/usr/bin', kinit_path_local = '/usr/bin/kinit',