AMBARI-13454. Make HDP 2.0 stack inactive for Ambari 2.1.3 (aonishuk)
Project: http://git-wip-us.apache.org/repos/asf/ambari/repo Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/e3958e2d Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/e3958e2d Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/e3958e2d Branch: refs/heads/branch-dev-patch-upgrade Commit: e3958e2da099f783ce637a57d752353fafd3f5b9 Parents: 8817569 Author: Andrew Onishuk <aonis...@hortonworks.com> Authored: Fri Oct 16 17:25:05 2015 +0300 Committer: Andrew Onishuk <aonis...@hortonworks.com> Committed: Fri Oct 16 17:25:05 2015 +0300 ---------------------------------------------------------------------- .../2.1.0.2.0/package/scripts/params_linux.py | 7 +----- .../0.12.0.2.0/package/scripts/params_linux.py | 20 +++++----------- .../4.0.0.2.0/package/scripts/params_linux.py | 2 +- .../2.0.6/hooks/after-INSTALL/scripts/params.py | 6 +---- .../2.0.6/hooks/before-ANY/scripts/params.py | 6 +---- .../2.0.6/hooks/before-START/scripts/params.py | 6 +---- .../main/resources/stacks/HDP/2.0/metainfo.xml | 2 +- .../stacks/2.0.6/HIVE/test_hcat_client.py | 8 +++---- .../stacks/2.0.6/HIVE/test_hive_metastore.py | 8 +++++++ .../stacks/2.0.6/HIVE/test_webhcat_server.py | 24 ++++++++++---------- 10 files changed, 36 insertions(+), 53 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py index 38eac2e..4cbce34 100644 --- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py +++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py @@ -106,12 +106,7 @@ hdfs_user_nofile_limit = default("/configurations/hadoop-env/hdfs_user_nofile_li hdfs_user_nproc_limit = default("/configurations/hadoop-env/hdfs_user_nproc_limit", "65536") create_lib_snappy_symlinks = not Script.is_hdp_stack_greater_or_equal("2.2") - -if Script.is_hdp_stack_greater_or_equal("2.0") and Script.is_hdp_stack_less_than("2.1") and not OSCheck.is_suse_family(): - # deprecated rhel jsvc_path - jsvc_path = "/usr/libexec/bigtop-utils" -else: - jsvc_path = "/usr/lib/bigtop-utils" +jsvc_path = "/usr/lib/bigtop-utils" execute_path = os.environ['PATH'] + os.pathsep + hadoop_bin_dir ulimit_cmd = "ulimit -c unlimited ; " http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py index e16df78..14cf1ac 100644 --- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py +++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py @@ -49,7 +49,7 @@ hostname = config["hostname"] # This is expected to be of the form #.#.#.# stack_version_unformatted = str(config['hostLevelParams']['stack_version']) hdp_stack_version_major = format_hdp_stack_version(stack_version_unformatted) -stack_is_hdp21 = Script.is_hdp_stack_greater_or_equal("2.0") and Script.is_hdp_stack_less_than("2.2") +stack_is_hdp21 = Script.is_hdp_stack_less_than("2.2") # this is not available on INSTALL action because hdp-select is not available hdp_stack_version = functions.get_hdp_version('hive-server2') @@ -83,10 +83,6 @@ sqoop_tar_file = '/usr/share/HDP-webhcat/sqoop*.tar.gz' hive_specific_configs_supported = False hive_etc_dir_prefix = "/etc/hive" limits_conf_dir = "/etc/security/limits.d" -hcat_conf_dir = '/etc/hcatalog/conf' -config_dir = '/etc/hcatalog/conf' -hcat_lib = '/usr/lib/hcatalog/share/hcatalog' -webhcat_bin_dir = '/usr/lib/hcatalog/sbin' hive_user_nofile_limit = default("/configurations/hive-env/hive_user_nofile_limit", "32000") hive_user_nproc_limit = default("/configurations/hive-env/hive_user_nproc_limit", "16000") @@ -101,11 +97,10 @@ hive_config_dir = status_params.hive_config_dir hive_client_conf_dir = status_params.hive_client_conf_dir hive_server_conf_dir = status_params.hive_server_conf_dir -if Script.is_hdp_stack_greater_or_equal("2.1"): - hcat_conf_dir = '/etc/hive-hcatalog/conf' - config_dir = '/etc/hive-webhcat/conf' - hcat_lib = '/usr/lib/hive-hcatalog/share/hcatalog' - webhcat_bin_dir = '/usr/lib/hive-hcatalog/sbin' +hcat_conf_dir = '/etc/hive-hcatalog/conf' +config_dir = '/etc/hive-webhcat/conf' +hcat_lib = '/usr/lib/hive-hcatalog/share/hcatalog' +webhcat_bin_dir = '/usr/lib/hive-hcatalog/sbin' # Starting from HDP2.3 drop should be executed with purge suffix purge_tables = "false" @@ -315,10 +310,7 @@ mysql_adduser_path = format("{tmp_dir}/addMysqlUser.sh") mysql_deluser_path = format("{tmp_dir}/removeMysqlUser.sh") ######## Metastore Schema -init_metastore_schema = False -if Script.is_hdp_stack_greater_or_equal("2.1"): - init_metastore_schema = True - +init_metastore_schema = True ########## HCAT hcat_dbroot = hcat_lib http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py index d27d10a..fd9a462 100644 --- a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py +++ b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py @@ -183,7 +183,7 @@ if https_port is not None: hdfs_site = config['configurations']['hdfs-site'] fs_root = config['configurations']['core-site']['fs.defaultFS'] -if Script.is_hdp_stack_greater_or_equal("2.0") and Script.is_hdp_stack_less_than("2.2"): +if Script.is_hdp_stack_less_than("2.2"): put_shared_lib_to_hdfs_cmd = format("hadoop --config {hadoop_conf_dir} dfs -put {oozie_shared_lib} {oozie_hdfs_user_dir}") # for newer else: http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py index 4aa5f8a..d7b18eb 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py @@ -62,11 +62,7 @@ hdfs_log_dir_prefix = config['configurations']['hadoop-env']['hdfs_log_dir_prefi hadoop_pid_dir_prefix = config['configurations']['hadoop-env']['hadoop_pid_dir_prefix'] hadoop_root_logger = config['configurations']['hadoop-env']['hadoop_root_logger'] -if Script.is_hdp_stack_greater_or_equal("2.0") and Script.is_hdp_stack_less_than("2.1") and not OSCheck.is_suse_family(): - # deprecated rhel jsvc_path - jsvc_path = "/usr/libexec/bigtop-utils" -else: - jsvc_path = "/usr/lib/bigtop-utils" +jsvc_path = "/usr/lib/bigtop-utils" hadoop_heapsize = config['configurations']['hadoop-env']['hadoop_heapsize'] namenode_heapsize = config['configurations']['hadoop-env']['namenode_heapsize'] http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py index 2593ee4..e1f33d4 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py @@ -132,11 +132,7 @@ hdfs_log_dir_prefix = config['configurations']['hadoop-env']['hdfs_log_dir_prefi hadoop_pid_dir_prefix = config['configurations']['hadoop-env']['hadoop_pid_dir_prefix'] hadoop_root_logger = config['configurations']['hadoop-env']['hadoop_root_logger'] -if hdp_stack_version != "" and compare_versions(hdp_stack_version, '2.0') >= 0 and compare_versions(hdp_stack_version, '2.1') < 0 and not OSCheck.is_suse_family(): - # deprecated rhel jsvc_path - jsvc_path = "/usr/libexec/bigtop-utils" -else: - jsvc_path = "/usr/lib/bigtop-utils" +jsvc_path = "/usr/lib/bigtop-utils" hadoop_heapsize = config['configurations']['hadoop-env']['hadoop_heapsize'] namenode_heapsize = config['configurations']['hadoop-env']['namenode_heapsize'] http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py index e771dd5..70c69fe 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py @@ -144,11 +144,7 @@ else: #hadoop-env.sh java_home = config['hostLevelParams']['java_home'] -if hdp_stack_version != "" and compare_versions(hdp_stack_version, '2.0') >= 0 and compare_versions(hdp_stack_version, '2.1') < 0 and not OSCheck.is_suse_family(): - # deprecated rhel jsvc_path - jsvc_path = "/usr/libexec/bigtop-utils" -else: - jsvc_path = "/usr/lib/bigtop-utils" +jsvc_path = "/usr/lib/bigtop-utils" hadoop_heapsize = config['configurations']['hadoop-env']['hadoop_heapsize'] namenode_heapsize = config['configurations']['hadoop-env']['namenode_heapsize'] http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml b/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml index 0d4c27f..0659ca4 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml +++ b/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml @@ -17,7 +17,7 @@ --> <metainfo> <versions> - <active>true</active> + <active>false</active> </versions> <extends>2.0.6</extends> <minJdk>1.7</minJdk> http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py index 6c9752a..e709693 100644 --- a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py +++ b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py @@ -37,7 +37,7 @@ class TestHcatClient(RMFTestCase): group = 'hadoop', recursive = True, ) - self.assertResourceCalled('Directory', '/etc/hcatalog/conf', + self.assertResourceCalled('Directory', '/etc/hive-hcatalog/conf', owner = 'hcat', group = 'hadoop', recursive = True, @@ -54,7 +54,7 @@ class TestHcatClient(RMFTestCase): configurations = self.getConfig()['configurations']['hive-site'], configuration_attributes = self.getConfig()['configuration_attributes']['hive-site'] ) - self.assertResourceCalled('File', '/etc/hcatalog/conf/hcat-env.sh', + self.assertResourceCalled('File', '/etc/hive-hcatalog/conf/hcat-env.sh', content = InlineTemplate(self.getConfig()['configurations']['hcat-env']['content']), owner = 'hcat', group = 'hadoop', @@ -76,7 +76,7 @@ class TestHcatClient(RMFTestCase): owner = 'hcat', group = 'hadoop', ) - self.assertResourceCalled('Directory', '/etc/hcatalog/conf', + self.assertResourceCalled('Directory', '/etc/hive-hcatalog/conf', recursive = True, owner = 'hcat', group = 'hadoop', @@ -93,7 +93,7 @@ class TestHcatClient(RMFTestCase): configurations = self.getConfig()['configurations']['hive-site'], configuration_attributes = self.getConfig()['configuration_attributes']['hive-site'] ) - self.assertResourceCalled('File', '/etc/hcatalog/conf/hcat-env.sh', + self.assertResourceCalled('File', '/etc/hive-hcatalog/conf/hcat-env.sh', content = InlineTemplate(self.getConfig()['configurations']['hcat-env']['content']), owner = 'hcat', group = 'hadoop', http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py index 328b52d..7f3efcb 100644 --- a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py +++ b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py @@ -227,6 +227,10 @@ class TestHiveMetastore(RMFTestCase): content = StaticFile('startMetastore.sh'), mode = 0755, ) + self.assertResourceCalled('Execute', 'export HIVE_CONF_DIR=/etc/hive/conf.server ; /usr/lib/hive/bin/schematool -initSchema -dbType mysql -userName hive -passWord \'!`"\'"\'"\' 1\'', + not_if = 'ambari-sudo.sh su hive -l -s /bin/bash -c \'[RMF_EXPORT_PLACEHOLDER]export HIVE_CONF_DIR=/etc/hive/conf.server ; /usr/lib/hive/bin/schematool -info -dbType mysql -userName hive -passWord \'"\'"\'!`"\'"\'"\'"\'"\'"\'"\'"\'"\' 1\'"\'"\'\'', + user = 'hive', + ) self.assertResourceCalled('Directory', '/var/run/hive', owner = 'hive', group = 'hadoop', @@ -331,6 +335,10 @@ class TestHiveMetastore(RMFTestCase): content = StaticFile('startMetastore.sh'), mode = 0755, ) + self.assertResourceCalled('Execute', 'export HIVE_CONF_DIR=/etc/hive/conf.server ; /usr/lib/hive/bin/schematool -initSchema -dbType mysql -userName hive -passWord \'!`"\'"\'"\' 1\'', + not_if = 'ambari-sudo.sh su hive -l -s /bin/bash -c \'[RMF_EXPORT_PLACEHOLDER]export HIVE_CONF_DIR=/etc/hive/conf.server ; /usr/lib/hive/bin/schematool -info -dbType mysql -userName hive -passWord \'"\'"\'!`"\'"\'"\'"\'"\'"\'"\'"\'"\' 1\'"\'"\'\'', + user = 'hive', + ) self.assertResourceCalled('Directory', '/var/run/hive', owner = 'hive', group = 'hadoop', http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py index c11c2d0..b0b34bc 100644 --- a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py +++ b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py @@ -48,7 +48,7 @@ class TestWebHCatServer(RMFTestCase): ) self.assert_configure_default() - self.assertResourceCalled('Execute', 'cd /var/run/webhcat ; /usr/lib/hcatalog/sbin/webhcat_server.sh start', + self.assertResourceCalled('Execute', 'cd /var/run/webhcat ; /usr/lib/hive-hcatalog/sbin/webhcat_server.sh start', environment = {'HADOOP_HOME': '/usr'}, not_if = "ambari-sudo.sh su hcat -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps -p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1'", user = 'hcat', @@ -64,7 +64,7 @@ class TestWebHCatServer(RMFTestCase): target = RMFTestCase.TARGET_COMMON_SERVICES ) - self.assertResourceCalled('Execute', '/usr/lib/hcatalog/sbin/webhcat_server.sh stop', + self.assertResourceCalled('Execute', '/usr/lib/hive-hcatalog/sbin/webhcat_server.sh stop', user = 'hcat', environment = {'HADOOP_HOME': '/usr' } ) @@ -105,7 +105,7 @@ class TestWebHCatServer(RMFTestCase): ) self.assert_configure_secured() - self.assertResourceCalled('Execute', 'cd /var/run/webhcat ; /usr/lib/hcatalog/sbin/webhcat_server.sh start', + self.assertResourceCalled('Execute', 'cd /var/run/webhcat ; /usr/lib/hive-hcatalog/sbin/webhcat_server.sh start', environment = {'HADOOP_HOME': '/usr'}, not_if = "ambari-sudo.sh su hcat -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps -p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1'", user = 'hcat', @@ -121,7 +121,7 @@ class TestWebHCatServer(RMFTestCase): target = RMFTestCase.TARGET_COMMON_SERVICES ) - self.assertResourceCalled('Execute', '/usr/lib/hcatalog/sbin/webhcat_server.sh stop', + self.assertResourceCalled('Execute', '/usr/lib/hive-hcatalog/sbin/webhcat_server.sh stop', user = 'hcat', environment = {'HADOOP_HOME': '/usr' } ) @@ -152,7 +152,7 @@ class TestWebHCatServer(RMFTestCase): recursive = True, mode = 0755, ) - self.assertResourceCalled('Directory', '/etc/hcatalog/conf', + self.assertResourceCalled('Directory', '/etc/hive-webhcat/conf', owner = 'hcat', group = 'hadoop', recursive = True, @@ -161,11 +161,11 @@ class TestWebHCatServer(RMFTestCase): self.assertResourceCalled('XmlConfig', 'webhcat-site.xml', owner = 'hcat', group = 'hadoop', - conf_dir = '/etc/hcatalog/conf', + conf_dir = '/etc/hive-webhcat/conf', configurations = self.getConfig()['configurations']['webhcat-site'], configuration_attributes = self.getConfig()['configuration_attributes']['webhcat-site'] ) - self.assertResourceCalled('File', '/etc/hcatalog/conf/webhcat-env.sh', + self.assertResourceCalled('File', '/etc/hive-webhcat/conf/webhcat-env.sh', content = InlineTemplate(self.getConfig()['configurations']['webhcat-env']['content']), owner = 'hcat', group = 'hadoop', @@ -174,7 +174,7 @@ class TestWebHCatServer(RMFTestCase): cd_access = 'a', recursive=True ) - self.assertResourceCalled('File', '/etc/hcatalog/conf/webhcat-log4j.properties', + self.assertResourceCalled('File', '/etc/hive-webhcat/conf/webhcat-log4j.properties', content = 'log4jproperties\nline2', owner = 'hcat', group = 'hadoop', @@ -194,7 +194,7 @@ class TestWebHCatServer(RMFTestCase): recursive = True, mode = 0755, ) - self.assertResourceCalled('Directory', '/etc/hcatalog/conf', + self.assertResourceCalled('Directory', '/etc/hive-webhcat/conf', owner = 'hcat', group = 'hadoop', recursive = True, @@ -207,11 +207,11 @@ class TestWebHCatServer(RMFTestCase): self.assertResourceCalled('XmlConfig', 'webhcat-site.xml', owner = 'hcat', group = 'hadoop', - conf_dir = '/etc/hcatalog/conf', + conf_dir = '/etc/hive-webhcat/conf', configurations = self.getConfig()['configurations']['webhcat-site'], configuration_attributes = self.getConfig()['configuration_attributes']['webhcat-site'] ) - self.assertResourceCalled('File', '/etc/hcatalog/conf/webhcat-env.sh', + self.assertResourceCalled('File', '/etc/hive-webhcat/conf/webhcat-env.sh', content = InlineTemplate(self.getConfig()['configurations']['webhcat-env']['content']), owner = 'hcat', group = 'hadoop', @@ -220,7 +220,7 @@ class TestWebHCatServer(RMFTestCase): cd_access = 'a', recursive=True ) - self.assertResourceCalled('File', '/etc/hcatalog/conf/webhcat-log4j.properties', + self.assertResourceCalled('File', '/etc/hive-webhcat/conf/webhcat-log4j.properties', content = 'log4jproperties\nline2', owner = 'hcat', group = 'hadoop',