AMBARI-15329: Code Cleanup: Remove hdp hardcodings in functions, variables etc. (jluniya)
Project: http://git-wip-us.apache.org/repos/asf/ambari/repo Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/456b4511 Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/456b4511 Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/456b4511 Branch: refs/heads/AMBARI-13364 Commit: 456b451180b427e89f52a5dc0a644e27fe346efb Parents: f5bd058 Author: Jayush Luniya <jlun...@hortonworks.com> Authored: Wed Mar 9 10:09:45 2016 -0800 Committer: Jayush Luniya <jlun...@hortonworks.com> Committed: Wed Mar 9 10:09:45 2016 -0800 ---------------------------------------------------------------------- .../src/main/python/amc_service.py | 2 +- .../src/main/python/main.py | 4 +- .../0.5.0.2.1/package/scripts/params_windows.py | 2 +- .../0.5.0.2.1/package/scripts/service_check.py | 2 +- .../1.4.0.2.0/package/scripts/flume_check.py | 2 +- .../1.4.0.2.0/package/scripts/params_windows.py | 2 +- .../package/scripts/params_windows.py | 2 +- .../0.96.0.2.0/package/scripts/service_check.py | 2 +- .../package/scripts/setup_ranger_hbase.py | 6 +- .../HBASE/0.96.0.2.0/package/scripts/upgrade.py | 4 +- .../package/scripts/setup_ranger_hdfs.py | 10 +-- .../2.1.0.2.0/package/scripts/status_params.py | 12 ++-- .../package/scripts/hcat_service_check.py | 2 +- .../package/scripts/hive_server_upgrade.py | 6 +- .../package/scripts/params_windows.py | 4 +- .../0.12.0.2.0/package/scripts/service_check.py | 2 +- .../package/scripts/setup_ranger_hive.py | 6 +- .../package/scripts/webhcat_service_check.py | 2 +- .../KAFKA/0.8.1.2.2/package/scripts/params.py | 4 +- .../0.5.0.2.2/package/scripts/params_windows.py | 4 +- .../package/scripts/setup_ranger_knox.py | 6 +- .../4.0.0.2.0/package/scripts/params_windows.py | 2 +- .../4.0.0.2.0/package/scripts/service_check.py | 2 +- .../package/scripts/params_windows.py | 4 +- .../0.12.0.2.0/package/scripts/service_check.py | 2 +- .../RANGER/0.4.0/package/scripts/upgrade.py | 6 +- .../0.5.0.2.3/package/scripts/kms_server.py | 4 +- .../0.5.0.2.3/package/scripts/upgrade.py | 6 +- .../package/scripts/params_windows.py | 4 +- .../0.60.0.2.2/package/scripts/service_check.py | 2 +- .../1.4.4.2.0/package/scripts/params_windows.py | 2 +- .../1.4.4.2.0/package/scripts/service_check.py | 2 +- .../0.9.1.2.1/package/scripts/params_windows.py | 2 +- .../0.9.1.2.1/package/scripts/service_check.py | 2 +- .../package/scripts/setup_ranger_storm.py | 6 +- .../0.4.0.2.1/package/scripts/params_windows.py | 4 +- .../0.4.0.2.1/package/scripts/service_check.py | 2 +- .../3.4.5.2.0/package/scripts/params_windows.py | 4 +- .../3.4.5.2.0/package/scripts/service_check.py | 2 +- .../resources/host_scripts/alert_disk_space.py | 20 +++--- .../0.8/hooks/after-INSTALL/scripts/hook.py | 2 +- .../0.8/hooks/after-INSTALL/scripts/params.py | 2 +- .../scripts/shared_initialization.py | 2 +- .../0.8/hooks/before-ANY/scripts/params.py | 2 +- .../HDFS/package/scripts/status_params.py | 12 ++-- .../2.0.6/hooks/after-INSTALL/scripts/hook.py | 4 +- .../2.0.6/hooks/after-INSTALL/scripts/params.py | 2 +- .../scripts/shared_initialization.py | 2 +- .../2.0.6/hooks/before-ANY/scripts/params.py | 2 +- .../2.1/hooks/after-INSTALL/scripts/params.py | 4 +- .../src/test/python/TestVersionSelectUtil.py | 8 +-- .../custom_actions/TestInstallPackages.py | 72 ++++++++++---------- 52 files changed, 139 insertions(+), 139 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-metrics/ambari-metrics-timelineservice/src/main/python/amc_service.py ---------------------------------------------------------------------- diff --git a/ambari-metrics/ambari-metrics-timelineservice/src/main/python/amc_service.py b/ambari-metrics/ambari-metrics-timelineservice/src/main/python/amc_service.py index f616fd0..a19fd11 100644 --- a/ambari-metrics/ambari-metrics-timelineservice/src/main/python/amc_service.py +++ b/ambari-metrics/ambari-metrics-timelineservice/src/main/python/amc_service.py @@ -165,7 +165,7 @@ def init_service_debug(options): sys.frozen = 'windows_exe' # Fake py2exe so we can debug -def ensure_hdp_service_soft_dependencies(): +def ensure_hadoop_service_soft_dependencies(): if SERVICE_STATUS_RUNNING != WinServiceController.QueryStatus(EMBEDDED_HBASE_MASTER_SERVICE): err = 'ERROR: Service "{0}" was not started.'.format(EMBEDDED_HBASE_MASTER_SERVICE) raise FatalException(1, err) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-metrics/ambari-metrics-timelineservice/src/main/python/main.py ---------------------------------------------------------------------- diff --git a/ambari-metrics/ambari-metrics-timelineservice/src/main/python/main.py b/ambari-metrics/ambari-metrics-timelineservice/src/main/python/main.py index 172861e..b6b4e0b 100644 --- a/ambari-metrics/ambari-metrics-timelineservice/src/main/python/main.py +++ b/ambari-metrics/ambari-metrics-timelineservice/src/main/python/main.py @@ -109,8 +109,8 @@ def server_process_main(options, scmStatus=None): #Ensure the 3 Hadoop services required are started on the local machine if not options.no_embedded_hbase: - from amc_service import ensure_hdp_service_soft_dependencies - ensure_hdp_service_soft_dependencies() + from amc_service import ensure_hadoop_service_soft_dependencies + ensure_hadoop_service_soft_dependencies() if scmStatus is not None: scmStatus.reportStartPending() http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/params_windows.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/params_windows.py index 9136957..b213028 100644 --- a/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/params_windows.py +++ b/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/params_windows.py @@ -35,7 +35,7 @@ falcon_home = None falcon_log_dir = "." if os.environ.has_key("HADOOP_HOME"): - hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], "..")) + stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], "..")) if os.environ.has_key("FALCON_CONF_DIR"): falcon_conf_dir = os.environ["FALCON_CONF_DIR"] http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/service_check.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/service_check.py index e633dcb..473b7c3 100644 --- a/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/service_check.py +++ b/ambari-server/src/main/resources/common-services/FALCON/0.5.0.2.1/package/scripts/service_check.py @@ -45,7 +45,7 @@ class FalconServiceCheckWindows(FalconServiceCheck): def service_check(self, env): import params env.set_params(params) - smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd") + smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd") service = "FALCON" Execute(format("cmd /C {smoke_cmd} {service}"), user=params.falcon_user, logoutput=True, tries = 3, try_sleep = 20) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_check.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_check.py b/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_check.py index ae166f8..25d9a08 100644 --- a/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_check.py +++ b/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/flume_check.py @@ -27,7 +27,7 @@ class FlumeServiceCheck(Script): def service_check(self, env): import params env.set_params(params) - smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd") + smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd") service = "FLUME" Execute(format("cmd /C {smoke_cmd} {service}"), logoutput=True, user=params.hdfs_user) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/params_windows.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/params_windows.py index 17177ec..66e9852 100644 --- a/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/params_windows.py +++ b/ambari-server/src/main/resources/common-services/FLUME/1.4.0.2.0/package/scripts/params_windows.py @@ -26,7 +26,7 @@ config = Script.get_config() hadoop_user = config["configurations"]["cluster-env"]["hadoop.user.name"] -hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],"..")) +stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],"..")) flume_home = os.environ['FLUME_HOME'] flume_conf_dir = os.path.join(flume_home, 'conf') flume_user = hadoop_user http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/params_windows.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/params_windows.py index c132503..7d634cb 100644 --- a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/params_windows.py +++ b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/params_windows.py @@ -27,7 +27,7 @@ config = Script.get_config() hbase_conf_dir = os.environ["HBASE_CONF_DIR"] hbase_bin_dir = os.path.join(os.environ["HBASE_HOME"],'bin') hbase_executable = os.path.join(hbase_bin_dir,"hbase.cmd") -hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],"..")) +stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],"..")) hadoop_user = config["configurations"]["cluster-env"]["hadoop.user.name"] hbase_user = hadoop_user http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/service_check.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/service_check.py index 97cdd32..6ecb58e 100644 --- a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/service_check.py +++ b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/service_check.py @@ -34,7 +34,7 @@ class HbaseServiceCheckWindows(HbaseServiceCheck): def service_check(self, env): import params env.set_params(params) - smoke_cmd = os.path.join(params.hdp_root, "Run-SmokeTests.cmd") + smoke_cmd = os.path.join(params.stack_root, "Run-SmokeTests.cmd") service = "HBASE" Execute(format("cmd /C {smoke_cmd} {service}"), user=params.hbase_user, logoutput=True) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/setup_ranger_hbase.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/setup_ranger_hbase.py b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/setup_ranger_hbase.py index 5c68583..ffd0715 100644 --- a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/setup_ranger_hbase.py +++ b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/setup_ranger_hbase.py @@ -29,10 +29,10 @@ def setup_ranger_hbase(upgrade_type=None): else: from resource_management.libraries.functions.setup_ranger_plugin import setup_ranger_plugin - hdp_version = None + stack_version = None if upgrade_type is not None: - hdp_version = params.version + stack_version = params.version if params.retryAble: Logger.info("HBase: Setup ranger: command retry enables thus retrying if ranger admin is down !") @@ -80,6 +80,6 @@ def setup_ranger_hbase(upgrade_type=None): component_list=['hbase-client', 'hbase-master', 'hbase-regionserver'], audit_db_is_enabled=params.xa_audit_db_is_enabled, credential_file=params.credential_file, xa_audit_db_password=params.xa_audit_db_password, ssl_truststore_password=params.ssl_truststore_password, ssl_keystore_password=params.ssl_keystore_password, - hdp_version_override = hdp_version, skip_if_rangeradmin_down= not params.retryAble) + stack_version_override = stack_version, skip_if_rangeradmin_down= not params.retryAble) else: Logger.info('Ranger admin not installed') http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py index c5ba682..92c0f70 100644 --- a/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py +++ b/ambari-server/src/main/resources/common-services/HBASE/0.96.0.2.0/package/scripts/upgrade.py @@ -26,12 +26,12 @@ from resource_management.libraries.functions import stack_select from resource_management.libraries.functions.version import compare_versions, format_stack_version from resource_management.libraries.functions.decorator import retry -def prestart(env, hdp_component): +def prestart(env, stack_component): import params if params.version and compare_versions(format_stack_version(params.version), '2.2.0.0') >= 0: conf_select.select(params.stack_name, "hbase", params.version) - stack_select.select(hdp_component, params.version) + stack_select.select(stack_component, params.version) def post_regionserver(env): import params http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/setup_ranger_hdfs.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/setup_ranger_hdfs.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/setup_ranger_hdfs.py index ff93e39..209ac91 100644 --- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/setup_ranger_hdfs.py +++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/setup_ranger_hdfs.py @@ -34,10 +34,10 @@ def setup_ranger_hdfs(upgrade_type=None): else: from resource_management.libraries.functions.setup_ranger_plugin import setup_ranger_plugin - hdp_version = None + stack_version = None if upgrade_type is not None: - hdp_version = params.version + stack_version = params.version if params.retryAble: Logger.info("HDFS: Setup ranger: command retry enables thus retrying if ranger admin is down !") @@ -58,11 +58,11 @@ def setup_ranger_hdfs(upgrade_type=None): component_list=['hadoop-client'], audit_db_is_enabled=params.xa_audit_db_is_enabled, credential_file=params.credential_file, xa_audit_db_password=params.xa_audit_db_password, ssl_truststore_password=params.ssl_truststore_password, ssl_keystore_password=params.ssl_keystore_password, - hdp_version_override = hdp_version, skip_if_rangeradmin_down= not params.retryAble) + stack_version_override = stack_version, skip_if_rangeradmin_down= not params.retryAble) - if hdp_version and params.upgrade_direction == Direction.UPGRADE: + if stack_version and params.upgrade_direction == Direction.UPGRADE: # when upgrading to 2.3+, this env file must be removed - if compare_versions(hdp_version, '2.3', format=True) > 0: + if compare_versions(stack_version, '2.3', format=True) > 0: source_file = os.path.join(params.hadoop_conf_dir, 'set-hdfs-plugin-env.sh') target_file = source_file + ".bak" Execute(("mv", source_file, target_file), sudo=True, only_if=format("test -f {source_file}")) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/status_params.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/status_params.py b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/status_params.py index 388fa59..cdb683b 100644 --- a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/status_params.py +++ b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/status_params.py @@ -36,12 +36,12 @@ if OSCheck.is_windows_family(): else: hadoop_pid_dir_prefix = config['configurations']['hadoop-env']['hadoop_pid_dir_prefix'] hdfs_user = config['configurations']['hadoop-env']['hdfs_user'] - hdp_pid_dir = format("{hadoop_pid_dir_prefix}/{hdfs_user}") - datanode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-datanode.pid") - namenode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-namenode.pid") - snamenode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-secondarynamenode.pid") - journalnode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-journalnode.pid") - zkfc_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-zkfc.pid") + hadoop_pid_dir = format("{hadoop_pid_dir_prefix}/{hdfs_user}") + datanode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-datanode.pid") + namenode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-namenode.pid") + snamenode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-secondarynamenode.pid") + journalnode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-journalnode.pid") + zkfc_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-zkfc.pid") nfsgateway_pid_file = format("{hadoop_pid_dir_prefix}/root/hadoop_privileged_nfs3.pid") # Security related/required params http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_service_check.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_service_check.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_service_check.py index 27ff29a..4153821 100644 --- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_service_check.py +++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hcat_service_check.py @@ -27,7 +27,7 @@ from ambari_commons import OSConst @OsFamilyFuncImpl(os_family=OSConst.WINSRV_FAMILY) def hcat_service_check(): import params - smoke_cmd = os.path.join(params.hdp_root, "Run-SmokeTests.cmd") + smoke_cmd = os.path.join(params.stack_root, "Run-SmokeTests.cmd") service = "HCatalog" Execute(format("cmd /C {smoke_cmd} {service}"), user=params.hcat_user, logoutput=True) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_upgrade.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_upgrade.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_upgrade.py index 664cafa..c3d15e5 100644 --- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_upgrade.py +++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/hive_server_upgrade.py @@ -121,7 +121,7 @@ def _get_current_hiveserver_version(): if formatted_source_version and compare_versions(formatted_source_version, "2.2") >= 0: version_hive_bin = format('/usr/hdp/{source_version}/hive/bin') command = format('{version_hive_bin}/hive --version') - return_code, hdp_output = shell.call(command, user=params.hive_user, path=hive_execute_path) + return_code, output = shell.call(command, user=params.hive_user, path=hive_execute_path) except Exception, e: Logger.error(str(e)) raise Fail('Unable to execute hive --version command to retrieve the hiveserver2 version.') @@ -129,12 +129,12 @@ def _get_current_hiveserver_version(): if return_code != 0: raise Fail('Unable to determine the current HiveServer2 version because of a non-zero return code of {0}'.format(str(return_code))) - match = re.search('^(Hive) ([0-9]+.[0-9]+.\S+)', hdp_output, re.MULTILINE) + match = re.search('^(Hive) ([0-9]+.[0-9]+.\S+)', output, re.MULTILINE) if match: current_hive_server_version = match.group(2) return current_hive_server_version else: - raise Fail('The extracted hiveserver2 version "{0}" does not matching any known pattern'.format(hdp_output)) + raise Fail('The extracted hiveserver2 version "{0}" does not matching any known pattern'.format(output)) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_windows.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_windows.py index 7c21b5f..5610019 100644 --- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_windows.py +++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_windows.py @@ -28,7 +28,7 @@ config = Script.get_config() stack_version_unformatted = str(config['hostLevelParams']['stack_version']) stack_version_formatted = format_stack_version(stack_version_unformatted) -hdp_root = None +stack_root = None hive_conf_dir = None hive_home = None hive_lib_dir = None @@ -39,7 +39,7 @@ hcat_config_dir = None hive_bin = None try: - hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],"..")) + stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],"..")) hive_conf_dir = os.environ["HIVE_CONF_DIR"] hive_home = os.environ["HIVE_HOME"] hive_lib_dir = os.environ["HIVE_LIB_DIR"] http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py index 0c254be..251e71f 100644 --- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py +++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/service_check.py @@ -37,7 +37,7 @@ class HiveServiceCheckWindows(HiveServiceCheck): def service_check(self, env): import params env.set_params(params) - smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd") + smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd") service = "HIVE" Execute(format("cmd /C {smoke_cmd} {service}"), user=params.hive_user, logoutput=True) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/setup_ranger_hive.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/setup_ranger_hive.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/setup_ranger_hive.py index 8b2e4e4..f51dbab 100644 --- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/setup_ranger_hive.py +++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/setup_ranger_hive.py @@ -29,10 +29,10 @@ def setup_ranger_hive(upgrade_type = None): else: from resource_management.libraries.functions.setup_ranger_plugin import setup_ranger_plugin - hdp_version = None + stack_version = None if upgrade_type is not None: - hdp_version = params.version + stack_version = params.version if params.retryAble: Logger.info("Hive: Setup ranger: command retry enables thus retrying if ranger admin is down !") @@ -72,6 +72,6 @@ def setup_ranger_hive(upgrade_type = None): component_list=['hive-client', 'hive-metastore', 'hive-server2'], audit_db_is_enabled=params.xa_audit_db_is_enabled, credential_file=params.credential_file, xa_audit_db_password=params.xa_audit_db_password, ssl_truststore_password=params.ssl_truststore_password, ssl_keystore_password=params.ssl_keystore_password, - hdp_version_override = hdp_version, skip_if_rangeradmin_down= not params.retryAble) + stack_version_override = stack_version, skip_if_rangeradmin_down= not params.retryAble) else: Logger.info('Ranger admin not installed') http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_service_check.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_service_check.py b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_service_check.py index 99f52a5..9965bfa9 100644 --- a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_service_check.py +++ b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/webhcat_service_check.py @@ -32,7 +32,7 @@ def webhcat_service_check(): # AMBARI-11633 [WinTP2] Webhcat service check fails # Hive doesn't pass the environment variables correctly to child processes, which fails the smoke test. # Reducing the amount of URLs checked to the minimum required. - #smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd") + #smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd") #service = "WEBHCAT" #Execute(format("cmd /C {smoke_cmd} {service}"), user=params.hcat_user, logoutput=True) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/params.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/params.py b/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/params.py index ba6857e..4e73730 100644 --- a/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/params.py +++ b/ambari-server/src/main/resources/common-services/KAFKA/0.8.1.2.2/package/scripts/params.py @@ -248,8 +248,8 @@ if has_ranger_admin and is_supported_kafka_ranger: ssl_truststore_password = unicode(config['configurations']['ranger-kafka-policymgr-ssl']['xasecure.policymgr.clientssl.truststore.password']) if xml_configurations_supported else None credential_file = format('/etc/ranger/{repo_name}/cred.jceks') if xml_configurations_supported else None - hdp_version = get_stack_version('kafka-broker') - setup_ranger_env_sh_source = format('/usr/hdp/{hdp_version}/ranger-kafka-plugin/install/conf.templates/enable/kafka-ranger-env.sh') + stack_version = get_stack_version('kafka-broker') + setup_ranger_env_sh_source = format('/usr/hdp/{stack_version}/ranger-kafka-plugin/install/conf.templates/enable/kafka-ranger-env.sh') setup_ranger_env_sh_target = format("{conf_dir}/kafka-ranger-env.sh") #For SQLA explicitly disable audit to DB for Ranger http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/params_windows.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/params_windows.py index e044d9a..e7a3747 100644 --- a/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/params_windows.py +++ b/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/params_windows.py @@ -25,7 +25,7 @@ from status_params import * # server configurations config = Script.get_config() -hdp_root = None +stack_root = None knox_home = None knox_conf_dir = None knox_logs_dir = None @@ -38,7 +38,7 @@ knox_master_secret_path = None knox_cert_store_path = None try: - hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],"..")) + stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],"..")) knox_home = os.environ['KNOX_HOME'] knox_conf_dir = os.environ['KNOX_CONF_DIR'] knox_logs_dir = os.environ['KNOX_LOG_DIR'] http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/setup_ranger_knox.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/setup_ranger_knox.py b/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/setup_ranger_knox.py index 1a08d54..13987c8 100644 --- a/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/setup_ranger_knox.py +++ b/ambari-server/src/main/resources/common-services/KNOX/0.5.0.2.2/package/scripts/setup_ranger_knox.py @@ -29,9 +29,9 @@ def setup_ranger_knox(upgrade_type=None): else: from resource_management.libraries.functions.setup_ranger_plugin import setup_ranger_plugin - hdp_version = None + stack_version = None if upgrade_type is not None: - hdp_version = params.version + stack_version = params.version if params.retryAble: Logger.info("Knox: Setup ranger: command retry enables thus retrying if ranger admin is down !") @@ -72,6 +72,6 @@ def setup_ranger_knox(upgrade_type=None): component_list=['knox-server'], audit_db_is_enabled=params.xa_audit_db_is_enabled, credential_file=params.credential_file, xa_audit_db_password=params.xa_audit_db_password, ssl_truststore_password=params.ssl_truststore_password, ssl_keystore_password=params.ssl_keystore_password, - hdp_version_override = hdp_version, skip_if_rangeradmin_down= not params.retryAble) + stack_version_override = stack_version, skip_if_rangeradmin_down= not params.retryAble) else: Logger.info('Ranger admin not installed') \ No newline at end of file http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_windows.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_windows.py index ec404c4..1f939d4 100644 --- a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_windows.py +++ b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_windows.py @@ -24,7 +24,7 @@ from status_params import * config = Script.get_config() hadoop_user = config["configurations"]["cluster-env"]["hadoop.user.name"] -hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], "..")) +stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], "..")) oozie_root = os.environ['OOZIE_ROOT'] oozie_home = os.environ['OOZIE_HOME'] oozie_conf_dir = os.path.join(oozie_home,'conf') http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/service_check.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/service_check.py index f0cdb2c..8d14836 100644 --- a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/service_check.py +++ b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/service_check.py @@ -130,7 +130,7 @@ class OozieServiceCheckWindows(OozieServiceCheck): import params env.set_params(params) - smoke_cmd = os.path.join(params.hdp_root, "Run-SmokeTests.cmd") + smoke_cmd = os.path.join(params.stack_root, "Run-SmokeTests.cmd") service = "OOZIE" Execute(format("cmd /C {smoke_cmd} {service}"), logoutput=True) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/params_windows.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/params_windows.py index 8a0a519..9cb0ce7 100644 --- a/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/params_windows.py +++ b/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/params_windows.py @@ -23,11 +23,11 @@ from resource_management import * # server configurations config = Script.get_config() -hdp_root = None +stack_root = None pig_home = None pig_conf_dir = None try: - hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],"..")) + stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],"..")) pig_home = os.environ['PIG_HOME'] pig_conf_dir = os.path.join(pig_home,'conf') except: http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/service_check.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/service_check.py index 155e63c..ec2946e 100644 --- a/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/service_check.py +++ b/ambari-server/src/main/resources/common-services/PIG/0.12.0.2.0/package/scripts/service_check.py @@ -125,7 +125,7 @@ class PigServiceCheckWindows(PigServiceCheck): def service_check(self, env): import params env.set_params(params) - smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd") + smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd") service = "PIG" Execute(format("cmd /C {smoke_cmd} {service}", smoke_cmd=smoke_cmd, service=service), logoutput=True, user=params.pig_user, timeout=300) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py b/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py index ed8b690..aa75949 100644 --- a/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py +++ b/ambari-server/src/main/resources/common-services/RANGER/0.4.0/package/scripts/upgrade.py @@ -23,9 +23,9 @@ from resource_management.libraries.functions import conf_select from resource_management.libraries.functions import stack_select from resource_management.libraries.functions.format import format -def prestart(env, hdp_component): +def prestart(env, stack_component): import params if params.version and params.stack_is_hdp22_or_further: - conf_select.select(params.stack_name, hdp_component, params.version) - stack_select.select(hdp_component, params.version) + conf_select.select(params.stack_name, stack_component, params.version) + stack_select.select(stack_component, params.version) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py b/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py index 9450184..42f1cb9 100755 --- a/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py +++ b/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/kms_server.py @@ -17,7 +17,7 @@ See the License for the specific language governing permissions and limitations under the License. """ -from resource_management.libraries.functions import hdp_select +from resource_management.libraries.functions import stack_select from resource_management.libraries.script import Script from resource_management.core.resources.system import Execute from resource_management.core.exceptions import ComponentIsNotRunning @@ -82,7 +82,7 @@ class KmsServer(Script): import params env.set_params(params) - upgrade_stack = hdp_select._get_upgrade_stack() + upgrade_stack = stack_select._get_upgrade_stack() if upgrade_stack is None: raise Fail('Unable to determine the stack and stack version') http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py b/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py index 798e8f7..315a417 100644 --- a/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py +++ b/ambari-server/src/main/resources/common-services/RANGER_KMS/0.5.0.2.3/package/scripts/upgrade.py @@ -22,9 +22,9 @@ from resource_management.libraries.functions import conf_select from resource_management.libraries.functions import stack_select from resource_management.libraries.functions.format import format -def prestart(env, hdp_component): +def prestart(env, stack_component): import params if params.version and params.stack_is_hdp23_or_further: - conf_select.select(params.stack_name, hdp_component, params.version) - stack_select.select(hdp_component, params.version) + conf_select.select(params.stack_name, stack_component, params.version) + stack_select.select(stack_component, params.version) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/params_windows.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/params_windows.py index 366a1c9..fdb8160 100644 --- a/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/params_windows.py +++ b/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/params_windows.py @@ -24,13 +24,13 @@ import os # server configurations config = Script.get_config() -hdp_root = None +stack_root = None slider_home = None slider_bin_dir = None slider_conf_dir = None storm_slider_conf_dir = None try: - hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],"..")) + stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],"..")) slider_home = os.environ['SLIDER_HOME'] slider_bin_dir = os.path.join(slider_home, 'bin') slider_conf_dir = os.path.join(slider_home, 'conf') http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/service_check.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/service_check.py index b93b0eb..a16d0e7 100644 --- a/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/service_check.py +++ b/ambari-server/src/main/resources/common-services/SLIDER/0.60.0.2.2/package/scripts/service_check.py @@ -29,7 +29,7 @@ class SliderServiceCheck(Script): def service_check(self, env): import params env.set_params(params) - smoke_cmd = os.path.join(params.hdp_root, "Run-SmokeTests.cmd") + smoke_cmd = os.path.join(params.stack_root, "Run-SmokeTests.cmd") service = "SLIDER" Execute(format("cmd /C {smoke_cmd} {service}"), logoutput=True, user=params.hdfs_user) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/params_windows.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/params_windows.py index e4b71f1..f930765 100644 --- a/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/params_windows.py +++ b/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/params_windows.py @@ -24,7 +24,7 @@ config = Script.get_config() sqoop_user = "sqoop" -hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], "..")) +stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], "..")) sqoop_env_cmd_template = config['configurations']['sqoop-env']['content'] sqoop_home_dir = os.environ["SQOOP_HOME"] sqoop_conf_dir = os.path.join(sqoop_home_dir, "conf") \ No newline at end of file http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/service_check.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/service_check.py index 426315f..c0d0e8c 100644 --- a/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/service_check.py +++ b/ambari-server/src/main/resources/common-services/SQOOP/1.4.4.2.0/package/scripts/service_check.py @@ -53,7 +53,7 @@ class SqoopServiceCheckWindows(SqoopServiceCheck): def service_check(self, env): import params env.set_params(params) - smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd") + smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd") service = "SQOOP" Execute(format("cmd /C {smoke_cmd} {service}"), logoutput=True) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/params_windows.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/params_windows.py index 90bc76d..88e6246 100644 --- a/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/params_windows.py +++ b/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/params_windows.py @@ -27,7 +27,7 @@ config = Script.get_config() stack_is_hdp23_or_further = Script.is_stack_greater_or_equal("2.3") -hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],"..")) +stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"],"..")) conf_dir = os.environ["STORM_CONF_DIR"] hadoop_user = config["configurations"]["cluster-env"]["hadoop.user.name"] storm_user = hadoop_user http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/service_check.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/service_check.py index 99a6439..80ea0f5 100644 --- a/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/service_check.py +++ b/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/service_check.py @@ -38,7 +38,7 @@ class ServiceCheckWindows(ServiceCheck): def service_check(self, env): import params env.set_params(params) - smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd") + smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd") service = "STORM" Execute(format("cmd /C {smoke_cmd} {service}", smoke_cmd=smoke_cmd, service=service), user=params.storm_user, logoutput=True) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/setup_ranger_storm.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/setup_ranger_storm.py b/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/setup_ranger_storm.py index a76457f..bef1f02 100644 --- a/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/setup_ranger_storm.py +++ b/ambari-server/src/main/resources/common-services/STORM/0.9.1.2.1/package/scripts/setup_ranger_storm.py @@ -32,9 +32,9 @@ def setup_ranger_storm(upgrade_type=None): else: from resource_management.libraries.functions.setup_ranger_plugin import setup_ranger_plugin - hdp_version = None + stack_version = None if upgrade_type is not None: - hdp_version = params.version + stack_version = params.version if params.retryAble: Logger.info("Storm: Setup ranger: command retry enables thus retrying if ranger admin is down !") @@ -75,6 +75,6 @@ def setup_ranger_storm(upgrade_type=None): component_list=['storm-client', 'storm-nimbus'], audit_db_is_enabled=params.xa_audit_db_is_enabled, credential_file=params.credential_file, xa_audit_db_password=params.xa_audit_db_password, ssl_truststore_password=params.ssl_truststore_password, ssl_keystore_password=params.ssl_keystore_password, - hdp_version_override = hdp_version, skip_if_rangeradmin_down= not params.retryAble) + stack_version_override = stack_version, skip_if_rangeradmin_down= not params.retryAble) else: Logger.info('Ranger admin not installed') http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/params_windows.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/params_windows.py index ad80830..dd732f5 100644 --- a/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/params_windows.py +++ b/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/params_windows.py @@ -36,9 +36,9 @@ except KeyError: stack_version_formatted = "" -hdp_root = None +stack_root = None try: - hdp_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], "..")) + stack_root = os.path.abspath(os.path.join(os.environ["HADOOP_HOME"], "..")) except: pass http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/service_check.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/service_check.py index b70256c..c0c66af 100644 --- a/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/service_check.py +++ b/ambari-server/src/main/resources/common-services/TEZ/0.4.0.2.1/package/scripts/service_check.py @@ -100,7 +100,7 @@ class TezServiceCheckWindows(TezServiceCheck): def service_check(self, env): import params env.set_params(params) - smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd") + smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd") service = "TEZ" Execute(format("cmd /C {smoke_cmd} {service}"), logoutput=True, user=params.tez_user) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/params_windows.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/params_windows.py b/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/params_windows.py index 480fc8b..c36e152 100644 --- a/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/params_windows.py +++ b/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/params_windows.py @@ -26,11 +26,11 @@ import status_params config = Script.get_config() config_dir = None -hdp_root = None +stack_root = None try: # not used zookeeper_home_dir = os.environ["ZOOKEEPER_HOME"] config_dir = os.environ["ZOOKEEPER_CONF_DIR"] - hdp_root = os.environ["HADOOP_NODE_INSTALL_ROOT"] + stack_root = os.environ["HADOOP_NODE_INSTALL_ROOT"] except: pass http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/service_check.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/service_check.py b/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/service_check.py index b53e5f5..622a5eb 100644 --- a/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/service_check.py +++ b/ambari-server/src/main/resources/common-services/ZOOKEEPER/3.4.5.2.0/package/scripts/service_check.py @@ -65,7 +65,7 @@ class ZookeeperServiceCheckWindows(ZookeeperServiceCheck): import params env.set_params(params) - smoke_cmd = os.path.join(params.hdp_root,"Run-SmokeTests.cmd") + smoke_cmd = os.path.join(params.stack_root,"Run-SmokeTests.cmd") service = "Zookeeper" Execute(format("cmd /C {smoke_cmd} {service}"), user=params.zk_user, logoutput=True, tries=3, try_sleep=20) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/host_scripts/alert_disk_space.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/host_scripts/alert_disk_space.py b/ambari-server/src/main/resources/host_scripts/alert_disk_space.py index 94f19d6..b7a7038 100644 --- a/ambari-server/src/main/resources/host_scripts/alert_disk_space.py +++ b/ambari-server/src/main/resources/host_scripts/alert_disk_space.py @@ -37,10 +37,10 @@ PERCENT_USED_WARNING_DEFAULT = 50 PERCENT_USED_CRITICAL_DEFAULT = 80 # the location where HDP installs components when using HDP 2.2+ -HDP_HOME_DIR = "/usr/hdp" +STACK_HOME_DIR = "/usr/hdp" # the location where HDP installs components when using HDP 2.0 to 2.1 -HDP_HOME_LEGACY_DIR = "/usr/lib" +STACK_HOME_LEGACY_DIR = "/usr/lib" def get_tokens(): """ @@ -66,14 +66,14 @@ def execute(configurations={}, parameters={}, host_name=None): """ # determine the location of HDP home - hdp_home = None - if os.path.isdir(HDP_HOME_DIR): - hdp_home = HDP_HOME_DIR - elif os.path.isdir(HDP_HOME_LEGACY_DIR): - hdp_home = HDP_HOME_LEGACY_DIR - - # if hdp home was found, use it; otherwise default to None - path = hdp_home if hdp_home is not None else None + stack_home = None + if os.path.isdir(STACK_HOME_DIR): + stack_home = STACK_HOME_DIR + elif os.path.isdir(STACK_HOME_LEGACY_DIR): + stack_home = STACK_HOME_LEGACY_DIR + + # if stack home was found, use it; otherwise default to None + path = stack_home if stack_home is not None else None try: disk_usage = _get_disk_usage(path) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/hook.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/hook.py b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/hook.py index 71ac3df..6c2806d 100644 --- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/hook.py +++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/hook.py @@ -28,7 +28,7 @@ class AfterInstallHook(Hook): import params env.set_params(params) - setup_hdp_install_directory() + setup_install_directory() setup_config() if __name__ == "__main__": http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/params.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/params.py b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/params.py index f5851aa..dd66d58 100644 --- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/params.py +++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/params.py @@ -35,7 +35,7 @@ else: hadoop_conf_dir = "/etc/hadoop/conf" hadoop_conf_empty_dir = "/etc/hadoop/conf.empty" -versioned_hdp_root = '/usr/bigtop/current' +versioned_stack_root = '/usr/bigtop/current' #security params security_enabled = config['configurations']['cluster-env']['security_enabled'] #java params http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/shared_initialization.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/shared_initialization.py b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/shared_initialization.py index a930f54..8094129 100644 --- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/shared_initialization.py +++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/after-INSTALL/scripts/shared_initialization.py @@ -19,7 +19,7 @@ limitations under the License. import os from resource_management import * -def setup_hdp_install_directory(): +def setup_install_directory(): import params if params.rpm_version: Execute(format('ambari-python-wrap /usr/bin/bigtop-select set all `ambari-python-wrap /usr/bin/bigtop-select versions | grep ^{rpm_version}- | tail -1`'), http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/before-ANY/scripts/params.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/before-ANY/scripts/params.py b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/before-ANY/scripts/params.py index 7a93d41..97e560a 100644 --- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/before-ANY/scripts/params.py +++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/hooks/before-ANY/scripts/params.py @@ -47,7 +47,7 @@ else: hadoop_conf_dir = "/etc/hadoop/conf" hadoop_conf_empty_dir = "/etc/hadoop/conf.empty" -versioned_hdp_root = '/usr/bigtop/current' +versioned_stack_root = '/usr/bigtop/current' #security params security_enabled = config['configurations']['cluster-env']['security_enabled'] http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/HDFS/package/scripts/status_params.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/HDFS/package/scripts/status_params.py b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/HDFS/package/scripts/status_params.py index 0027a4c..c2adadb 100644 --- a/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/HDFS/package/scripts/status_params.py +++ b/ambari-server/src/main/resources/stacks/BIGTOP/0.8/services/HDFS/package/scripts/status_params.py @@ -23,9 +23,9 @@ config = Script.get_config() hadoop_pid_dir_prefix = config['configurations']['hadoop-env']['hadoop_pid_dir_prefix'] hdfs_user = config['configurations']['hadoop-env']['hdfs_user'] -hdp_pid_dir = format("{hadoop_pid_dir_prefix}/{hdfs_user}") -datanode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-datanode.pid") -namenode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-namenode.pid") -snamenode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-secondarynamenode.pid") -journalnode_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-journalnode.pid") -zkfc_pid_file = format("{hdp_pid_dir}/hadoop-{hdfs_user}-zkfc.pid") +hadoop_pid_dir = format("{hadoop_pid_dir_prefix}/{hdfs_user}") +datanode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-datanode.pid") +namenode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-namenode.pid") +snamenode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-secondarynamenode.pid") +journalnode_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-journalnode.pid") +zkfc_pid_file = format("{hadoop_pid_dir}/hadoop-{hdfs_user}-zkfc.pid") http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py index 73412e0..8a583b3 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/hook.py @@ -20,7 +20,7 @@ limitations under the License. from resource_management.libraries.script.hook import Hook from shared_initialization import link_configs from shared_initialization import setup_config -from shared_initialization import setup_hdp_symlinks +from shared_initialization import setup_stack_symlinks class AfterInstallHook(Hook): @@ -28,7 +28,7 @@ class AfterInstallHook(Hook): import params env.set_params(params) - setup_hdp_symlinks() + setup_stack_symlinks() setup_config() link_configs(self.stroutfile) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py index ab7dea5..cd04492 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py @@ -49,7 +49,7 @@ if Script.is_stack_greater_or_equal("2.2"): # not supported in HDP 2.2+ hadoop_conf_empty_dir = None -versioned_hdp_root = '/usr/hdp/current' +versioned_stack_root = '/usr/hdp/current' #security params security_enabled = config['configurations']['cluster-env']['security_enabled'] http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py index 8cf75cb..96dc104 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/shared_initialization.py @@ -28,7 +28,7 @@ from resource_management.libraries.resources.xml_config import XmlConfig from resource_management.libraries.script import Script -def setup_hdp_symlinks(): +def setup_stack_symlinks(): """ Invokes hdp-select set all against a calculated fully-qualified, "normalized" version based on a stack version, such as "2.3". This should always be called after a component has been http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py index 7e37873..e4cad39 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py @@ -103,7 +103,7 @@ hadoop_libexec_dir = stack_select.get_hadoop_dir("libexec", force_latest_on_upgr hadoop_conf_empty_dir = "/etc/hadoop/conf.empty" hadoop_secure_dn_user = hdfs_user hadoop_dir = "/etc/hadoop" -versioned_hdp_root = '/usr/hdp/current' +versioned_stack_root = '/usr/hdp/current' hadoop_java_io_tmpdir = os.path.join(tmp_dir, "hadoop_java_io_tmpdir") datanode_max_locked_memory = config['configurations']['hdfs-site']['dfs.datanode.max.locked.memory'] is_datanode_max_locked_memory_set = not is_empty(config['configurations']['hdfs-site']['dfs.datanode.max.locked.memory']) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/main/resources/stacks/HDPWIN/2.1/hooks/after-INSTALL/scripts/params.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDPWIN/2.1/hooks/after-INSTALL/scripts/params.py b/ambari-server/src/main/resources/stacks/HDPWIN/2.1/hooks/after-INSTALL/scripts/params.py index bfc605c..d541a15 100644 --- a/ambari-server/src/main/resources/stacks/HDPWIN/2.1/hooks/after-INSTALL/scripts/params.py +++ b/ambari-server/src/main/resources/stacks/HDPWIN/2.1/hooks/after-INSTALL/scripts/params.py @@ -39,8 +39,8 @@ if os.environ.has_key("HADOOP_NODE_INSTALL_ROOT"): hadoop_install_root = os.environ["HADOOP_NODE_INSTALL_ROOT"] -hdp_log_dir = "c:\\hadoop\\logs" -hdp_data_dir = "c:\\hadoop" +stack_log_dir = "c:\\hadoop\\logs" +stack_data_dir = "c:\\hadoop" db_flavor = "MSSQL" db_hostname = "localhost" db_port = "1433" http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/test/python/TestVersionSelectUtil.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/test/python/TestVersionSelectUtil.py b/ambari-server/src/test/python/TestVersionSelectUtil.py index f753023..9fa24cc 100644 --- a/ambari-server/src/test/python/TestVersionSelectUtil.py +++ b/ambari-server/src/test/python/TestVersionSelectUtil.py @@ -39,7 +39,7 @@ class TestVersionSelectUtil(TestCase): @patch('__builtin__.open') @patch("resource_management.core.shell.call") def test_get_component_version(self, call_mock, open_mock): - hdp_expected_version = "2.2.1.0-2175" + stack_expected_version = "2.2.1.0-2175" # Mock classes for reading from a file class MagicFile(object): @@ -49,7 +49,7 @@ class TestVersionSelectUtil(TestCase): "zookeeper-client" ]) def read(self, value): - return (value + " - " + hdp_expected_version) if value in self.allowed_names else ("ERROR: Invalid package - " + value) + return (value + " - " + stack_expected_version) if value in self.allowed_names else ("ERROR: Invalid package - " + value) def __exit__(self, exc_type, exc_val, exc_tb): pass @@ -88,6 +88,6 @@ class TestVersionSelectUtil(TestCase): # Pass version = self.module.get_component_version("HDP", "hadoop-hdfs-namenode") - self.assertEquals(version, hdp_expected_version) + self.assertEquals(version, stack_expected_version) version = self.module.get_component_version("HDP", "hadoop-hdfs-datanode") - self.assertEquals(version, hdp_expected_version) + self.assertEquals(version, stack_expected_version) http://git-wip-us.apache.org/repos/asf/ambari/blob/456b4511/ambari-server/src/test/python/custom_actions/TestInstallPackages.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/test/python/custom_actions/TestInstallPackages.py b/ambari-server/src/test/python/custom_actions/TestInstallPackages.py index 80ba480..f022c80 100644 --- a/ambari-server/src/test/python/custom_actions/TestInstallPackages.py +++ b/ambari-server/src/test/python/custom_actions/TestInstallPackages.py @@ -70,9 +70,9 @@ class TestInstallPackages(RMFTestCase): def test_normal_flow_rhel(self, write_actual_version_to_history_file_mock, read_actual_version_from_history_file_mock, - hdp_versions_mock, + stack_versions_mock, put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock): - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [], # before installation attempt [VERSION_STUB] ] @@ -129,10 +129,10 @@ class TestInstallPackages(RMFTestCase): @patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file") def test_normal_flow_sles(self, write_actual_version_to_history_file_mock, read_actual_version_from_history_file_mock, - hdp_versions_mock, put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock, is_suse_family_mock): + stack_versions_mock, put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock, is_suse_family_mock): is_suse_family_mock = True Script.stack_version_from_distro_select = VERSION_STUB - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [], # before installation attempt [VERSION_STUB] ] @@ -190,10 +190,10 @@ class TestInstallPackages(RMFTestCase): @patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file") def test_exclude_existing_repo(self, write_actual_version_to_history_file_mock, read_actual_version_from_history_file_mock, - hdp_versions_mock, + stack_versions_mock, allInstalledPackages_mock, put_structured_out_mock, is_redhat_family_mock, list_ambari_managed_repos_mock): - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [], # before installation attempt [VERSION_STUB] ] @@ -324,11 +324,11 @@ class TestInstallPackages(RMFTestCase): @patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file") def test_format_package_name(self, write_actual_version_to_history_file_mock, read_actual_version_from_history_file_mock, - hdp_versions_mock, + stack_versions_mock, allInstalledPackages_mock, put_structured_out_mock, package_mock, is_suse_family_mock): Script.stack_version_from_distro_select = VERSION_STUB - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [], # before installation attempt [VERSION_STUB] ] @@ -387,9 +387,9 @@ class TestInstallPackages(RMFTestCase): def test_version_reporting__build_number_defined(self, write_actual_version_to_history_file_mock, read_actual_version_from_history_file_mock, - hdp_versions_mock, + stack_versions_mock, put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock): - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [OLD_VERSION_STUB], # before installation attempt [OLD_VERSION_STUB, VERSION_STUB] ] @@ -419,12 +419,12 @@ class TestInstallPackages(RMFTestCase): self.assertTrue(write_actual_version_to_history_file_mock.called) self.assertEquals(write_actual_version_to_history_file_mock.call_args[0], (VERSION_STUB_WITHOUT_BUILD_NUMBER, VERSION_STUB)) - hdp_versions_mock.reset_mock() + stack_versions_mock.reset_mock() write_actual_version_to_history_file_mock.reset_mock() put_structured_out_mock.reset_mock() # Test retrying install again - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [OLD_VERSION_STUB, VERSION_STUB], [OLD_VERSION_STUB, VERSION_STUB] ] @@ -464,14 +464,14 @@ class TestInstallPackages(RMFTestCase): @patch("resource_management.libraries.functions.repo_version_history.read_actual_version_from_history_file") @patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file") @patch("os.path.exists") - def test_version_reporting__build_number_not_defined__usr_hdp_present__no_components_installed(self, + def test_version_reporting__build_number_not_defined_stack_root_present__no_components_installed(self, exists_mock, write_actual_version_to_history_file_mock, read_actual_version_from_history_file_mock, - hdp_versions_mock, + stack_versions_mock, put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock): exists_mock.return_value = True - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [], # before installation attempt [] ] @@ -508,7 +508,7 @@ class TestInstallPackages(RMFTestCase): self.assertFalse(write_actual_version_to_history_file_mock.called) - hdp_versions_mock.reset_mock() + stack_versions_mock.reset_mock() write_actual_version_to_history_file_mock.reset_mock() put_structured_out_mock.reset_mock() @@ -520,14 +520,14 @@ class TestInstallPackages(RMFTestCase): @patch("resource_management.libraries.functions.repo_version_history.read_actual_version_from_history_file") @patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file") @patch("os.path.exists") - def test_version_reporting__build_number_not_defined__usr_hdp_absent(self, + def test_version_reporting__build_number_not_defined_stack_root_absent(self, exists_mock, write_actual_version_to_history_file_mock, read_actual_version_from_history_file_mock, - hdp_versions_mock, + stack_versions_mock, put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock): exists_mock.return_value = False - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [], # before installation attempt [] ] @@ -562,13 +562,13 @@ class TestInstallPackages(RMFTestCase): self.assertFalse(write_actual_version_to_history_file_mock.called) - hdp_versions_mock.reset_mock() + stack_versions_mock.reset_mock() write_actual_version_to_history_file_mock.reset_mock() put_structured_out_mock.reset_mock() # Test retrying install again (correct build number, provided by other nodes, is now received from server) - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [], # before installation attempt [] ] @@ -611,12 +611,12 @@ class TestInstallPackages(RMFTestCase): @patch("resource_management.libraries.functions.stack_select.get_stack_versions") @patch("resource_management.libraries.functions.repo_version_history.read_actual_version_from_history_file") @patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file") - def test_version_reporting__build_number_not_defined__usr_hdp_present(self, + def test_version_reporting__build_number_not_defined_stack_root_present(self, write_actual_version_to_history_file_mock, read_actual_version_from_history_file_mock, - hdp_versions_mock, + stack_versions_mock, put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock): - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [OLD_VERSION_STUB], # before installation attempt [OLD_VERSION_STUB, VERSION_STUB] ] @@ -646,12 +646,12 @@ class TestInstallPackages(RMFTestCase): self.assertTrue(write_actual_version_to_history_file_mock.called) self.assertEquals(write_actual_version_to_history_file_mock.call_args[0], (VERSION_STUB_WITHOUT_BUILD_NUMBER, VERSION_STUB)) - hdp_versions_mock.reset_mock() + stack_versions_mock.reset_mock() write_actual_version_to_history_file_mock.reset_mock() put_structured_out_mock.reset_mock() # Test retrying install again - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [OLD_VERSION_STUB, VERSION_STUB], [OLD_VERSION_STUB, VERSION_STUB] ] @@ -689,12 +689,12 @@ class TestInstallPackages(RMFTestCase): @patch("resource_management.libraries.functions.stack_select.get_stack_versions") @patch("resource_management.libraries.functions.repo_version_history.read_actual_version_from_history_file") @patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file") - def test_version_reporting__wrong_build_number_specified__usr_hdp_present(self, + def test_version_reporting__wrong_build_number_specified_stack_root_present(self, write_actual_version_to_history_file_mock, read_actual_version_from_history_file_mock, - hdp_versions_mock, + stack_versions_mock, put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock): - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [OLD_VERSION_STUB], # before installation attempt [OLD_VERSION_STUB, VERSION_STUB] ] @@ -724,12 +724,12 @@ class TestInstallPackages(RMFTestCase): self.assertTrue(write_actual_version_to_history_file_mock.called) self.assertEquals(write_actual_version_to_history_file_mock.call_args[0], ('2.2.0.1', VERSION_STUB)) - hdp_versions_mock.reset_mock() + stack_versions_mock.reset_mock() write_actual_version_to_history_file_mock.reset_mock() put_structured_out_mock.reset_mock() # Test retrying install again - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [OLD_VERSION_STUB, VERSION_STUB], [OLD_VERSION_STUB, VERSION_STUB] ] @@ -768,14 +768,14 @@ class TestInstallPackages(RMFTestCase): @patch("resource_management.libraries.functions.repo_version_history.read_actual_version_from_history_file") @patch("resource_management.libraries.functions.repo_version_history.write_actual_version_to_history_file") @patch("os.path.exists") - def test_version_reporting__wrong_build_number_specified__usr_hdp_absent(self, + def test_version_reporting__wrong_build_number_specified_stack_root_absent(self, exists_mock, write_actual_version_to_history_file_mock, read_actual_version_from_history_file_mock, - hdp_versions_mock, + stack_versions_mock, put_structured_out_mock, allInstalledPackages_mock, list_ambari_managed_repos_mock): exists_mock.return_value = False - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [], # before installation attempt [] ] @@ -810,13 +810,13 @@ class TestInstallPackages(RMFTestCase): self.assertFalse(write_actual_version_to_history_file_mock.called) - hdp_versions_mock.reset_mock() + stack_versions_mock.reset_mock() write_actual_version_to_history_file_mock.reset_mock() put_structured_out_mock.reset_mock() # Test retrying install again (correct build number, provided by other nodes, is now received from server) - hdp_versions_mock.side_effect = [ + stack_versions_mock.side_effect = [ [], # before installation attempt [] ]