AMBARI-13454. Make HDP 2.0 stack inactive for Ambari 2.1.3 (aonishuk)

Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/e3958e2d
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/e3958e2d
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/e3958e2d

Branch: refs/heads/branch-dev-patch-upgrade
Commit: e3958e2da099f783ce637a57d752353fafd3f5b9
Parents: 8817569
Author: Andrew Onishuk <aonis...@hortonworks.com>
Authored: Fri Oct 16 17:25:05 2015 +0300
Committer: Andrew Onishuk <aonis...@hortonworks.com>
Committed: Fri Oct 16 17:25:05 2015 +0300

----------------------------------------------------------------------
 .../2.1.0.2.0/package/scripts/params_linux.py   |  7 +-----
 .../0.12.0.2.0/package/scripts/params_linux.py  | 20 +++++-----------
 .../4.0.0.2.0/package/scripts/params_linux.py   |  2 +-
 .../2.0.6/hooks/after-INSTALL/scripts/params.py |  6 +----
 .../2.0.6/hooks/before-ANY/scripts/params.py    |  6 +----
 .../2.0.6/hooks/before-START/scripts/params.py  |  6 +----
 .../main/resources/stacks/HDP/2.0/metainfo.xml  |  2 +-
 .../stacks/2.0.6/HIVE/test_hcat_client.py       |  8 +++----
 .../stacks/2.0.6/HIVE/test_hive_metastore.py    |  8 +++++++
 .../stacks/2.0.6/HIVE/test_webhcat_server.py    | 24 ++++++++++----------
 10 files changed, 36 insertions(+), 53 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py
 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py
index 38eac2e..4cbce34 100644
--- 
a/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py
+++ 
b/ambari-server/src/main/resources/common-services/HDFS/2.1.0.2.0/package/scripts/params_linux.py
@@ -106,12 +106,7 @@ hdfs_user_nofile_limit = 
default("/configurations/hadoop-env/hdfs_user_nofile_li
 hdfs_user_nproc_limit = 
default("/configurations/hadoop-env/hdfs_user_nproc_limit", "65536")
 
 create_lib_snappy_symlinks = not Script.is_hdp_stack_greater_or_equal("2.2")
-
-if Script.is_hdp_stack_greater_or_equal("2.0") and 
Script.is_hdp_stack_less_than("2.1") and not OSCheck.is_suse_family():
-  # deprecated rhel jsvc_path
-  jsvc_path = "/usr/libexec/bigtop-utils"
-else:
-  jsvc_path = "/usr/lib/bigtop-utils"
+jsvc_path = "/usr/lib/bigtop-utils"
 
 execute_path = os.environ['PATH'] + os.pathsep + hadoop_bin_dir
 ulimit_cmd = "ulimit -c unlimited ; "

http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py
 
b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py
index e16df78..14cf1ac 100644
--- 
a/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py
+++ 
b/ambari-server/src/main/resources/common-services/HIVE/0.12.0.2.0/package/scripts/params_linux.py
@@ -49,7 +49,7 @@ hostname = config["hostname"]
 # This is expected to be of the form #.#.#.#
 stack_version_unformatted = str(config['hostLevelParams']['stack_version'])
 hdp_stack_version_major = format_hdp_stack_version(stack_version_unformatted)
-stack_is_hdp21 = Script.is_hdp_stack_greater_or_equal("2.0") and 
Script.is_hdp_stack_less_than("2.2")
+stack_is_hdp21 = Script.is_hdp_stack_less_than("2.2")
 
 # this is not available on INSTALL action because hdp-select is not available
 hdp_stack_version = functions.get_hdp_version('hive-server2')
@@ -83,10 +83,6 @@ sqoop_tar_file = '/usr/share/HDP-webhcat/sqoop*.tar.gz'
 hive_specific_configs_supported = False
 hive_etc_dir_prefix = "/etc/hive"
 limits_conf_dir = "/etc/security/limits.d"
-hcat_conf_dir = '/etc/hcatalog/conf'
-config_dir = '/etc/hcatalog/conf'
-hcat_lib = '/usr/lib/hcatalog/share/hcatalog'
-webhcat_bin_dir = '/usr/lib/hcatalog/sbin'
 
 hive_user_nofile_limit = 
default("/configurations/hive-env/hive_user_nofile_limit", "32000")
 hive_user_nproc_limit = 
default("/configurations/hive-env/hive_user_nproc_limit", "16000")
@@ -101,11 +97,10 @@ hive_config_dir = status_params.hive_config_dir
 hive_client_conf_dir = status_params.hive_client_conf_dir
 hive_server_conf_dir = status_params.hive_server_conf_dir
 
-if Script.is_hdp_stack_greater_or_equal("2.1"):
-  hcat_conf_dir = '/etc/hive-hcatalog/conf'
-  config_dir = '/etc/hive-webhcat/conf'
-  hcat_lib = '/usr/lib/hive-hcatalog/share/hcatalog'
-  webhcat_bin_dir = '/usr/lib/hive-hcatalog/sbin'
+hcat_conf_dir = '/etc/hive-hcatalog/conf'
+config_dir = '/etc/hive-webhcat/conf'
+hcat_lib = '/usr/lib/hive-hcatalog/share/hcatalog'
+webhcat_bin_dir = '/usr/lib/hive-hcatalog/sbin'
 
 # Starting from HDP2.3 drop should be executed with purge suffix
 purge_tables = "false"
@@ -315,10 +310,7 @@ mysql_adduser_path = format("{tmp_dir}/addMysqlUser.sh")
 mysql_deluser_path = format("{tmp_dir}/removeMysqlUser.sh")
 
 ######## Metastore Schema
-init_metastore_schema = False
-if Script.is_hdp_stack_greater_or_equal("2.1"):
-  init_metastore_schema = True
-
+init_metastore_schema = True
 
 ########## HCAT
 hcat_dbroot = hcat_lib

http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py
 
b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py
index d27d10a..fd9a462 100644
--- 
a/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py
+++ 
b/ambari-server/src/main/resources/common-services/OOZIE/4.0.0.2.0/package/scripts/params_linux.py
@@ -183,7 +183,7 @@ if https_port is not None:
 hdfs_site = config['configurations']['hdfs-site']
 fs_root = config['configurations']['core-site']['fs.defaultFS']
 
-if Script.is_hdp_stack_greater_or_equal("2.0") and 
Script.is_hdp_stack_less_than("2.2"):
+if Script.is_hdp_stack_less_than("2.2"):
   put_shared_lib_to_hdfs_cmd = format("hadoop --config {hadoop_conf_dir} dfs 
-put {oozie_shared_lib} {oozie_hdfs_user_dir}")
 # for newer
 else:

http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py
 
b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py
index 4aa5f8a..d7b18eb 100644
--- 
a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py
+++ 
b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/after-INSTALL/scripts/params.py
@@ -62,11 +62,7 @@ hdfs_log_dir_prefix = 
config['configurations']['hadoop-env']['hdfs_log_dir_prefi
 hadoop_pid_dir_prefix = 
config['configurations']['hadoop-env']['hadoop_pid_dir_prefix']
 hadoop_root_logger = 
config['configurations']['hadoop-env']['hadoop_root_logger']
 
-if Script.is_hdp_stack_greater_or_equal("2.0") and 
Script.is_hdp_stack_less_than("2.1") and not OSCheck.is_suse_family():
-  # deprecated rhel jsvc_path
-  jsvc_path = "/usr/libexec/bigtop-utils"
-else:
-  jsvc_path = "/usr/lib/bigtop-utils"
+jsvc_path = "/usr/lib/bigtop-utils"
 
 hadoop_heapsize = config['configurations']['hadoop-env']['hadoop_heapsize']
 namenode_heapsize = config['configurations']['hadoop-env']['namenode_heapsize']

http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py
 
b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py
index 2593ee4..e1f33d4 100644
--- 
a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py
+++ 
b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-ANY/scripts/params.py
@@ -132,11 +132,7 @@ hdfs_log_dir_prefix = 
config['configurations']['hadoop-env']['hdfs_log_dir_prefi
 hadoop_pid_dir_prefix = 
config['configurations']['hadoop-env']['hadoop_pid_dir_prefix']
 hadoop_root_logger = 
config['configurations']['hadoop-env']['hadoop_root_logger']
 
-if hdp_stack_version != "" and compare_versions(hdp_stack_version, '2.0') >= 0 
and compare_versions(hdp_stack_version, '2.1') < 0 and not 
OSCheck.is_suse_family():
-  # deprecated rhel jsvc_path
-  jsvc_path = "/usr/libexec/bigtop-utils"
-else:
-  jsvc_path = "/usr/lib/bigtop-utils"
+jsvc_path = "/usr/lib/bigtop-utils"
 
 hadoop_heapsize = config['configurations']['hadoop-env']['hadoop_heapsize']
 namenode_heapsize = config['configurations']['hadoop-env']['namenode_heapsize']

http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py
 
b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py
index e771dd5..70c69fe 100644
--- 
a/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py
+++ 
b/ambari-server/src/main/resources/stacks/HDP/2.0.6/hooks/before-START/scripts/params.py
@@ -144,11 +144,7 @@ else:
 #hadoop-env.sh
 java_home = config['hostLevelParams']['java_home']
 
-if hdp_stack_version != "" and compare_versions(hdp_stack_version, '2.0') >= 0 
and compare_versions(hdp_stack_version, '2.1') < 0 and not 
OSCheck.is_suse_family():
-  # deprecated rhel jsvc_path
-  jsvc_path = "/usr/libexec/bigtop-utils"
-else:
-  jsvc_path = "/usr/lib/bigtop-utils"
+jsvc_path = "/usr/lib/bigtop-utils"
 
 hadoop_heapsize = config['configurations']['hadoop-env']['hadoop_heapsize']
 namenode_heapsize = config['configurations']['hadoop-env']['namenode_heapsize']

http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml
----------------------------------------------------------------------
diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml 
b/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml
index 0d4c27f..0659ca4 100644
--- a/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml
+++ b/ambari-server/src/main/resources/stacks/HDP/2.0/metainfo.xml
@@ -17,7 +17,7 @@
 -->
 <metainfo>
     <versions>
-      <active>true</active>
+      <active>false</active>
     </versions>
     <extends>2.0.6</extends>
     <minJdk>1.7</minJdk>

http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py 
b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py
index 6c9752a..e709693 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hcat_client.py
@@ -37,7 +37,7 @@ class TestHcatClient(RMFTestCase):
                               group = 'hadoop',
                               recursive = True,
     )
-    self.assertResourceCalled('Directory', '/etc/hcatalog/conf',
+    self.assertResourceCalled('Directory', '/etc/hive-hcatalog/conf',
       owner = 'hcat',
       group = 'hadoop',
       recursive = True,
@@ -54,7 +54,7 @@ class TestHcatClient(RMFTestCase):
       configurations = self.getConfig()['configurations']['hive-site'],
       configuration_attributes = 
self.getConfig()['configuration_attributes']['hive-site']
     )
-    self.assertResourceCalled('File', '/etc/hcatalog/conf/hcat-env.sh',
+    self.assertResourceCalled('File', '/etc/hive-hcatalog/conf/hcat-env.sh',
                               content = 
InlineTemplate(self.getConfig()['configurations']['hcat-env']['content']),
                               owner = 'hcat',
                               group = 'hadoop',
@@ -76,7 +76,7 @@ class TestHcatClient(RMFTestCase):
                               owner = 'hcat',
                               group = 'hadoop',
     )
-    self.assertResourceCalled('Directory', '/etc/hcatalog/conf',
+    self.assertResourceCalled('Directory', '/etc/hive-hcatalog/conf',
       recursive = True,
       owner = 'hcat',
       group = 'hadoop',
@@ -93,7 +93,7 @@ class TestHcatClient(RMFTestCase):
       configurations = self.getConfig()['configurations']['hive-site'],
       configuration_attributes = 
self.getConfig()['configuration_attributes']['hive-site']
     )
-    self.assertResourceCalled('File', '/etc/hcatalog/conf/hcat-env.sh',
+    self.assertResourceCalled('File', '/etc/hive-hcatalog/conf/hcat-env.sh',
                               content = 
InlineTemplate(self.getConfig()['configurations']['hcat-env']['content']),
                               owner = 'hcat',
                               group = 'hadoop',

http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py 
b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py
index 328b52d..7f3efcb 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_hive_metastore.py
@@ -227,6 +227,10 @@ class TestHiveMetastore(RMFTestCase):
                               content = StaticFile('startMetastore.sh'),
                               mode = 0755,
                               )
+    self.assertResourceCalled('Execute', 'export 
HIVE_CONF_DIR=/etc/hive/conf.server ; /usr/lib/hive/bin/schematool -initSchema 
-dbType mysql -userName hive -passWord \'!`"\'"\'"\' 1\'',
+        not_if = 'ambari-sudo.sh su hive -l -s /bin/bash -c 
\'[RMF_EXPORT_PLACEHOLDER]export HIVE_CONF_DIR=/etc/hive/conf.server ; 
/usr/lib/hive/bin/schematool -info -dbType mysql -userName hive -passWord 
\'"\'"\'!`"\'"\'"\'"\'"\'"\'"\'"\'"\' 1\'"\'"\'\'',
+        user = 'hive',
+    )
     self.assertResourceCalled('Directory', '/var/run/hive',
                               owner = 'hive',
                               group = 'hadoop',
@@ -331,6 +335,10 @@ class TestHiveMetastore(RMFTestCase):
                               content = StaticFile('startMetastore.sh'),
                               mode = 0755,
                               )
+    self.assertResourceCalled('Execute', 'export 
HIVE_CONF_DIR=/etc/hive/conf.server ; /usr/lib/hive/bin/schematool -initSchema 
-dbType mysql -userName hive -passWord \'!`"\'"\'"\' 1\'',
+        not_if = 'ambari-sudo.sh su hive -l -s /bin/bash -c 
\'[RMF_EXPORT_PLACEHOLDER]export HIVE_CONF_DIR=/etc/hive/conf.server ; 
/usr/lib/hive/bin/schematool -info -dbType mysql -userName hive -passWord 
\'"\'"\'!`"\'"\'"\'"\'"\'"\'"\'"\'"\' 1\'"\'"\'\'',
+        user = 'hive',
+    )
     self.assertResourceCalled('Directory', '/var/run/hive',
                               owner = 'hive',
                               group = 'hadoop',

http://git-wip-us.apache.org/repos/asf/ambari/blob/e3958e2d/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py
----------------------------------------------------------------------
diff --git 
a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py 
b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py
index c11c2d0..b0b34bc 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py
+++ b/ambari-server/src/test/python/stacks/2.0.6/HIVE/test_webhcat_server.py
@@ -48,7 +48,7 @@ class TestWebHCatServer(RMFTestCase):
     )
 
     self.assert_configure_default()
-    self.assertResourceCalled('Execute', 'cd /var/run/webhcat ; 
/usr/lib/hcatalog/sbin/webhcat_server.sh start',
+    self.assertResourceCalled('Execute', 'cd /var/run/webhcat ; 
/usr/lib/hive-hcatalog/sbin/webhcat_server.sh start',
         environment = {'HADOOP_HOME': '/usr'},
         not_if = "ambari-sudo.sh su hcat -l -s /bin/bash -c 
'[RMF_EXPORT_PLACEHOLDER]ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps 
-p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1'",
         user = 'hcat',
@@ -64,7 +64,7 @@ class TestWebHCatServer(RMFTestCase):
                        target = RMFTestCase.TARGET_COMMON_SERVICES
     )
 
-    self.assertResourceCalled('Execute', 
'/usr/lib/hcatalog/sbin/webhcat_server.sh stop',
+    self.assertResourceCalled('Execute', 
'/usr/lib/hive-hcatalog/sbin/webhcat_server.sh stop',
                               user = 'hcat',
                               environment = {'HADOOP_HOME': '/usr' }
                               )
@@ -105,7 +105,7 @@ class TestWebHCatServer(RMFTestCase):
     )
 
     self.assert_configure_secured()
-    self.assertResourceCalled('Execute', 'cd /var/run/webhcat ; 
/usr/lib/hcatalog/sbin/webhcat_server.sh start',
+    self.assertResourceCalled('Execute', 'cd /var/run/webhcat ; 
/usr/lib/hive-hcatalog/sbin/webhcat_server.sh start',
         environment = {'HADOOP_HOME': '/usr'},
         not_if = "ambari-sudo.sh su hcat -l -s /bin/bash -c 
'[RMF_EXPORT_PLACEHOLDER]ls /var/run/webhcat/webhcat.pid >/dev/null 2>&1 && ps 
-p `cat /var/run/webhcat/webhcat.pid` >/dev/null 2>&1'",
         user = 'hcat',
@@ -121,7 +121,7 @@ class TestWebHCatServer(RMFTestCase):
                        target = RMFTestCase.TARGET_COMMON_SERVICES
     )
 
-    self.assertResourceCalled('Execute', 
'/usr/lib/hcatalog/sbin/webhcat_server.sh stop',
+    self.assertResourceCalled('Execute', 
'/usr/lib/hive-hcatalog/sbin/webhcat_server.sh stop',
                               user = 'hcat',
                               environment = {'HADOOP_HOME': '/usr' }
                               )
@@ -152,7 +152,7 @@ class TestWebHCatServer(RMFTestCase):
                               recursive = True,
                               mode = 0755,
                               )
-    self.assertResourceCalled('Directory', '/etc/hcatalog/conf',
+    self.assertResourceCalled('Directory', '/etc/hive-webhcat/conf',
                               owner = 'hcat',
                               group = 'hadoop',
                               recursive = True,
@@ -161,11 +161,11 @@ class TestWebHCatServer(RMFTestCase):
     self.assertResourceCalled('XmlConfig', 'webhcat-site.xml',
                               owner = 'hcat',
                               group = 'hadoop',
-                              conf_dir = '/etc/hcatalog/conf',
+                              conf_dir = '/etc/hive-webhcat/conf',
                               configurations = 
self.getConfig()['configurations']['webhcat-site'],
                               configuration_attributes = 
self.getConfig()['configuration_attributes']['webhcat-site']
     )
-    self.assertResourceCalled('File', '/etc/hcatalog/conf/webhcat-env.sh',
+    self.assertResourceCalled('File', '/etc/hive-webhcat/conf/webhcat-env.sh',
                               content = 
InlineTemplate(self.getConfig()['configurations']['webhcat-env']['content']),
                               owner = 'hcat',
                               group = 'hadoop',
@@ -174,7 +174,7 @@ class TestWebHCatServer(RMFTestCase):
         cd_access = 'a',
         recursive=True
     )
-    self.assertResourceCalled('File', 
'/etc/hcatalog/conf/webhcat-log4j.properties',
+    self.assertResourceCalled('File', 
'/etc/hive-webhcat/conf/webhcat-log4j.properties',
                               content = 'log4jproperties\nline2',
                               owner = 'hcat',
                               group = 'hadoop',
@@ -194,7 +194,7 @@ class TestWebHCatServer(RMFTestCase):
                               recursive = True,
                               mode = 0755,
                               )
-    self.assertResourceCalled('Directory', '/etc/hcatalog/conf',
+    self.assertResourceCalled('Directory', '/etc/hive-webhcat/conf',
                               owner = 'hcat',
                               group = 'hadoop',
                               recursive = True,
@@ -207,11 +207,11 @@ class TestWebHCatServer(RMFTestCase):
     self.assertResourceCalled('XmlConfig', 'webhcat-site.xml',
                               owner = 'hcat',
                               group = 'hadoop',
-                              conf_dir = '/etc/hcatalog/conf',
+                              conf_dir = '/etc/hive-webhcat/conf',
                               configurations = 
self.getConfig()['configurations']['webhcat-site'],
                               configuration_attributes = 
self.getConfig()['configuration_attributes']['webhcat-site']
     )
-    self.assertResourceCalled('File', '/etc/hcatalog/conf/webhcat-env.sh',
+    self.assertResourceCalled('File', '/etc/hive-webhcat/conf/webhcat-env.sh',
                               content = 
InlineTemplate(self.getConfig()['configurations']['webhcat-env']['content']),
                               owner = 'hcat',
                               group = 'hadoop',
@@ -220,7 +220,7 @@ class TestWebHCatServer(RMFTestCase):
         cd_access = 'a',
         recursive=True
     )
-    self.assertResourceCalled('File', 
'/etc/hcatalog/conf/webhcat-log4j.properties',
+    self.assertResourceCalled('File', 
'/etc/hive-webhcat/conf/webhcat-log4j.properties',
                               content = 'log4jproperties\nline2',
                               owner = 'hcat',
                               group = 'hadoop',

Reply via email to