Hi,

I use ambari 1.6.1 to install a single node cluster, I can see ambari installed 
the lasted version 2.1.7 of HDP, but the hive service failed to start with the 
following messages:





stderr:   /var/lib/ambari-agent/data/errors-56.txt
2014-11-04 16:46:08,931 - Error while executing command 'start': Traceback 
(most recent call last):   File 
"/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
 line 111, in execute     method(env)   File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HIVE/package/scripts/hive_metastore.py",
 line 42, in start     self.configure(env) # FOR SECURITY   File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HIVE/package/scripts/hive_metastore.py",
 line 37, in configure     hive(name='metastore')   File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HIVE/package/scripts/hive.py",
 line 108, in hive     not_if = check_schema_created_cmd   File 
"/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 148, 
in __init__     self.env.run()   File 
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
line 149, in run     self.run_action(resource, action)   File 
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
line 115, in run_action     provider_action()   File 
"/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py",
 line 239, in action_run     raise ex Fail: Execution of 'export 
HIVE_CONF_DIR=/etc/hive/conf.server ; /usr/lib/hive/bin/schematool -initSchema 
-dbType mysql -userName hive -passWord [PROTECTED]' returned 1. Metastore 
connection URL:         
jdbc:mysql://lix1.bh.com/hive?createDatabaseIfNotExist=true Metastore 
Connection Driver :       com.mysql.jdbc.Driver Metastore connection User:      
  hive org.apache.hadoop.hive.metastore.HiveMetaException: Failed to get schema 
version. *** schemaTool failed ***
stdout:   /var/lib/ambari-agent/data/output-56.txt
2014-11-04 16:45:55,983 - Execute['mkdir -p /tmp/HDP-artifacts/;     curl -kf 
-x "" --retry 10     
http://ambari.bh.com:8080/resources//UnlimitedJCEPolicyJDK7.zip -o 
/tmp/HDP-artifacts//UnlimitedJCEPolicyJDK7.zip'] {'environment': ..., 'not_if': 
'test -e /tmp/HDP-artifacts//UnlimitedJCEPolicyJDK7.zip', 'ignore_failures': 
True, 'path': ['/bin', '/usr/bin/']} 2014-11-04 16:45:56,001 - Skipping 
Execute['mkdir -p /tmp/HDP-artifacts/;     curl -kf -x "" --retry 10     
http://ambari.bh.com:8080/resources//UnlimitedJCEPolicyJDK7.zip -o 
/tmp/HDP-artifacts//UnlimitedJCEPolicyJDK7.zip'] due to not_if 2014-11-04 
16:45:56,115 - Directory['/etc/hadoop/conf.empty'] {'owner': 'root', 'group': 
'root', 'recursive': True} 2014-11-04 16:45:56,116 - Link['/etc/hadoop/conf'] 
{'not_if': 'ls /etc/hadoop/conf', 'to': '/etc/hadoop/conf.empty'} 2014-11-04 
16:45:56,137 - Skipping Link['/etc/hadoop/conf'] due to not_if 2014-11-04 
16:45:56,152 - File['/etc/hadoop/conf/hadoop-env.sh'] {'content': 
Template('hadoop-env.sh.j2'), 'owner': 'hdfs'} 2014-11-04 16:45:56,153 - 
XmlConfig['core-site.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': 
'/etc/hadoop/conf', 'configurations': ...} 2014-11-04 16:45:56,159 - Generating 
config: /etc/hadoop/conf/core-site.xml 2014-11-04 16:45:56,160 - 
File['/etc/hadoop/conf/core-site.xml'] {'owner': 'hdfs', 'content': 
InlineTemplate(...), 'group': 'hadoop', 'mode': None} 2014-11-04 16:45:56,160 - 
Writing File['/etc/hadoop/conf/core-site.xml'] because contents don't match 
2014-11-04 16:45:56,177 - Execute['/bin/echo 0 > /selinux/enforce'] {'only_if': 
'test -f /selinux/enforce'} 2014-11-04 16:45:56,216 - Execute['mkdir -p 
/usr/lib/hadoop/lib/native/Linux-i386-32; ln -sf /usr/lib/libsnappy.so 
/usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so'] {} 2014-11-04 
16:45:56,241 - Execute['mkdir -p /usr/lib/hadoop/lib/native/Linux-amd64-64; ln 
-sf /usr/lib64/libsnappy.so 
/usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so'] {} 2014-11-04 
16:45:56,262 - Directory['/var/log/hadoop'] {'owner': 'root', 'group': 'root', 
'recursive': True} 2014-11-04 16:45:56,263 - Directory['/var/run/hadoop'] 
{'owner': 'root', 'group': 'root', 'recursive': True} 2014-11-04 16:45:56,265 - 
Directory['/tmp/hadoop-hdfs'] {'owner': 'hdfs', 'recursive': True} 2014-11-04 
16:45:56,274 - File['/etc/hadoop/conf/commons-logging.properties'] {'content': 
Template('commons-logging.properties.j2'), 'owner': 'hdfs'} 2014-11-04 
16:45:56,276 - File['/etc/hadoop/conf/health_check'] {'content': 
Template('health_check-v2.j2'), 'owner': 'hdfs'} 2014-11-04 16:45:56,278 - 
File['/etc/hadoop/conf/log4j.properties'] {'content': '...', 'owner': 'hdfs', 
'group': 'hadoop', 'mode': 0644} 2014-11-04 16:45:56,282 - 
File['/etc/hadoop/conf/hadoop-metrics2.properties'] {'content': 
Template('hadoop-metrics2.properties.j2'), 'owner': 'hdfs'} 2014-11-04 
16:45:56,283 - File['/etc/hadoop/conf/task-log4j.properties'] {'content': 
StaticFile('task-log4j.properties'), 'mode': 0755} 2014-11-04 16:45:56,284 - 
File['/etc/hadoop/conf/configuration.xsl'] {'owner': 'hdfs', 'group': 'hadoop'} 
2014-11-04 16:45:56,457 - Execute['hive mkdir -p /tmp/HDP-artifacts/ ; cp 
/usr/share/java/mysql-connector-java.jar 
/usr/lib/hive/lib//mysql-connector-java.jar'] {'creates': 
'/usr/lib/hive/lib//mysql-connector-java.jar', 'path': ['/bin', '/usr/bin/'], 
'not_if': 'test -f /usr/lib/hive/lib//mysql-connector-java.jar'} 2014-11-04 
16:46:00,436 - Directory['/etc/hive/conf.server'] {'owner': 'hive', 'group': 
'hadoop', 'recursive': True} 2014-11-04 16:46:00,437 - Creating directory 
Directory['/etc/hive/conf.server'] 2014-11-04 16:46:00,440 - Changing owner for 
/etc/hive/conf.server from 0 to hive 2014-11-04 16:46:00,440 - Changing group 
for /etc/hive/conf.server from 0 to hadoop 2014-11-04 16:46:00,440 - 
XmlConfig['mapred-site.xml'] {'owner': 'hive', 'group': 'hadoop', 'mode': 0600, 
'conf_dir': '/etc/hive/conf.server', 'configurations': ...} 2014-11-04 
16:46:00,447 - Generating config: /etc/hive/conf.server/mapred-site.xml 
2014-11-04 16:46:00,447 - File['/etc/hive/conf.server/mapred-site.xml'] 
{'owner': 'hive', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 
0600} 2014-11-04 16:46:00,448 - Writing 
File['/etc/hive/conf.server/mapred-site.xml'] because it doesn't exist 
2014-11-04 16:46:00,448 - Changing permission for 
/etc/hive/conf.server/mapred-site.xml from 644 to 600 2014-11-04 16:46:00,448 - 
Changing owner for /etc/hive/conf.server/mapred-site.xml from 0 to hive 
2014-11-04 16:46:00,448 - Changing group for 
/etc/hive/conf.server/mapred-site.xml from 0 to hadoop 2014-11-04 16:46:00,449 
- XmlConfig['hive-site.xml'] {'owner': 'hive', 'group': 'hadoop', 'mode': 0600, 
'conf_dir': '/etc/hive/conf.server', 'configurations': ...} 2014-11-04 
16:46:00,451 - Generating config: /etc/hive/conf.server/hive-site.xml 
2014-11-04 16:46:00,451 - File['/etc/hive/conf.server/hive-site.xml'] {'owner': 
'hive', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0600} 
2014-11-04 16:46:00,452 - Writing File['/etc/hive/conf.server/hive-site.xml'] 
because it doesn't exist 2014-11-04 16:46:00,452 - Changing permission for 
/etc/hive/conf.server/hive-site.xml from 644 to 600 2014-11-04 16:46:00,452 - 
Changing owner for /etc/hive/conf.server/hive-site.xml from 0 to hive 
2014-11-04 16:46:00,452 - Changing group for 
/etc/hive/conf.server/hive-site.xml from 0 to hadoop 2014-11-04 16:46:00,452 - 
Execute['/bin/sh -c 'cd /usr/lib/ambari-agent/ && curl -kf -x "" --retry 5 
http://ambari.bh.com:8080/resources/DBConnectionVerification.jar -o 
DBConnectionVerification.jar''] {'environment': ..., 'not_if': '[ -f 
DBConnectionVerification.jar]'} 2014-11-04 16:46:00,513 - 
File['/etc/hive/conf.server/hive-env.sh'] {'content': 
Template('hive-env.sh.j2'), 'owner': 'hive', 'group': 'hadoop'} 2014-11-04 
16:46:00,513 - Writing File['/etc/hive/conf.server/hive-env.sh'] because it 
doesn't exist 2014-11-04 16:46:00,514 - Changing owner for 
/etc/hive/conf.server/hive-env.sh from 0 to hive 2014-11-04 16:46:00,514 - 
Changing group for /etc/hive/conf.server/hive-env.sh from 0 to hadoop 
2014-11-04 16:46:00,514 - File['/tmp/start_metastore_script'] {'content': 
StaticFile('startMetastore.sh'), 'mode': 0755} 2014-11-04 16:46:00,515 - 
Execute['export HIVE_CONF_DIR=/etc/hive/conf.server ; 
/usr/lib/hive/bin/schematool -initSchema -dbType mysql -userName hive -passWord 
[PROTECTED]'] {'not_if': 'export HIVE_CONF_DIR=/etc/hive/conf.server ; 
/usr/lib/hive/bin/schematool -info -dbType mysql -userName hive -passWord 
[PROTECTED]'} 2014-11-04 16:46:08,931 - Error while executing command 'start': 
Traceback (most recent call last):   File 
"/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py",
 line 111, in execute     method(env)   File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HIVE/package/scripts/hive_metastore.py",
 line 42, in start     self.configure(env) # FOR SECURITY   File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HIVE/package/scripts/hive_metastore.py",
 line 37, in configure     hive(name='metastore')   File 
"/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/HIVE/package/scripts/hive.py",
 line 108, in hive     not_if = check_schema_created_cmd   File 
"/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 148, 
in __init__     self.env.run()   File 
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
line 149, in run     self.run_action(resource, action)   File 
"/usr/lib/python2.6/site-packages/resource_management/core/environment.py", 
line 115, in run_action     provider_action()   File 
"/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py",
 line 239, in action_run     raise ex Fail: Execution of 'export 
HIVE_CONF_DIR=/etc/hive/conf.server ; /usr/lib/hive/bin/schematool -initSchema 
-dbType mysql -userName hive -passWord [PROTECTED]' returned 1. Metastore 
connection URL:      
jdbc:mysql://lix1.bh.com/hive?createDatabaseIfNotExist=true Metastore 
Connection Driver :       com.mysql.jdbc.Driver Metastore connection User:      
  hive org.apache.hadoop.hive.metastore.HiveMetaException: Failed to get schema 
version. *** schemaTool failed ***

Reply via email to