Did you attempt to start the services manually using root as I see some folders owned by root?
On Mon, Nov 3, 2014 at 6:18 PM, guxiaobo1982 <[email protected]> wrote: > there are some error messages in the yarn-yarn-historyserver-lix1.bh.com.log > file: > > > ************************************************************/ > > 2014-11-03 12:07:34,974 INFO > applicationhistoryservice.ApplicationHistoryServer > (StringUtils.java:startupShutdownMessage(614)) - STARTUP_MSG: > > /************************************************************ > > STARTUP_MSG: Starting ApplicationHistoryServer > > STARTUP_MSG: host = lix1.bh.com/192.168.100.3 > > STARTUP_MSG: args = [] > > STARTUP_MSG: version = 2.4.0.2.1.5.0-695 > > STARTUP_MSG: classpath = > /etc/hadoop/conf:/etc/hadoop/conf:/etc/hadoop/conf:/usr/lib/hadoop/lib/jersey-core-1.9.jar:/usr/lib/hadoop/lib/mockito-all > > -1.8.5.jar: > > > ....... > > > STARTUP_MSG: build = [email protected]:hortonworks/hadoop.git -r > c11220208321e1835912fde828f1038eedb1afae; compiled by 'jenkins' on > 2014-08-28T03:10Z > > STARTUP_MSG: java = 1.7.0_45 > > ************************************************************/ > > 2014-11-03 12:07:35,039 INFO > applicationhistoryservice.ApplicationHistoryServer > (SignalLogger.java:register(91)) - registered UNIX signal handlers f > > or [TERM, HUP, INT] > > 2014-11-03 12:07:36,977 INFO impl.MetricsConfig > (MetricsConfig.java:loadFirst(111)) - loaded properties from > hadoop-metrics2.properties > > 2014-11-03 12:07:37,208 INFO impl.MetricsSystemImpl > (MetricsSystemImpl.java:startTimer(355)) - Scheduled snapshot period at 60 > second(s). > > 2014-11-03 12:07:37,208 INFO impl.MetricsSystemImpl > (MetricsSystemImpl.java:start(183)) - ApplicationHistoryServer metrics > system started > > 2014-11-03 12:07:37,230 INFO > applicationhistoryservice.ApplicationHistoryManagerImpl > (ApplicationHistoryManagerImpl.java:serviceInit(61)) - Applicat > > ionHistory Init > > 2014-11-03 12:07:37,615 INFO timeline.LeveldbTimelineStore > (LeveldbTimelineStore.java:serviceInit(194)) - Using leveldb path > /hadoop/yarn/timeline/l > > eveldb-timeline-store.ldb > > 2014-11-03 12:07:37,647 INFO service.AbstractService > (AbstractService.java:noteFailure(272)) - Service > org.apache.hadoop.yarn.server.timeline.Leveld > > bTimelineStore failed in state INITED; cause: > org.fusesource.leveldbjni.internal.NativeDB$DBException: IO error: > /hadoop/yarn/timeline/leveldb-timeli > > ne-store.ldb/LOCK: Permission denied > > org.fusesource.leveldbjni.internal.NativeDB$DBException: IO error: > /hadoop/yarn/timeline/leveldb-timeline-store.ldb/LOCK: Permission denied > > at > org.fusesource.leveldbjni.internal.NativeDB.checkStatus(NativeDB.java:200) > > at org.fusesource.leveldbjni.internal.NativeDB.open(NativeDB.java:218) > > at org.fusesource.leveldbjni.JniDBFactory.open(JniDBFactory.java:168) > > at > org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore.serviceInit(LeveldbTimelineStore.java:195) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > > at > org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:88) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:145) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:155) > > 2014-11-03 12:07:37,654 INFO service.AbstractService > (AbstractService.java:noteFailure(272)) - Service > org.apache.hadoop.yarn.server.applicationhist > > oryservice.ApplicationHistoryServer failed in state INITED; cause: > org.apache.hadoop.service.ServiceStateException: > org.fusesource.leveldbjni.interna > > l.NativeDB$DBException: IO error: > /hadoop/yarn/timeline/leveldb-timeline-store.ldb/LOCK: Permission denied > > org.apache.hadoop.service.ServiceStateException: > org.fusesource.leveldbjni.internal.NativeDB$DBException: IO error: > /hadoop/yarn/timeline/leveldb-tim > > eline-store.ldb/LOCK: Permission denied > > at > org.apache.hadoop.service.ServiceStateException.convert(ServiceStateException.java:59) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:172) > > at > org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:88) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:145) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:155) > > Caused by: org.fusesource.leveldbjni.internal.NativeDB$DBException: IO > error: /hadoop/yarn/timeline/leveldb-timeline-store.ldb/LOCK: Permission > denie > > d > > at > org.fusesource.leveldbjni.internal.NativeDB.checkStatus(NativeDB.java:200) > > at org.fusesource.leveldbjni.internal.NativeDB.open(NativeDB.java:218) > > at org.fusesource.leveldbjni.JniDBFactory.open(JniDBFactory.java:168) > > at > org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore.serviceInit(LeveldbTimelineStore.java:195) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > > ... 5 more > > 2014-11-03 12:07:37,655 INFO impl.MetricsSystemImpl > (MetricsSystemImpl.java:stop(200)) - Stopping ApplicationHistoryServer > metrics system... > > 2014-11-03 12:07:37,656 INFO impl.MetricsSystemImpl > (MetricsSystemImpl.java:stop(206)) - ApplicationHistoryServer metrics > system stopped. > > 2014-11-03 12:07:37,656 INFO impl.MetricsSystemImpl > (MetricsSystemImpl.java:shutdown(583)) - ApplicationHistoryServer metrics > system shutdown complete. > > 2014-11-03 12:07:37,656 INFO > applicationhistoryservice.ApplicationHistoryManagerImpl > (ApplicationHistoryManagerImpl.java:serviceStop(78)) - Stopping > ApplicationHistory > > 2014-11-03 12:07:37,657 FATAL > applicationhistoryservice.ApplicationHistoryServer > (ApplicationHistoryServer.java:launchAppHistoryServer(148)) - Error > starting ApplicationHistoryServer > > org.apache.hadoop.service.ServiceStateException: > org.fusesource.leveldbjni.internal.NativeDB$DBException: IO error: > /hadoop/yarn/timeline/leveldb-timeline-store.ldb/LOCK: Permission denied > > at > org.apache.hadoop.service.ServiceStateException.convert(ServiceStateException.java:59) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:172) > > at > org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:88) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:145) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:155) > > Caused by: org.fusesource.leveldbjni.internal.NativeDB$DBException: IO > error: /hadoop/yarn/timeline/leveldb-timeline-store.ldb/LOCK: Permission > denied > > at > org.fusesource.leveldbjni.internal.NativeDB.checkStatus(NativeDB.java:200) > > at org.fusesource.leveldbjni.internal.NativeDB.open(NativeDB.java:218) > > at org.fusesource.leveldbjni.JniDBFactory.open(JniDBFactory.java:168) > > at > org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore.serviceInit(LeveldbTimelineStore.java:195) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > > ... 5 more > > 2014-11-03 12:07:37,664 INFO util.ExitUtil (ExitUtil.java:terminate(124)) > - Exiting with status -1 > > 2014-11-03 12:07:37,668 INFO > applicationhistoryservice.ApplicationHistoryServer > (StringUtils.java:run(640)) - SHUTDOWN_MSG: > > /************************************************************ > > SHUTDOWN_MSG: Shutting down ApplicationHistoryServer at > lix1.bh.com/192.168.100.3 > > > And I list the ownership for the /hadoop path > > > [root@lix1 hadoop]# pwd > > /hadoop > > [root@lix1 hadoop]# ls > > falcon hbase hdfs oozie storm yarn zookeeper > > [root@lix1 hadoop]# ls -al > > 总用量 36 > > drwxr-xr-x. 9 root root 4096 11月 3 11:58 . > > dr-xr-xr-x. 28 root root 4096 11月 3 18:01 .. > > drwxr-xr-x. 4 falcon root 4096 11月 3 11:54 falcon > > drwxr-xr-x. 3 hbase root 4096 11月 3 10:34 hbase > > drwxr-xr-x. 5 root root 4096 11月 3 11:58 hdfs > > drwxr-xr-x. 3 root root 4096 11月 3 11:58 oozie > > drwxr-xr-x. 5 storm hadoop 4096 11月 3 12:05 storm > > drwxr-xr-x. 5 root root 4096 10月 14 18:34 yarn > > drwxr-xr-x. 3 zookeeper hadoop 4096 11月 3 11:51 zookeeper > > [root@lix1 hadoop]# ls -al yarn/ > > 总用量 20 > > drwxr-xr-x. 5 root root 4096 10月 14 18:34 . > > drwxr-xr-x. 9 root root 4096 11月 3 11:58 .. > > drwxr-xr-x. 6 yarn root 4096 11月 3 12:58 local > > drwxr-xr-x. 2 yarn root 4096 11月 3 12:09 log > > drwxr-xr-x. 3 yarn hadoop 4096 10月 14 18:34 timeline > > [root@lix1 hadoop]# > > > > > ------------------ Original ------------------ > *From: * "Mingjiang Shi";<[email protected]>; > *Send time:* Monday, Nov 3, 2014 4:36 PM > *To:* "[email protected]"<[email protected]>; > *Subject: * Re: timeline service installed by ambari can't start > > Hi, > Could you check the timeline server log located at > /var/log/hadoop-yarn/yarn/yarn-yarn-timelineserver*.log to see what problem > caused the failure? > > On Mon, Nov 3, 2014 at 3:56 PM, guxiaobo1982 <[email protected]> wrote: > >> The HDFS installed is of version >> Version:2.4.0.2.1.5.0-695, rc11220208321e1835912fde828f1038eedb1afae >> >> >> ------------------ Original ------------------ >> *From: * "guxiaobo1982";<[email protected]>; >> *Send time:* Monday, Nov 3, 2014 3:48 PM >> *To:* "user"<[email protected]>; >> *Subject: * timeline service installed by ambari can't start >> >> Hi, >> >> I use Ambari 16.1 installed HDP 2.1 Single node deployment, but the >> timeline service can't start with the following error: >> stderr: /var/lib/ambari-agent/data/errors-96.txt >> >> 2014-11-03 13:28:03,199 - Error while executing command 'restart': >> Traceback (most recent call last): >> File >> "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", >> line 111, in execute >> method(env) >> File >> "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", >> line 212, in restart >> self.start(env) >> File >> "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/application_timeline_server.py", >> line 42, in start >> service('historyserver', action='start') >> File >> "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/service.py", >> line 51, in service >> initial_wait=5 >> File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", >> line 148, in __init__ >> self.env.run() >> File >> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", >> line 149, in run >> self.run_action(resource, action) >> File >> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", >> line 115, in run_action >> provider_action() >> File >> "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", >> line 239, in action_run >> raise ex >> Fail: Execution of 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid >> >/dev/null 2>&1 && ps `cat >> /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid` >/dev/null 2>&1' >> returned 1. >> >> stdout: /var/lib/ambari-agent/data/output-96.txt >> >> 2014-11-03 13:27:56,524 - Execute['mkdir -p /tmp/HDP-artifacts/; curl >> -kf -x "" --retry 10 >> http://ambari.bh.com:8080/resources//UnlimitedJCEPolicyJDK7.zip -o >> /tmp/HDP-artifacts//UnlimitedJCEPolicyJDK7.zip'] {'environment': ..., >> 'not_if': 'test -e /tmp/HDP-artifacts//UnlimitedJCEPolicyJDK7.zip', >> 'ignore_failures': True, 'path': ['/bin', '/usr/bin/']} >> 2014-11-03 13:27:56,543 - Skipping Execute['mkdir -p /tmp/HDP-artifacts/; >> curl -kf -x "" --retry 10 >> http://ambari.bh.com:8080/resources//UnlimitedJCEPolicyJDK7.zip -o >> /tmp/HDP-artifacts//UnlimitedJCEPolicyJDK7.zip'] due to not_if >> 2014-11-03 13:27:56,618 - Directory['/etc/hadoop/conf.empty'] {'owner': >> 'root', 'group': 'root', 'recursive': True} >> 2014-11-03 13:27:56,620 - Link['/etc/hadoop/conf'] {'not_if': 'ls >> /etc/hadoop/conf', 'to': '/etc/hadoop/conf.empty'} >> 2014-11-03 13:27:56,634 - Skipping Link['/etc/hadoop/conf'] due to not_if >> 2014-11-03 13:27:56,644 - File['/etc/hadoop/conf/hadoop-env.sh'] {'content': >> Template('hadoop-env.sh.j2'), 'owner': 'hdfs'} >> 2014-11-03 13:27:56,646 - XmlConfig['core-site.xml'] {'owner': 'hdfs', >> 'group': 'hadoop', 'conf_dir': '/etc/hadoop/conf', 'configurations': ...} >> 2014-11-03 13:27:56,650 - Generating config: /etc/hadoop/conf/core-site.xml >> 2014-11-03 13:27:56,650 - File['/etc/hadoop/conf/core-site.xml'] {'owner': >> 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None} >> 2014-11-03 13:27:56,651 - Writing File['/etc/hadoop/conf/core-site.xml'] >> because contents don't match >> 2014-11-03 13:27:56,662 - Execute['/bin/echo 0 > /selinux/enforce'] >> {'only_if': 'test -f /selinux/enforce'} >> 2014-11-03 13:27:56,683 - Execute['mkdir -p >> /usr/lib/hadoop/lib/native/Linux-i386-32; ln -sf /usr/lib/libsnappy.so >> /usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so'] {} >> 2014-11-03 13:27:56,698 - Execute['mkdir -p >> /usr/lib/hadoop/lib/native/Linux-amd64-64; ln -sf /usr/lib64/libsnappy.so >> /usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so'] {} >> 2014-11-03 13:27:56,709 - Directory['/var/log/hadoop'] {'owner': 'root', >> 'group': 'root', 'recursive': True} >> 2014-11-03 13:27:56,710 - Directory['/var/run/hadoop'] {'owner': 'root', >> 'group': 'root', 'recursive': True} >> 2014-11-03 13:27:56,710 - Directory['/tmp/hadoop-hdfs'] {'owner': 'hdfs', >> 'recursive': True} >> 2014-11-03 13:27:56,714 - >> File['/etc/hadoop/conf/commons-logging.properties'] {'content': >> Template('commons-logging.properties.j2'), 'owner': 'hdfs'} >> 2014-11-03 13:27:56,716 - File['/etc/hadoop/conf/health_check'] {'content': >> Template('health_check-v2.j2'), 'owner': 'hdfs'} >> 2014-11-03 13:27:56,717 - File['/etc/hadoop/conf/log4j.properties'] >> {'content': '...', 'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644} >> 2014-11-03 13:27:56,720 - >> File['/etc/hadoop/conf/hadoop-metrics2.properties'] {'content': >> Template('hadoop-metrics2.properties.j2'), 'owner': 'hdfs'} >> 2014-11-03 13:27:56,720 - File['/etc/hadoop/conf/task-log4j.properties'] >> {'content': StaticFile('task-log4j.properties'), 'mode': 0755} >> 2014-11-03 13:27:56,721 - File['/etc/hadoop/conf/configuration.xsl'] >> {'owner': 'hdfs', 'group': 'hadoop'} >> 2014-11-03 13:27:56,803 - Execute['export >> HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && >> /usr/lib/hadoop-yarn/sbin/yarn-daemon.sh --config /etc/hadoop/conf stop >> historyserver'] {'user': 'yarn'} >> 2014-11-03 13:27:56,924 - Execute['rm -f >> /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid'] {'user': 'yarn'} >> 2014-11-03 13:27:56,955 - Directory['/var/run/hadoop-yarn/yarn'] {'owner': >> 'yarn', 'group': 'hadoop', 'recursive': True} >> 2014-11-03 13:27:56,956 - Directory['/var/log/hadoop-yarn/yarn'] {'owner': >> 'yarn', 'group': 'hadoop', 'recursive': True} >> 2014-11-03 13:27:56,956 - Directory['/var/run/hadoop-mapreduce/mapred'] >> {'owner': 'mapred', 'group': 'hadoop', 'recursive': True} >> 2014-11-03 13:27:56,956 - Directory['/var/log/hadoop-mapreduce/mapred'] >> {'owner': 'mapred', 'group': 'hadoop', 'recursive': True} >> 2014-11-03 13:27:56,956 - Directory['/hadoop/yarn/local'] {'owner': 'yarn', >> 'ignore_failures': True, 'recursive': True} >> 2014-11-03 13:27:56,956 - Directory['/hadoop/yarn/log'] {'owner': 'yarn', >> 'ignore_failures': True, 'recursive': True} >> 2014-11-03 13:27:56,957 - Directory['/var/log/hadoop-yarn'] {'owner': >> 'yarn', 'ignore_failures': True, 'recursive': True} >> 2014-11-03 13:27:56,957 - XmlConfig['core-site.xml'] {'owner': 'hdfs', >> 'group': 'hadoop', 'mode': 0644, 'conf_dir': '/etc/hadoop/conf', >> 'configurations': ...} >> 2014-11-03 13:27:56,963 - Generating config: /etc/hadoop/conf/core-site.xml >> 2014-11-03 13:27:56,963 - File['/etc/hadoop/conf/core-site.xml'] {'owner': >> 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644} >> 2014-11-03 13:27:56,963 - XmlConfig['mapred-site.xml'] {'owner': 'yarn', >> 'group': 'hadoop', 'mode': 0644, 'conf_dir': '/etc/hadoop/conf', >> 'configurations': ...} >> 2014-11-03 13:27:56,966 - Generating config: /etc/hadoop/conf/mapred-site.xml >> 2014-11-03 13:27:56,966 - File['/etc/hadoop/conf/mapred-site.xml'] {'owner': >> 'yarn', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644} >> 2014-11-03 13:27:56,967 - Writing File['/etc/hadoop/conf/mapred-site.xml'] >> because contents don't match >> 2014-11-03 13:27:56,967 - Changing owner for >> /etc/hadoop/conf/mapred-site.xml from 1022 to yarn >> 2014-11-03 13:27:56,967 - XmlConfig['yarn-site.xml'] {'owner': 'yarn', >> 'group': 'hadoop', 'mode': 0644, 'conf_dir': '/etc/hadoop/conf', >> 'configurations': ...} >> 2014-11-03 13:27:56,969 - Generating config: /etc/hadoop/conf/yarn-site.xml >> 2014-11-03 13:27:56,969 - File['/etc/hadoop/conf/yarn-site.xml'] {'owner': >> 'yarn', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644} >> 2014-11-03 13:27:56,970 - Writing File['/etc/hadoop/conf/yarn-site.xml'] >> because contents don't match >> 2014-11-03 13:27:56,971 - XmlConfig['capacity-scheduler.xml'] {'owner': >> 'yarn', 'group': 'hadoop', 'mode': 0644, 'conf_dir': '/etc/hadoop/conf', >> 'configurations': ...} >> 2014-11-03 13:27:56,974 - Generating config: >> /etc/hadoop/conf/capacity-scheduler.xml >> 2014-11-03 13:27:56,974 - File['/etc/hadoop/conf/capacity-scheduler.xml'] >> {'owner': 'yarn', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': >> 0644} >> 2014-11-03 13:27:56,975 - Writing >> File['/etc/hadoop/conf/capacity-scheduler.xml'] because contents don't match >> 2014-11-03 13:27:56,975 - Changing owner for >> /etc/hadoop/conf/capacity-scheduler.xml from 1021 to yarn >> 2014-11-03 13:27:56,975 - Directory['/hadoop/yarn/timeline'] {'owner': >> 'yarn', 'group': 'hadoop', 'recursive': True} >> 2014-11-03 13:27:56,975 - File['/etc/hadoop/conf/yarn.exclude'] {'owner': >> 'yarn', 'group': 'hadoop'} >> 2014-11-03 13:27:56,977 - File['/etc/security/limits.d/yarn.conf'] >> {'content': Template('yarn.conf.j2'), 'mode': 0644} >> 2014-11-03 13:27:56,980 - File['/etc/security/limits.d/mapreduce.conf'] >> {'content': Template('mapreduce.conf.j2'), 'mode': 0644} >> 2014-11-03 13:27:56,982 - File['/etc/hadoop/conf/yarn-env.sh'] {'content': >> Template('yarn-env.sh.j2'), 'owner': 'yarn', 'group': 'hadoop', 'mode': 0755} >> 2014-11-03 13:27:56,984 - File['/etc/hadoop/conf/mapred-env.sh'] {'content': >> Template('mapred-env.sh.j2'), 'owner': 'hdfs'} >> 2014-11-03 13:27:56,985 - File['/etc/hadoop/conf/taskcontroller.cfg'] >> {'content': Template('taskcontroller.cfg.j2'), 'owner': 'hdfs'} >> 2014-11-03 13:27:56,986 - XmlConfig['mapred-site.xml'] {'owner': 'mapred', >> 'group': 'hadoop', 'conf_dir': '/etc/hadoop/conf', 'configurations': ...} >> 2014-11-03 13:27:56,988 - Generating config: /etc/hadoop/conf/mapred-site.xml >> 2014-11-03 13:27:56,988 - File['/etc/hadoop/conf/mapred-site.xml'] {'owner': >> 'mapred', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None} >> 2014-11-03 13:27:56,988 - Changing owner for >> /etc/hadoop/conf/mapred-site.xml from 1020 to mapred >> 2014-11-03 13:27:56,988 - XmlConfig['capacity-scheduler.xml'] {'owner': >> 'hdfs', 'group': 'hadoop', 'conf_dir': '/etc/hadoop/conf', 'configurations': >> ...} >> 2014-11-03 13:27:56,991 - Generating config: >> /etc/hadoop/conf/capacity-scheduler.xml >> 2014-11-03 13:27:56,991 - File['/etc/hadoop/conf/capacity-scheduler.xml'] >> {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': >> None} >> 2014-11-03 13:27:56,992 - Changing owner for >> /etc/hadoop/conf/capacity-scheduler.xml from 1020 to hdfs >> 2014-11-03 13:27:56,992 - File['/etc/hadoop/conf/ssl-client.xml.example'] >> {'owner': 'mapred', 'group': 'hadoop'} >> 2014-11-03 13:27:56,992 - File['/etc/hadoop/conf/ssl-server.xml.example'] >> {'owner': 'mapred', 'group': 'hadoop'} >> 2014-11-03 13:27:56,993 - Execute['export >> HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && >> /usr/lib/hadoop-yarn/sbin/yarn-daemon.sh --config /etc/hadoop/conf start >> historyserver'] {'not_if': 'ls >> /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid >/dev/null 2>&1 && ps >> `cat /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid` >/dev/null >> 2>&1', 'user': 'yarn'} >> 2014-11-03 13:27:58,089 - Execute['ls >> /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid >/dev/null 2>&1 && ps >> `cat /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid` >/dev/null >> 2>&1'] {'initial_wait': 5, 'not_if': 'ls >> /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid >/dev/null 2>&1 && ps >> `cat /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid` >/dev/null >> 2>&1', 'user': 'yarn'} >> 2014-11-03 13:28:03,199 - Error while executing command 'restart': >> Traceback (most recent call last): >> File >> "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", >> line 111, in execute >> method(env) >> File >> "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", >> line 212, in restart >> self.start(env) >> File >> "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/application_timeline_server.py", >> line 42, in start >> service('historyserver', action='start') >> File >> "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/service.py", >> line 51, in service >> initial_wait=5 >> File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", >> line 148, in __init__ >> self.env.run() >> File >> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", >> line 149, in run >> self.run_action(resource, action) >> File >> "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", >> line 115, in run_action >> provider_action() >> File >> "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", >> line 239, in action_run >> raise ex >> Fail: Execution of 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid >> >/dev/null 2>&1 && ps `cat >> /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid` >/dev/null 2>&1' >> returned 1. >> >> It seems this is a known issue according to >> http://docs.hortonworks.com/HDPDocuments/Ambari-1.6.1.0/bk_releasenotes_ambari_1.6.1/content/ch_relnotes-ambari-1.6.1.0-knownissues.html >> , >> >> I checked my environment, it is configured with the default value of >> org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore >> for yarn.timeline-service.store-class, and I can't determine which >> version of HDP ambari-server has installed, so I tried >> org.apache.hadoop.yarn.server.applicationhistoryservice.timeline.LeveldbTimelineStore >> and org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore, but >> both failed with the same problem, can you help with this, and another >> questions is how can determine which version of HDP is installed? >> >> Thanks >> >> > > > -- > Cheers > -MJ > -- Cheers -MJ
