If you have not made this working yet you can use this blueprint - https://raw.githubusercontent.com/sequenceiq/sequenceiq-samples/master/timeline-server/blueprints/multi-node-hdfs-yarn-tez-timeline-service.json Also if you are looking for further information you can check this post (http://blog.sequenceiq.com/blog/2014/11/04/yarn-timeline-service-tez/) - it’s in the context of Tez but same applies to MR2.
Janos On 04 Nov 2014, at 08:39, Mingjiang Shi <m...@pivotal.io> wrote: > Did you attempt to start the services manually using root as I see some > folders owned by root? > > On Mon, Nov 3, 2014 at 6:18 PM, guxiaobo1982 <guxiaobo1...@qq.com> wrote: > there are some error messages in the yarn-yarn-historyserver-lix1.bh.com.log > file: > > > > ************************************************************/ > > 2014-11-03 12:07:34,974 INFO > applicationhistoryservice.ApplicationHistoryServer > (StringUtils.java:startupShutdownMessage(614)) - STARTUP_MSG: > > /************************************************************ > > STARTUP_MSG: Starting ApplicationHistoryServer > > STARTUP_MSG: host = lix1.bh.com/192.168.100.3 > > STARTUP_MSG: args = [] > > STARTUP_MSG: version = 2.4.0.2.1.5.0-695 > > STARTUP_MSG: classpath = > /etc/hadoop/conf:/etc/hadoop/conf:/etc/hadoop/conf:/usr/lib/hadoop/lib/jersey-core-1.9.jar:/usr/lib/hadoop/lib/mockito-all > > -1.8.5.jar: > > > > ....... > > > > STARTUP_MSG: build = g...@github.com:hortonworks/hadoop.git -r > c11220208321e1835912fde828f1038eedb1afae; compiled by 'jenkins' on > 2014-08-28T03:10Z > > STARTUP_MSG: java = 1.7.0_45 > > ************************************************************/ > > 2014-11-03 12:07:35,039 INFO > applicationhistoryservice.ApplicationHistoryServer > (SignalLogger.java:register(91)) - registered UNIX signal handlers f > > or [TERM, HUP, INT] > > 2014-11-03 12:07:36,977 INFO impl.MetricsConfig > (MetricsConfig.java:loadFirst(111)) - loaded properties from > hadoop-metrics2.properties > > 2014-11-03 12:07:37,208 INFO impl.MetricsSystemImpl > (MetricsSystemImpl.java:startTimer(355)) - Scheduled snapshot period at 60 > second(s). > > 2014-11-03 12:07:37,208 INFO impl.MetricsSystemImpl > (MetricsSystemImpl.java:start(183)) - ApplicationHistoryServer metrics system > started > > 2014-11-03 12:07:37,230 INFO > applicationhistoryservice.ApplicationHistoryManagerImpl > (ApplicationHistoryManagerImpl.java:serviceInit(61)) - Applicat > > ionHistory Init > > 2014-11-03 12:07:37,615 INFO timeline.LeveldbTimelineStore > (LeveldbTimelineStore.java:serviceInit(194)) - Using leveldb path > /hadoop/yarn/timeline/l > > eveldb-timeline-store.ldb > > 2014-11-03 12:07:37,647 INFO service.AbstractService > (AbstractService.java:noteFailure(272)) - Service > org.apache.hadoop.yarn.server.timeline.Leveld > > bTimelineStore failed in state INITED; cause: > org.fusesource.leveldbjni.internal.NativeDB$DBException: IO error: > /hadoop/yarn/timeline/leveldb-timeli > > ne-store.ldb/LOCK: Permission denied > > org.fusesource.leveldbjni.internal.NativeDB$DBException: IO error: > /hadoop/yarn/timeline/leveldb-timeline-store.ldb/LOCK: Permission denied > > at org.fusesource.leveldbjni.internal.NativeDB.checkStatus(NativeDB.java:200) > > at org.fusesource.leveldbjni.internal.NativeDB.open(NativeDB.java:218) > > at org.fusesource.leveldbjni.JniDBFactory.open(JniDBFactory.java:168) > > at > org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore.serviceInit(LeveldbTimelineStore.java:195) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > > at > org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:88) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:145) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:155) > > 2014-11-03 12:07:37,654 INFO service.AbstractService > (AbstractService.java:noteFailure(272)) - Service > org.apache.hadoop.yarn.server.applicationhist > > oryservice.ApplicationHistoryServer failed in state INITED; cause: > org.apache.hadoop.service.ServiceStateException: > org.fusesource.leveldbjni.interna > > l.NativeDB$DBException: IO error: > /hadoop/yarn/timeline/leveldb-timeline-store.ldb/LOCK: Permission denied > > org.apache.hadoop.service.ServiceStateException: > org.fusesource.leveldbjni.internal.NativeDB$DBException: IO error: > /hadoop/yarn/timeline/leveldb-tim > > eline-store.ldb/LOCK: Permission denied > > at > org.apache.hadoop.service.ServiceStateException.convert(ServiceStateException.java:59) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:172) > > at > org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:88) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:145) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:155) > > Caused by: org.fusesource.leveldbjni.internal.NativeDB$DBException: IO error: > /hadoop/yarn/timeline/leveldb-timeline-store.ldb/LOCK: Permission denie > > d > > at org.fusesource.leveldbjni.internal.NativeDB.checkStatus(NativeDB.java:200) > > at org.fusesource.leveldbjni.internal.NativeDB.open(NativeDB.java:218) > > at org.fusesource.leveldbjni.JniDBFactory.open(JniDBFactory.java:168) > > at > org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore.serviceInit(LeveldbTimelineStore.java:195) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > > ... 5 more > > 2014-11-03 12:07:37,655 INFO impl.MetricsSystemImpl > (MetricsSystemImpl.java:stop(200)) - Stopping ApplicationHistoryServer > metrics system... > > 2014-11-03 12:07:37,656 INFO impl.MetricsSystemImpl > (MetricsSystemImpl.java:stop(206)) - ApplicationHistoryServer metrics system > stopped. > > 2014-11-03 12:07:37,656 INFO impl.MetricsSystemImpl > (MetricsSystemImpl.java:shutdown(583)) - ApplicationHistoryServer metrics > system shutdown complete. > > 2014-11-03 12:07:37,656 INFO > applicationhistoryservice.ApplicationHistoryManagerImpl > (ApplicationHistoryManagerImpl.java:serviceStop(78)) - Stopping > ApplicationHistory > > 2014-11-03 12:07:37,657 FATAL > applicationhistoryservice.ApplicationHistoryServer > (ApplicationHistoryServer.java:launchAppHistoryServer(148)) - Error starting > ApplicationHistoryServer > > org.apache.hadoop.service.ServiceStateException: > org.fusesource.leveldbjni.internal.NativeDB$DBException: IO error: > /hadoop/yarn/timeline/leveldb-timeline-store.ldb/LOCK: Permission denied > > at > org.apache.hadoop.service.ServiceStateException.convert(ServiceStateException.java:59) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:172) > > at > org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:88) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:145) > > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:155) > > Caused by: org.fusesource.leveldbjni.internal.NativeDB$DBException: IO error: > /hadoop/yarn/timeline/leveldb-timeline-store.ldb/LOCK: Permission denied > > at org.fusesource.leveldbjni.internal.NativeDB.checkStatus(NativeDB.java:200) > > at org.fusesource.leveldbjni.internal.NativeDB.open(NativeDB.java:218) > > at org.fusesource.leveldbjni.JniDBFactory.open(JniDBFactory.java:168) > > at > org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore.serviceInit(LeveldbTimelineStore.java:195) > > at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > > ... 5 more > > 2014-11-03 12:07:37,664 INFO util.ExitUtil (ExitUtil.java:terminate(124)) - > Exiting with status -1 > > 2014-11-03 12:07:37,668 INFO > applicationhistoryservice.ApplicationHistoryServer > (StringUtils.java:run(640)) - SHUTDOWN_MSG: > > /************************************************************ > > SHUTDOWN_MSG: Shutting down ApplicationHistoryServer at > lix1.bh.com/192.168.100.3 > > > > And I list the ownership for the /hadoop path > > > > [root@lix1 hadoop]# pwd > > /hadoop > > [root@lix1 hadoop]# ls > > falcon hbase hdfs oozie storm yarn zookeeper > > [root@lix1 hadoop]# ls -al > > 总用量 36 > > drwxr-xr-x. 9 root root 4096 11月 3 11:58 . > > dr-xr-xr-x. 28 root root 4096 11月 3 18:01 .. > > drwxr-xr-x. 4 falcon root 4096 11月 3 11:54 falcon > > drwxr-xr-x. 3 hbase root 4096 11月 3 10:34 hbase > > drwxr-xr-x. 5 root root 4096 11月 3 11:58 hdfs > > drwxr-xr-x. 3 root root 4096 11月 3 11:58 oozie > > drwxr-xr-x. 5 storm hadoop 4096 11月 3 12:05 storm > > drwxr-xr-x. 5 root root 4096 10月 14 18:34 yarn > > drwxr-xr-x. 3 zookeeper hadoop 4096 11月 3 11:51 zookeeper > > [root@lix1 hadoop]# ls -al yarn/ > > 总用量 20 > > drwxr-xr-x. 5 root root 4096 10月 14 18:34 . > > drwxr-xr-x. 9 root root 4096 11月 3 11:58 .. > > drwxr-xr-x. 6 yarn root 4096 11月 3 12:58 local > > drwxr-xr-x. 2 yarn root 4096 11月 3 12:09 log > > drwxr-xr-x. 3 yarn hadoop 4096 10月 14 18:34 timeline > > > [root@lix1 hadoop]# > > > > > > ------------------ Original ------------------ > From: "Mingjiang Shi";<m...@pivotal.io>; > Send time: Monday, Nov 3, 2014 4:36 PM > To: "user@ambari.apache.org"<user@ambari.apache.org>; > Subject: Re: timeline service installed by ambari can't start > > Hi, > Could you check the timeline server log located at > /var/log/hadoop-yarn/yarn/yarn-yarn-timelineserver*.log to see what problem > caused the failure? > > On Mon, Nov 3, 2014 at 3:56 PM, guxiaobo1982 <guxiaobo1...@qq.com> wrote: > The HDFS installed is of version > Version: 2.4.0.2.1.5.0-695, rc11220208321e1835912fde828f1038eedb1afae > > > ------------------ Original ------------------ > From: "guxiaobo1982";<guxiaobo1...@qq.com>; > Send time: Monday, Nov 3, 2014 3:48 PM > To: "user"<user@ambari.apache.org>; > Subject: timeline service installed by ambari can't start > > Hi, > > I use Ambari 16.1 installed HDP 2.1 Single node deployment, but the timeline > service can't start with the following error: > stderr: /var/lib/ambari-agent/data/errors-96.txt > > 2014-11-03 13:28:03,199 - Error while executing command 'restart': > Traceback (most recent call last): > File > "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", > line 111, in execute > method(env) > File > "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", > line 212, in restart > self.start(env) > File > "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/application_timeline_server.py", > line 42, in start > service('historyserver', action='start') > File > "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/service.py", > line 51, in service > initial_wait=5 > File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", > line 148, in __init__ > self.env.run() > File > "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", > line 149, in run > self.run_action(resource, action) > File > "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", > line 115, in run_action > provider_action() > File > "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", > line 239, in action_run > raise ex > Fail: Execution of 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid > >/dev/null 2>&1 && ps `cat > /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid` >/dev/null 2>&1' > returned 1. > stdout: /var/lib/ambari-agent/data/output-96.txt > > 2014-11-03 13:27:56,524 - Execute['mkdir -p /tmp/HDP-artifacts/; curl -kf > -x "" --retry 10 > http://ambari.bh.com:8080/resources//UnlimitedJCEPolicyJDK7.zip -o > /tmp/HDP-artifacts//UnlimitedJCEPolicyJDK7.zip'] {'environment': ..., > 'not_if': 'test -e /tmp/HDP-artifacts//UnlimitedJCEPolicyJDK7.zip', > 'ignore_failures': True, 'path': ['/bin', '/usr/bin/']} > 2014-11-03 13:27:56,543 - Skipping Execute['mkdir -p /tmp/HDP-artifacts/; > curl -kf -x "" --retry 10 > http://ambari.bh.com:8080/resources//UnlimitedJCEPolicyJDK7.zip -o > /tmp/HDP-artifacts//UnlimitedJCEPolicyJDK7.zip'] due to not_if > 2014-11-03 13:27:56,618 - Directory['/etc/hadoop/conf.empty'] {'owner': > 'root', 'group': 'root', 'recursive': True} > 2014-11-03 13:27:56,620 - Link['/etc/hadoop/conf'] {'not_if': 'ls > /etc/hadoop/conf', 'to': '/etc/hadoop/conf.empty'} > 2014-11-03 13:27:56,634 - Skipping Link['/etc/hadoop/conf'] due to not_if > 2014-11-03 13:27:56,644 - File['/etc/hadoop/conf/hadoop-env.sh'] {'content': > Template('hadoop-env.sh.j2'), 'owner': 'hdfs'} > 2014-11-03 13:27:56,646 - XmlConfig['core-site.xml'] {'owner': 'hdfs', > 'group': 'hadoop', 'conf_dir': '/etc/hadoop/conf', 'configurations': ...} > 2014-11-03 13:27:56,650 - Generating config: /etc/hadoop/conf/core-site.xml > 2014-11-03 13:27:56,650 - File['/etc/hadoop/conf/core-site.xml'] {'owner': > 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None} > 2014-11-03 13:27:56,651 - Writing File['/etc/hadoop/conf/core-site.xml'] > because contents don't match > 2014-11-03 13:27:56,662 - Execute['/bin/echo 0 > /selinux/enforce'] > {'only_if': 'test -f /selinux/enforce'} > 2014-11-03 13:27:56,683 - Execute['mkdir -p > /usr/lib/hadoop/lib/native/Linux-i386-32; ln -sf /usr/lib/libsnappy.so > /usr/lib/hadoop/lib/native/Linux-i386-32/libsnappy.so'] {} > 2014-11-03 13:27:56,698 - Execute['mkdir -p > /usr/lib/hadoop/lib/native/Linux-amd64-64; ln -sf /usr/lib64/libsnappy.so > /usr/lib/hadoop/lib/native/Linux-amd64-64/libsnappy.so'] {} > 2014-11-03 13:27:56,709 - Directory['/var/log/hadoop'] {'owner': 'root', > 'group': 'root', 'recursive': True} > 2014-11-03 13:27:56,710 - Directory['/var/run/hadoop'] {'owner': 'root', > 'group': 'root', 'recursive': True} > 2014-11-03 13:27:56,710 - Directory['/tmp/hadoop-hdfs'] {'owner': 'hdfs', > 'recursive': True} > 2014-11-03 13:27:56,714 - File['/etc/hadoop/conf/commons-logging.properties'] > {'content': Template('commons-logging.properties.j2'), 'owner': 'hdfs'} > 2014-11-03 13:27:56,716 - File['/etc/hadoop/conf/health_check'] {'content': > Template('health_check-v2.j2'), 'owner': 'hdfs'} > 2014-11-03 13:27:56,717 - File['/etc/hadoop/conf/log4j.properties'] > {'content': '...', 'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644} > 2014-11-03 13:27:56,720 - File['/etc/hadoop/conf/hadoop-metrics2.properties'] > {'content': Template('hadoop-metrics2.properties.j2'), 'owner': 'hdfs'} > 2014-11-03 13:27:56,720 - File['/etc/hadoop/conf/task-log4j.properties'] > {'content': StaticFile('task-log4j.properties'), 'mode': 0755} > 2014-11-03 13:27:56,721 - File['/etc/hadoop/conf/configuration.xsl'] > {'owner': 'hdfs', 'group': 'hadoop'} > 2014-11-03 13:27:56,803 - Execute['export > HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && > /usr/lib/hadoop-yarn/sbin/yarn-daemon.sh --config /etc/hadoop/conf stop > historyserver'] {'user': 'yarn'} > 2014-11-03 13:27:56,924 - Execute['rm -f > /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid'] {'user': 'yarn'} > 2014-11-03 13:27:56,955 - Directory['/var/run/hadoop-yarn/yarn'] {'owner': > 'yarn', 'group': 'hadoop', 'recursive': True} > 2014-11-03 13:27:56,956 - Directory['/var/log/hadoop-yarn/yarn'] {'owner': > 'yarn', 'group': 'hadoop', 'recursive': True} > 2014-11-03 13:27:56,956 - Directory['/var/run/hadoop-mapreduce/mapred'] > {'owner': 'mapred', 'group': 'hadoop', 'recursive': True} > 2014-11-03 13:27:56,956 - Directory['/var/log/hadoop-mapreduce/mapred'] > {'owner': 'mapred', 'group': 'hadoop', 'recursive': True} > 2014-11-03 13:27:56,956 - Directory['/hadoop/yarn/local'] {'owner': 'yarn', > 'ignore_failures': True, 'recursive': True} > 2014-11-03 13:27:56,956 - Directory['/hadoop/yarn/log'] {'owner': 'yarn', > 'ignore_failures': True, 'recursive': True} > 2014-11-03 13:27:56,957 - Directory['/var/log/hadoop-yarn'] {'owner': 'yarn', > 'ignore_failures': True, 'recursive': True} > 2014-11-03 13:27:56,957 - XmlConfig['core-site.xml'] {'owner': 'hdfs', > 'group': 'hadoop', 'mode': 0644, 'conf_dir': '/etc/hadoop/conf', > 'configurations': ...} > 2014-11-03 13:27:56,963 - Generating config: /etc/hadoop/conf/core-site.xml > 2014-11-03 13:27:56,963 - File['/etc/hadoop/conf/core-site.xml'] {'owner': > 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644} > 2014-11-03 13:27:56,963 - XmlConfig['mapred-site.xml'] {'owner': 'yarn', > 'group': 'hadoop', 'mode': 0644, 'conf_dir': '/etc/hadoop/conf', > 'configurations': ...} > 2014-11-03 13:27:56,966 - Generating config: /etc/hadoop/conf/mapred-site.xml > 2014-11-03 13:27:56,966 - File['/etc/hadoop/conf/mapred-site.xml'] {'owner': > 'yarn', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644} > 2014-11-03 13:27:56,967 - Writing File['/etc/hadoop/conf/mapred-site.xml'] > because contents don't match > 2014-11-03 13:27:56,967 - Changing owner for /etc/hadoop/conf/mapred-site.xml > from 1022 to yarn > 2014-11-03 13:27:56,967 - XmlConfig['yarn-site.xml'] {'owner': 'yarn', > 'group': 'hadoop', 'mode': 0644, 'conf_dir': '/etc/hadoop/conf', > 'configurations': ...} > 2014-11-03 13:27:56,969 - Generating config: /etc/hadoop/conf/yarn-site.xml > 2014-11-03 13:27:56,969 - File['/etc/hadoop/conf/yarn-site.xml'] {'owner': > 'yarn', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644} > 2014-11-03 13:27:56,970 - Writing File['/etc/hadoop/conf/yarn-site.xml'] > because contents don't match > 2014-11-03 13:27:56,971 - XmlConfig['capacity-scheduler.xml'] {'owner': > 'yarn', 'group': 'hadoop', 'mode': 0644, 'conf_dir': '/etc/hadoop/conf', > 'configurations': ...} > 2014-11-03 13:27:56,974 - Generating config: > /etc/hadoop/conf/capacity-scheduler.xml > 2014-11-03 13:27:56,974 - File['/etc/hadoop/conf/capacity-scheduler.xml'] > {'owner': 'yarn', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': > 0644} > 2014-11-03 13:27:56,975 - Writing > File['/etc/hadoop/conf/capacity-scheduler.xml'] because contents don't match > 2014-11-03 13:27:56,975 - Changing owner for > /etc/hadoop/conf/capacity-scheduler.xml from 1021 to yarn > 2014-11-03 13:27:56,975 - Directory['/hadoop/yarn/timeline'] {'owner': > 'yarn', 'group': 'hadoop', 'recursive': True} > 2014-11-03 13:27:56,975 - File['/etc/hadoop/conf/yarn.exclude'] {'owner': > 'yarn', 'group': 'hadoop'} > 2014-11-03 13:27:56,977 - File['/etc/security/limits.d/yarn.conf'] > {'content': Template('yarn.conf.j2'), 'mode': 0644} > 2014-11-03 13:27:56,980 - File['/etc/security/limits.d/mapreduce.conf'] > {'content': Template('mapreduce.conf.j2'), 'mode': 0644} > 2014-11-03 13:27:56,982 - File['/etc/hadoop/conf/yarn-env.sh'] {'content': > Template('yarn-env.sh.j2'), 'owner': 'yarn', 'group': 'hadoop', 'mode': 0755} > 2014-11-03 13:27:56,984 - File['/etc/hadoop/conf/mapred-env.sh'] {'content': > Template('mapred-env.sh.j2'), 'owner': 'hdfs'} > 2014-11-03 13:27:56,985 - File['/etc/hadoop/conf/taskcontroller.cfg'] > {'content': Template('taskcontroller.cfg.j2'), 'owner': 'hdfs'} > 2014-11-03 13:27:56,986 - XmlConfig['mapred-site.xml'] {'owner': 'mapred', > 'group': 'hadoop', 'conf_dir': '/etc/hadoop/conf', 'configurations': ...} > 2014-11-03 13:27:56,988 - Generating config: /etc/hadoop/conf/mapred-site.xml > 2014-11-03 13:27:56,988 - File['/etc/hadoop/conf/mapred-site.xml'] {'owner': > 'mapred', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None} > 2014-11-03 13:27:56,988 - Changing owner for /etc/hadoop/conf/mapred-site.xml > from 1020 to mapred > 2014-11-03 13:27:56,988 - XmlConfig['capacity-scheduler.xml'] {'owner': > 'hdfs', 'group': 'hadoop', 'conf_dir': '/etc/hadoop/conf', 'configurations': > ...} > 2014-11-03 13:27:56,991 - Generating config: > /etc/hadoop/conf/capacity-scheduler.xml > 2014-11-03 13:27:56,991 - File['/etc/hadoop/conf/capacity-scheduler.xml'] > {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': > None} > 2014-11-03 13:27:56,992 - Changing owner for > /etc/hadoop/conf/capacity-scheduler.xml from 1020 to hdfs > 2014-11-03 13:27:56,992 - File['/etc/hadoop/conf/ssl-client.xml.example'] > {'owner': 'mapred', 'group': 'hadoop'} > 2014-11-03 13:27:56,992 - File['/etc/hadoop/conf/ssl-server.xml.example'] > {'owner': 'mapred', 'group': 'hadoop'} > 2014-11-03 13:27:56,993 - Execute['export > HADOOP_LIBEXEC_DIR=/usr/lib/hadoop/libexec && > /usr/lib/hadoop-yarn/sbin/yarn-daemon.sh --config /etc/hadoop/conf start > historyserver'] {'not_if': 'ls > /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid >/dev/null 2>&1 && ps > `cat /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid` >/dev/null 2>&1', > 'user': 'yarn'} > 2014-11-03 13:27:58,089 - Execute['ls > /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid >/dev/null 2>&1 && ps > `cat /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid` >/dev/null 2>&1'] > {'initial_wait': 5, 'not_if': 'ls > /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid >/dev/null 2>&1 && ps > `cat /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid` >/dev/null 2>&1', > 'user': 'yarn'} > 2014-11-03 13:28:03,199 - Error while executing command 'restart': > Traceback (most recent call last): > File > "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", > line 111, in execute > method(env) > File > "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", > line 212, in restart > self.start(env) > File > "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/application_timeline_server.py", > line 42, in start > service('historyserver', action='start') > File > "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/service.py", > line 51, in service > initial_wait=5 > File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", > line 148, in __init__ > self.env.run() > File > "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", > line 149, in run > self.run_action(resource, action) > File > "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", > line 115, in run_action > provider_action() > File > "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", > line 239, in action_run > raise ex > Fail: Execution of 'ls /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid > >/dev/null 2>&1 && ps `cat > /var/run/hadoop-yarn/yarn/yarn-yarn-historyserver.pid` >/dev/null 2>&1' > returned 1. > It seems this is a known issue according to > http://docs.hortonworks.com/HDPDocuments/Ambari-1.6.1.0/bk_releasenotes_ambari_1.6.1/content/ch_relnotes-ambari-1.6.1.0-knownissues.html, > > I checked my environment, it is configured with the default value of > org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore for > yarn.timeline-service.store-class, and I can't determine which version of HDP > ambari-server has installed, so I tried > org.apache.hadoop.yarn.server.applicationhistoryservice.timeline.LeveldbTimelineStore > and org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore, but both > failed with the same problem, can you help with this, and another questions > is how can determine which version of HDP is installed? > > Thanks > > > > > -- > Cheers > -MJ > > > > -- > Cheers > -MJ