Re: How to submit the patch MAPREDUCE-4490.patch which works for branch-1.2, not trunk?

2014-02-16 Thread sam liu
Hi Arpit,

Thanks for your guide!  As a new contributor, I still have following two
questions need your help:

1) So I guess I should run 'test-patch.sh' on my local environment against
branch-1.2, not on Apache Hadoop test server, right?
2) On branch-1.2, I found the 'test-patch.sh' is on
./src/test/bin/test-patch.sh, not ./dev-support/test-patch.sh. My command
is 'sh ./src/test/bin/test-patch.sh MAPREDUCE-4490.patch', however failed
with message 'ERROR: usage ./src/test/bin/test-patch.sh HUDSON [args] |
DEVELOPER [args]'. What's the correct way to manually run 'test-patch.sh'?




2014-02-15 5:25 GMT+08:00 Arpit Agarwal aagar...@hortonworks.com:

 Hi Sam,

 Hadoop Jenkins does not accept patches for 1.x.

 You can manually run 'test-patch.sh' to verify there are no regressions
 introduced by your patch and copy-paste the results into a Jira comment.


 On Thu, Feb 13, 2014 at 10:50 PM, sam liu samliuhad...@gmail.com wrote:

 Hi Experts,

 I have been working on the JIRA
 https://issues.apache.org/jira/browse/MAPREDUCE-4490 and attached
 MAPREDUCE-4490.patch which could fix this jira. I would like to contribute
 my patch to community, but encountered some issues.

 MAPREDUCE-4490 is an issue on Hadoop-1.x versions, and my patch based on
 the  latest code of origin/branch-1.2. However, current trunk bases on Yarn
 and does not has such issue any more. So my patch could not be applied on
 current trunk code, and it's actually no need to generate a similar patch
 on trunk at all.

 How to submit the patch MAPREDUCE-4490.patch only to origin/branch-1.2,
 not trunk? Is it allowed by Apache Hadoop?

 Thanks!



 CONFIDENTIALITY NOTICE
 NOTICE: This message is intended for the use of the individual or entity
 to which it is addressed and may contain information that is confidential,
 privileged and exempt from disclosure under applicable law. If the reader
 of this message is not the intended recipient, you are hereby notified that
 any printing, copying, dissemination, distribution, disclosure or
 forwarding of this communication is strictly prohibited. If you have
 received this communication in error, please contact the sender immediately
 and delete it from your system. Thank You.


Start hadoop service

2014-02-16 Thread EdwardKing
I have install hadoop-2.2.0 under two machine,one is master and other is 
slave,then I start hadoop service under master machine.
[hadoop@master ~]$./start-dfs.sh
[hadoop@master ~]$./start-yarn.sh
[hadoop@master ~]$./mr-jobhistory-daemon.sh start historyserver

My question is whether I need start hadoop service under slave machine again? 
Thanks.
[hadoop@slave ~]$./start-dfs.sh
[hadoop@slave ~]$./start-yarn.sh
[hadoop@slave ~]$./mr-jobhistory-daemon.sh start historyserver





---
Confidentiality Notice: The information contained in this e-mail and any 
accompanying attachment(s) 
is intended only for the use of the intended recipient and may be confidential 
and/or privileged of 
Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of 
this communication is 
not the intended recipient, unauthorized use, forwarding, printing,  storing, 
disclosure or copying 
is strictly prohibited, and may be unlawful.If you have received this 
communication in error,please 
immediately notify the sender by return e-mail, and delete the original message 
and all copies from 
your system. Thank you. 
---


Re: Start hadoop service

2014-02-16 Thread Anand Mundada
No. 
You don't need to.
Master will start all required daemons on slave.

Check all daemons using jps command.

Sent from my iPhone

 On Feb 16, 2014, at 7:03 PM, EdwardKing zhan...@neusoft.com wrote:
 
 I have install hadoop-2.2.0 under two machine,one is master and other is 
 slave,then I start hadoop service under master machine.
 [hadoop@master ~]$./start-dfs.sh
 [hadoop@master ~]$./start-yarn.sh
 [hadoop@master ~]$./mr-jobhistory-daemon.sh start historyserver
  
 My question is whether I need start hadoop service under slave machine again? 
 Thanks.
 [hadoop@slave ~]$./start-dfs.sh
 [hadoop@slave ~]$./start-yarn.sh
 [hadoop@slave ~]$./mr-jobhistory-daemon.sh start historyserver
  
  
  
  
  
  
 ---
 Confidentiality Notice: The information contained in this e-mail and any 
 accompanying attachment(s) 
 is intended only for the use of the intended recipient and may be 
 confidential and/or privileged of 
 Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of 
 this communication is 
 not the intended recipient, unauthorized use, forwarding, printing,  storing, 
 disclosure or copying 
 is strictly prohibited, and may be unlawful.If you have received this 
 communication in error,please 
 immediately notify the sender by return e-mail, and delete the original 
 message and all copies from 
 your system. Thank you. 
 ---


Re: Start hadoop service

2014-02-16 Thread EdwardKing
My OS is CenterOS 2.6.18, master IP is 172.11.12.6 and slave IP is 172.11.12.7
First I start hadoop service under master,like follows:
[hadoop@master ~]$ cd /home/software/hadoop-2.2.0/sbin
[hadoop@master ~]$./start-dfs.sh
14/02/16 17:24:16 WARN util.NativeCodeLoader: Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
Starting namenodes on [master]
master: starting namenode, logging to 
/home/software/hadoop-2.2.0/logs/hadoop-hadoop-namenode-master.out
master: starting datanode, logging to 
/home/software/hadoop-2.2.0/logs/hadoop-hadoop-datanode-master.out
Starting secondary namenodes [master]
master: starting secondarynamenode, logging to 
/home/software/hadoop-2.2.0/logs/hadoop-hadoop-secondarynamenode-master.out
14/02/16 17:24:43 WARN util.NativeCodeLoader: Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable

[hadoop@master ~]$./start-yarn.sh
starting yarn daemons
starting resourcemanager, logging to 
/home/software/hadoop-2.2.0/logs/yarn-hadoop-resourcemanager-master.out
master: starting nodemanager, logging to 
/home/software/hadoop-2.2.0/logs/yarn-hadoop-nodemanager-master.out

[hadoop@master ~]$./mr-jobhistory-daemon.sh start historyserver
starting historyserver, logging to 
/home/software/hadoop-2.2.0/logs/mapred-hadoop-historyserver-master.out

[hadoop@master ~]$jps
4439 DataNode
5400 Jps
4884 NodeManager
4331 NameNode
5342 JobHistoryServer
4595 SecondaryNameNode

Then I open firefox to view hadoop service,like follows:
http://172.11.12.6:9002/dfshealth.jsp
NameNode 'master:9000' (active)
..
Live Nodes  : 1 (Decommissioned: 0)


http://172.11.12.6:8088/cluster
Firefox can't establish a connection to the server at 172.11.12.6:8088.

http://172.11.12.6:9002/dfsnodelist.jsp?whatNodes=LIVE
Live Datanodes : 1
master 172.11.12.6:50010 ..

If I start hadoop service under slave, then I open firefox to view hadoop 
service correct,like follows:
http://172.11.12.6:9002/dfsnodelist.jsp?whatNodes=LIVE
Live Datanodes : 2
master 172.11.12.6:50010 ..
slave  172.11.12.7:

Why need I start hadoop service again under slave? Which configuration file is 
wrong? Thanks



- Original Message - 
From: Anand Mundada 
To: user@hadoop.apache.org 
Cc: user@hadoop.apache.org 
Sent: Monday, February 17, 2014 9:12 AM
Subject: Re: Start hadoop service 


No. 
You don't need to.
Master will start all required daemons on slave.


Check all daemons using jps command.

Sent from my iPhone

On Feb 16, 2014, at 7:03 PM, EdwardKing zhan...@neusoft.com wrote:


I have install hadoop-2.2.0 under two machine,one is master and other is 
slave,then I start hadoop service under master machine.
[hadoop@master ~]$./start-dfs.sh
[hadoop@master ~]$./start-yarn.sh
[hadoop@master ~]$./mr-jobhistory-daemon.sh start historyserver

My question is whether I need start hadoop service under slave machine again? 
Thanks.
[hadoop@slave ~]$./start-dfs.sh
[hadoop@slave ~]$./start-yarn.sh
[hadoop@slave ~]$./mr-jobhistory-daemon.sh start historyserver






---
Confidentiality Notice: The information contained in this e-mail and any 
accompanying attachment(s) 
is intended only for the use of the intended recipient and may be confidential 
and/or privileged of 
Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of 
this communication is 
not the intended recipient, unauthorized use, forwarding, printing,  storing, 
disclosure or copying 
is strictly prohibited, and may be unlawful.If you have received this 
communication in error,please 
immediately notify the sender by return e-mail, and delete the original message 
and all copies from 
your system. Thank you. 
---
---
Confidentiality Notice: The information contained in this e-mail and any 
accompanying attachment(s) 
is intended only for the use of the intended recipient and may be confidential 
and/or privileged of 
Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of 
this communication is 
not the intended recipient, unauthorized use, forwarding, printing,  storing, 
disclosure or copying 
is strictly prohibited, and may be unlawful.If you have received this 
communication in error,please 
immediately notify the sender by return e-mail, and delete the original message 
and all copies from 
your system. Thank you. 
---


Aggregation service start

2014-02-16 Thread EdwardKing
hadoop 2.2.0, I want to view Tracking UI,so I visit 
http://172.11.12.6:8088/cluster,
then I click History of Completed Job,such as follows:

MapReduce Job job_1392601388579_0001
Attempt Number  Start Time NodeLogs
1   Sun Feb 16 17:44:57 PST 2014  master:8042  logs

Then I click logs,but it failed.
Aggregation is not enabled. Try the nodemanager at master:8994

I guess it must a service don't start, which command I need to execute under 
home/software/hadoop-2.2.0/sbin ?  Thanks.
[hadoop@node1 sbin]$ ls
distribute-exclude.shstart-all.cmdstop-all.sh
hadoop-daemon.sh start-all.sh stop-balancer.sh
hadoop-daemons.shstart-balancer.shstop-dfs.cmd
hdfs-config.cmd  start-dfs.cmdstop-dfs.sh
hdfs-config.sh   start-dfs.sh stop-secure-dns.sh
httpfs.shstart-secure-dns.sh  stop-yarn.cmd
mr-jobhistory-daemon.sh  start-yarn.cmd   stop-yarn.sh
refresh-namenodes.sh start-yarn.shyarn-daemon.sh
slaves.shstop-all.cmd yarn-daemons.sh







---
Confidentiality Notice: The information contained in this e-mail and any 
accompanying attachment(s) 
is intended only for the use of the intended recipient and may be confidential 
and/or privileged of 
Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of 
this communication is 
not the intended recipient, unauthorized use, forwarding, printing,  storing, 
disclosure or copying 
is strictly prohibited, and may be unlawful.If you have received this 
communication in error,please 
immediately notify the sender by return e-mail, and delete the original message 
and all copies from 
your system. Thank you. 
---


Re: Aggregation service start

2014-02-16 Thread Zhijie Shen
Please  set

property
nameyarn.log-aggregation-enable/name
valuetrue/value
/property

in yarn-site.xml to enable log aggregation.

-Zhijie
 On Feb 16, 2014 6:15 PM, EdwardKing zhan...@neusoft.com wrote:

 hadoop 2.2.0, I want to view Tracking UI,so I visit
 http://172.11.12.6:8088/cluster,
 then I click History of Completed Job,such as follows:

 MapReduce Job job_1392601388579_0001
 Attempt Number  Start Time NodeLogs
 1   Sun Feb 16 17:44:57 PST 2014  master:8042  logs

 Then I click logs,but it failed.
 Aggregation is not enabled. Try the nodemanager at master:8994

 I guess it must a service don't start, which command I need to execute
 under home/software/hadoop-2.2.0/sbin ?  Thanks.
 [hadoop@node1 sbin]$ ls
 distribute-exclude.shstart-all.cmdstop-all.sh
 hadoop-daemon.sh start-all.sh stop-balancer.sh
 hadoop-daemons.shstart-balancer.shstop-dfs.cmd
 hdfs-config.cmd  start-dfs.cmdstop-dfs.sh
 hdfs-config.sh   start-dfs.sh stop-secure-dns.sh
 httpfs.shstart-secure-dns.sh  stop-yarn.cmd
 mr-jobhistory-daemon.sh  start-yarn.cmd   stop-yarn.sh
 refresh-namenodes.sh start-yarn.shyarn-daemon.sh
 slaves.shstop-all.cmd yarn-daemons.sh








 ---
 Confidentiality Notice: The information contained in this e-mail and any
 accompanying attachment(s)
 is intended only for the use of the intended recipient and may be
 confidential and/or privileged of
 Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader
 of this communication is
 not the intended recipient, unauthorized use, forwarding, printing,
  storing, disclosure or copying
 is strictly prohibited, and may be unlawful.If you have received this
 communication in error,please
 immediately notify the sender by return e-mail, and delete the original
 message and all copies from
 your system. Thank you.

 ---


-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.


RE: Bigdata - MapR, Cloudera, Hortonworks, OracleBigdata appliances

2014-02-16 Thread Nirmal Kumar
Hi All,

This information and comparison in the inadvertently shared  link 
(https://docs.google.com/spreadsheet/ccc?key=0AjfuzftCi_w7dFNVQXBnSHUtc25NV1UxSHppN2dvckE#gid=0)
  has NOT been written or is endorsed by me or anybody in Impetus - my 
employer. I and my employer have not participated in creation of this content 
and nor do we vouch for the accuracy of the contents. This report is from a 
publicly available document on the internet by a company called flux7.com - 
please connect to flux7.com for any concerns around the report contents. I 
apologize if the validity  of the shared information has caused concerns to  
anybody.

Please also note that all information and opinions shared in this user list are 
my own and not a representation of my employers opinion or thought process in 
any way.

Thanks,
-Nirmal

From: Nirmal Kumar
Sent: Saturday, February 15, 2014 1:44 PM
To: 'user@hadoop.apache.org'
Cc: 'vjshal...@gmail.com'
Subject: RE: Bigdata - MapR, Cloudera, Hortonworks, OracleBigdata appliances

All,

Apologies for any inconvenience caused.

Request kindly to *ignore* the below info provided by me.

From: Nirmal Kumar
Sent: Friday, February 14, 2014 6:01 PM
To: Hadoop User Mailer List
Cc: 'vjshal...@gmail.com'
Subject: RE: Bigdata - MapR, Cloudera, Hortonworks, OracleBigdata appliances

Detailed comparison of various Hadoop distros:

https://docs.google.com/spreadsheet/ccc?key=0AjfuzftCi_w7dFNVQXBnSHUtc25NV1UxSHppN2dvckE#gid=0

-Nirmal

From: VJ Shalish [mailto:vjshal...@gmail.com]
Sent: Friday, February 14, 2014 5:20 PM
To: Hadoop User Mailer List
Subject: Bigdata - MapR, Cloudera, Hortonworks, OracleBigdata appliances

Can anyone send me valid comparison points or links taking into consideration
Bigdata - MapR, Cloudera, Hortonworks, OracleBigdata appliances etc
Thanks
Shalish.








NOTE: This message may contain information that is confidential, proprietary, 
privileged or otherwise protected by law. The message is intended solely for 
the named addressee. If received in error, please destroy and notify the 
sender. Any use of this email is prohibited when received in error. Impetus 
does not represent, warrant and/or guarantee, that the integrity of this 
communication has been maintained nor that the communication is free of errors, 
virus, interception or interference.


Re: Yarn - specify hosts in ContainerRequest

2014-02-16 Thread Krishna Kishore Bonagiri
Hi Anand,

  Which version of Hadoop are you using? It works from 2.2.0

Try like this, and it should work. I am using this feature on 2.2.0

   String[] hosts = new String[1];
   hosts[0] = node_name;
   ContainerRequest request = new ContainerRequest(capability, hosts,
null, p, false);


Thanks,
Kishore


On Fri, Feb 14, 2014 at 11:43 PM, Anand Mundada anandmund...@ymail.comwrote:

 Hi All,

 How can I launch container on a particular host?
 I tried specifying host name in
 *new ContainerRequest()*

 Thanks,
 Anand



Re: Aggregation service start

2014-02-16 Thread EdwardKing
Thanks for you help. I set yarn-site.xml as you told me,like follows:

[hadoop@master hadoop]$ cat yarn-site.xml
?xml version=1.0?
!--
  Licensed under the Apache License, Version 2.0 (the License);
  you may not use this file except in compliance with the License.
  You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

  Unless required by applicable law or agreed to in writing, software
  distributed under the License is distributed on an AS IS BASIS,
  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  See the License for the specific language governing permissions and
  limitations under the License. See accompanying LICENSE file.
--
configuration

!-- Site specific YARN configuration properties --
property
  nameyarn.resourcemanager.resource-tracker.address/name
  valuemaster:8990/value
  descriptionhost is the hostname of the resource manager and port is the 
port on which the NodeManagers contact the Resource Manager.
  /description
/property
property
  nameyarn.resourcemanager.scheduler.address/name
  valuemaster:8991/value
  descriptionhost is the hostname of the resourcemanager and port is the port 
on which the Applications in the cluster talk to the Resource Manager.
  /description
/property
property
  nameyarn.resourcemanager.scheduler.class/name
  
valueorg.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler/value
  descriptionIn case you do not want to use the default 
scheduler/description
/property
property
  nameyarn.resourcemanager.address/name
  valuemaster:8993/value
  descriptionthe host is the hostname of the ResourceManager and port is the 
port on which the clients can talk to the Resource Manager /description
/property
property
  nameyarn.nodemanager.local-dirs/name
  value/home/software/tmp/node/value
  descriptionthe local directions used by the nodemanager/description
/property
property
  nameyarn.nodemanager.address/name
  valuemaster:8994/value
  descriptionthe nodemanagers bind to this port/description
/property
property
  nameyarn.nodemanager.resource.memory-mb/name
  value5120/value
  descriptionthe amount of memory on the NodeManager in GB/description
/property
property
  nameyarn.nodemanager.remote-app-log-dir/name
  value/home/software/tmp/app-logs/value
  descriptiondirectory on hdfs where the application logs are moved to 
/description
/property
property
  nameyarn.nodemanager.log-dirs/name
  value/home/software/tmp/node/value
  descriptionthe directories used by Nodemanager as log 
directories/description
/property
property
  nameyarn.nodemanager.aux-services/name
  valuemapreduce_shuffle/value
  descriptionshuffle service that needs to be set for Map Reduce to run 
/description
/property
  
property
  nameyarn.log-aggregation-enable/name
  valuetrue/value
/property

/configuration

Then I submit a job,when this job is running, I click History of Tracking UI 
http://172.11.12.6:8088/cluster
I can view all log information. It runs ok.
But when the job complete, then I click History of Tracking UI 
http://172.11.12.6:8088/cluster again, it raise following error:

Firefox can't establish a connection to the server at master:19888.

Do I missing some configuration information in my xml file?  How to correct?  
Thanks in advance.








- Original Message - 
From: Zhijie Shen 
To: user@hadoop.apache.org 
Sent: Monday, February 17, 2014 11:11 AM
Subject: Re: Aggregation service start


Please  set
property
nameyarn.log-aggregation-enable/name
valuetrue/value
/property 
in yarn-site.xml to enable log aggregation.
-Zhijie

On Feb 16, 2014 6:15 PM, EdwardKing zhan...@neusoft.com wrote:

hadoop 2.2.0, I want to view Tracking UI,so I visit 
http://172.11.12.6:8088/cluster,
then I click History of Completed Job,such as follows:

MapReduce Job job_1392601388579_0001
Attempt Number  Start Time NodeLogs
1   Sun Feb 16 17:44:57 PST 2014  master:8042  logs

Then I click logs,but it failed.
Aggregation is not enabled. Try the nodemanager at master:8994

I guess it must a service don't start, which command I need to execute under 
home/software/hadoop-2.2.0/sbin ?  Thanks.
[hadoop@node1 sbin]$ ls
distribute-exclude.shstart-all.cmdstop-all.sh
hadoop-daemon.sh start-all.sh stop-balancer.sh
hadoop-daemons.shstart-balancer.shstop-dfs.cmd
hdfs-config.cmd  start-dfs.cmdstop-dfs.sh
hdfs-config.sh   start-dfs.sh stop-secure-dns.sh
httpfs.shstart-secure-dns.sh  stop-yarn.cmd
mr-jobhistory-daemon.sh  start-yarn.cmd   stop-yarn.sh
refresh-namenodes.sh start-yarn.shyarn-daemon.sh
slaves.shstop-all.cmd yarn-daemons.sh







---
Confidentiality Notice: The information contained in this e-mail and any 
accompanying attachment(s)
is intended only 

Re: Aggregation service start

2014-02-16 Thread Zhijie Shen
But when the job complete, then I click History of Tracking UI
http://172.11.12.6:8088/cluster again, it raise following error:

Firefox can't establish a connection to the server at master:19888.

This is a problem other than log aggregation. After an MapReduce job
completes, the tracking URL is pointing to the MapReduce history server. It
is very likely that the history server hasn't been running on your machine,
such that you didn't get response from it.

- Zhijie


On Sun, Feb 16, 2014 at 10:03 PM, EdwardKing zhan...@neusoft.com wrote:

 Thanks for you help. I set yarn-site.xml as you told me,like follows:

 [hadoop@master hadoop]$ cat yarn-site.xml
 ?xml version=1.0?
 !--
   Licensed under the Apache License, Version 2.0 (the License);
   you may not use this file except in compliance with the License.
   You may obtain a copy of the License at

 http://www.apache.org/licenses/LICENSE-2.0

   Unless required by applicable law or agreed to in writing, software
   distributed under the License is distributed on an AS IS BASIS,
   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
   See the License for the specific language governing permissions and
   limitations under the License. See accompanying LICENSE file.
 --
 configuration

 !-- Site specific YARN configuration properties --
 property
   nameyarn.resourcemanager.resource-tracker.address/name
   valuemaster:8990/value
   descriptionhost is the hostname of the resource manager and port is
 the port on which the NodeManagers contact the Resource Manager.
   /description
 /property
 property
   nameyarn.resourcemanager.scheduler.address/name
   valuemaster:8991/value
   descriptionhost is the hostname of the resourcemanager and port is the
 port on which the Applications in the cluster talk to the Resource Manager.
   /description
 /property
 property
   nameyarn.resourcemanager.scheduler.class/name

 valueorg.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler/value
   descriptionIn case you do not want to use the default
 scheduler/description
 /property
 property
   nameyarn.resourcemanager.address/name
   valuemaster:8993/value
   descriptionthe host is the hostname of the ResourceManager and port is
 the port on which the clients can talk to the Resource Manager
 /description
 /property
 property
   nameyarn.nodemanager.local-dirs/name
   value/home/software/tmp/node/value
   descriptionthe local directions used by the nodemanager/description
 /property
 property
   nameyarn.nodemanager.address/name
   valuemaster:8994/value
   descriptionthe nodemanagers bind to this port/description
 /property
 property
   nameyarn.nodemanager.resource.memory-mb/name
   value5120/value
   descriptionthe amount of memory on the NodeManager in GB/description
 /property
 property
   nameyarn.nodemanager.remote-app-log-dir/name
   value/home/software/tmp/app-logs/value
   descriptiondirectory on hdfs where the application logs are moved to
 /description
 /property
 property
   nameyarn.nodemanager.log-dirs/name
   value/home/software/tmp/node/value
   descriptionthe directories used by Nodemanager as log
 directories/description
 /property
 property
   nameyarn.nodemanager.aux-services/name
   valuemapreduce_shuffle/value
   descriptionshuffle service that needs to be set for Map Reduce to run
 /description
 /property

 property
   nameyarn.log-aggregation-enable/name
   valuetrue/value
 /property

 /configuration

 Then I submit a job,when this job is running, I click History of Tracking
 UI http://172.11.12.6:8088/cluster
 I can view all log information. It runs ok.
 But when the job complete, then I click History of Tracking UI
 http://172.11.12.6:8088/cluster again, it raise following error:

 Firefox can't establish a connection to the server at master:19888.

 Do I missing some configuration information in my xml file?  How to
 correct?  Thanks in advance.








 - Original Message -
 From: Zhijie Shen
 To: user@hadoop.apache.org
 Sent: Monday, February 17, 2014 11:11 AM
 Subject: Re: Aggregation service start


 Please  set
 property
 nameyarn.log-aggregation-enable/name
 valuetrue/value
 /property
 in yarn-site.xml to enable log aggregation.
 -Zhijie

 On Feb 16, 2014 6:15 PM, EdwardKing zhan...@neusoft.com wrote:

 hadoop 2.2.0, I want to view Tracking UI,so I visit
 http://172.11.12.6:8088/cluster,
 then I click History of Completed Job,such as follows:

 MapReduce Job job_1392601388579_0001
 Attempt Number  Start Time NodeLogs
 1   Sun Feb 16 17:44:57 PST 2014  master:8042  logs

 Then I click logs,but it failed.
 Aggregation is not enabled. Try the nodemanager at master:8994

 I guess it must a service don't start, which command I need to execute
 under home/software/hadoop-2.2.0/sbin ?  Thanks.
 [hadoop@node1 sbin]$ ls
 distribute-exclude.shstart-all.cmdstop-all.sh
 hadoop-daemon.sh start-all.sh stop-balancer.sh