[jira] [Commented] (MAPREDUCE-3532) When 0 is provided as port number in yarn.nodemanager.webapp.address, NMs webserver component picks up random port, NM keeps on Reporting 0 port to RM

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186187#comment-13186187
 ] 

Hudson commented on MAPREDUCE-3532:
---

Integrated in Hadoop-Hdfs-trunk #925 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/925/])
MAPREDUCE-3532. Modified NM to report correct http address when an 
ephemeral web port is configured. Contributed by Bhallamudi Venkata Siva Kamesh.

vinodkv : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231342
Files : 
* /hadoop/common/trunk/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/webapp/WebServer.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestNMWebServer.java


 When 0 is provided as port number in yarn.nodemanager.webapp.address, NMs 
 webserver component picks up random port, NM keeps on Reporting 0 port to RM
 --

 Key: MAPREDUCE-3532
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3532
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv2, nodemanager
Affects Versions: 0.23.1
Reporter: Karam Singh
Assignee: Bhallamudi Venkata Siva Kamesh
Priority: Critical
 Fix For: 0.23.1

 Attachments: MAPREDUCE-3532-1.patch, MAPREDUCE-3532.patch


 I tried following -:
 yarn.nodemanager.address=0.0.0.0:0
 yarn.nodemanager.webapp.address=0.0.0.0:0
 yarn.nodemanager.localizer.address=0.0.0.0:0
 mapreduce.shuffle.port=0
 When 0 is provided as number in yarn.nodemanager.webapp.address. 
 NM instantiate WebServer as 0 piort e.g.
 {code}
 2011-12-08 11:33:02,467 INFO 
 org.apache.hadoop.yarn.server.nodemanager.webapp.WebServer: Instantiating 
 NMWebApp at 0.0.0.0:0
 {code}
 After that WebServer pick up some random port e.g.
 {code}
 2011-12-08 11:33:02,562 INFO org.apache.hadoop.http.HttpServer: Jetty bound 
 to port 36272
 2011-12-08 11:33:02,562 INFO org.mortbay.log: jetty-6.1.26
 2011-12-08 11:33:02,831 INFO org.mortbay.log: Started 
 SelectChannelConnector@0.0.0.0:36272
 2011-12-08 11:33:02,831 INFO org.apache.hadoop.yarn.webapp.WebApps: Web app 
 /node started at 36272
 {code}
 And NM WebServer responds correctly but
  RM's cluster/Nodes page shows the following -:
 {code}
 /Rack RUNNING NM:57963 NM:0 Healthy 8-Dec-2011 11:33:01 Healthy 8 12 GB 0 KB
 {code}
 Whereas NM:0 is not clickable.
 Seems even NM's webserver pick random port but it never gets updated and so 
 NM report 0 as HTTP port to RM causing NM Hyperlinks un-clickable
 But verified that MR job runs successfully with random.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (MAPREDUCE-3656) Sort job on 350 scale is consistently failing with latest MRV2 code

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3656?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186189#comment-13186189
 ] 

Hudson commented on MAPREDUCE-3656:
---

Integrated in Hadoop-Hdfs-trunk #925 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/925/])
MAPREDUCE-3656. Fixed a race condition in MR AM which is failing the sort 
benchmark consistently. Contributed by Siddarth Seth.

vinodkv : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231314
Files : 
* /hadoop/common/trunk/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapred/TaskAttemptListenerImpl.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/TaskAttemptListener.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/TaskHeartbeatHandler.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/job/impl/TaskAttemptImpl.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapred/TestTaskAttemptListenerImpl.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapreduce/v2/app/MRApp.java


 Sort job on 350 scale is consistently failing with latest MRV2 code 
 

 Key: MAPREDUCE-3656
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3656
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster, mrv2, resourcemanager
Affects Versions: 0.23.1
Reporter: Karam Singh
Assignee: Siddharth Seth
Priority: Blocker
 Fix For: 0.23.1

 Attachments: MR3656.txt, MR3656.txt, MR3656.txt


 With the code checked out on last two days. 
 Sort Job on 350 node scale with 16800 maps and 680 reduces consistently 
 failing for around last 6 runs
 When around 50% of maps are completed, suddenly job jumps to failed state.
 On looking at NM log, found RM sent Stop Container Request to NM for AM 
 container.
 But at INFO level from RM log not able find why RM is killing AM when job is 
 not killed manually.
 One thing found common on failed AM logs is -:
 org.apache.hadoop.yarn.state.InvalidStateTransitonException
 With with different.
 For e.g. One log says -:
 {code}
 org.apache.hadoop.yarn.state.InvalidStateTransitonException: Invalid event: 
 TA_UPDATE at ASSIGNED 
 {code}
 Whereas other logs says -:
 {code}
 org.apache.hadoop.yarn.state.InvalidStateTransitonException: Invalid event: 
 JOB_COUNTER_UPDATE at ERROR
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (MAPREDUCE-3596) Sort benchmark got hang after completion of 99% map phase

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3596?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186188#comment-13186188
 ] 

Hudson commented on MAPREDUCE-3596:
---

Integrated in Hadoop-Hdfs-trunk #925 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/925/])
MAPREDUCE-3596. Fix scheduler to handle cleaned up containers, which NMs 
may subsequently report as running. (Contributed by Vinod Kumar Vavilapalli)

sseth : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231297
Files : 
* /hadoop/common/trunk/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/BuilderUtils.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/container/ContainerImpl.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceManager.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceTrackerService.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmnode/RMNode.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmnode/RMNodeImpl.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/SchedulerApp.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CapacityScheduler.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fifo/FifoScheduler.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockNM.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockNodes.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestApplicationCleanup.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestResourceTrackerService.java


 Sort benchmark got hang after completion of 99% map phase
 -

 Key: MAPREDUCE-3596
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3596
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster, mrv2
Affects Versions: 0.23.0
Reporter: Ravi Prakash
Assignee: Vinod Kumar Vavilapalli
Priority: Blocker
 Fix For: 0.23.1

 Attachments: MAPREDUCE-3596-20120111.1.txt, 
 MAPREDUCE-3596-20120111.txt, MAPREDUCE-3596-20120112.1.txt, 
 MAPREDUCE-3596-20120112.txt, logs.tar.bz2, logs.tar.bz2


 Courtesy [~vinaythota]
 {quote}
 Ran sort benchmark couple of times and every time the job got hang after 
 completion 99% map phase. There are some map tasks failed. Also it's not 
 scheduled some of the pending map tasks.
 Cluster size is 350 nodes.
 Build Details:
 ==
 Compiled:   Fri Dec 9 16:25:27 PST 2011 by someone from 
 branches/branch-0.23/hadoop-common-project/hadoop-common 
 ResourceManager version:revision 1212681 by someone source checksum 
 on Fri Dec 9 16:52:07 PST 2011
 Hadoop version: revision 1212592 by someone Fri Dec 9 16:25:27 PST 
 2011
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (MAPREDUCE-3404) Speculative Execution: speculative map tasks launched even if -Dmapreduce.map.speculative=false

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3404?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186190#comment-13186190
 ] 

Hudson commented on MAPREDUCE-3404:
---

Integrated in Hadoop-Hdfs-trunk #925 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/925/])
MAPREDUCE-3404. Corrected MR AM to honor speculative configuration and 
enable speculating either maps or reduces. Contributed by Eric Payne.

vinodkv : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231395
Files : 
* /hadoop/common/trunk/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/MRAppMaster.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/v2/TestSpeculativeExecution.java


 Speculative Execution: speculative map tasks launched even if 
 -Dmapreduce.map.speculative=false
 ---

 Key: MAPREDUCE-3404
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3404
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: job submission, mrv2
Affects Versions: 0.23.0
 Environment: Hadoop version is: Hadoop 0.23.0.1110031628
 10 node test cluster
Reporter: patrick white
Assignee: Eric Payne
Priority: Critical
 Fix For: 0.23.1, 0.24.0

 Attachments: MAPREDUCE-3404.1.txt, MAPREDUCE-3404.2.txt


 When forcing a mapper to take significantly longer than other map tasks, 
 speculative map tasks are
 launched even if the mapreduce.job.maps.speculative.execution parameter is 
 set to 'false'.
 Testcase: ran default WordCount job with spec execution set to false for both 
 map and reduce but still saw a fifth mapper
 task launch, ran job as follows:
 hadoop --config config  jar   /tmp/testphw/wordcount.jar   WordCount  
 -Dmapreduce.job.maps.speculative.execution=false  
 -Dmapreduce.job.reduces.speculative.execution=false 
 /tmp/test_file_of_words* /tmp/file_of_words.out
 Input data was 4 text files hdfs blocksize, with same word pattern plus one 
 diff text line in each file, fourth
 file was 4 times as large as others:
 hadoop --config config  fs -ls  /tmp
 Found 5 items
 drwxr-xr-x   - user hdfs  0 2011-10-20 16:17 /tmp/file_of_words.out
 -rw-r--r--   3 user hdfs   62800021 2011-10-20 14:45 /tmp/test_file_of_words1
 -rw-r--r--   3 user hdfs   62800024 2011-10-20 14:46 /tmp/test_file_of_words2
 -rw-r--r--   3 user hdfs   62800024 2011-10-20 14:46 /tmp/test_file_of_words3
 -rw-r--r--   3 user hdfs  271708312 2011-10-20 15:50 /tmp/test_file_of_words4
 Job launched 5 mappers despite spec exec set to false, output snippet:
 org.apache.hadoop.mapreduce.JobCounter
 NUM_FAILED_MAPS=1
 TOTAL_LAUNCHED_MAPS=5
 TOTAL_LAUNCHED_REDUCES=1
 RACK_LOCAL_MAPS=5
 SLOTS_MILLIS_MAPS=273540
 SLOTS_MILLIS_REDUCES=212876
 Reran same case as above only set both spec exec params to 'true', same 
 results only this time the fifth task being
 launched is expected since spec exec = true.
 job run:
 hadoop --config config  jar   /tmp/testphw/wordcount.jar   WordCount  
 -Dmapreduce.job.maps.speculative.execution=true  
 -Dmapreduce.job.reduces.speculative.execution=true 
 /tmp/test_file_of_words* /tmp/file_of_words.out
 output snippet:
 org.apache.hadoop.mapreduce.JobCounter
 NUM_FAILED_MAPS=1
 TOTAL_LAUNCHED_MAPS=5
 TOTAL_LAUNCHED_REDUCES=1
 RACK_LOCAL_MAPS=5
 SLOTS_MILLIS_MAPS=279653
 SLOTS_MILLIS_REDUCES=211474

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (MAPREDUCE-3532) When 0 is provided as port number in yarn.nodemanager.webapp.address, NMs webserver component picks up random port, NM keeps on Reporting 0 port to RM

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186191#comment-13186191
 ] 

Hudson commented on MAPREDUCE-3532:
---

Integrated in Hadoop-Hdfs-0.23-Build #138 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-0.23-Build/138/])
MAPREDUCE-3532. Modified NM to report correct http address when an 
ephemeral web port is configured. Contributed by Bhallamudi Venkata Siva Kamesh.
svn merge --ignore-ancestry -c 1231342 ../../trunk/

vinodkv : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231344
Files : 
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/webapp/WebServer.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestNMWebServer.java


 When 0 is provided as port number in yarn.nodemanager.webapp.address, NMs 
 webserver component picks up random port, NM keeps on Reporting 0 port to RM
 --

 Key: MAPREDUCE-3532
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3532
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv2, nodemanager
Affects Versions: 0.23.1
Reporter: Karam Singh
Assignee: Bhallamudi Venkata Siva Kamesh
Priority: Critical
 Fix For: 0.23.1

 Attachments: MAPREDUCE-3532-1.patch, MAPREDUCE-3532.patch


 I tried following -:
 yarn.nodemanager.address=0.0.0.0:0
 yarn.nodemanager.webapp.address=0.0.0.0:0
 yarn.nodemanager.localizer.address=0.0.0.0:0
 mapreduce.shuffle.port=0
 When 0 is provided as number in yarn.nodemanager.webapp.address. 
 NM instantiate WebServer as 0 piort e.g.
 {code}
 2011-12-08 11:33:02,467 INFO 
 org.apache.hadoop.yarn.server.nodemanager.webapp.WebServer: Instantiating 
 NMWebApp at 0.0.0.0:0
 {code}
 After that WebServer pick up some random port e.g.
 {code}
 2011-12-08 11:33:02,562 INFO org.apache.hadoop.http.HttpServer: Jetty bound 
 to port 36272
 2011-12-08 11:33:02,562 INFO org.mortbay.log: jetty-6.1.26
 2011-12-08 11:33:02,831 INFO org.mortbay.log: Started 
 SelectChannelConnector@0.0.0.0:36272
 2011-12-08 11:33:02,831 INFO org.apache.hadoop.yarn.webapp.WebApps: Web app 
 /node started at 36272
 {code}
 And NM WebServer responds correctly but
  RM's cluster/Nodes page shows the following -:
 {code}
 /Rack RUNNING NM:57963 NM:0 Healthy 8-Dec-2011 11:33:01 Healthy 8 12 GB 0 KB
 {code}
 Whereas NM:0 is not clickable.
 Seems even NM's webserver pick random port but it never gets updated and so 
 NM report 0 as HTTP port to RM causing NM Hyperlinks un-clickable
 But verified that MR job runs successfully with random.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (MAPREDUCE-3596) Sort benchmark got hang after completion of 99% map phase

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3596?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186192#comment-13186192
 ] 

Hudson commented on MAPREDUCE-3596:
---

Integrated in Hadoop-Hdfs-0.23-Build #138 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-0.23-Build/138/])
merge MAPREDUCE-3596 from trunk. Fix scheduler to handle cleaned up 
containers, which NMs may subsequently report as running. (Contributed by Vinod 
Kumar Vavilapalli)

sseth : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231303
Files : 
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/BuilderUtils.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/container/ContainerImpl.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceManager.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceTrackerService.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmnode/RMNode.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmnode/RMNodeImpl.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/SchedulerApp.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CapacityScheduler.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fifo/FifoScheduler.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockNM.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockNodes.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestApplicationCleanup.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestResourceTrackerService.java


 Sort benchmark got hang after completion of 99% map phase
 -

 Key: MAPREDUCE-3596
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3596
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster, mrv2
Affects Versions: 0.23.0
Reporter: Ravi Prakash
Assignee: Vinod Kumar Vavilapalli
Priority: Blocker
 Fix For: 0.23.1

 Attachments: MAPREDUCE-3596-20120111.1.txt, 
 MAPREDUCE-3596-20120111.txt, MAPREDUCE-3596-20120112.1.txt, 
 MAPREDUCE-3596-20120112.txt, logs.tar.bz2, logs.tar.bz2


 Courtesy [~vinaythota]
 {quote}
 Ran sort benchmark couple of times and every time the job got hang after 
 completion 99% map phase. There are some map tasks failed. Also it's not 
 scheduled some of the pending map tasks.
 Cluster size is 350 nodes.
 Build Details:
 ==
 Compiled:   Fri Dec 9 16:25:27 PST 2011 by someone from 
 branches/branch-0.23/hadoop-common-project/hadoop-common 
 ResourceManager version:revision 1212681 by someone source checksum 
 on Fri Dec 9 16:52:07 PST 2011
 Hadoop version: revision 1212592 by someone Fri Dec 9 16:25:27 PST 
 2011
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on 

[jira] [Commented] (MAPREDUCE-3404) Speculative Execution: speculative map tasks launched even if -Dmapreduce.map.speculative=false

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3404?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186194#comment-13186194
 ] 

Hudson commented on MAPREDUCE-3404:
---

Integrated in Hadoop-Hdfs-0.23-Build #138 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-0.23-Build/138/])
MAPREDUCE-3404. Corrected MR AM to honor speculative configuration and 
enable speculating either maps or reduces. Contributed by Eric Payne.
svn merge --ignore-ancestry -c 1231395 ../../trunk/

vinodkv : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231397
Files : 
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/MRAppMaster.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/v2/TestSpeculativeExecution.java


 Speculative Execution: speculative map tasks launched even if 
 -Dmapreduce.map.speculative=false
 ---

 Key: MAPREDUCE-3404
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3404
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: job submission, mrv2
Affects Versions: 0.23.0
 Environment: Hadoop version is: Hadoop 0.23.0.1110031628
 10 node test cluster
Reporter: patrick white
Assignee: Eric Payne
Priority: Critical
 Fix For: 0.23.1, 0.24.0

 Attachments: MAPREDUCE-3404.1.txt, MAPREDUCE-3404.2.txt


 When forcing a mapper to take significantly longer than other map tasks, 
 speculative map tasks are
 launched even if the mapreduce.job.maps.speculative.execution parameter is 
 set to 'false'.
 Testcase: ran default WordCount job with spec execution set to false for both 
 map and reduce but still saw a fifth mapper
 task launch, ran job as follows:
 hadoop --config config  jar   /tmp/testphw/wordcount.jar   WordCount  
 -Dmapreduce.job.maps.speculative.execution=false  
 -Dmapreduce.job.reduces.speculative.execution=false 
 /tmp/test_file_of_words* /tmp/file_of_words.out
 Input data was 4 text files hdfs blocksize, with same word pattern plus one 
 diff text line in each file, fourth
 file was 4 times as large as others:
 hadoop --config config  fs -ls  /tmp
 Found 5 items
 drwxr-xr-x   - user hdfs  0 2011-10-20 16:17 /tmp/file_of_words.out
 -rw-r--r--   3 user hdfs   62800021 2011-10-20 14:45 /tmp/test_file_of_words1
 -rw-r--r--   3 user hdfs   62800024 2011-10-20 14:46 /tmp/test_file_of_words2
 -rw-r--r--   3 user hdfs   62800024 2011-10-20 14:46 /tmp/test_file_of_words3
 -rw-r--r--   3 user hdfs  271708312 2011-10-20 15:50 /tmp/test_file_of_words4
 Job launched 5 mappers despite spec exec set to false, output snippet:
 org.apache.hadoop.mapreduce.JobCounter
 NUM_FAILED_MAPS=1
 TOTAL_LAUNCHED_MAPS=5
 TOTAL_LAUNCHED_REDUCES=1
 RACK_LOCAL_MAPS=5
 SLOTS_MILLIS_MAPS=273540
 SLOTS_MILLIS_REDUCES=212876
 Reran same case as above only set both spec exec params to 'true', same 
 results only this time the fifth task being
 launched is expected since spec exec = true.
 job run:
 hadoop --config config  jar   /tmp/testphw/wordcount.jar   WordCount  
 -Dmapreduce.job.maps.speculative.execution=true  
 -Dmapreduce.job.reduces.speculative.execution=true 
 /tmp/test_file_of_words* /tmp/file_of_words.out
 output snippet:
 org.apache.hadoop.mapreduce.JobCounter
 NUM_FAILED_MAPS=1
 TOTAL_LAUNCHED_MAPS=5
 TOTAL_LAUNCHED_REDUCES=1
 RACK_LOCAL_MAPS=5
 SLOTS_MILLIS_MAPS=279653
 SLOTS_MILLIS_REDUCES=211474

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (MAPREDUCE-3656) Sort job on 350 scale is consistently failing with latest MRV2 code

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3656?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186193#comment-13186193
 ] 

Hudson commented on MAPREDUCE-3656:
---

Integrated in Hadoop-Hdfs-0.23-Build #138 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-0.23-Build/138/])
MAPREDUCE-3656. Fixed a race condition in MR AM which is failing the sort 
benchmark consistently. Contributed by Siddarth Seth.
svn merge --ignore-ancestry -c 1231314 ../../trunk/

vinodkv : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231316
Files : 
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapred/TaskAttemptListenerImpl.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/TaskAttemptListener.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/TaskHeartbeatHandler.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/job/impl/TaskAttemptImpl.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapred/TestTaskAttemptListenerImpl.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapreduce/v2/app/MRApp.java


 Sort job on 350 scale is consistently failing with latest MRV2 code 
 

 Key: MAPREDUCE-3656
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3656
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster, mrv2, resourcemanager
Affects Versions: 0.23.1
Reporter: Karam Singh
Assignee: Siddharth Seth
Priority: Blocker
 Fix For: 0.23.1

 Attachments: MR3656.txt, MR3656.txt, MR3656.txt


 With the code checked out on last two days. 
 Sort Job on 350 node scale with 16800 maps and 680 reduces consistently 
 failing for around last 6 runs
 When around 50% of maps are completed, suddenly job jumps to failed state.
 On looking at NM log, found RM sent Stop Container Request to NM for AM 
 container.
 But at INFO level from RM log not able find why RM is killing AM when job is 
 not killed manually.
 One thing found common on failed AM logs is -:
 org.apache.hadoop.yarn.state.InvalidStateTransitonException
 With with different.
 For e.g. One log says -:
 {code}
 org.apache.hadoop.yarn.state.InvalidStateTransitonException: Invalid event: 
 TA_UPDATE at ASSIGNED 
 {code}
 Whereas other logs says -:
 {code}
 org.apache.hadoop.yarn.state.InvalidStateTransitonException: Invalid event: 
 JOB_COUNTER_UPDATE at ERROR
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (MAPREDUCE-3596) Sort benchmark got hang after completion of 99% map phase

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3596?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186198#comment-13186198
 ] 

Hudson commented on MAPREDUCE-3596:
---

Integrated in Hadoop-Mapreduce-0.23-Build #160 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-0.23-Build/160/])
merge MAPREDUCE-3596 from trunk. Fix scheduler to handle cleaned up 
containers, which NMs may subsequently report as running. (Contributed by Vinod 
Kumar Vavilapalli)

sseth : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231303
Files : 
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/BuilderUtils.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/container/ContainerImpl.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceManager.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceTrackerService.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmnode/RMNode.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmnode/RMNodeImpl.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/SchedulerApp.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CapacityScheduler.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fifo/FifoScheduler.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockNM.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockNodes.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestApplicationCleanup.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestResourceTrackerService.java


 Sort benchmark got hang after completion of 99% map phase
 -

 Key: MAPREDUCE-3596
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3596
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster, mrv2
Affects Versions: 0.23.0
Reporter: Ravi Prakash
Assignee: Vinod Kumar Vavilapalli
Priority: Blocker
 Fix For: 0.23.1

 Attachments: MAPREDUCE-3596-20120111.1.txt, 
 MAPREDUCE-3596-20120111.txt, MAPREDUCE-3596-20120112.1.txt, 
 MAPREDUCE-3596-20120112.txt, logs.tar.bz2, logs.tar.bz2


 Courtesy [~vinaythota]
 {quote}
 Ran sort benchmark couple of times and every time the job got hang after 
 completion 99% map phase. There are some map tasks failed. Also it's not 
 scheduled some of the pending map tasks.
 Cluster size is 350 nodes.
 Build Details:
 ==
 Compiled:   Fri Dec 9 16:25:27 PST 2011 by someone from 
 branches/branch-0.23/hadoop-common-project/hadoop-common 
 ResourceManager version:revision 1212681 by someone source checksum 
 on Fri Dec 9 16:52:07 PST 2011
 Hadoop version: revision 1212592 by someone Fri Dec 9 16:25:27 PST 
 2011
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more 

[jira] [Commented] (MAPREDUCE-3532) When 0 is provided as port number in yarn.nodemanager.webapp.address, NMs webserver component picks up random port, NM keeps on Reporting 0 port to RM

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186197#comment-13186197
 ] 

Hudson commented on MAPREDUCE-3532:
---

Integrated in Hadoop-Mapreduce-0.23-Build #160 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-0.23-Build/160/])
MAPREDUCE-3532. Modified NM to report correct http address when an 
ephemeral web port is configured. Contributed by Bhallamudi Venkata Siva Kamesh.
svn merge --ignore-ancestry -c 1231342 ../../trunk/

vinodkv : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231344
Files : 
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/webapp/WebServer.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestNMWebServer.java


 When 0 is provided as port number in yarn.nodemanager.webapp.address, NMs 
 webserver component picks up random port, NM keeps on Reporting 0 port to RM
 --

 Key: MAPREDUCE-3532
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3532
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv2, nodemanager
Affects Versions: 0.23.1
Reporter: Karam Singh
Assignee: Bhallamudi Venkata Siva Kamesh
Priority: Critical
 Fix For: 0.23.1

 Attachments: MAPREDUCE-3532-1.patch, MAPREDUCE-3532.patch


 I tried following -:
 yarn.nodemanager.address=0.0.0.0:0
 yarn.nodemanager.webapp.address=0.0.0.0:0
 yarn.nodemanager.localizer.address=0.0.0.0:0
 mapreduce.shuffle.port=0
 When 0 is provided as number in yarn.nodemanager.webapp.address. 
 NM instantiate WebServer as 0 piort e.g.
 {code}
 2011-12-08 11:33:02,467 INFO 
 org.apache.hadoop.yarn.server.nodemanager.webapp.WebServer: Instantiating 
 NMWebApp at 0.0.0.0:0
 {code}
 After that WebServer pick up some random port e.g.
 {code}
 2011-12-08 11:33:02,562 INFO org.apache.hadoop.http.HttpServer: Jetty bound 
 to port 36272
 2011-12-08 11:33:02,562 INFO org.mortbay.log: jetty-6.1.26
 2011-12-08 11:33:02,831 INFO org.mortbay.log: Started 
 SelectChannelConnector@0.0.0.0:36272
 2011-12-08 11:33:02,831 INFO org.apache.hadoop.yarn.webapp.WebApps: Web app 
 /node started at 36272
 {code}
 And NM WebServer responds correctly but
  RM's cluster/Nodes page shows the following -:
 {code}
 /Rack RUNNING NM:57963 NM:0 Healthy 8-Dec-2011 11:33:01 Healthy 8 12 GB 0 KB
 {code}
 Whereas NM:0 is not clickable.
 Seems even NM's webserver pick random port but it never gets updated and so 
 NM report 0 as HTTP port to RM causing NM Hyperlinks un-clickable
 But verified that MR job runs successfully with random.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (MAPREDUCE-3656) Sort job on 350 scale is consistently failing with latest MRV2 code

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3656?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186199#comment-13186199
 ] 

Hudson commented on MAPREDUCE-3656:
---

Integrated in Hadoop-Mapreduce-0.23-Build #160 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-0.23-Build/160/])
MAPREDUCE-3656. Fixed a race condition in MR AM which is failing the sort 
benchmark consistently. Contributed by Siddarth Seth.
svn merge --ignore-ancestry -c 1231314 ../../trunk/

vinodkv : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231316
Files : 
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapred/TaskAttemptListenerImpl.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/TaskAttemptListener.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/TaskHeartbeatHandler.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/job/impl/TaskAttemptImpl.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapred/TestTaskAttemptListenerImpl.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapreduce/v2/app/MRApp.java


 Sort job on 350 scale is consistently failing with latest MRV2 code 
 

 Key: MAPREDUCE-3656
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3656
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster, mrv2, resourcemanager
Affects Versions: 0.23.1
Reporter: Karam Singh
Assignee: Siddharth Seth
Priority: Blocker
 Fix For: 0.23.1

 Attachments: MR3656.txt, MR3656.txt, MR3656.txt


 With the code checked out on last two days. 
 Sort Job on 350 node scale with 16800 maps and 680 reduces consistently 
 failing for around last 6 runs
 When around 50% of maps are completed, suddenly job jumps to failed state.
 On looking at NM log, found RM sent Stop Container Request to NM for AM 
 container.
 But at INFO level from RM log not able find why RM is killing AM when job is 
 not killed manually.
 One thing found common on failed AM logs is -:
 org.apache.hadoop.yarn.state.InvalidStateTransitonException
 With with different.
 For e.g. One log says -:
 {code}
 org.apache.hadoop.yarn.state.InvalidStateTransitonException: Invalid event: 
 TA_UPDATE at ASSIGNED 
 {code}
 Whereas other logs says -:
 {code}
 org.apache.hadoop.yarn.state.InvalidStateTransitonException: Invalid event: 
 JOB_COUNTER_UPDATE at ERROR
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (MAPREDUCE-3404) Speculative Execution: speculative map tasks launched even if -Dmapreduce.map.speculative=false

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3404?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186200#comment-13186200
 ] 

Hudson commented on MAPREDUCE-3404:
---

Integrated in Hadoop-Mapreduce-0.23-Build #160 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-0.23-Build/160/])
MAPREDUCE-3404. Corrected MR AM to honor speculative configuration and 
enable speculating either maps or reduces. Contributed by Eric Payne.
svn merge --ignore-ancestry -c 1231395 ../../trunk/

vinodkv : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231397
Files : 
* /hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/MRAppMaster.java
* 
/hadoop/common/branches/branch-0.23/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/v2/TestSpeculativeExecution.java


 Speculative Execution: speculative map tasks launched even if 
 -Dmapreduce.map.speculative=false
 ---

 Key: MAPREDUCE-3404
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3404
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: job submission, mrv2
Affects Versions: 0.23.0
 Environment: Hadoop version is: Hadoop 0.23.0.1110031628
 10 node test cluster
Reporter: patrick white
Assignee: Eric Payne
Priority: Critical
 Fix For: 0.23.1, 0.24.0

 Attachments: MAPREDUCE-3404.1.txt, MAPREDUCE-3404.2.txt


 When forcing a mapper to take significantly longer than other map tasks, 
 speculative map tasks are
 launched even if the mapreduce.job.maps.speculative.execution parameter is 
 set to 'false'.
 Testcase: ran default WordCount job with spec execution set to false for both 
 map and reduce but still saw a fifth mapper
 task launch, ran job as follows:
 hadoop --config config  jar   /tmp/testphw/wordcount.jar   WordCount  
 -Dmapreduce.job.maps.speculative.execution=false  
 -Dmapreduce.job.reduces.speculative.execution=false 
 /tmp/test_file_of_words* /tmp/file_of_words.out
 Input data was 4 text files hdfs blocksize, with same word pattern plus one 
 diff text line in each file, fourth
 file was 4 times as large as others:
 hadoop --config config  fs -ls  /tmp
 Found 5 items
 drwxr-xr-x   - user hdfs  0 2011-10-20 16:17 /tmp/file_of_words.out
 -rw-r--r--   3 user hdfs   62800021 2011-10-20 14:45 /tmp/test_file_of_words1
 -rw-r--r--   3 user hdfs   62800024 2011-10-20 14:46 /tmp/test_file_of_words2
 -rw-r--r--   3 user hdfs   62800024 2011-10-20 14:46 /tmp/test_file_of_words3
 -rw-r--r--   3 user hdfs  271708312 2011-10-20 15:50 /tmp/test_file_of_words4
 Job launched 5 mappers despite spec exec set to false, output snippet:
 org.apache.hadoop.mapreduce.JobCounter
 NUM_FAILED_MAPS=1
 TOTAL_LAUNCHED_MAPS=5
 TOTAL_LAUNCHED_REDUCES=1
 RACK_LOCAL_MAPS=5
 SLOTS_MILLIS_MAPS=273540
 SLOTS_MILLIS_REDUCES=212876
 Reran same case as above only set both spec exec params to 'true', same 
 results only this time the fifth task being
 launched is expected since spec exec = true.
 job run:
 hadoop --config config  jar   /tmp/testphw/wordcount.jar   WordCount  
 -Dmapreduce.job.maps.speculative.execution=true  
 -Dmapreduce.job.reduces.speculative.execution=true 
 /tmp/test_file_of_words* /tmp/file_of_words.out
 output snippet:
 org.apache.hadoop.mapreduce.JobCounter
 NUM_FAILED_MAPS=1
 TOTAL_LAUNCHED_MAPS=5
 TOTAL_LAUNCHED_REDUCES=1
 RACK_LOCAL_MAPS=5
 SLOTS_MILLIS_MAPS=279653
 SLOTS_MILLIS_REDUCES=211474

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (MAPREDUCE-3532) When 0 is provided as port number in yarn.nodemanager.webapp.address, NMs webserver component picks up random port, NM keeps on Reporting 0 port to RM

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186204#comment-13186204
 ] 

Hudson commented on MAPREDUCE-3532:
---

Integrated in Hadoop-Mapreduce-trunk #958 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/958/])
MAPREDUCE-3532. Modified NM to report correct http address when an 
ephemeral web port is configured. Contributed by Bhallamudi Venkata Siva Kamesh.

vinodkv : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231342
Files : 
* /hadoop/common/trunk/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/webapp/WebServer.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestNMWebServer.java


 When 0 is provided as port number in yarn.nodemanager.webapp.address, NMs 
 webserver component picks up random port, NM keeps on Reporting 0 port to RM
 --

 Key: MAPREDUCE-3532
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3532
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv2, nodemanager
Affects Versions: 0.23.1
Reporter: Karam Singh
Assignee: Bhallamudi Venkata Siva Kamesh
Priority: Critical
 Fix For: 0.23.1

 Attachments: MAPREDUCE-3532-1.patch, MAPREDUCE-3532.patch


 I tried following -:
 yarn.nodemanager.address=0.0.0.0:0
 yarn.nodemanager.webapp.address=0.0.0.0:0
 yarn.nodemanager.localizer.address=0.0.0.0:0
 mapreduce.shuffle.port=0
 When 0 is provided as number in yarn.nodemanager.webapp.address. 
 NM instantiate WebServer as 0 piort e.g.
 {code}
 2011-12-08 11:33:02,467 INFO 
 org.apache.hadoop.yarn.server.nodemanager.webapp.WebServer: Instantiating 
 NMWebApp at 0.0.0.0:0
 {code}
 After that WebServer pick up some random port e.g.
 {code}
 2011-12-08 11:33:02,562 INFO org.apache.hadoop.http.HttpServer: Jetty bound 
 to port 36272
 2011-12-08 11:33:02,562 INFO org.mortbay.log: jetty-6.1.26
 2011-12-08 11:33:02,831 INFO org.mortbay.log: Started 
 SelectChannelConnector@0.0.0.0:36272
 2011-12-08 11:33:02,831 INFO org.apache.hadoop.yarn.webapp.WebApps: Web app 
 /node started at 36272
 {code}
 And NM WebServer responds correctly but
  RM's cluster/Nodes page shows the following -:
 {code}
 /Rack RUNNING NM:57963 NM:0 Healthy 8-Dec-2011 11:33:01 Healthy 8 12 GB 0 KB
 {code}
 Whereas NM:0 is not clickable.
 Seems even NM's webserver pick random port but it never gets updated and so 
 NM report 0 as HTTP port to RM causing NM Hyperlinks un-clickable
 But verified that MR job runs successfully with random.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (MAPREDUCE-3596) Sort benchmark got hang after completion of 99% map phase

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3596?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186205#comment-13186205
 ] 

Hudson commented on MAPREDUCE-3596:
---

Integrated in Hadoop-Mapreduce-trunk #958 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/958/])
MAPREDUCE-3596. Fix scheduler to handle cleaned up containers, which NMs 
may subsequently report as running. (Contributed by Vinod Kumar Vavilapalli)

sseth : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231297
Files : 
* /hadoop/common/trunk/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/BuilderUtils.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/container/ContainerImpl.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceManager.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceTrackerService.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmnode/RMNode.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmnode/RMNodeImpl.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/SchedulerApp.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CapacityScheduler.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fifo/FifoScheduler.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockNM.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockNodes.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestApplicationCleanup.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestResourceTrackerService.java


 Sort benchmark got hang after completion of 99% map phase
 -

 Key: MAPREDUCE-3596
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3596
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster, mrv2
Affects Versions: 0.23.0
Reporter: Ravi Prakash
Assignee: Vinod Kumar Vavilapalli
Priority: Blocker
 Fix For: 0.23.1

 Attachments: MAPREDUCE-3596-20120111.1.txt, 
 MAPREDUCE-3596-20120111.txt, MAPREDUCE-3596-20120112.1.txt, 
 MAPREDUCE-3596-20120112.txt, logs.tar.bz2, logs.tar.bz2


 Courtesy [~vinaythota]
 {quote}
 Ran sort benchmark couple of times and every time the job got hang after 
 completion 99% map phase. There are some map tasks failed. Also it's not 
 scheduled some of the pending map tasks.
 Cluster size is 350 nodes.
 Build Details:
 ==
 Compiled:   Fri Dec 9 16:25:27 PST 2011 by someone from 
 branches/branch-0.23/hadoop-common-project/hadoop-common 
 ResourceManager version:revision 1212681 by someone source checksum 
 on Fri Dec 9 16:52:07 PST 2011
 Hadoop version: revision 1212592 by someone Fri Dec 9 16:25:27 PST 
 2011
 {quote}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (MAPREDUCE-3656) Sort job on 350 scale is consistently failing with latest MRV2 code

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3656?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186206#comment-13186206
 ] 

Hudson commented on MAPREDUCE-3656:
---

Integrated in Hadoop-Mapreduce-trunk #958 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/958/])
MAPREDUCE-3656. Fixed a race condition in MR AM which is failing the sort 
benchmark consistently. Contributed by Siddarth Seth.

vinodkv : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231314
Files : 
* /hadoop/common/trunk/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapred/TaskAttemptListenerImpl.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/TaskAttemptListener.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/TaskHeartbeatHandler.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/job/impl/TaskAttemptImpl.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapred/TestTaskAttemptListenerImpl.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapreduce/v2/app/MRApp.java


 Sort job on 350 scale is consistently failing with latest MRV2 code 
 

 Key: MAPREDUCE-3656
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3656
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster, mrv2, resourcemanager
Affects Versions: 0.23.1
Reporter: Karam Singh
Assignee: Siddharth Seth
Priority: Blocker
 Fix For: 0.23.1

 Attachments: MR3656.txt, MR3656.txt, MR3656.txt


 With the code checked out on last two days. 
 Sort Job on 350 node scale with 16800 maps and 680 reduces consistently 
 failing for around last 6 runs
 When around 50% of maps are completed, suddenly job jumps to failed state.
 On looking at NM log, found RM sent Stop Container Request to NM for AM 
 container.
 But at INFO level from RM log not able find why RM is killing AM when job is 
 not killed manually.
 One thing found common on failed AM logs is -:
 org.apache.hadoop.yarn.state.InvalidStateTransitonException
 With with different.
 For e.g. One log says -:
 {code}
 org.apache.hadoop.yarn.state.InvalidStateTransitonException: Invalid event: 
 TA_UPDATE at ASSIGNED 
 {code}
 Whereas other logs says -:
 {code}
 org.apache.hadoop.yarn.state.InvalidStateTransitonException: Invalid event: 
 JOB_COUNTER_UPDATE at ERROR
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (MAPREDUCE-3404) Speculative Execution: speculative map tasks launched even if -Dmapreduce.map.speculative=false

2012-01-14 Thread Hudson (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3404?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186207#comment-13186207
 ] 

Hudson commented on MAPREDUCE-3404:
---

Integrated in Hadoop-Mapreduce-trunk #958 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/958/])
MAPREDUCE-3404. Corrected MR AM to honor speculative configuration and 
enable speculating either maps or reduces. Contributed by Eric Payne.

vinodkv : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1231395
Files : 
* /hadoop/common/trunk/hadoop-mapreduce-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/MRAppMaster.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/v2/TestSpeculativeExecution.java


 Speculative Execution: speculative map tasks launched even if 
 -Dmapreduce.map.speculative=false
 ---

 Key: MAPREDUCE-3404
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3404
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: job submission, mrv2
Affects Versions: 0.23.0
 Environment: Hadoop version is: Hadoop 0.23.0.1110031628
 10 node test cluster
Reporter: patrick white
Assignee: Eric Payne
Priority: Critical
 Fix For: 0.23.1, 0.24.0

 Attachments: MAPREDUCE-3404.1.txt, MAPREDUCE-3404.2.txt


 When forcing a mapper to take significantly longer than other map tasks, 
 speculative map tasks are
 launched even if the mapreduce.job.maps.speculative.execution parameter is 
 set to 'false'.
 Testcase: ran default WordCount job with spec execution set to false for both 
 map and reduce but still saw a fifth mapper
 task launch, ran job as follows:
 hadoop --config config  jar   /tmp/testphw/wordcount.jar   WordCount  
 -Dmapreduce.job.maps.speculative.execution=false  
 -Dmapreduce.job.reduces.speculative.execution=false 
 /tmp/test_file_of_words* /tmp/file_of_words.out
 Input data was 4 text files hdfs blocksize, with same word pattern plus one 
 diff text line in each file, fourth
 file was 4 times as large as others:
 hadoop --config config  fs -ls  /tmp
 Found 5 items
 drwxr-xr-x   - user hdfs  0 2011-10-20 16:17 /tmp/file_of_words.out
 -rw-r--r--   3 user hdfs   62800021 2011-10-20 14:45 /tmp/test_file_of_words1
 -rw-r--r--   3 user hdfs   62800024 2011-10-20 14:46 /tmp/test_file_of_words2
 -rw-r--r--   3 user hdfs   62800024 2011-10-20 14:46 /tmp/test_file_of_words3
 -rw-r--r--   3 user hdfs  271708312 2011-10-20 15:50 /tmp/test_file_of_words4
 Job launched 5 mappers despite spec exec set to false, output snippet:
 org.apache.hadoop.mapreduce.JobCounter
 NUM_FAILED_MAPS=1
 TOTAL_LAUNCHED_MAPS=5
 TOTAL_LAUNCHED_REDUCES=1
 RACK_LOCAL_MAPS=5
 SLOTS_MILLIS_MAPS=273540
 SLOTS_MILLIS_REDUCES=212876
 Reran same case as above only set both spec exec params to 'true', same 
 results only this time the fifth task being
 launched is expected since spec exec = true.
 job run:
 hadoop --config config  jar   /tmp/testphw/wordcount.jar   WordCount  
 -Dmapreduce.job.maps.speculative.execution=true  
 -Dmapreduce.job.reduces.speculative.execution=true 
 /tmp/test_file_of_words* /tmp/file_of_words.out
 output snippet:
 org.apache.hadoop.mapreduce.JobCounter
 NUM_FAILED_MAPS=1
 TOTAL_LAUNCHED_MAPS=5
 TOTAL_LAUNCHED_REDUCES=1
 RACK_LOCAL_MAPS=5
 SLOTS_MILLIS_MAPS=279653
 SLOTS_MILLIS_REDUCES=211474

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (MAPREDUCE-3641) CapacityScheduler should be more conservative assigning off-switch requests

2012-01-14 Thread Arun C Murthy (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-3641?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-3641:
-

Status: Patch Available  (was: Open)

Writing unit tests now...

 CapacityScheduler should be more conservative assigning off-switch requests
 ---

 Key: MAPREDUCE-3641
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3641
 Project: Hadoop Map/Reduce
  Issue Type: Sub-task
  Components: mrv2, scheduler
Affects Versions: 0.23.0
Reporter: Arun C Murthy
Assignee: Arun C Murthy
Priority: Blocker
 Attachments: MAPREDUCE-3641.patch, MAPREDUCE-3641.patch, 
 MAPREDUCE-3641.patch


 In hadoop-1, the CS is very conservative handing out off-switch assignments, 
 we need to do the same in YARN.
 We noticed performance regressions due to this, particularly for reduces.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (MAPREDUCE-3641) CapacityScheduler should be more conservative assigning off-switch requests

2012-01-14 Thread Arun C Murthy (Updated) (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-3641?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-3641:
-

Attachment: MAPREDUCE-3641.patch

Ok, tested patch on secure cluster.

 CapacityScheduler should be more conservative assigning off-switch requests
 ---

 Key: MAPREDUCE-3641
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3641
 Project: Hadoop Map/Reduce
  Issue Type: Sub-task
  Components: mrv2, scheduler
Affects Versions: 0.23.0
Reporter: Arun C Murthy
Assignee: Arun C Murthy
Priority: Blocker
 Attachments: MAPREDUCE-3641.patch, MAPREDUCE-3641.patch, 
 MAPREDUCE-3641.patch


 In hadoop-1, the CS is very conservative handing out off-switch assignments, 
 we need to do the same in YARN.
 We noticed performance regressions due to this, particularly for reduces.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (MAPREDUCE-3641) CapacityScheduler should be more conservative assigning off-switch requests

2012-01-14 Thread Hadoop QA (Commented) (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-3641?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13186400#comment-13186400
 ] 

Hadoop QA commented on MAPREDUCE-3641:
--

-1 overall.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12510612/MAPREDUCE-3641.patch
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

-1 tests included.  The patch doesn't appear to include any new or modified 
tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 eclipse:eclipse.  The patch built with eclipse:eclipse.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

-1 core tests.  The patch failed these unit tests:
  
org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.TestParentQueue

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/1611//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/1611//console

This message is automatically generated.

 CapacityScheduler should be more conservative assigning off-switch requests
 ---

 Key: MAPREDUCE-3641
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3641
 Project: Hadoop Map/Reduce
  Issue Type: Sub-task
  Components: mrv2, scheduler
Affects Versions: 0.23.0
Reporter: Arun C Murthy
Assignee: Arun C Murthy
Priority: Blocker
 Attachments: MAPREDUCE-3641.patch, MAPREDUCE-3641.patch, 
 MAPREDUCE-3641.patch


 In hadoop-1, the CS is very conservative handing out off-switch assignments, 
 we need to do the same in YARN.
 We noticed performance regressions due to this, particularly for reduces.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Moved] (MAPREDUCE-3673) org.apache.hadoop.mapreduce.lib.chain.ChainMapper missing on 1.0

2012-01-14 Thread Harsh J (Moved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-3673?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J moved HADOOP-7969 to MAPREDUCE-3673:


  Component/s: (was: build)
   mrv1
 Target Version/s: 1.1.0  (was: 1.0.0)
Affects Version/s: (was: 1.0.0)
   1.0.0
   Issue Type: Wish  (was: New Feature)
  Key: MAPREDUCE-3673  (was: HADOOP-7969)
  Project: Hadoop Map/Reduce  (was: Hadoop Common)

 org.apache.hadoop.mapreduce.lib.chain.ChainMapper missing on 1.0
 

 Key: MAPREDUCE-3673
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3673
 Project: Hadoop Map/Reduce
  Issue Type: Wish
  Components: mrv1
Affects Versions: 1.0.0
 Environment: All
Reporter: surajz
Priority: Minor

 org.apache.hadoop.mapreduce.lib.chain.ChainMapper is missing on 1.0.
 if we are using new context api, do we still the ChainMapper/ChainReducer 
 from the mapred package.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Resolved] (MAPREDUCE-3673) org.apache.hadoop.mapreduce.lib.chain.ChainMapper missing on 1.0

2012-01-14 Thread Harsh J (Resolved) (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-3673?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J resolved MAPREDUCE-3673.


Resolution: Duplicate

1.0 does not contain the entire set of mapred.lib.* ported to new API.

Would you be interested in contributing a backport patch for these classes you 
demand? Please provide/request on MAPREDUCE-3607 instead, in any case.

Marking as duplicate.

 org.apache.hadoop.mapreduce.lib.chain.ChainMapper missing on 1.0
 

 Key: MAPREDUCE-3673
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-3673
 Project: Hadoop Map/Reduce
  Issue Type: Wish
  Components: mrv1
Affects Versions: 1.0.0
 Environment: All
Reporter: surajz
Priority: Minor

 org.apache.hadoop.mapreduce.lib.chain.ChainMapper is missing on 1.0.
 if we are using new context api, do we still the ChainMapper/ChainReducer 
 from the mapred package.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira