Modified: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestFileInputFormat.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestFileInputFormat.java?rev=1390763&r1=1390762&r2=1390763&view=diff ============================================================================== --- hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestFileInputFormat.java (original) +++ hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestFileInputFormat.java Wed Sep 26 22:55:00 2012 @@ -23,6 +23,7 @@ import static org.mockito.Mockito.when; import java.io.DataOutputStream; import java.io.IOException; +import java.util.concurrent.TimeoutException; import junit.framework.TestCase; @@ -95,7 +96,7 @@ public class TestFileInputFormat extends } private void createInputs(FileSystem fs, Path inDir, String fileName) - throws IOException { + throws IOException, TimeoutException, InterruptedException { // create a multi-block file on hdfs Path path = new Path(inDir, fileName); final short replication = 2; @@ -157,7 +158,7 @@ public class TestFileInputFormat extends } } - public void testMultiLevelInput() throws IOException { + public void testMultiLevelInput() throws Exception { JobConf job = new JobConf(conf); job.setBoolean("dfs.replication.considerLoad", false); @@ -291,7 +292,8 @@ public class TestFileInputFormat extends } static void writeFile(Configuration conf, Path name, - short replication, int numBlocks) throws IOException { + short replication, int numBlocks) + throws IOException, TimeoutException, InterruptedException { FileSystem fileSys = FileSystem.get(conf); FSDataOutputStream stm = fileSys.create(name, true,
Modified: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestMultipleLevelCaching.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestMultipleLevelCaching.java?rev=1390763&r1=1390762&r2=1390763&view=diff ============================================================================== --- hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestMultipleLevelCaching.java (original) +++ hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/TestMultipleLevelCaching.java Wed Sep 26 22:55:00 2012 @@ -71,13 +71,13 @@ public class TestMultipleLevelCaching ex return rack.toString(); } - public void testMultiLevelCaching() throws IOException { + public void testMultiLevelCaching() throws Exception { for (int i = 1 ; i <= MAX_LEVEL; ++i) { testCachingAtLevel(i); } } - private void testCachingAtLevel(int level) throws IOException { + private void testCachingAtLevel(int level) throws Exception { String namenode = null; MiniDFSCluster dfs = null; MiniMRCluster mr = null; Modified: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/UtilsForTests.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/UtilsForTests.java?rev=1390763&r1=1390762&r2=1390763&view=diff ============================================================================== --- hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/UtilsForTests.java (original) +++ hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/UtilsForTests.java Wed Sep 26 22:55:00 2012 @@ -31,6 +31,7 @@ import java.util.Enumeration; import java.util.Iterator; import java.util.List; import java.util.Properties; +import java.util.concurrent.TimeoutException; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; @@ -449,11 +450,14 @@ public class UtilsForTests { static void signalTasks(MiniDFSCluster dfs, FileSystem fileSys, String mapSignalFile, String reduceSignalFile, int replication) - throws IOException { - writeFile(dfs.getNameNode(), fileSys.getConf(), new Path(mapSignalFile), - (short)replication); - writeFile(dfs.getNameNode(), fileSys.getConf(), new Path(reduceSignalFile), - (short)replication); + throws IOException, TimeoutException { + try { + writeFile(dfs.getNameNode(), fileSys.getConf(), new Path(mapSignalFile), + (short)replication); + writeFile(dfs.getNameNode(), fileSys.getConf(), new Path(reduceSignalFile), (short)replication); + } catch (InterruptedException ie) { + // Ignore + } } /** @@ -462,12 +466,16 @@ public class UtilsForTests { static void signalTasks(MiniDFSCluster dfs, FileSystem fileSys, boolean isMap, String mapSignalFile, String reduceSignalFile) - throws IOException { - // signal the maps to complete - writeFile(dfs.getNameNode(), fileSys.getConf(), - isMap - ? new Path(mapSignalFile) - : new Path(reduceSignalFile), (short)1); + throws IOException, TimeoutException { + try { + // signal the maps to complete + writeFile(dfs.getNameNode(), fileSys.getConf(), + isMap + ? new Path(mapSignalFile) + : new Path(reduceSignalFile), (short)1); + } catch (InterruptedException ie) { + // Ignore + } } static String getSignalFile(Path dir) { @@ -483,7 +491,8 @@ public class UtilsForTests { } static void writeFile(NameNode namenode, Configuration conf, Path name, - short replication) throws IOException { + short replication) + throws IOException, TimeoutException, InterruptedException { FileSystem fileSys = FileSystem.get(conf); SequenceFile.Writer writer = SequenceFile.createWriter(fileSys, conf, name, Modified: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestCombineFileInputFormat.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestCombineFileInputFormat.java?rev=1390763&r1=1390762&r2=1390763&view=diff ============================================================================== --- hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestCombineFileInputFormat.java (original) +++ hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/lib/input/TestCombineFileInputFormat.java Wed Sep 26 22:55:00 2012 @@ -23,6 +23,7 @@ import java.net.URI; import java.util.List; import java.util.ArrayList; import java.util.zip.GZIPOutputStream; +import java.util.concurrent.TimeoutException; import junit.framework.TestCase; @@ -278,7 +279,7 @@ public class TestCombineFileInputFormat assertFalse(rr.nextKeyValue()); } - public void testSplitPlacement() throws IOException { + public void testSplitPlacement() throws Exception { MiniDFSCluster dfs = null; FileSystem fileSys = null; try { @@ -678,7 +679,8 @@ public class TestCombineFileInputFormat } static void writeFile(Configuration conf, Path name, - short replication, int numBlocks) throws IOException { + short replication, int numBlocks) + throws IOException, TimeoutException, InterruptedException { FileSystem fileSys = FileSystem.get(conf); FSDataOutputStream stm = fileSys.create(name, true, @@ -689,7 +691,8 @@ public class TestCombineFileInputFormat // Creates the gzip file and return the FileStatus static FileStatus writeGzipFile(Configuration conf, Path name, - short replication, int numBlocks) throws IOException { + short replication, int numBlocks) + throws IOException, TimeoutException, InterruptedException { FileSystem fileSys = FileSystem.get(conf); GZIPOutputStream out = new GZIPOutputStream(fileSys.create(name, true, conf @@ -699,7 +702,8 @@ public class TestCombineFileInputFormat } private static void writeDataAndSetReplication(FileSystem fileSys, Path name, - OutputStream out, short replication, int numBlocks) throws IOException { + OutputStream out, short replication, int numBlocks) + throws IOException, TimeoutException, InterruptedException { for (int i = 0; i < numBlocks; i++) { out.write(databuf); } @@ -707,7 +711,7 @@ public class TestCombineFileInputFormat DFSTestUtil.waitReplication(fileSys, name, replication); } - public void testSplitPlacementForCompressedFiles() throws IOException { + public void testSplitPlacementForCompressedFiles() throws Exception { MiniDFSCluster dfs = null; FileSystem fileSys = null; try { @@ -1058,7 +1062,7 @@ public class TestCombineFileInputFormat /** * Test that CFIF can handle missing blocks. */ - public void testMissingBlocks() throws IOException { + public void testMissingBlocks() throws Exception { String namenode = null; MiniDFSCluster dfs = null; FileSystem fileSys = null; Modified: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/security/TestUmbilicalProtocolWithJobToken.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/security/TestUmbilicalProtocolWithJobToken.java?rev=1390763&r1=1390762&r2=1390763&view=diff ============================================================================== --- hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/security/TestUmbilicalProtocolWithJobToken.java (original) +++ hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/security/TestUmbilicalProtocolWithJobToken.java Wed Sep 26 22:55:00 2012 @@ -88,8 +88,10 @@ public class TestUmbilicalProtocolWithJo .when(mockTT).getProtocolSignature(anyString(), anyLong(), anyInt()); JobTokenSecretManager sm = new JobTokenSecretManager(); - final Server server = RPC.getServer(TaskUmbilicalProtocol.class, mockTT, - ADDRESS, 0, 5, true, conf, sm); + final Server server = new RPC.Builder(conf) + .setProtocol(TaskUmbilicalProtocol.class).setInstance(mockTT) + .setBindAddress(ADDRESS).setPort(0).setNumHandlers(5).setVerbose(true) + .setSecretManager(sm).build(); server.start(); Modified: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml?rev=1390763&r1=1390762&r2=1390763&view=diff ============================================================================== --- hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml (original) +++ hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml Wed Sep 26 22:55:00 2012 @@ -172,6 +172,18 @@ <effort>Max</effort> </configuration> </plugin> + <plugin> + <groupId>org.apache.maven.plugins</groupId> + <artifactId>maven-surefire-plugin</artifactId> + <configuration> + <properties> + <property> + <name>listener</name> + <value>org.apache.hadoop.test.TimedOutTestsListener</value> + </property> + </properties> + </configuration> + </plugin> </plugins> </build> Modified: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/DBCountPageView.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/DBCountPageView.java?rev=1390763&r1=1390762&r2=1390763&view=diff ============================================================================== --- hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/DBCountPageView.java (original) +++ hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/DBCountPageView.java Wed Sep 26 22:55:00 2012 @@ -27,7 +27,6 @@ import java.sql.PreparedStatement; import java.sql.ResultSet; import java.sql.SQLException; import java.sql.Statement; -import java.util.Iterator; import java.util.Random; import org.apache.commons.logging.Log; @@ -82,6 +81,7 @@ public class DBCountPageView extends Con private Connection connection; private boolean initialized = false; + private boolean isOracle = false; private static final String[] AccessFieldNames = {"url", "referrer", "time"}; private static final String[] PageviewFieldNames = {"url", "pageview"}; @@ -102,7 +102,9 @@ public class DBCountPageView extends Con private void createConnection(String driverClassName , String url) throws Exception { - + if(driverClassName.toLowerCase().contains("oracle")) { + isOracle = true; + } Class.forName(driverClassName); connection = DriverManager.getConnection(url); connection.setAutoCommit(false); @@ -142,7 +144,7 @@ public class DBCountPageView extends Con } private void dropTables() { - String dropAccess = "DROP TABLE Access"; + String dropAccess = "DROP TABLE HAccess"; String dropPageview = "DROP TABLE Pageview"; Statement st = null; try { @@ -157,18 +159,21 @@ public class DBCountPageView extends Con } private void createTables() throws SQLException { - + String dataType = "BIGINT NOT NULL"; + if(isOracle) { + dataType = "NUMBER(19) NOT NULL"; + } String createAccess = "CREATE TABLE " + - "Access(url VARCHAR(100) NOT NULL," + + "HAccess(url VARCHAR(100) NOT NULL," + " referrer VARCHAR(100)," + - " time BIGINT NOT NULL, " + + " time " + dataType + ", " + " PRIMARY KEY (url, time))"; String createPageview = "CREATE TABLE " + "Pageview(url VARCHAR(100) NOT NULL," + - " pageview BIGINT NOT NULL, " + + " pageview " + dataType + ", " + " PRIMARY KEY (url))"; Statement st = connection.createStatement(); @@ -189,7 +194,7 @@ public class DBCountPageView extends Con PreparedStatement statement = null ; try { statement = connection.prepareStatement( - "INSERT INTO Access(url, referrer, time)" + + "INSERT INTO HAccess(url, referrer, time)" + " VALUES (?, ?, ?)"); Random random = new Random(); @@ -248,7 +253,7 @@ public class DBCountPageView extends Con /**Verifies the results are correct */ private boolean verify() throws SQLException { //check total num pageview - String countAccessQuery = "SELECT COUNT(*) FROM Access"; + String countAccessQuery = "SELECT COUNT(*) FROM HAccess"; String sumPageviewQuery = "SELECT SUM(pageview) FROM Pageview"; Statement st = null; ResultSet rs = null; @@ -396,7 +401,7 @@ public class DBCountPageView extends Con DBConfiguration.configureDB(conf, driverClassName, url); - Job job = new Job(conf); + Job job = Job.getInstance(conf); job.setJobName("Count Pageviews of URLs"); job.setJarByClass(DBCountPageView.class); @@ -404,7 +409,7 @@ public class DBCountPageView extends Con job.setCombinerClass(LongSumReducer.class); job.setReducerClass(PageviewReducer.class); - DBInputFormat.setInput(job, AccessRecord.class, "Access" + DBInputFormat.setInput(job, AccessRecord.class, "HAccess" , null, "url", AccessFieldNames); DBOutputFormat.setOutput(job, "Pageview", PageviewFieldNames); Modified: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/SecondarySort.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/SecondarySort.java?rev=1390763&r1=1390762&r2=1390763&view=diff ============================================================================== --- hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/SecondarySort.java (original) +++ hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/SecondarySort.java Wed Sep 26 22:55:00 2012 @@ -211,7 +211,7 @@ public class SecondarySort { Configuration conf = new Configuration(); String[] otherArgs = new GenericOptionsParser(conf, args).getRemainingArgs(); if (otherArgs.length != 2) { - System.err.println("Usage: secondarysrot <in> <out>"); + System.err.println("Usage: secondarysort <in> <out>"); System.exit(2); } Job job = new Job(conf, "secondary sort"); Modified: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/pom.xml URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/pom.xml?rev=1390763&r1=1390762&r2=1390763&view=diff ============================================================================== --- hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/pom.xml (original) +++ hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/pom.xml Wed Sep 26 22:55:00 2012 @@ -220,6 +220,18 @@ </includes> </configuration> </plugin> + <plugin> + <groupId>org.apache.maven.plugins</groupId> + <artifactId>maven-surefire-plugin</artifactId> + <configuration> + <properties> + <property> + <name>listener</name> + <value>org.apache.hadoop.test.TimedOutTestsListener</value> + </property> + </properties> + </configuration> + </plugin> </plugins> </build> Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/c++/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/c++:r1379224-1390762 Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/contrib/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/contrib:r1379224-1390762 Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/contrib/block_forensics/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/contrib/block_forensics:r1379224-1390762 Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/contrib/build-contrib.xml ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/contrib/build-contrib.xml:r1379224-1390762 Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/contrib/build.xml ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/contrib/build.xml:r1379224-1390762 Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/contrib/data_join/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/contrib/data_join:r1379224-1390762 Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/contrib/eclipse-plugin/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/contrib/eclipse-plugin:r1379224-1390762 Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/contrib/index/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/contrib/index:r1379224-1390762 Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/contrib/vaidya/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/contrib/vaidya:r1379224-1390762 Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/examples/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/examples:r1379224-1390762 Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/java/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/java:r1379224-1390762 Modified: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/java/mapred-default.xml URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/java/mapred-default.xml?rev=1390763&r1=1390762&r2=1390763&view=diff ============================================================================== --- hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/java/mapred-default.xml (original) +++ hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/java/mapred-default.xml Wed Sep 26 22:55:00 2012 @@ -701,7 +701,7 @@ <property> <name>mapreduce.job.maxtaskfailures.per.tracker</name> - <value>4</value> + <value>3</value> <description>The number of task-failures on a tasktracker of a given job after which new tasks of that job aren't assigned to it. </description> Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/test/mapred/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/test/mapred:r1379224-1390762 Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/fs/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/fs:r1379224-1390762 Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/hdfs/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/hdfs:r1379224-1390762 Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/ipc/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/test/mapred/org/apache/hadoop/ipc:r1379224-1390762 Propchange: hadoop/common/branches/branch-trunk-win/hadoop-mapreduce-project/src/webapps/job/ ------------------------------------------------------------------------------ Merged /hadoop/common/trunk/hadoop-mapreduce-project/src/webapps/job:r1379224-1390762
