Author: tucu
Date: Thu Jan 10 17:27:58 2013
New Revision: 1431510

URL: http://svn.apache.org/viewvc?rev=1431510&view=rev
Log:
MAPREDUCE-4907. Ammendment, forgot to svn add testcase in original commit

Added:
    
hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/filecache/
    
hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/filecache/TestClientDistributedCacheManager.java

Added: 
hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/filecache/TestClientDistributedCacheManager.java
URL: 
http://svn.apache.org/viewvc/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/filecache/TestClientDistributedCacheManager.java?rev=1431510&view=auto
==============================================================================
--- 
hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/filecache/TestClientDistributedCacheManager.java
 (added)
+++ 
hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/filecache/TestClientDistributedCacheManager.java
 Thu Jan 10 17:27:58 2013
@@ -0,0 +1,114 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.mapreduce.filecache;
+
+import java.io.File;
+import java.io.IOException;
+import java.net.URI;
+import java.util.HashMap;
+import java.util.Map;
+
+import org.apache.commons.logging.Log;
+import org.apache.commons.logging.LogFactory;
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.FileStatus;
+import org.apache.hadoop.fs.FileSystem;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.io.SequenceFile;
+import org.apache.hadoop.io.Text;
+import org.apache.hadoop.io.SequenceFile.CompressionType;
+import org.apache.hadoop.mapreduce.Job;
+import org.apache.hadoop.mapreduce.MRJobConfig;
+import org.junit.After;
+import org.junit.Assert;
+import org.junit.Before;
+import org.junit.Test;
+
+public class TestClientDistributedCacheManager {
+  private static final Log LOG = LogFactory.getLog(
+      TestClientDistributedCacheManager.class);
+  
+  private static final String TEST_ROOT_DIR = 
+      new File(System.getProperty("test.build.data", "/tmp")).toURI()
+      .toString().replace(' ', '+');
+  
+  private FileSystem fs;
+  private Path firstCacheFile;
+  private Path secondCacheFile;
+  private Configuration conf;
+  
+  @Before
+  public void setup() throws IOException {
+    conf = new Configuration();
+    fs = FileSystem.get(conf);
+    firstCacheFile = new Path(TEST_ROOT_DIR, "firstcachefile");
+    secondCacheFile = new Path(TEST_ROOT_DIR, "secondcachefile");
+    createTempFile(firstCacheFile, conf);
+    createTempFile(secondCacheFile, conf);
+  }
+  
+  @After
+  public void tearDown() throws IOException {
+    if (!fs.delete(firstCacheFile, false)) {
+      LOG.warn("Failed to delete firstcachefile");
+    }
+    if (!fs.delete(secondCacheFile, false)) {
+      LOG.warn("Failed to delete secondcachefile");
+    }
+  }
+  
+  @Test
+  public void testDetermineTimestamps() throws IOException {
+    Job job = Job.getInstance(conf);
+    job.addCacheFile(firstCacheFile.toUri());
+    job.addCacheFile(secondCacheFile.toUri());
+    Configuration jobConf = job.getConfiguration();
+    
+    Map<URI, FileStatus> statCache = new HashMap<URI, FileStatus>();
+    ClientDistributedCacheManager.determineTimestamps(jobConf, statCache);
+    
+    FileStatus firstStatus = statCache.get(firstCacheFile.toUri());
+    FileStatus secondStatus = statCache.get(secondCacheFile.toUri());
+    
+    Assert.assertNotNull(firstStatus);
+    Assert.assertNotNull(secondStatus);
+    Assert.assertEquals(2, statCache.size());
+    String expected = firstStatus.getModificationTime() + ","
+        + secondStatus.getModificationTime();
+    Assert.assertEquals(expected, 
jobConf.get(MRJobConfig.CACHE_FILE_TIMESTAMPS));
+  }
+  
+  @SuppressWarnings("deprecation")
+  void createTempFile(Path p, Configuration conf) throws IOException {
+    SequenceFile.Writer writer = null;
+    try {
+      writer = SequenceFile.createWriter(fs, conf, p,
+                                         Text.class, Text.class,
+                                         CompressionType.NONE);
+      writer.append(new Text("text"), new Text("moretext"));
+    } catch(Exception e) {
+      throw new IOException(e.getLocalizedMessage());
+    } finally {
+      if (writer != null) {
+        writer.close();
+      }
+      writer = null;
+    }
+    LOG.info("created: " + p);
+  }
+}


Reply via email to