hadoop-common-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From d...@apache.org
Subject svn commit: r642260 - /hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestFileInputFormatPathFilter.java
Date Fri, 28 Mar 2008 15:26:51 GMT
Author: ddas
Date: Fri Mar 28 08:26:47 2008
New Revision: 642260

URL: http://svn.apache.org/viewvc?rev=642260&view=rev
Log:
HADOOP-2055. Adding the testcase missed in the earlier commit of this issue.

Added:
    hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestFileInputFormatPathFilter.java

Added: hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestFileInputFormatPathFilter.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestFileInputFormatPathFilter.java?rev=642260&view=auto
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestFileInputFormatPathFilter.java
(added)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/mapred/TestFileInputFormatPathFilter.java
Fri Mar 28 08:26:47 2008
@@ -0,0 +1,149 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.mapred;
+
+import junit.framework.TestCase;
+import org.apache.hadoop.fs.FileSystem;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.fs.PathFilter;
+
+import java.io.IOException;
+import java.io.Writer;
+import java.io.OutputStreamWriter;
+import java.util.Set;
+import java.util.HashSet;
+
+public class TestFileInputFormatPathFilter extends TestCase {
+
+  public static class DummyFileInputFormat extends FileInputFormat {
+
+    public RecordReader getRecordReader(InputSplit split, JobConf job,
+                                        Reporter reporter) throws IOException {
+      return null;
+    }
+
+    public Path[] listPaths(JobConf job) throws IOException {
+      return super.listPaths(job);
+    }
+
+  }
+
+  private static FileSystem localFs = null;
+
+  static {
+    try {
+      localFs = FileSystem.getLocal(new JobConf());
+    } catch (IOException e) {
+      throw new RuntimeException("init failure", e);
+    }
+  }
+
+  private static Path workDir =
+      new Path(new Path(System.getProperty("test.build.data", "."), "data"),
+          "TestFileInputFormatPathFilter");
+
+
+  public void setUp() throws Exception {
+    tearDown();
+    localFs.mkdirs(workDir);
+  }
+
+  public void tearDown() throws Exception {
+    if (localFs.exists(workDir)) {
+      localFs.delete(workDir, true);
+    }
+  }
+
+  protected Path createFile(String fileName) throws IOException {
+    Path file = new Path(workDir, fileName);
+    Writer writer = new OutputStreamWriter(localFs.create(file));
+    writer.write("");
+    writer.close();
+    return localFs.makeQualified(file);
+  }
+
+  protected Set<Path> createFiles() throws IOException {
+    Set<Path> files = new HashSet<Path>();
+    files.add(createFile("a"));
+    files.add(createFile("b"));
+    files.add(createFile("aa"));
+    files.add(createFile("bb"));
+    files.add(createFile("_hello"));
+    files.add(createFile(".hello"));
+    return files;
+  }
+
+
+  public static class TestPathFilter implements PathFilter {
+
+    public boolean accept(Path path) {
+      return path.getName().length() == 1;
+    }
+  }
+
+  private void _testInputFiles(boolean withFilter, boolean withGlob) throws Exception {
+    Set<Path> createdFiles = createFiles();
+    JobConf conf = new JobConf();
+
+    Path inputDir = (withGlob) ? new Path(workDir, "a*") : workDir;
+    conf.setInputPath(inputDir);
+    conf.setInputFormat(DummyFileInputFormat.class);
+
+    if (withFilter) {
+      FileInputFormat.setInputPathFilter(conf, TestPathFilter.class);
+    }
+
+    DummyFileInputFormat inputFormat =
+        (DummyFileInputFormat) conf.getInputFormat();
+    Path[] listPaths = inputFormat.listPaths(conf);
+    Set<Path> computedFiles = new HashSet<Path>();
+    for (Path path : listPaths) {
+      computedFiles.add(path);
+    }
+
+    createdFiles.remove(localFs.makeQualified(new Path(workDir, "_hello")));
+    createdFiles.remove(localFs.makeQualified(new Path(workDir, ".hello")));
+    
+    if (withFilter) {
+      createdFiles.remove(localFs.makeQualified(new Path(workDir, "aa")));
+      createdFiles.remove(localFs.makeQualified(new Path(workDir, "bb")));
+    }
+
+    if (withGlob) {
+      createdFiles.remove(localFs.makeQualified(new Path(workDir, "b")));
+      createdFiles.remove(localFs.makeQualified(new Path(workDir, "bb")));
+    }
+    assertEquals(createdFiles, computedFiles);
+  }
+
+  public void testWithoutPathFilterWithoutGlob() throws Exception {
+    _testInputFiles(false, false);
+  }
+
+  public void testWithoutPathFilterWithGlob() throws Exception {
+    _testInputFiles(false, true);
+  }
+
+  public void testWithPathFilterWithoutGlob() throws Exception {
+    _testInputFiles(true, false);
+  }
+
+  public void testWithPathFilterWithGlob() throws Exception {
+    _testInputFiles(true, true);
+  }
+}



Mime
View raw message