beam-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From tg...@apache.org
Subject [1/2] beam git commit: Revert "Add HadoopResourceId"
Date Tue, 25 Apr 2017 18:37:00 GMT
Repository: beam
Updated Branches:
  refs/heads/master 1a79635fe -> ed97edd2e


Revert "Add HadoopResourceId"

This reverts commit 512b2c0abb68c777c207dcec39de5c44e6ff34da.

Breaks FindBugs, RAT


Project: http://git-wip-us.apache.org/repos/asf/beam/repo
Commit: http://git-wip-us.apache.org/repos/asf/beam/commit/46c4caf4
Tree: http://git-wip-us.apache.org/repos/asf/beam/tree/46c4caf4
Diff: http://git-wip-us.apache.org/repos/asf/beam/diff/46c4caf4

Branch: refs/heads/master
Commit: 46c4caf4bd188f0da752b5320c424a65521d1c02
Parents: 1a79635
Author: Thomas Groh <tgroh@google.com>
Authored: Tue Apr 25 11:35:16 2017 -0700
Committer: Thomas Groh <tgroh@google.com>
Committed: Tue Apr 25 11:36:11 2017 -0700

----------------------------------------------------------------------
 .../beam/sdk/io/hdfs/HadoopResourceId.java      | 47 +-------------
 .../beam/sdk/io/hdfs/HadoopResourceIdTest.java  | 66 --------------------
 2 files changed, 2 insertions(+), 111 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/beam/blob/46c4caf4/sdks/java/io/hdfs/src/main/java/org/apache/beam/sdk/io/hdfs/HadoopResourceId.java
----------------------------------------------------------------------
diff --git a/sdks/java/io/hdfs/src/main/java/org/apache/beam/sdk/io/hdfs/HadoopResourceId.java
b/sdks/java/io/hdfs/src/main/java/org/apache/beam/sdk/io/hdfs/HadoopResourceId.java
index 5a66ac9..2a29bb9 100644
--- a/sdks/java/io/hdfs/src/main/java/org/apache/beam/sdk/io/hdfs/HadoopResourceId.java
+++ b/sdks/java/io/hdfs/src/main/java/org/apache/beam/sdk/io/hdfs/HadoopResourceId.java
@@ -17,69 +17,26 @@
  */
 package org.apache.beam.sdk.io.hdfs;
 
-import static com.google.common.base.Preconditions.checkArgument;
-import static com.google.common.base.Preconditions.checkNotNull;
-
 import org.apache.beam.sdk.io.fs.ResolveOptions;
 import org.apache.beam.sdk.io.fs.ResourceId;
-import org.apache.hadoop.fs.Path;
 
 /**
  * {@link ResourceId} implementation for the {@link HadoopFileSystem}.
  */
 public class HadoopResourceId implements ResourceId {
 
-  private final Path path;
-
-  /**
-   * Constructs a HadoopResourceId from the provided absolute path. If only a relative path
is
-   * available, you can create a {@link HadoopResourceId} from the absolute path of the root
of the
-   * server, and then use resolve to add the relative path to the root.
-   */
-  public static HadoopResourceId fromPath(Path path) {
-    checkNotNull(path, "path must not be null");
-    checkArgument(path.isAbsolute(), "path must be absolute");
-    return new HadoopResourceId(path);
-  }
-
-  private HadoopResourceId(Path path) {
-    this.path = path;
-  }
-
   @Override
   public ResourceId resolve(String other, ResolveOptions resolveOptions) {
-    checkArgument(
-        resolveOptions.equals(ResolveOptions.StandardResolveOptions.RESOLVE_FILE)
-            || resolveOptions.equals(ResolveOptions.StandardResolveOptions.RESOLVE_DIRECTORY),
-        String.format("ResolveOptions: [%s] is not supported. "
-            + "Supported ResolveOptions are RESOLVE_FILE and RESOLVE_DIRECTORY.", resolveOptions));
-    if (resolveOptions.equals(ResolveOptions.StandardResolveOptions.RESOLVE_FILE)) {
-      checkArgument(
-          !other.endsWith("/"),
-          "ResolveOptions: [%s] ends with '/', which is not supported for RESOLVE_FILE.",
-          other);
-    }
-    return new HadoopResourceId(new Path(path, other));
+    throw new UnsupportedOperationException();
   }
 
   @Override
   public ResourceId getCurrentDirectory() {
-    // See BEAM-2069. Possible workaround: inject FileSystem into this class, and call
-    // org.apache.hadoop.fs.FileSystem#isDirectory.
     throw new UnsupportedOperationException();
   }
 
   @Override
   public String getScheme() {
-    return path.toUri().getScheme();
-  }
-
-  public Path getPath() {
-    return path;
-  }
-
-  @Override
-  public String toString() {
-    return path.toString();
+    throw new UnsupportedOperationException();
   }
 }

http://git-wip-us.apache.org/repos/asf/beam/blob/46c4caf4/sdks/java/io/hdfs/src/test/java/org/apache/beam/sdk/io/hdfs/HadoopResourceIdTest.java
----------------------------------------------------------------------
diff --git a/sdks/java/io/hdfs/src/test/java/org/apache/beam/sdk/io/hdfs/HadoopResourceIdTest.java
b/sdks/java/io/hdfs/src/test/java/org/apache/beam/sdk/io/hdfs/HadoopResourceIdTest.java
deleted file mode 100644
index 724bcd9..0000000
--- a/sdks/java/io/hdfs/src/test/java/org/apache/beam/sdk/io/hdfs/HadoopResourceIdTest.java
+++ /dev/null
@@ -1,66 +0,0 @@
-package org.apache.beam.sdk.io.hdfs;
-
-import static org.junit.Assert.assertEquals;
-
-import org.apache.beam.sdk.io.fs.ResolveOptions;
-import org.apache.hadoop.fs.Path;
-import org.junit.Test;
-
-/**
- * Tests for the HadoopResourceId class.
- */
-public class HadoopResourceIdTest {
-  @Test
-  public void fromAndToPath() {
-    // Directory path without slash
-    Path dirPathWithoutSlash = new Path("hdfs://myhost/mydir");
-    HadoopResourceId resourceDirWithoutSlash = HadoopResourceId.fromPath(dirPathWithoutSlash);
-    assertEquals("hdfs://myhost/mydir",
-        resourceDirWithoutSlash.toString());
-    assertEquals(dirPathWithoutSlash, resourceDirWithoutSlash.getPath());
-
-    // Directory path with slash
-    Path dirPathWithSlash = new Path("hdfs://myhost/mydir/");
-    HadoopResourceId resourceDirWithSlash = HadoopResourceId.fromPath(dirPathWithSlash);
-    assertEquals("hdfs://myhost/mydir",
-        resourceDirWithSlash.toString());
-    assertEquals(dirPathWithSlash, resourceDirWithSlash.getPath());
-
-    // File path
-    Path filePath = new Path("hdfs://myhost/mydir/myfile.txt");
-    HadoopResourceId resourceFile = HadoopResourceId.fromPath(filePath);
-    assertEquals("hdfs://myhost/mydir/myfile.txt",
-        resourceFile.toString());
-    assertEquals(filePath, resourceFile.getPath());
-  }
-
-  @Test
-  public void handlesRelativePathsAddedToDir() {
-    // Directory + file - slash on Directory
-    HadoopResourceId dirWithSlash = HadoopResourceId.fromPath(new Path("hdfs://myhost/mydir/"));
-    assertEquals("hdfs://myhost/mydir/myfile.txt",
-        dirWithSlash.resolve("myfile.txt",
-            ResolveOptions.StandardResolveOptions.RESOLVE_FILE).toString());
-
-    // Directory + Directory
-    assertEquals("hdfs://myhost/mydir/2nddir",
-        dirWithSlash.resolve("2nddir",
-            ResolveOptions.StandardResolveOptions.RESOLVE_DIRECTORY).toString());
-    assertEquals("hdfs://myhost/mydir/2nddir",
-        dirWithSlash.resolve("2nddir/",
-            ResolveOptions.StandardResolveOptions.RESOLVE_DIRECTORY).toString());
-
-
-    // Directory + File - no slash on either
-    HadoopResourceId dirWithoutSlash = HadoopResourceId.fromPath(new Path("hdfs://myhost/mydir"));
-    assertEquals("hdfs://myhost/mydir/myfile.txt",
-        dirWithoutSlash.resolve("myfile.txt",
-            ResolveOptions.StandardResolveOptions.RESOLVE_FILE).toString());
-  }
-
-  @Test
-  public void testScheme() {
-    assertEquals("hdfs",
-        HadoopResourceId.fromPath(new Path("hdfs://myhost/mydir/file.txt")).getScheme());
-  }
-}


Mime
View raw message