hadoop-common-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From whe...@apache.org
Subject hadoop git commit: HDFS-7404. Remove o.a.h.hdfs.server.datanode.web.resources. Contributed by Li Lu.
Date Tue, 18 Nov 2014 01:36:01 GMT
Repository: hadoop
Updated Branches:
  refs/heads/trunk dcb8e2442 -> 9dd5d673c


HDFS-7404. Remove o.a.h.hdfs.server.datanode.web.resources. Contributed by Li Lu.


Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/9dd5d673
Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/9dd5d673
Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/9dd5d673

Branch: refs/heads/trunk
Commit: 9dd5d673c92db3d6c4cf2783b3507fa9af083560
Parents: dcb8e24
Author: Haohui Mai <wheat9@apache.org>
Authored: Mon Nov 17 17:35:36 2014 -0800
Committer: Haohui Mai <wheat9@apache.org>
Committed: Mon Nov 17 17:35:36 2014 -0800

----------------------------------------------------------------------
 hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt     |   3 +
 .../web/resources/DatanodeWebHdfsMethods.java   | 477 -------------------
 .../datanode/web/resources/OpenEntity.java      |  92 ----
 .../resources/TestDatanodeWebHdfsMethods.java   |  54 ---
 .../web/webhdfs/TestParameterParser.java        |  55 +++
 .../hdfs/web/TestFSMainOperationsWebHdfs.java   |   2 -
 6 files changed, 58 insertions(+), 625 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hadoop/blob/9dd5d673/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
index 903b890..45cd6b9 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
+++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
@@ -367,6 +367,9 @@ Release 2.7.0 - UNRELEASED
 
     HDFS-7279. Use netty to implement DatanodeWebHdfsMethods. (wheat9)
 
+    HDFS-7404. Remove o.a.h.hdfs.server.datanode.web.resources.
+    (Li Lu via wheat9)
+
   OPTIMIZATIONS
 
   BUG FIXES

http://git-wip-us.apache.org/repos/asf/hadoop/blob/9dd5d673/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/web/resources/DatanodeWebHdfsMethods.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/web/resources/DatanodeWebHdfsMethods.java
b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/web/resources/DatanodeWebHdfsMethods.java
deleted file mode 100644
index 0f0f3be..0000000
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/web/resources/DatanodeWebHdfsMethods.java
+++ /dev/null
@@ -1,477 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.hadoop.hdfs.server.datanode.web.resources;
-
-import java.io.IOException;
-import java.io.InputStream;
-import java.net.URI;
-import java.net.URISyntaxException;
-import java.security.PrivilegedExceptionAction;
-import java.util.EnumSet;
-
-import javax.servlet.ServletContext;
-import javax.servlet.http.HttpServletRequest;
-import javax.servlet.http.HttpServletResponse;
-import javax.ws.rs.Consumes;
-import javax.ws.rs.DefaultValue;
-import javax.ws.rs.GET;
-import javax.ws.rs.POST;
-import javax.ws.rs.PUT;
-import javax.ws.rs.Path;
-import javax.ws.rs.PathParam;
-import javax.ws.rs.Produces;
-import javax.ws.rs.QueryParam;
-import javax.ws.rs.core.Context;
-import javax.ws.rs.core.MediaType;
-import javax.ws.rs.core.Response;
-
-import com.google.common.annotations.VisibleForTesting;
-import org.apache.commons.logging.Log;
-import org.apache.commons.logging.LogFactory;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.CreateFlag;
-import org.apache.hadoop.fs.FSDataOutputStream;
-import org.apache.hadoop.fs.MD5MD5CRC32FileChecksum;
-import org.apache.hadoop.fs.permission.FsPermission;
-import org.apache.hadoop.hdfs.DFSClient;
-import org.apache.hadoop.hdfs.HAUtil;
-import org.apache.hadoop.hdfs.client.HdfsDataInputStream;
-import org.apache.hadoop.hdfs.protocol.HdfsConstants;
-import org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenIdentifier;
-import org.apache.hadoop.hdfs.server.datanode.DataNode;
-import org.apache.hadoop.hdfs.web.JsonUtil;
-import org.apache.hadoop.hdfs.web.ParamFilter;
-import org.apache.hadoop.hdfs.web.SWebHdfsFileSystem;
-import org.apache.hadoop.hdfs.web.WebHdfsFileSystem;
-import org.apache.hadoop.hdfs.web.resources.BlockSizeParam;
-import org.apache.hadoop.hdfs.web.resources.BufferSizeParam;
-import org.apache.hadoop.hdfs.web.resources.DelegationParam;
-import org.apache.hadoop.hdfs.web.resources.GetOpParam;
-import org.apache.hadoop.hdfs.web.resources.HttpOpParam;
-import org.apache.hadoop.hdfs.web.resources.LengthParam;
-import org.apache.hadoop.hdfs.web.resources.NamenodeAddressParam;
-import org.apache.hadoop.hdfs.web.resources.OffsetParam;
-import org.apache.hadoop.hdfs.web.resources.OverwriteParam;
-import org.apache.hadoop.hdfs.web.resources.Param;
-import org.apache.hadoop.hdfs.web.resources.PermissionParam;
-import org.apache.hadoop.hdfs.web.resources.PostOpParam;
-import org.apache.hadoop.hdfs.web.resources.PutOpParam;
-import org.apache.hadoop.hdfs.web.resources.ReplicationParam;
-import org.apache.hadoop.hdfs.web.resources.UriFsPathParam;
-import org.apache.hadoop.io.IOUtils;
-import org.apache.hadoop.security.SecurityUtil;
-import org.apache.hadoop.security.UserGroupInformation;
-import org.apache.hadoop.security.token.Token;
-
-import com.sun.jersey.spi.container.ResourceFilters;
-
-/** Web-hdfs DataNode implementation. */
-@Path("")
-@ResourceFilters(ParamFilter.class)
-public class DatanodeWebHdfsMethods {
-  public static final Log LOG = LogFactory.getLog(DatanodeWebHdfsMethods.class);
-
-  private static final UriFsPathParam ROOT = new UriFsPathParam("");
-
-  private @Context ServletContext context;
-  private @Context HttpServletRequest request;
-  private @Context HttpServletResponse response;
-
-  private void init(final UserGroupInformation ugi,
-      final DelegationParam delegation, final String nnId,
-      final UriFsPathParam path, final HttpOpParam<?> op,
-      final Param<?, ?>... parameters) throws IOException {
-    if (LOG.isTraceEnabled()) {
-      LOG.trace("HTTP " + op.getValue().getType() + ": " + op + ", " + path
-          + ", ugi=" + ugi + Param.toSortedString(", ", parameters));
-    }
-    if (nnId == null) {
-      throw new IllegalArgumentException(NamenodeAddressParam.NAME
-          + " is not specified.");
-    }
-
-    //clear content type
-    response.setContentType(null);
-    
-    if (UserGroupInformation.isSecurityEnabled()) {
-      //add a token for RPC.
-      final Token<DelegationTokenIdentifier> token = deserializeToken
-              (delegation.getValue(), nnId);
-      ugi.addToken(token);
-    }
-  }
-
-  @VisibleForTesting
-  Token<DelegationTokenIdentifier> deserializeToken
-          (String delegation,String nnId) throws IOException {
-    final DataNode datanode = (DataNode) context.getAttribute("datanode");
-    final Configuration conf = datanode.getConf();
-    final Token<DelegationTokenIdentifier> token = new
-            Token<DelegationTokenIdentifier>();
-    token.decodeFromUrlString(delegation);
-    URI nnUri = URI.create(HdfsConstants.HDFS_URI_SCHEME +
-            "://" + nnId);
-    boolean isLogical = HAUtil.isLogicalUri(conf, nnUri);
-    if (isLogical) {
-      token.setService(HAUtil.buildTokenServiceForLogicalUri(nnUri,
-          HdfsConstants.HDFS_URI_SCHEME));
-    } else {
-      token.setService(SecurityUtil.buildTokenService(nnUri));
-    }
-    token.setKind(DelegationTokenIdentifier.HDFS_DELEGATION_KIND);
-    return token;
-  }
-
-  /** Handle HTTP PUT request for the root. */
-  @PUT
-  @Path("/")
-  @Consumes({"*/*"})
-  @Produces({MediaType.APPLICATION_OCTET_STREAM, MediaType.APPLICATION_JSON})
-  public Response putRoot(
-      final InputStream in,
-      @Context final UserGroupInformation ugi,
-      @QueryParam(DelegationParam.NAME) @DefaultValue(DelegationParam.DEFAULT)
-          final DelegationParam delegation,
-      @QueryParam(NamenodeAddressParam.NAME)
-      @DefaultValue(NamenodeAddressParam.DEFAULT)
-          final NamenodeAddressParam namenode,
-      @QueryParam(PutOpParam.NAME) @DefaultValue(PutOpParam.DEFAULT)
-          final PutOpParam op,
-      @QueryParam(PermissionParam.NAME) @DefaultValue(PermissionParam.DEFAULT)
-          final PermissionParam permission,
-      @QueryParam(OverwriteParam.NAME) @DefaultValue(OverwriteParam.DEFAULT)
-          final OverwriteParam overwrite,
-      @QueryParam(BufferSizeParam.NAME) @DefaultValue(BufferSizeParam.DEFAULT)
-          final BufferSizeParam bufferSize,
-      @QueryParam(ReplicationParam.NAME) @DefaultValue(ReplicationParam.DEFAULT)
-          final ReplicationParam replication,
-      @QueryParam(BlockSizeParam.NAME) @DefaultValue(BlockSizeParam.DEFAULT)
-          final BlockSizeParam blockSize
-      ) throws IOException, InterruptedException {
-    return put(in, ugi, delegation, namenode, ROOT, op, permission,
-            overwrite, bufferSize, replication, blockSize);
-  }
-
-  /** Handle HTTP PUT request. */
-  @PUT
-  @Path("{" + UriFsPathParam.NAME + ":.*}")
-  @Consumes({"*/*"})
-  @Produces({MediaType.APPLICATION_OCTET_STREAM, MediaType.APPLICATION_JSON})
-  public Response put(
-      final InputStream in,
-      @Context final UserGroupInformation ugi,
-      @QueryParam(DelegationParam.NAME) @DefaultValue(DelegationParam.DEFAULT)
-          final DelegationParam delegation,
-      @QueryParam(NamenodeAddressParam.NAME)
-      @DefaultValue(NamenodeAddressParam.DEFAULT)
-          final NamenodeAddressParam namenode,
-      @PathParam(UriFsPathParam.NAME) final UriFsPathParam path,
-      @QueryParam(PutOpParam.NAME) @DefaultValue(PutOpParam.DEFAULT)
-          final PutOpParam op,
-      @QueryParam(PermissionParam.NAME) @DefaultValue(PermissionParam.DEFAULT)
-          final PermissionParam permission,
-      @QueryParam(OverwriteParam.NAME) @DefaultValue(OverwriteParam.DEFAULT)
-          final OverwriteParam overwrite,
-      @QueryParam(BufferSizeParam.NAME) @DefaultValue(BufferSizeParam.DEFAULT)
-          final BufferSizeParam bufferSize,
-      @QueryParam(ReplicationParam.NAME) @DefaultValue(ReplicationParam.DEFAULT)
-          final ReplicationParam replication,
-      @QueryParam(BlockSizeParam.NAME) @DefaultValue(BlockSizeParam.DEFAULT)
-          final BlockSizeParam blockSize
-      ) throws IOException, InterruptedException {
-
-    final String nnId = namenode.getValue();
-    init(ugi, delegation, nnId, path, op, permission,
-        overwrite, bufferSize, replication, blockSize);
-
-    return ugi.doAs(new PrivilegedExceptionAction<Response>() {
-      @Override
-      public Response run() throws IOException, URISyntaxException {
-        return put(in, nnId, path.getAbsolutePath(), op,
-                permission, overwrite, bufferSize, replication, blockSize);
-      }
-    });
-  }
-
-  private Response put(
-      final InputStream in,
-      final String nnId,
-      final String fullpath,
-      final PutOpParam op,
-      final PermissionParam permission,
-      final OverwriteParam overwrite,
-      final BufferSizeParam bufferSize,
-      final ReplicationParam replication,
-      final BlockSizeParam blockSize
-      ) throws IOException, URISyntaxException {
-    final DataNode datanode = (DataNode)context.getAttribute("datanode");
-
-    switch(op.getValue()) {
-    case CREATE:
-    {
-      final Configuration conf = new Configuration(datanode.getConf());
-      conf.set(FsPermission.UMASK_LABEL, "000");
-
-      final int b = bufferSize.getValue(conf);
-      DFSClient dfsclient = newDfsClient(nnId, conf);
-      FSDataOutputStream out = null;
-      try {
-        out = dfsclient.createWrappedOutputStream(dfsclient.create(
-            fullpath, permission.getFsPermission(), 
-            overwrite.getValue() ?
-                EnumSet.of(CreateFlag.CREATE, CreateFlag.OVERWRITE) :
-                EnumSet.of(CreateFlag.CREATE),
-            replication.getValue(conf), blockSize.getValue(conf), null,
-            b, null), null);
-        IOUtils.copyBytes(in, out, b);
-        out.close();
-        out = null;
-        dfsclient.close();
-        dfsclient = null;
-      } finally {
-        IOUtils.cleanup(LOG, out);
-        IOUtils.cleanup(LOG, dfsclient);
-      }
-      final String scheme = "http".equals(request.getScheme()) ?
-      WebHdfsFileSystem.SCHEME : SWebHdfsFileSystem.SCHEME;
-      final URI uri = new URI(scheme, nnId, fullpath, null, null);
-      return Response.created(uri).type(MediaType.APPLICATION_OCTET_STREAM).build();
-    }
-    default:
-      throw new UnsupportedOperationException(op + " is not supported");
-    }
-  }
-
-  /** Handle HTTP POST request for the root for the root. */
-  @POST
-  @Path("/")
-  @Consumes({"*/*"})
-  @Produces({MediaType.APPLICATION_OCTET_STREAM, MediaType.APPLICATION_JSON})
-  public Response postRoot(
-      final InputStream in,
-      @Context final UserGroupInformation ugi,
-      @QueryParam(DelegationParam.NAME) @DefaultValue(DelegationParam.DEFAULT)
-          final DelegationParam delegation,
-      @QueryParam(NamenodeAddressParam.NAME)
-      @DefaultValue(NamenodeAddressParam.DEFAULT)
-          final NamenodeAddressParam namenode,
-      @QueryParam(PostOpParam.NAME) @DefaultValue(PostOpParam.DEFAULT)
-          final PostOpParam op,
-      @QueryParam(BufferSizeParam.NAME) @DefaultValue(BufferSizeParam.DEFAULT)
-          final BufferSizeParam bufferSize
-      ) throws IOException, InterruptedException {
-    return post(in, ugi, delegation, namenode, ROOT, op, bufferSize);
-  }
-
-  /** Handle HTTP POST request. */
-  @POST
-  @Path("{" + UriFsPathParam.NAME + ":.*}")
-  @Consumes({"*/*"})
-  @Produces({MediaType.APPLICATION_OCTET_STREAM, MediaType.APPLICATION_JSON})
-  public Response post(
-      final InputStream in,
-      @Context final UserGroupInformation ugi,
-      @QueryParam(DelegationParam.NAME) @DefaultValue(DelegationParam.DEFAULT)
-          final DelegationParam delegation,
-      @QueryParam(NamenodeAddressParam.NAME)
-      @DefaultValue(NamenodeAddressParam.DEFAULT)
-          final NamenodeAddressParam namenode,
-      @PathParam(UriFsPathParam.NAME) final UriFsPathParam path,
-      @QueryParam(PostOpParam.NAME) @DefaultValue(PostOpParam.DEFAULT)
-          final PostOpParam op,
-      @QueryParam(BufferSizeParam.NAME) @DefaultValue(BufferSizeParam.DEFAULT)
-          final BufferSizeParam bufferSize
-      ) throws IOException, InterruptedException {
-
-    final String nnId = namenode.getValue();
-    init(ugi, delegation, nnId, path, op, bufferSize);
-
-    return ugi.doAs(new PrivilegedExceptionAction<Response>() {
-      @Override
-      public Response run() throws IOException {
-        return post(in, nnId, path.getAbsolutePath(), op,
-                bufferSize);
-      }
-    });
-  }
-
-  private Response post(
-      final InputStream in,
-      final String nnId,
-      final String fullpath,
-      final PostOpParam op,
-      final BufferSizeParam bufferSize
-      ) throws IOException {
-    final DataNode datanode = (DataNode)context.getAttribute("datanode");
-
-    switch(op.getValue()) {
-    case APPEND:
-    {
-      final Configuration conf = new Configuration(datanode.getConf());
-      final int b = bufferSize.getValue(conf);
-      DFSClient dfsclient = newDfsClient(nnId, conf);
-      FSDataOutputStream out = null;
-      try {
-        out = dfsclient.append(fullpath, b, null, null);
-        IOUtils.copyBytes(in, out, b);
-        out.close();
-        out = null;
-        dfsclient.close();
-        dfsclient = null;
-      } finally {
-        IOUtils.cleanup(LOG, out);
-        IOUtils.cleanup(LOG, dfsclient);
-      }
-      return Response.ok().type(MediaType.APPLICATION_OCTET_STREAM).build();
-    }
-    default:
-      throw new UnsupportedOperationException(op + " is not supported");
-    }
-  }
-
-  /** Handle HTTP GET request for the root. */
-  @GET
-  @Path("/")
-  @Produces({MediaType.APPLICATION_OCTET_STREAM, MediaType.APPLICATION_JSON})
-  public Response getRoot(
-      @Context final UserGroupInformation ugi,
-      @QueryParam(DelegationParam.NAME) @DefaultValue(DelegationParam.DEFAULT)
-          final DelegationParam delegation,
-      @QueryParam(NamenodeAddressParam.NAME)
-      @DefaultValue(NamenodeAddressParam.DEFAULT)
-          final NamenodeAddressParam namenode,
-      @QueryParam(GetOpParam.NAME) @DefaultValue(GetOpParam.DEFAULT)
-          final GetOpParam op,
-      @QueryParam(OffsetParam.NAME) @DefaultValue(OffsetParam.DEFAULT)
-          final OffsetParam offset,
-      @QueryParam(LengthParam.NAME) @DefaultValue(LengthParam.DEFAULT)
-          final LengthParam length,
-      @QueryParam(BufferSizeParam.NAME) @DefaultValue(BufferSizeParam.DEFAULT)
-          final BufferSizeParam bufferSize
-      ) throws IOException, InterruptedException {
-    return get(ugi, delegation, namenode, ROOT, op, offset, length,
-        bufferSize);
-  }
-
-  /** Handle HTTP GET request. */
-  @GET
-  @Path("{" + UriFsPathParam.NAME + ":.*}")
-  @Produces({MediaType.APPLICATION_OCTET_STREAM, MediaType.APPLICATION_JSON})
-  public Response get(
-      @Context final UserGroupInformation ugi,
-      @QueryParam(DelegationParam.NAME) @DefaultValue(DelegationParam.DEFAULT)
-          final DelegationParam delegation,
-      @QueryParam(NamenodeAddressParam.NAME)
-      @DefaultValue(NamenodeAddressParam.DEFAULT)
-          final NamenodeAddressParam namenode,
-      @PathParam(UriFsPathParam.NAME) final UriFsPathParam path,
-      @QueryParam(GetOpParam.NAME) @DefaultValue(GetOpParam.DEFAULT)
-          final GetOpParam op,
-      @QueryParam(OffsetParam.NAME) @DefaultValue(OffsetParam.DEFAULT)
-          final OffsetParam offset,
-      @QueryParam(LengthParam.NAME) @DefaultValue(LengthParam.DEFAULT)
-          final LengthParam length,
-      @QueryParam(BufferSizeParam.NAME) @DefaultValue(BufferSizeParam.DEFAULT)
-          final BufferSizeParam bufferSize
-      ) throws IOException, InterruptedException {
-
-    final String nnId = namenode.getValue();
-    init(ugi, delegation, nnId, path, op, offset, length, bufferSize);
-
-    return ugi.doAs(new PrivilegedExceptionAction<Response>() {
-      @Override
-      public Response run() throws IOException {
-        return get(nnId, path.getAbsolutePath(), op, offset,
-                length, bufferSize);
-      }
-    });
-  }
-
-  private Response get(
-      final String nnId,
-      final String fullpath,
-      final GetOpParam op,
-      final OffsetParam offset,
-      final LengthParam length,
-      final BufferSizeParam bufferSize
-      ) throws IOException {
-    final DataNode datanode = (DataNode)context.getAttribute("datanode");
-    final Configuration conf = new Configuration(datanode.getConf());
-
-    switch(op.getValue()) {
-    case OPEN:
-    {
-      final int b = bufferSize.getValue(conf);
-      final DFSClient dfsclient = newDfsClient(nnId, conf);
-      HdfsDataInputStream in = null;
-      try {
-        in = dfsclient.createWrappedInputStream(
-            dfsclient.open(fullpath, b, true));
-        in.seek(offset.getValue());
-      } catch(IOException ioe) {
-        IOUtils.cleanup(LOG, in);
-        IOUtils.cleanup(LOG, dfsclient);
-        throw ioe;
-      }
-      
-      final long n = length.getValue() != null ?
-        Math.min(length.getValue(), in.getVisibleLength() - offset.getValue()) :
-        in.getVisibleLength() - offset.getValue();
-
-      // jetty 6 reserves 12 bytes in the out buffer for chunked responses
-      // (file length > 2GB) which causes extremely poor performance when
-      // 12 bytes of the output spill into another buffer which results
-      // in a big and little write
-      int outBufferSize = response.getBufferSize();
-      if (n > Integer.MAX_VALUE) {
-        outBufferSize -= 12;
-      }
-      /**
-       * Allow the Web UI to perform an AJAX request to get the data.
-       */
-      return Response.ok(new OpenEntity(in, n, outBufferSize, dfsclient))
-          .type(MediaType.APPLICATION_OCTET_STREAM)
-          .header("Access-Control-Allow-Methods", "GET")
-          .header("Access-Control-Allow-Origin", "*")
-          .build();
-    }
-    case GETFILECHECKSUM:
-    {
-      MD5MD5CRC32FileChecksum checksum = null;
-      DFSClient dfsclient = newDfsClient(nnId, conf);
-      try {
-        checksum = dfsclient.getFileChecksum(fullpath, Long.MAX_VALUE);
-        dfsclient.close();
-        dfsclient = null;
-      } finally {
-        IOUtils.cleanup(LOG, dfsclient);
-      }
-      final String js = JsonUtil.toJsonString(checksum);
-      return Response.ok(js).type(MediaType.APPLICATION_JSON).build();
-    }
-    default:
-      throw new UnsupportedOperationException(op + " is not supported");
-    }
-  }
-
-  private static DFSClient newDfsClient(String nnId,
-                                        Configuration conf) throws IOException {
-    URI uri = URI.create(HdfsConstants.HDFS_URI_SCHEME + "://" + nnId);
-    return new DFSClient(uri, conf);
-  }
-}

http://git-wip-us.apache.org/repos/asf/hadoop/blob/9dd5d673/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/web/resources/OpenEntity.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/web/resources/OpenEntity.java
b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/web/resources/OpenEntity.java
deleted file mode 100644
index 1596f3d..0000000
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/web/resources/OpenEntity.java
+++ /dev/null
@@ -1,92 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.hadoop.hdfs.server.datanode.web.resources;
-
-import org.apache.hadoop.hdfs.DFSClient;
-import org.apache.hadoop.hdfs.client.HdfsDataInputStream;
-import org.apache.hadoop.io.IOUtils;
-
-import javax.ws.rs.core.MediaType;
-import javax.ws.rs.core.MultivaluedMap;
-import javax.ws.rs.ext.MessageBodyWriter;
-import javax.ws.rs.ext.Provider;
-import java.io.IOException;
-import java.io.OutputStream;
-import java.lang.annotation.Annotation;
-import java.lang.reflect.Type;
-
-/**
- * A response entity for a HdfsDataInputStream.
- */
-public class OpenEntity {
-  private final HdfsDataInputStream in;
-  private final long length;
-  private final int outBufferSize;
-  private final DFSClient dfsclient;
-  
-  OpenEntity(final HdfsDataInputStream in, final long length,
-      final int outBufferSize, final DFSClient dfsclient) {
-    this.in = in;
-    this.length = length;
-    this.outBufferSize = outBufferSize;
-    this.dfsclient = dfsclient;
-  }
-  
-  /**
-   * A {@link MessageBodyWriter} for {@link OpenEntity}.
-   */
-  @Provider
-  public static class Writer implements MessageBodyWriter<OpenEntity> {
-
-    @Override
-    public boolean isWriteable(Class<?> clazz, Type genericType,
-        Annotation[] annotations, MediaType mediaType) {
-      return clazz == OpenEntity.class
-          && MediaType.APPLICATION_OCTET_STREAM_TYPE.isCompatible(mediaType);
-    }
-
-    @Override
-    public long getSize(OpenEntity e, Class<?> type, Type genericType,
-        Annotation[] annotations, MediaType mediaType) {
-      return e.length;
-    }
-
-    @Override
-    public void writeTo(OpenEntity e, Class<?> type, Type genericType,
-        Annotation[] annotations, MediaType mediaType,
-        MultivaluedMap<String, Object> httpHeaders, OutputStream out
-        ) throws IOException {
-      try {
-        byte[] buf = new byte[e.outBufferSize];
-        long remaining = e.length;
-        while (remaining > 0) {
-          int read = e.in.read(buf, 0, (int)Math.min(buf.length, remaining));
-          if (read == -1) { // EOF
-            break;
-          }
-          out.write(buf, 0, read);
-          out.flush();
-          remaining -= read;
-        }
-      } finally {
-        IOUtils.cleanup(DatanodeWebHdfsMethods.LOG, e.in);
-        IOUtils.cleanup(DatanodeWebHdfsMethods.LOG, e.dfsclient);
-      }
-    }
-  }
-}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hadoop/blob/9dd5d673/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/web/resources/TestDatanodeWebHdfsMethods.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/web/resources/TestDatanodeWebHdfsMethods.java
b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/web/resources/TestDatanodeWebHdfsMethods.java
deleted file mode 100644
index 5b45054..0000000
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/web/resources/TestDatanodeWebHdfsMethods.java
+++ /dev/null
@@ -1,54 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.hadoop.hdfs.server.datanode.web.resources;
-
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.hdfs.DFSTestUtil;
-import org.apache.hadoop.hdfs.HAUtil;
-import org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenIdentifier;
-import org.apache.hadoop.hdfs.server.datanode.DataNode;
-import org.apache.hadoop.security.token.Token;
-import org.junit.Assert;
-import org.junit.Test;
-import org.mockito.internal.util.reflection.Whitebox;
-
-import javax.servlet.ServletContext;
-import java.io.IOException;
-
-import static org.mockito.Mockito.doReturn;
-import static org.mockito.Mockito.mock;
-
-public class TestDatanodeWebHdfsMethods {
-  private static final String LOGICAL_NAME = "minidfs";
-
-  @Test
-  public void testDeserializeHAToken() throws IOException {
-    Configuration conf = DFSTestUtil.newHAConfiguration(LOGICAL_NAME);
-    DataNode dn = mock(DataNode.class);
-    doReturn(conf).when(dn).getConf();
-    ServletContext context = mock(ServletContext.class);
-    doReturn(dn).when(context).getAttribute("datanode");
-    final Token<DelegationTokenIdentifier> token = new
-            Token<DelegationTokenIdentifier>();
-    DatanodeWebHdfsMethods method = new DatanodeWebHdfsMethods();
-    Whitebox.setInternalState(method, "context", context);
-    final Token<DelegationTokenIdentifier> tok2 = method.deserializeToken
-            (token.encodeToUrlString(), LOGICAL_NAME);
-    Assert.assertTrue(HAUtil.isTokenForLogicalUri(tok2));
-  }
-}

http://git-wip-us.apache.org/repos/asf/hadoop/blob/9dd5d673/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/web/webhdfs/TestParameterParser.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/web/webhdfs/TestParameterParser.java
b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/web/webhdfs/TestParameterParser.java
new file mode 100644
index 0000000..8b4235b
--- /dev/null
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/web/webhdfs/TestParameterParser.java
@@ -0,0 +1,55 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hdfs.server.datanode.web.webhdfs;
+
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.hdfs.DFSTestUtil;
+import org.apache.hadoop.hdfs.HAUtil;
+import org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenIdentifier;
+import org.apache.hadoop.hdfs.web.resources.DelegationParam;
+import org.apache.hadoop.hdfs.web.resources.NamenodeAddressParam;
+import org.apache.hadoop.security.token.Token;
+import org.junit.Assert;
+import org.junit.Test;
+
+import io.netty.handler.codec.http.QueryStringDecoder;
+
+import javax.servlet.ServletContext;
+
+import java.io.IOException;
+
+import static org.mockito.Mockito.doReturn;
+import static org.mockito.Mockito.mock;
+
+public class TestParameterParser {
+  private static final String LOGICAL_NAME = "minidfs";
+
+  @Test
+  public void testDeserializeHAToken() throws IOException {
+    Configuration conf = DFSTestUtil.newHAConfiguration(LOGICAL_NAME);
+    final Token<DelegationTokenIdentifier> token = new
+        Token<DelegationTokenIdentifier>();
+    QueryStringDecoder decoder = new QueryStringDecoder(
+      WebHdfsHandler.WEBHDFS_PREFIX + "/?"
+      + NamenodeAddressParam.NAME + "=" + LOGICAL_NAME + "&"
+      + DelegationParam.NAME + "=" + token.encodeToUrlString());
+    ParameterParser testParser = new ParameterParser(decoder, conf);
+    final Token<DelegationTokenIdentifier> tok2 = testParser.delegationToken();
+    Assert.assertTrue(HAUtil.isTokenForLogicalUri(tok2));
+  }
+}

http://git-wip-us.apache.org/repos/asf/hadoop/blob/9dd5d673/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestFSMainOperationsWebHdfs.java
----------------------------------------------------------------------
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestFSMainOperationsWebHdfs.java
b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestFSMainOperationsWebHdfs.java
index ad2352b..b4216f0 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestFSMainOperationsWebHdfs.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestFSMainOperationsWebHdfs.java
@@ -31,7 +31,6 @@ import org.apache.hadoop.fs.permission.FsPermission;
 import org.apache.hadoop.hdfs.DFSConfigKeys;
 import org.apache.hadoop.hdfs.DFSTestUtil;
 import org.apache.hadoop.hdfs.MiniDFSCluster;
-import org.apache.hadoop.hdfs.server.datanode.web.resources.DatanodeWebHdfsMethods;
 import org.apache.hadoop.hdfs.web.resources.ExceptionHandler;
 import org.apache.hadoop.security.AccessControlException;
 import org.apache.hadoop.security.UserGroupInformation;
@@ -44,7 +43,6 @@ import org.junit.Test;
 public class TestFSMainOperationsWebHdfs extends FSMainOperationsBaseTest {
   {
     ((Log4JLogger)ExceptionHandler.LOG).getLogger().setLevel(Level.ALL);
-    ((Log4JLogger)DatanodeWebHdfsMethods.LOG).getLogger().setLevel(Level.ALL);
   }
 
   private static MiniDFSCluster cluster = null;


Mime
View raw message