Return-Path: X-Original-To: apmail-hadoop-hdfs-commits-archive@minotaur.apache.org Delivered-To: apmail-hadoop-hdfs-commits-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 3CA9C10F42 for ; Tue, 12 Nov 2013 00:22:20 +0000 (UTC) Received: (qmail 37240 invoked by uid 500); 12 Nov 2013 00:22:20 -0000 Delivered-To: apmail-hadoop-hdfs-commits-archive@hadoop.apache.org Received: (qmail 37196 invoked by uid 500); 12 Nov 2013 00:22:20 -0000 Mailing-List: contact hdfs-commits-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: hdfs-dev@hadoop.apache.org Delivered-To: mailing list hdfs-commits@hadoop.apache.org Received: (qmail 37188 invoked by uid 99); 12 Nov 2013 00:22:20 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 12 Nov 2013 00:22:20 +0000 X-ASF-Spam-Status: No, hits=-2000.0 required=5.0 tests=ALL_TRUSTED X-Spam-Check-By: apache.org Received: from [140.211.11.4] (HELO eris.apache.org) (140.211.11.4) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 12 Nov 2013 00:22:18 +0000 Received: from eris.apache.org (localhost [127.0.0.1]) by eris.apache.org (Postfix) with ESMTP id 4EB36238883D; Tue, 12 Nov 2013 00:21:58 +0000 (UTC) Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit Subject: svn commit: r1540894 - in /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs: CHANGES.txt src/test/java/org/apache/hadoop/hdfs/web/TestHftpFileSystem.java src/test/java/org/apache/hadoop/hdfs/web/TestHftpURLTimeouts.java Date: Tue, 12 Nov 2013 00:21:58 -0000 To: hdfs-commits@hadoop.apache.org From: jing9@apache.org X-Mailer: svnmailer-1.0.9 Message-Id: <20131112002158.4EB36238883D@eris.apache.org> X-Virus-Checked: Checked by ClamAV on apache.org Author: jing9 Date: Tue Nov 12 00:21:57 2013 New Revision: 1540894 URL: http://svn.apache.org/r1540894 Log: HDFS-5488. Clean up TestHftpURLTimeout. Contributed by Haohui Mai. Removed: hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestHftpURLTimeouts.java Modified: hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestHftpFileSystem.java Modified: hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt URL: http://svn.apache.org/viewvc/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt?rev=1540894&r1=1540893&r2=1540894&view=diff ============================================================================== --- hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt (original) +++ hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt Tue Nov 12 00:21:57 2013 @@ -524,6 +524,8 @@ Release 2.3.0 - UNRELEASED HDFS-5325. Remove WebHdfsFileSystem#ConnRunner. (Haohui Mai via jing9) + HDFS-5488. Clean up TestHftpURLTimeout. (Haohui Mai via jing9) + Release 2.2.1 - UNRELEASED INCOMPATIBLE CHANGES Modified: hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestHftpFileSystem.java URL: http://svn.apache.org/viewvc/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestHftpFileSystem.java?rev=1540894&r1=1540893&r2=1540894&view=diff ============================================================================== --- hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestHftpFileSystem.java (original) +++ hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestHftpFileSystem.java Tue Nov 12 00:21:57 2013 @@ -28,6 +28,7 @@ import java.net.HttpURLConnection; import java.net.URI; import java.net.URISyntaxException; import java.net.URL; +import java.net.URLConnection; import java.util.Random; import org.apache.commons.logging.impl.Log4JLogger; @@ -40,15 +41,16 @@ import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hdfs.DFSConfigKeys; import org.apache.hadoop.hdfs.MiniDFSCluster; -import org.apache.hadoop.hdfs.MiniDFSCluster.Builder; import org.apache.hadoop.hdfs.server.datanode.DataNode; import org.apache.hadoop.hdfs.server.datanode.DataNodeTestUtils; import org.apache.hadoop.hdfs.server.protocol.DatanodeRegistration; -import org.apache.hadoop.hdfs.web.HftpFileSystem; -import org.apache.hadoop.hdfs.web.HsftpFileSystem; import org.apache.hadoop.util.ServletUtil; import org.apache.log4j.Level; -import org.junit.*; +import org.junit.After; +import org.junit.AfterClass; +import org.junit.Before; +import org.junit.BeforeClass; +import org.junit.Test; public class TestHftpFileSystem { private static final Random RAN = new Random(); @@ -65,32 +67,24 @@ public class TestHftpFileSystem { new Path("/foo;bar"), // URI does not encode, Request#getPathInfo returns verbatim - new Path("/foo+"), - new Path("/foo+bar/foo+bar"), - new Path("/foo=bar/foo=bar"), - new Path("/foo,bar/foo,bar"), - new Path("/foo@bar/foo@bar"), - new Path("/foo&bar/foo&bar"), - new Path("/foo$bar/foo$bar"), - new Path("/foo_bar/foo_bar"), - new Path("/foo~bar/foo~bar"), - new Path("/foo.bar/foo.bar"), - new Path("/foo../bar/foo../bar"), - new Path("/foo.../bar/foo.../bar"), + new Path("/foo+"), new Path("/foo+bar/foo+bar"), + new Path("/foo=bar/foo=bar"), new Path("/foo,bar/foo,bar"), + new Path("/foo@bar/foo@bar"), new Path("/foo&bar/foo&bar"), + new Path("/foo$bar/foo$bar"), new Path("/foo_bar/foo_bar"), + new Path("/foo~bar/foo~bar"), new Path("/foo.bar/foo.bar"), + new Path("/foo../bar/foo../bar"), new Path("/foo.../bar/foo.../bar"), new Path("/foo'bar/foo'bar"), new Path("/foo#bar/foo#bar"), new Path("/foo!bar/foo!bar"), // HDFS file names may not contain ":" // URI percent encodes, Request#getPathInfo decodes - new Path("/foo bar/foo bar"), - new Path("/foo?bar/foo?bar"), - new Path("/foo\">bar/foo\">bar"), - }; + new Path("/foo bar/foo bar"), new Path("/foo?bar/foo?bar"), + new Path("/foo\">bar/foo\">bar"), }; @BeforeClass public static void setUp() throws IOException { - ((Log4JLogger)HftpFileSystem.LOG).getLogger().setLevel(Level.ALL); + ((Log4JLogger) HftpFileSystem.LOG).getLogger().setLevel(Level.ALL); final long seed = RAN.nextLong(); System.out.println("seed=" + seed); @@ -99,8 +93,8 @@ public class TestHftpFileSystem { config = new Configuration(); cluster = new MiniDFSCluster.Builder(config).numDataNodes(2).build(); blockPoolId = cluster.getNamesystem().getBlockPoolId(); - hftpUri = - "hftp://" + config.get(DFSConfigKeys.DFS_NAMENODE_HTTP_ADDRESS_KEY); + hftpUri = "hftp://" + + config.get(DFSConfigKeys.DFS_NAMENODE_HTTP_ADDRESS_KEY); } @AfterClass @@ -140,7 +134,8 @@ public class TestHftpFileSystem { // Check the file status matches the path. Hftp returns a FileStatus // with the entire URI, extract the path part. - assertEquals(p, new Path(hftpFs.getFileStatus(p).getPath().toUri().getPath())); + assertEquals(p, new Path(hftpFs.getFileStatus(p).getPath().toUri() + .getPath())); // Test list status (listPath servlet) assertEquals(1, hftpFs.listStatus(p).length); @@ -158,21 +153,20 @@ public class TestHftpFileSystem { if (hdfs.exists(path)) { hdfs.delete(path, true); } - FSDataOutputStream out = hdfs.create(path, (short)1); + FSDataOutputStream out = hdfs.create(path, (short) 1); out.writeBytes("0123456789"); out.close(); // Get the path's block location so we can determine // if we were redirected to the right DN. - BlockLocation[] locations = - hdfs.getFileBlockLocations(path, 0, 10); + BlockLocation[] locations = hdfs.getFileBlockLocations(path, 0, 10); String xferAddr = locations[0].getNames()[0]; // Connect to the NN to get redirected URL u = hftpFs.getNamenodeURL( "/data" + ServletUtil.encodePath(path.toUri().getPath()), "ugi=userx,groupy"); - HttpURLConnection conn = (HttpURLConnection)u.openConnection(); + HttpURLConnection conn = (HttpURLConnection) u.openConnection(); HttpURLConnection.setFollowRedirects(true); conn.connect(); conn.getInputStream(); @@ -181,15 +175,15 @@ public class TestHftpFileSystem { // Find the datanode that has the block according to locations // and check that the URL was redirected to this DN's info port for (DataNode node : cluster.getDataNodes()) { - DatanodeRegistration dnR = - DataNodeTestUtils.getDNRegistrationForBP(node, blockPoolId); + DatanodeRegistration dnR = DataNodeTestUtils.getDNRegistrationForBP(node, + blockPoolId); if (dnR.getXferAddr().equals(xferAddr)) { checked = true; assertEquals(dnR.getInfoPort(), conn.getURL().getPort()); } } - assertTrue("The test never checked that location of " + - "the block and hftp desitnation are the same", checked); + assertTrue("The test never checked that location of " + + "the block and hftp desitnation are the same", checked); } /** @@ -260,7 +254,7 @@ public class TestHftpFileSystem { os.writeBytes("0123456789"); os.close(); - // ByteRangeInputStream delays opens until reads. Make sure it doesn't + // ByteRangeInputStream delays opens until reads. Make sure it doesn't // open a closed stream that has never been opened FSDataInputStream in = hftpFs.open(testFile); in.close(); @@ -298,16 +292,15 @@ public class TestHftpFileSystem { URI uri = URI.create("hftp://localhost"); HftpFileSystem fs = (HftpFileSystem) FileSystem.get(uri, conf); - assertEquals(DFSConfigKeys.DFS_NAMENODE_HTTP_PORT_DEFAULT, fs.getDefaultPort()); + assertEquals(DFSConfigKeys.DFS_NAMENODE_HTTP_PORT_DEFAULT, + fs.getDefaultPort()); assertEquals(uri, fs.getUri()); // HFTP uses http to get the token so canonical service name should // return the http port. - assertEquals( - "127.0.0.1:" + DFSConfigKeys.DFS_NAMENODE_HTTP_PORT_DEFAULT, - fs.getCanonicalServiceName() - ); + assertEquals("127.0.0.1:" + DFSConfigKeys.DFS_NAMENODE_HTTP_PORT_DEFAULT, + fs.getCanonicalServiceName()); } @Test @@ -324,10 +317,7 @@ public class TestHftpFileSystem { // HFTP uses http to get the token so canonical service name should // return the http port. - assertEquals( - "127.0.0.1:123", - fs.getCanonicalServiceName() - ); + assertEquals("127.0.0.1:123", fs.getCanonicalServiceName()); } @Test @@ -336,13 +326,11 @@ public class TestHftpFileSystem { URI uri = URI.create("hftp://localhost:123"); HftpFileSystem fs = (HftpFileSystem) FileSystem.get(uri, conf); - assertEquals(DFSConfigKeys.DFS_NAMENODE_HTTP_PORT_DEFAULT, fs.getDefaultPort()); + assertEquals(DFSConfigKeys.DFS_NAMENODE_HTTP_PORT_DEFAULT, + fs.getDefaultPort()); assertEquals(uri, fs.getUri()); - assertEquals( - "127.0.0.1:123", - fs.getCanonicalServiceName() - ); + assertEquals("127.0.0.1:123", fs.getCanonicalServiceName()); } @Test @@ -356,13 +344,20 @@ public class TestHftpFileSystem { assertEquals(123, fs.getDefaultPort()); assertEquals(uri, fs.getUri()); - assertEquals( - "127.0.0.1:789", - fs.getCanonicalServiceName() - ); + assertEquals("127.0.0.1:789", fs.getCanonicalServiceName()); } - /// + @Test + public void testTimeout() throws IOException { + Configuration conf = new Configuration(); + URI uri = URI.create("hftp://localhost"); + HftpFileSystem fs = (HftpFileSystem) FileSystem.get(uri, conf); + URLConnection conn = fs.connectionFactory.openConnection(new URL("http://localhost")); + assertEquals(URLConnectionFactory.DEFAULT_SOCKET_TIMEOUT, conn.getConnectTimeout()); + assertEquals(URLConnectionFactory.DEFAULT_SOCKET_TIMEOUT, conn.getReadTimeout()); + } + + // / @Test public void testHsftpDefaultPorts() throws IOException { @@ -370,13 +365,12 @@ public class TestHftpFileSystem { URI uri = URI.create("hsftp://localhost"); HsftpFileSystem fs = (HsftpFileSystem) FileSystem.get(uri, conf); - assertEquals(DFSConfigKeys.DFS_NAMENODE_HTTPS_PORT_DEFAULT, fs.getDefaultPort()); + assertEquals(DFSConfigKeys.DFS_NAMENODE_HTTPS_PORT_DEFAULT, + fs.getDefaultPort()); assertEquals(uri, fs.getUri()); - assertEquals( - "127.0.0.1:"+DFSConfigKeys.DFS_NAMENODE_HTTPS_PORT_DEFAULT, - fs.getCanonicalServiceName() - ); + assertEquals("127.0.0.1:" + DFSConfigKeys.DFS_NAMENODE_HTTPS_PORT_DEFAULT, + fs.getCanonicalServiceName()); } @Test @@ -391,10 +385,7 @@ public class TestHftpFileSystem { assertEquals(456, fs.getDefaultPort()); assertEquals(uri, fs.getUri()); - assertEquals( - "127.0.0.1:456", - fs.getCanonicalServiceName() - ); + assertEquals("127.0.0.1:456", fs.getCanonicalServiceName()); } @Test @@ -403,13 +394,11 @@ public class TestHftpFileSystem { URI uri = URI.create("hsftp://localhost:123"); HsftpFileSystem fs = (HsftpFileSystem) FileSystem.get(uri, conf); - assertEquals(DFSConfigKeys.DFS_NAMENODE_HTTPS_PORT_DEFAULT, fs.getDefaultPort()); + assertEquals(DFSConfigKeys.DFS_NAMENODE_HTTPS_PORT_DEFAULT, + fs.getDefaultPort()); assertEquals(uri, fs.getUri()); - assertEquals( - "127.0.0.1:123", - fs.getCanonicalServiceName() - ); + assertEquals("127.0.0.1:123", fs.getCanonicalServiceName()); } @Test @@ -424,9 +413,6 @@ public class TestHftpFileSystem { assertEquals(456, fs.getDefaultPort()); assertEquals(uri, fs.getUri()); - assertEquals( - "127.0.0.1:789", - fs.getCanonicalServiceName() - ); + assertEquals("127.0.0.1:789", fs.getCanonicalServiceName()); } }