Return-Path: X-Original-To: apmail-hadoop-hdfs-commits-archive@minotaur.apache.org Delivered-To: apmail-hadoop-hdfs-commits-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 0FCDCD36C for ; Sat, 30 Jun 2012 05:06:08 +0000 (UTC) Received: (qmail 26816 invoked by uid 500); 30 Jun 2012 05:06:07 -0000 Delivered-To: apmail-hadoop-hdfs-commits-archive@hadoop.apache.org Received: (qmail 26771 invoked by uid 500); 30 Jun 2012 05:06:07 -0000 Mailing-List: contact hdfs-commits-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: hdfs-dev@hadoop.apache.org Delivered-To: mailing list hdfs-commits@hadoop.apache.org Received: (qmail 26744 invoked by uid 99); 30 Jun 2012 05:06:07 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 30 Jun 2012 05:06:07 +0000 X-ASF-Spam-Status: No, hits=-2000.0 required=5.0 tests=ALL_TRUSTED X-Spam-Check-By: apache.org Received: from [140.211.11.4] (HELO eris.apache.org) (140.211.11.4) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 30 Jun 2012 05:06:05 +0000 Received: from eris.apache.org (localhost [127.0.0.1]) by eris.apache.org (Postfix) with ESMTP id 5E35B23889BB; Sat, 30 Jun 2012 05:05:45 +0000 (UTC) Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit Subject: svn commit: r1355637 - /hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSShell.java Date: Sat, 30 Jun 2012 05:05:45 -0000 To: hdfs-commits@hadoop.apache.org From: harsh@apache.org X-Mailer: svnmailer-1.0.8-patched Message-Id: <20120630050545.5E35B23889BB@eris.apache.org> X-Virus-Checked: Checked by ClamAV on apache.org Author: harsh Date: Sat Jun 30 05:05:41 2012 New Revision: 1355637 URL: http://svn.apache.org/viewvc?rev=1355637&view=rev Log: HADOOP-8449. hadoop fs -text fails with compressed sequence files with the codec file extension (backported from trunk) (harsh) Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSShell.java Modified: hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSShell.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSShell.java?rev=1355637&r1=1355636&r2=1355637&view=diff ============================================================================== --- hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSShell.java (original) +++ hadoop/common/branches/branch-2/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSShell.java Sat Jun 30 05:05:41 2012 @@ -48,6 +48,8 @@ import org.apache.hadoop.hdfs.server.dat import org.apache.hadoop.hdfs.server.datanode.DataNodeTestUtils; import org.apache.hadoop.hdfs.tools.DFSAdmin; import org.apache.hadoop.io.IOUtils; +import org.apache.hadoop.io.SequenceFile; +import org.apache.hadoop.io.Text; import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.util.StringUtils; import org.apache.hadoop.util.ToolRunner; @@ -545,7 +547,7 @@ public class TestDFSShell extends TestCa textTest(new Path("/texttest").makeQualified(dfs.getUri(), dfs.getWorkingDirectory()), conf); - conf.set("fs.default.name", dfs.getUri().toString()); + conf.set("fs.defaultFS", dfs.getUri().toString()); final FileSystem lfs = FileSystem.getLocal(conf); textTest(new Path(TEST_ROOT_DIR, "texttest").makeQualified(lfs.getUri(), lfs.getWorkingDirectory()), conf); @@ -564,6 +566,7 @@ public class TestDFSShell extends TestCa OutputStream zout = new GZIPOutputStream( fs.create(new Path(root, "file.gz"))); Random r = new Random(); + bak = System.out; ByteArrayOutputStream file = new ByteArrayOutputStream(); for (int i = 0; i < 1024; ++i) { char c = Character.forDigit(r.nextInt(26) + 10, 36); @@ -572,7 +575,6 @@ public class TestDFSShell extends TestCa } zout.close(); - bak = System.out; ByteArrayOutputStream out = new ByteArrayOutputStream(); System.setOut(new PrintStream(out)); @@ -581,10 +583,28 @@ public class TestDFSShell extends TestCa argv[1] = new Path(root, "file.gz").toString(); int ret = ToolRunner.run(new FsShell(conf), argv); assertEquals("'-text " + argv[1] + " returned " + ret, 0, ret); - file.reset(); - out.reset(); assertTrue("Output doesn't match input", Arrays.equals(file.toByteArray(), out.toByteArray())); + + // Create a sequence file with a gz extension, to test proper + // container detection + SequenceFile.Writer writer = SequenceFile.createWriter( + conf, + SequenceFile.Writer.file(new Path(root, "file.gz")), + SequenceFile.Writer.keyClass(Text.class), + SequenceFile.Writer.valueClass(Text.class)); + writer.append(new Text("Foo"), new Text("Bar")); + writer.close(); + out = new ByteArrayOutputStream(); + System.setOut(new PrintStream(out)); + argv = new String[2]; + argv[0] = "-text"; + argv[1] = new Path(root, "file.gz").toString(); + ret = ToolRunner.run(new FsShell(conf), argv); + assertEquals("'-text " + argv[1] + " returned " + ret, 0, ret); + assertTrue("Output doesn't match input", + Arrays.equals("Foo\tBar\n".getBytes(), out.toByteArray())); + out.reset(); } finally { if (null != bak) { System.setOut(bak);