Return-Path: X-Original-To: apmail-hadoop-common-commits-archive@www.apache.org Delivered-To: apmail-hadoop-common-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id D12259FB7 for ; Mon, 28 Nov 2011 21:20:38 +0000 (UTC) Received: (qmail 2235 invoked by uid 500); 28 Nov 2011 21:20:38 -0000 Delivered-To: apmail-hadoop-common-commits-archive@hadoop.apache.org Received: (qmail 2205 invoked by uid 500); 28 Nov 2011 21:20:38 -0000 Mailing-List: contact common-commits-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: common-dev@hadoop.apache.org Delivered-To: mailing list common-commits@hadoop.apache.org Received: (qmail 2198 invoked by uid 99); 28 Nov 2011 21:20:38 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 28 Nov 2011 21:20:38 +0000 X-ASF-Spam-Status: No, hits=-2000.0 required=5.0 tests=ALL_TRUSTED X-Spam-Check-By: apache.org Received: from [140.211.11.4] (HELO eris.apache.org) (140.211.11.4) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 28 Nov 2011 21:20:37 +0000 Received: from eris.apache.org (localhost [127.0.0.1]) by eris.apache.org (Postfix) with ESMTP id 1F28623889B3 for ; Mon, 28 Nov 2011 21:20:17 +0000 (UTC) Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit Subject: svn commit: r1207617 - in /hadoop/common/branches/branch-1: CHANGES.txt src/core/org/apache/hadoop/security/UserGroupInformation.java src/test/org/apache/hadoop/hdfs/TestFileAppend2.java src/test/org/apache/hadoop/hdfs/TestFileConcurrentReader.java Date: Mon, 28 Nov 2011 21:20:16 -0000 To: common-commits@hadoop.apache.org From: jitendra@apache.org X-Mailer: svnmailer-1.0.8-patched Message-Id: <20111128212017.1F28623889B3@eris.apache.org> Author: jitendra Date: Mon Nov 28 21:20:14 2011 New Revision: 1207617 URL: http://svn.apache.org/viewvc?rev=1207617&view=rev Log: HADOOP-7865. Test Failures in 1.0.0 hdfs/common. Modified: hadoop/common/branches/branch-1/CHANGES.txt hadoop/common/branches/branch-1/src/core/org/apache/hadoop/security/UserGroupInformation.java hadoop/common/branches/branch-1/src/test/org/apache/hadoop/hdfs/TestFileAppend2.java hadoop/common/branches/branch-1/src/test/org/apache/hadoop/hdfs/TestFileConcurrentReader.java Modified: hadoop/common/branches/branch-1/CHANGES.txt URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-1/CHANGES.txt?rev=1207617&r1=1207616&r2=1207617&view=diff ============================================================================== --- hadoop/common/branches/branch-1/CHANGES.txt (original) +++ hadoop/common/branches/branch-1/CHANGES.txt Mon Nov 28 21:20:14 2011 @@ -236,6 +236,8 @@ Release 1.0.0 - unreleased HADOOP-7854. UGI getCurrentUser is not synchronized. (Daryn Sharp via jitendra) + HADOOP-7865. Test Failures in 1.0 hdfs/common. (jitendra) + Release 0.20.205.0 - 2011.10.06 NEW FEATURES Modified: hadoop/common/branches/branch-1/src/core/org/apache/hadoop/security/UserGroupInformation.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-1/src/core/org/apache/hadoop/security/UserGroupInformation.java?rev=1207617&r1=1207616&r2=1207617&view=diff ============================================================================== --- hadoop/common/branches/branch-1/src/core/org/apache/hadoop/security/UserGroupInformation.java (original) +++ hadoop/common/branches/branch-1/src/core/org/apache/hadoop/security/UserGroupInformation.java Mon Nov 28 21:20:14 2011 @@ -998,10 +998,11 @@ public class UserGroupInformation { */ @Override public String toString() { - String me = (getRealUser() != null) - ? getUserName() + " via " + getRealUser().toString() - : getUserName(); - return me + " (auth:"+getAuthenticationMethod()+")"; + if (getRealUser() != null) { + return getUserName() + " via " + getRealUser().toString(); + } else { + return getUserName(); + } } /** Modified: hadoop/common/branches/branch-1/src/test/org/apache/hadoop/hdfs/TestFileAppend2.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-1/src/test/org/apache/hadoop/hdfs/TestFileAppend2.java?rev=1207617&r1=1207616&r2=1207617&view=diff ============================================================================== --- hadoop/common/branches/branch-1/src/test/org/apache/hadoop/hdfs/TestFileAppend2.java (original) +++ hadoop/common/branches/branch-1/src/test/org/apache/hadoop/hdfs/TestFileAppend2.java Mon Nov 28 21:20:14 2011 @@ -29,32 +29,17 @@ import org.apache.hadoop.fs.FSDataOutput import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.permission.FsPermission; -import org.apache.hadoop.hdfs.server.datanode.DataNode; import org.apache.hadoop.hdfs.server.datanode.SimulatedFSDataset; -import org.apache.hadoop.hdfs.server.namenode.FSNamesystem; -import org.apache.hadoop.hdfs.server.namenode.LeaseManager; -import org.apache.hadoop.hdfs.server.namenode.NameNode; import org.apache.hadoop.io.IOUtils; import org.apache.hadoop.security.AccessControlException; import org.apache.hadoop.security.UserGroupInformation; -import org.apache.commons.logging.impl.Log4JLogger; -import org.apache.log4j.Level; - /** * This class tests the building blocks that are needed to * support HDFS appends. */ public class TestFileAppend2 extends TestCase { - { - ((Log4JLogger)NameNode.stateChangeLog).getLogger().setLevel(Level.ALL); - ((Log4JLogger)LeaseManager.LOG).getLogger().setLevel(Level.ALL); - ((Log4JLogger)FSNamesystem.LOG).getLogger().setLevel(Level.ALL); - ((Log4JLogger)DataNode.LOG).getLogger().setLevel(Level.ALL); - ((Log4JLogger)DFSClient.LOG).getLogger().setLevel(Level.ALL); - } - static final int blockSize = 1024; static final int numBlocks = 5; static final int fileSize = numBlocks * blockSize + 1; Modified: hadoop/common/branches/branch-1/src/test/org/apache/hadoop/hdfs/TestFileConcurrentReader.java URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-1/src/test/org/apache/hadoop/hdfs/TestFileConcurrentReader.java?rev=1207617&r1=1207616&r2=1207617&view=diff ============================================================================== --- hadoop/common/branches/branch-1/src/test/org/apache/hadoop/hdfs/TestFileConcurrentReader.java (original) +++ hadoop/common/branches/branch-1/src/test/org/apache/hadoop/hdfs/TestFileConcurrentReader.java Mon Nov 28 21:20:14 2011 @@ -17,7 +17,10 @@ */ package org.apache.hadoop.hdfs; -import org.apache.commons.logging.impl.Log4JLogger; +import java.io.IOException; +import java.util.Arrays; +import java.util.concurrent.atomic.AtomicBoolean; + import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.BlockLocation; import org.apache.hadoop.fs.ChecksumException; @@ -25,17 +28,10 @@ import org.apache.hadoop.fs.FSDataInputS import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; -import org.apache.hadoop.hdfs.server.namenode.FSNamesystem; -import org.apache.hadoop.hdfs.server.namenode.LeaseManager; import org.apache.hadoop.io.IOUtils; -import org.apache.log4j.Level; import org.apache.log4j.Logger; import org.junit.Before; -import java.io.IOException; -import java.util.*; -import java.util.concurrent.atomic.*; - /** * This class tests the cases of a concurrent reads/writes to a file; @@ -52,12 +48,6 @@ public class TestFileConcurrentReader ex private static final Logger LOG = Logger.getLogger(TestFileConcurrentReader.class); - - { - ((Log4JLogger) LeaseManager.LOG).getLogger().setLevel(Level.ALL); - ((Log4JLogger) FSNamesystem.LOG).getLogger().setLevel(Level.ALL); - ((Log4JLogger) DFSClient.LOG).getLogger().setLevel(Level.ALL); - } static final long seed = 0xDEADBEEFL; static final int blockSize = 8192; @@ -67,16 +57,6 @@ public class TestFileConcurrentReader ex private MiniDFSCluster cluster; private FileSystem fileSystem; - // creates a file but does not close it - private FSDataOutputStream createFile(FileSystem fileSys, Path name, int repl) - throws IOException { - System.out.println("createFile: Created " + name + " with " + repl + " replica."); - FSDataOutputStream stm = fileSys.create(name, true, - fileSys.getConf().getInt("io.file.buffer.size", 4096), - (short) repl, (long) blockSize); - return stm; - } - @Before protected void setUp() throws IOException { conf = new Configuration();