Return-Path: X-Original-To: apmail-hadoop-common-commits-archive@www.apache.org Delivered-To: apmail-hadoop-common-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 5BCCC194D5 for ; Sat, 16 Apr 2016 00:07:52 +0000 (UTC) Received: (qmail 11550 invoked by uid 500); 16 Apr 2016 00:07:43 -0000 Delivered-To: apmail-hadoop-common-commits-archive@hadoop.apache.org Received: (qmail 11265 invoked by uid 500); 16 Apr 2016 00:07:42 -0000 Mailing-List: contact common-commits-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: common-dev@hadoop.apache.org Delivered-To: mailing list common-commits@hadoop.apache.org Received: (qmail 10226 invoked by uid 99); 16 Apr 2016 00:07:42 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 16 Apr 2016 00:07:42 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id 6C147E027F; Sat, 16 Apr 2016 00:07:42 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: wangda@apache.org To: common-commits@hadoop.apache.org Date: Sat, 16 Apr 2016 00:08:09 -0000 Message-Id: <637cff6c9349453e9776a97dd2db45ed@git.apache.org> In-Reply-To: <76120e36f37945e6862773579cc3e424@git.apache.org> References: <76120e36f37945e6862773579cc3e424@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [29/50] [abbrv] hadoop git commit: HDFS-10281. TestPendingCorruptDnMessages fails intermittently. Contributed by Mingliang Liu. HDFS-10281. TestPendingCorruptDnMessages fails intermittently. Contributed by Mingliang Liu. Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/b9c9d035 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/b9c9d035 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/b9c9d035 Branch: refs/heads/YARN-3368 Commit: b9c9d03591a49be31f3fbc738d01a31700bfdbc4 Parents: 2c155af Author: Kihwal Lee Authored: Thu Apr 14 15:24:39 2016 -0500 Committer: Kihwal Lee Committed: Thu Apr 14 15:24:39 2016 -0500 ---------------------------------------------------------------------- .../ha/TestPendingCorruptDnMessages.java | 51 +++++++++++--------- 1 file changed, 28 insertions(+), 23 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hadoop/blob/b9c9d035/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestPendingCorruptDnMessages.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestPendingCorruptDnMessages.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestPendingCorruptDnMessages.java index 5f116d9..5063acd 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestPendingCorruptDnMessages.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/ha/TestPendingCorruptDnMessages.java @@ -18,12 +18,14 @@ package org.apache.hadoop.hdfs.server.namenode.ha; import static org.junit.Assert.assertEquals; +import static org.junit.Assert.assertNotNull; import static org.junit.Assert.assertTrue; import java.io.IOException; import java.io.OutputStream; import java.net.URISyntaxException; import java.util.List; +import java.util.concurrent.TimeoutException; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -37,19 +39,22 @@ import org.apache.hadoop.hdfs.protocol.ExtendedBlock; import org.apache.hadoop.hdfs.protocol.HdfsConstants.DatanodeReportType; import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeDescriptor; import org.apache.hadoop.hdfs.server.datanode.DataNodeTestUtils; -import org.apache.hadoop.util.ThreadUtil; +import org.apache.hadoop.test.GenericTestUtils; + +import com.google.common.base.Supplier; + import org.junit.Test; public class TestPendingCorruptDnMessages { private static final Path filePath = new Path("/foo.txt"); - @Test + @Test (timeout = 60000) public void testChangedStorageId() throws IOException, URISyntaxException, - InterruptedException { + InterruptedException, TimeoutException { HdfsConfiguration conf = new HdfsConfiguration(); conf.setInt(DFSConfigKeys.DFS_HA_TAILEDITS_PERIOD_KEY, 1); - MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf) + final MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf) .numDataNodes(1) .nnTopology(MiniDFSNNTopology.simpleHATopology()) .build(); @@ -83,27 +88,27 @@ public class TestPendingCorruptDnMessages { // Wait until the standby NN queues up the corrupt block in the pending DN // message queue. - while (cluster.getNamesystem(1).getBlockManager() - .getPendingDataNodeMessageCount() < 1) { - ThreadUtil.sleepAtLeastIgnoreInterrupts(1000); - } - - assertEquals(1, cluster.getNamesystem(1).getBlockManager() - .getPendingDataNodeMessageCount()); - String oldStorageId = getRegisteredDatanodeUid(cluster, 1); + GenericTestUtils.waitFor(new Supplier() { + @Override + public Boolean get() { + return cluster.getNamesystem(1).getBlockManager() + .getPendingDataNodeMessageCount() == 1; + } + }, 1000, 30000); + + final String oldStorageId = getRegisteredDatanodeUid(cluster, 1); + assertNotNull(oldStorageId); // Reformat/restart the DN. assertTrue(wipeAndRestartDn(cluster, 0)); - // Give the DN time to start up and register, which will cause the - // DatanodeManager to dissociate the old storage ID from the DN xfer addr. - String newStorageId = ""; - do { - ThreadUtil.sleepAtLeastIgnoreInterrupts(1000); - newStorageId = getRegisteredDatanodeUid(cluster, 1); - System.out.println("====> oldStorageId: " + oldStorageId + - " newStorageId: " + newStorageId); - } while (newStorageId.equals(oldStorageId)); + GenericTestUtils.waitFor(new Supplier() { + @Override + public Boolean get() { + final String newStorageId = getRegisteredDatanodeUid(cluster, 1); + return newStorageId != null && !newStorageId.equals(oldStorageId); + } + }, 1000, 30000); assertEquals(0, cluster.getNamesystem(1).getBlockManager() .getPendingDataNodeMessageCount()); @@ -121,8 +126,8 @@ public class TestPendingCorruptDnMessages { List registeredDatanodes = cluster.getNamesystem(nnIndex) .getBlockManager().getDatanodeManager() .getDatanodeListForReport(DatanodeReportType.ALL); - assertEquals(1, registeredDatanodes.size()); - return registeredDatanodes.get(0).getDatanodeUuid(); + return registeredDatanodes.isEmpty() ? null : + registeredDatanodes.get(0).getDatanodeUuid(); } private static boolean wipeAndRestartDn(MiniDFSCluster cluster, int dnIndex)