Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 00BA6200B36 for ; Wed, 22 Jun 2016 02:26:17 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id F36C6160A69; Wed, 22 Jun 2016 00:26:16 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 1F2C6160A60 for ; Wed, 22 Jun 2016 02:26:15 +0200 (CEST) Received: (qmail 65233 invoked by uid 500); 22 Jun 2016 00:26:01 -0000 Mailing-List: contact common-commits-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list common-commits@hadoop.apache.org Received: (qmail 62577 invoked by uid 99); 22 Jun 2016 00:25:59 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 22 Jun 2016 00:25:59 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id D4E1DE07FE; Wed, 22 Jun 2016 00:25:59 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: subru@apache.org To: common-commits@hadoop.apache.org Date: Wed, 22 Jun 2016 00:26:45 -0000 Message-Id: In-Reply-To: <85a9aecfdfad441186f9f7f2137dd235@git.apache.org> References: <85a9aecfdfad441186f9f7f2137dd235@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [47/51] [abbrv] hadoop git commit: HDFS-9530. ReservedSpace is not cleared for abandoned Blocks (Contributed by Brahma Reddy Battula) archived-at: Wed, 22 Jun 2016 00:26:17 -0000 HDFS-9530. ReservedSpace is not cleared for abandoned Blocks (Contributed by Brahma Reddy Battula) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/f2ac132d Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/f2ac132d Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/f2ac132d Branch: refs/heads/YARN-2915 Commit: f2ac132d6a21c215093b7f87acf2843ac8123716 Parents: 46f1602 Author: Brahma Reddy Battula Authored: Tue Jun 21 15:42:28 2016 +0530 Committer: Brahma Reddy Battula Committed: Tue Jun 21 15:42:28 2016 +0530 ---------------------------------------------------------------------- .../server/datanode/DataNodeFaultInjector.java | 2 + .../hdfs/server/datanode/DataXceiver.java | 3 ++ .../datanode/fsdataset/impl/FsDatasetImpl.java | 4 ++ .../fsdataset/impl/TestSpaceReservation.java | 48 ++++++++++++++++++++ 4 files changed, 57 insertions(+) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hadoop/blob/f2ac132d/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNodeFaultInjector.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNodeFaultInjector.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNodeFaultInjector.java index 7327420..4ecbdc0 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNodeFaultInjector.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNodeFaultInjector.java @@ -53,4 +53,6 @@ public class DataNodeFaultInjector { public void stopSendingPacketDownstream() throws IOException {} public void noRegistration() throws IOException { } + + public void failMirrorConnection() throws IOException { } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/f2ac132d/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java index d5dc328..829badd 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java @@ -739,6 +739,9 @@ class DataXceiver extends Receiver implements Runnable { mirrorTarget = NetUtils.createSocketAddr(mirrorNode); mirrorSock = datanode.newSocket(); try { + + DataNodeFaultInjector.get().failMirrorConnection(); + int timeoutValue = dnConf.socketTimeout + (HdfsConstants.READ_TIMEOUT_EXTENSION * targets.length); int writeTimeout = dnConf.socketWriteTimeout + http://git-wip-us.apache.org/repos/asf/hadoop/blob/f2ac132d/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java index 243a0e2..b042297 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java @@ -79,6 +79,7 @@ import org.apache.hadoop.hdfs.server.datanode.ReplicaAlreadyExistsException; import org.apache.hadoop.hdfs.server.datanode.ReplicaBeingWritten; import org.apache.hadoop.hdfs.server.datanode.ReplicaHandler; import org.apache.hadoop.hdfs.server.datanode.ReplicaInPipeline; +import org.apache.hadoop.hdfs.server.datanode.ReplicaInPipelineInterface; import org.apache.hadoop.hdfs.server.datanode.ReplicaInfo; import org.apache.hadoop.hdfs.server.datanode.ReplicaNotFoundException; import org.apache.hadoop.hdfs.server.datanode.ReplicaUnderRecovery; @@ -1955,6 +1956,9 @@ class FsDatasetImpl implements FsDatasetSpi { LOG.debug("Block file " + removing.getBlockFile().getName() + " is to be deleted"); } + if (removing instanceof ReplicaInPipelineInterface) { + ((ReplicaInPipelineInterface) removing).releaseAllBytesReserved(); + } } if (v.isTransientStorage()) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/f2ac132d/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestSpaceReservation.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestSpaceReservation.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestSpaceReservation.java index 6dbd299..fad5216 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestSpaceReservation.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/TestSpaceReservation.java @@ -36,6 +36,7 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.hdfs.*; import org.apache.hadoop.hdfs.protocol.Block; import org.apache.hadoop.hdfs.server.datanode.DataNode; +import org.apache.hadoop.hdfs.server.datanode.DataNodeFaultInjector; import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi; import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeReference; import org.apache.hadoop.ipc.RemoteException; @@ -77,6 +78,7 @@ public class TestSpaceReservation { private DFSClient client = null; FsVolumeReference singletonVolumeRef = null; FsVolumeImpl singletonVolume = null; + private DataNodeFaultInjector old = null; private static Random rand = new Random(); @@ -146,6 +148,9 @@ public class TestSpaceReservation { cluster.shutdown(); cluster = null; } + if (old != null) { + DataNodeFaultInjector.set(old); + } } private void createFileAndTestSpaceReservation( @@ -613,6 +618,49 @@ public class TestSpaceReservation { checkReservedSpace(expectedFile2Reserved); } + @Test(timeout = 30000) + public void testReservedSpaceForPipelineRecovery() throws Exception { + final short replication = 3; + startCluster(BLOCK_SIZE, replication, -1); + + final String methodName = GenericTestUtils.getMethodName(); + final Path file = new Path("/" + methodName + ".01.dat"); + + old = DataNodeFaultInjector.get(); + // Fault injector to fail connection to mirror first time. + DataNodeFaultInjector.set(new DataNodeFaultInjector() { + private int tries = 0; + + @Override + public void failMirrorConnection() throws IOException { + if (tries++ == 0) { + throw new IOException("Failing Mirror for space reservation"); + } + } + }); + // Write 1 byte to the file and kill the writer. + FSDataOutputStream os = fs.create(file, replication); + os.write(new byte[1]); + os.close(); + // Ensure all space reserved for the replica was released on each + // DataNode. + cluster.triggerBlockReports(); + for (final DataNode dn : cluster.getDataNodes()) { + try (FsDatasetSpi.FsVolumeReferences volumes = + dn.getFSDataset().getFsVolumeReferences()) { + final FsVolumeImpl volume = (FsVolumeImpl) volumes.get(0); + GenericTestUtils.waitFor(new Supplier() { + @Override + public Boolean get() { + LOG.info("dn " + dn.getDisplayName() + " space : " + + volume.getReservedForReplicas()); + return (volume.getReservedForReplicas() == 0); + } + }, 100, Integer.MAX_VALUE); // Wait until the test times out. + } + } + } + private void checkReservedSpace(final long expectedReserved) throws TimeoutException, InterruptedException, IOException { for (final DataNode dn : cluster.getDataNodes()) { --------------------------------------------------------------------- To unsubscribe, e-mail: common-commits-unsubscribe@hadoop.apache.org For additional commands, e-mail: common-commits-help@hadoop.apache.org