Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 38F7F200B66 for ; Thu, 18 Aug 2016 10:12:23 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id 3775A160AAE; Thu, 18 Aug 2016 08:12:23 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 5E5D2160AAB for ; Thu, 18 Aug 2016 10:12:22 +0200 (CEST) Received: (qmail 23156 invoked by uid 500); 18 Aug 2016 08:12:21 -0000 Mailing-List: contact hdfs-issues-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list hdfs-issues@hadoop.apache.org Received: (qmail 23044 invoked by uid 99); 18 Aug 2016 08:12:21 -0000 Received: from arcas.apache.org (HELO arcas) (140.211.11.28) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 18 Aug 2016 08:12:21 +0000 Received: from arcas.apache.org (localhost [127.0.0.1]) by arcas (Postfix) with ESMTP id BE49F2C02B3 for ; Thu, 18 Aug 2016 08:12:20 +0000 (UTC) Date: Thu, 18 Aug 2016 08:12:20 +0000 (UTC) From: "Hadoop QA (JIRA)" To: hdfs-issues@hadoop.apache.org Message-ID: In-Reply-To: References: Subject: [jira] [Commented] (HDFS-10605) Can not synchronized call method of object and Mockito.spy(object), So UT:testRemoveVolumeBeingWritten passed but maybe deadlock online MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 archived-at: Thu, 18 Aug 2016 08:12:23 -0000 [ https://issues.apache.org/jira/browse/HDFS-10605?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15426073#comment-15426073 ] Hadoop QA commented on HDFS-10605: ---------------------------------- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 21s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 33s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 46s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 25s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 53s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 12s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 43s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 58s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 52s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 48s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 48s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 23s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 53s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 10s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 51s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 54s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 72m 53s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 18s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 93m 12s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.namenode.TestCacheDirectives | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12824291/TestDataNodeHotSwapVolumes.java.patch | | JIRA Issue | HDFS-10605 | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 5bf5b6a18db7 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 20f0eb8 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | unit | https://builds.apache.org/job/PreCommit-HDFS-Build/16468/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt | | Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/16468/testReport/ | | modules | C: hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project/hadoop-hdfs | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/16468/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Can not synchronized call method of object and Mockito.spy(object), So UT:testRemoveVolumeBeingWritten passed but maybe deadlock online > --------------------------------------------------------------------------------------------------------------------------------------- > > Key: HDFS-10605 > URL: https://issues.apache.org/jira/browse/HDFS-10605 > Project: Hadoop HDFS > Issue Type: Bug > Components: datanode > Affects Versions: 2.7.0, 2.8.0, 2.7.1, 2.7.2 > Reporter: ade > Labels: test > Attachments: TestDataNodeHotSwapVolumes.java.patch > > > The UT: TestDataNodeHotSwapVolumes.testRemoveVolumeBeingWritten can be ran successful, but deadlock like HDFS-9874 maybe happen online. > * UT: > {code:title=TestDataNodeHotSwapVolumes.java|borderStyle=solid} > final FsDatasetSpi data = dn.data; > dn.data = Mockito.spy(data); > LOG.info("data hash:" + data.hashCode() + "; dn.data hash:" + dn.data.hashCode()); > doAnswer(new Answer() { > public Object answer(InvocationOnMock invocation) > throws IOException, InterruptedException { > Thread.sleep(1000); > // Bypass the argument to FsDatasetImpl#finalizeBlock to verify that > // the block is not removed, since the volume reference should not > // be released at this point. > data.finalizeBlock((ExtendedBlock) invocation.getArguments()[0]); > return null; > } > }).when(dn.data).finalizeBlock(any(ExtendedBlock.class)); > {code} > Two thread can run synchronized method dn.data.removeVolumes and data.finalizeBlock concurrently because dn.data(mocked) and data is not the same object(hash 1903955157 and 1508483764). > {noformat} > 2016-07-11 16:16:07,788 INFO [Thread-0] datanode.TestDataNodeHotSwapVolumes (TestDataNodeHotSwapVolumes.java:testRemoveVolumeBeingWrittenForDatanode(599)) - data hash:1903955157; dn.data hash:1508483764 > 2016-07-11 16:16:07,801 INFO [Thread-157] datanode.DataNode (DataNode.java:reconfigurePropertyImpl(456)) - Reconfiguring dfs.datanode.data.dir to [DISK]file:/Users/ade/workspace/Hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/data/data2 > 2016-07-11 16:16:07,810 WARN [Thread-157] common.Util (Util.java:stringAsURI(56)) - Path /Users/ade/workspace/Hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/data/data1 should be specified as a URI in configuration files. Please update hdfs configuration. > 2016-07-11 16:16:07,811 INFO [Thread-157] datanode.DataNode (DataNode.java:removeVolumes(674)) - Deactivating volumes (clear failure=true): /Users/ade/workspace/Hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/data/data1 > 2016-07-11 16:16:07,836 INFO [Thread-157] impl.FsDatasetImpl (FsDatasetImpl.java:removeVolumes(459)) - Removing /Users/ade/workspace/Hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/data/data1 from FsDataset. > 2016-07-11 16:16:07,836 INFO [Thread-157] impl.FsDatasetImpl (FsDatasetImpl.java:removeVolumes(463)) - removeVolumes of object hash:1508483764 > 2016-07-11 16:16:07,836 INFO [Thread-157] datanode.BlockScanner (BlockScanner.java:removeVolumeScanner(243)) - Removing scanner for volume /Users/ade/workspace/Hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/data/data1 (StorageID DS-f4df3404-9f02-470e-b202-75f5a4de29cb) > 2016-07-11 16:16:07,836 INFO [VolumeScannerThread(/Users/ade/workspace/Hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/data/data1)] datanode.VolumeScanner (VolumeScanner.java:run(630)) - VolumeScanner(/Users/ade/workspace/Hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/data/data1, DS-f4df3404-9f02-470e-b202-75f5a4de29cb) exiting. > 2016-07-11 16:16:07,891 INFO [IPC Server handler 7 on 63546] blockmanagement.DatanodeDescriptor (DatanodeDescriptor.java:pruneStorageMap(517)) - Removed storage [DISK]DS-f4df3404-9f02-470e-b202-75f5a4de29cb:NORMAL:127.0.0.1:63548 from DataNode127.0.0.1:63548 > 2016-07-11 16:16:07,908 INFO [IPC Server handler 9 on 63546] blockmanagement.DatanodeDescriptor (DatanodeDescriptor.java:updateStorage(866)) - Adding new storage ID DS-f4df3404-9f02-470e-b202-75f5a4de29cb for DN 127.0.0.1:63548 > 2016-07-11 16:16:08,845 INFO [PacketResponder: BP-1077872064-127.0.0.1-1468224964600:blk_1073741825_1001, type=LAST_IN_PIPELINE, downstreams=0:[]] impl.FsDatasetImpl (FsDatasetImpl.java:finalizeBlock(1559)) - finalizeBlock of object hash:1903955157 > 2016-07-11 16:16:12,933 INFO [DataXceiver for client at /127.0.0.1:63574 [Receiving block BP-1077872064-127.0.0.1-1468224964600:blk_1073741825_1001]] impl.FsDatasetImpl (FsDatasetImpl.java:finalizeBlock(1559)) - finalizeBlock of object hash:1903955157 > {noformat} > The UT ran passed. > * Online > When dn.data.removeVolumes the thread run in FsVolumeImpl.closeAndWait() with dn.data lock and wait referenceCount() = 0, but the other DataXceiver thread maybe blocked by dn.data lock and with referencing volume. This can be happened like HDFS-9874. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-issues-unsubscribe@hadoop.apache.org For additional commands, e-mail: hdfs-issues-help@hadoop.apache.org