Return-Path: X-Original-To: apmail-hbase-issues-archive@www.apache.org Delivered-To: apmail-hbase-issues-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id BBA1017E99 for ; Thu, 16 Oct 2014 22:44:34 +0000 (UTC) Received: (qmail 96905 invoked by uid 500); 16 Oct 2014 22:44:34 -0000 Delivered-To: apmail-hbase-issues-archive@hbase.apache.org Received: (qmail 96860 invoked by uid 500); 16 Oct 2014 22:44:34 -0000 Mailing-List: contact issues-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list issues@hbase.apache.org Received: (qmail 96848 invoked by uid 99); 16 Oct 2014 22:44:34 -0000 Received: from arcas.apache.org (HELO arcas.apache.org) (140.211.11.28) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 16 Oct 2014 22:44:34 +0000 Date: Thu, 16 Oct 2014 22:44:34 +0000 (UTC) From: "Hudson (JIRA)" To: issues@hbase.apache.org Message-ID: In-Reply-To: References: Subject: [jira] [Commented] (HBASE-12241) The crash of regionServer when taking deadserver's replication queue breaks replication MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 [ https://issues.apache.org/jira/browse/HBASE-12241?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14174389#comment-14174389 ] Hudson commented on HBASE-12241: -------------------------------- FAILURE: Integrated in HBase-1.0 #323 (See [https://builds.apache.org/job/HBase-1.0/323/]) HBASE-12241 The crash of regionServer when taking deadserver's replication queue breaks replication (Shaohui) (tedyu: rev 5b3f6fb1a70133918f6b982b538b4a910aeb5633) * hbase-common/src/main/resources/hbase-default.xml > The crash of regionServer when taking deadserver's replication queue breaks replication > --------------------------------------------------------------------------------------- > > Key: HBASE-12241 > URL: https://issues.apache.org/jira/browse/HBASE-12241 > Project: HBase > Issue Type: Bug > Components: Replication > Reporter: Liu Shaohui > Assignee: Liu Shaohui > Priority: Critical > Fix For: 2.0.0, 0.99.2 > > Attachments: HBASE-12241-trunk-v1.diff > > > When a regionserver crash, another regionserver will try to take over the replication hlogs queue and help the the the dead regionserver to finish the replcation.See NodeFailoverWorker in ReplicationSourceManager > Currently hbase.zookeeper.useMulti is false in default configuration. The operation of taking over replication queue is not atomic. The ReplicationSourceManager firstly lock the replication node of dead regionserver and then copy the replication queue, and delete replication node of dead regionserver at last. The operation of the lockOtherRS just creates a persistent zk node named "lock" which prevent other regionserver taking over the replication queue. > See: > {code} > public boolean lockOtherRS(String znode) { > try { > String parent = ZKUtil.joinZNode(this.rsZNode, znode); > if (parent.equals(rsServerNameZnode)) { > LOG.warn("Won't lock because this is us, we're dead!"); > return false; > } > String p = ZKUtil.joinZNode(parent, RS_LOCK_ZNODE); > ZKUtil.createAndWatch(this.zookeeper, p, Bytes.toBytes(rsServerNameZnode)); > } catch (KeeperException e) { > ... > return false; > } > return true; > } > {code} > But if a regionserver crashed after creating this "lock" zk node and before coping the replication queue to its replication queue, the "lock" zk node will be left forever and > no other regionserver can take over the replication queue. > In out production cluster, we encounter this problem. We found the replication queue was there and no regionserver took over it and a "lock" zk node left there. > {quote} > hbase.32561.log:2014-09-24,14:09:28,790 INFO org.apache.hadoop.hbase.replication.ReplicationZookeeper: Won't transfer the queue, another RS took care of it because of: KeeperErrorCode = NoNode for /hbase/hhsrv-micloud/replication/rs/hh-hadoop-srv-st09.bj,12610,1410937824255/lock > hbase.32561.log:2014-09-24,14:14:45,148 INFO org.apache.hadoop.hbase.replication.ReplicationZookeeper: Won't transfer the queue, another RS took care of it because of: KeeperErrorCode = NoNode for /hbase/hhsrv-micloud/replication/rs/hh-hadoop-srv-st10.bj,12600,1410937795685/lock > {quote} > A quick solution is that the lock operation just create an ephemeral "lock" zookeeper node and when the lock node is deleted, other regionserver will be notified to check if there are replication queue left. > Suggestions are welcomed! Thanks. -- This message was sent by Atlassian JIRA (v6.3.4#6332)