hadoop-mapreduce-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Bryan Beaudreault <bbeaudrea...@hubspot.com>
Subject Re: Juggling or swaping out the standby NameNode in a QJM / HA configuration
Date Fri, 01 Aug 2014 13:46:33 GMT
Also you shouldn't format the new standby. You only format a namenode for a
brand new cluster. Once a cluster is live you should just use the bootstrap
on the new namenodes and never format again. Bootstrap is basically a
special format that just creates the dirs and copies an active fsimage to
the host.

If format fails (it's buggy imo) just rsync from the active namenode. It
will catch up by replaying the edits from the QJM when it is started.

On Friday, August 1, 2014, Bryan Beaudreault <bbeaudreault@hubspot.com>
wrote:

> You should first replace the namenode, then when that is completely
> finished move on to replacing any journal nodes. That part is easy:
>
> 1) bootstrap new JN (rsync from an existing)
> 2) Start new JN
> 3) push hdfs-site.xml to both namenodes
> 4) restart standby namenode
> 5) verify logs and admin ui show new JN
> 6) restart active namenode
> 7) verify both namenodes (failover should have happened and old standby
> should be writing to the new JN)
>
> You can remove an existing JN at the same time if you want, just be
> careful to preserve the majority of the quorum during the whole operation
> (I.e only replace 1 at a time).
>
> Also I think it is best to do hdfs dfsadmin -rollEdits after each replaced
> journalnode. IIRC there is a JIRA open about rolling restarting journal
> nodes not being safe unless you roll edits. So that would go for replacing
> too.
>
> On Friday, August 1, 2014, Colin Kincaid Williams <discord@uw.edu
> <javascript:_e(%7B%7D,'cvml','discord@uw.edu');>> wrote:
>
>> I will run through the procedure again tomorrow. It was late in the day
>> before I had a chance to test the procedure.
>>
>> If I recall correctly I had an issue formatting the New standby, before
>> bootstrapping.  I think either at that point, or during the Zookeeper
>> format command,  I was queried  to format the journal to the 3 hosts in the
>> quorum.I was unable to proceed without exception unless choosing this
>> option .
>>
>> Are there any concerns adding another journal node to the new standby?
>> On Jul 31, 2014 9:44 PM, "Bryan Beaudreault" <bbeaudreault@hubspot.com>
>> wrote:
>>
>>> This shouldn't have affected the journalnodes at all -- they are mostly
>>> unaware of the zkfc and active/standby state.  Did you do something else
>>> that may have impacted the journalnodes? (i.e. shut down 1 or more of them,
>>> or something else)
>>>
>>> For your previous 2 emails, reporting errors/warns when doing -formatZK:
>>>
>>> The WARN is fine.  It's true that you could get in a weird state if you
>>> had multiple namenodes up.  But with just 1 namenode up, you should be
>>> safe. What you are trying to avoid is a split brain or standby/standby
>>> state, but that is impossible with just 1 namenode alive.  Similarly, the
>>> ERROR is a sanity check to make sure you don't screw yourself by formatting
>>> a running cluster.  I should have mentioned you need to use the -force
>>> argument to get around that.
>>>
>>>
>>> On Fri, Aug 1, 2014 at 12:35 AM, Colin Kincaid Williams <discord@uw.edu>
>>> wrote:
>>>
>>>> However continuing with the process my QJM eventually error'd out and
>>>> my Active NameNode went down.
>>>>
>>>> 2014-07-31 20:59:33,944 WARN  [Logger channel to rhel6.local/
>>>> 10.120.5.247:8485] client.QuorumJournalManager
>>>> (IPCLoggerChannel.java:call(357)) - Remote journal 10.120.5.247:8485
>>>> failed to write txns 9635-9635. Will try to write to this JN again after
>>>> the next log roll.
>>>> org.apache.hadoop.ipc.RemoteException(java.io.IOException): IPC's epoch
>>>> 5 is not the current writer epoch  0
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.server.Journal.checkWriteRequest(Journal.java:430)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.server.Journal.journal(Journal.java:331)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.journal(JournalNodeRpcServer.java:142)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolServerSideTranslatorPB.journal(QJournalProtocolServerSideTranslatorPB.java:132)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos$QJournalProtocolService$2.callBlockingMethod(QJournalProtocolProtos.java:14018)
>>>>  at
>>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453)
>>>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:898)
>>>>  at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1693)
>>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1689)
>>>>  at java.security.AccessController.doPrivileged(Native Method)
>>>> at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>  at
>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1332)
>>>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1687)
>>>>
>>>> at org.apache.hadoop.ipc.Client.call(Client.java:1224)
>>>> at
>>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202)
>>>>  at com.sun.proxy.$Proxy9.journal(Unknown Source)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolTranslatorPB.journal(QJournalProtocolTranslatorPB.java:156)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.client.IPCLoggerChannel$7.call(IPCLoggerChannel.java:354)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.client.IPCLoggerChannel$7.call(IPCLoggerChannel.java:347)
>>>>  at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>>> at
>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>>  at
>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>> at java.lang.Thread.run(Thread.java:745)
>>>> 2014-07-31 20:59:33,954 WARN  [Logger channel to rhel1.local/
>>>> 10.120.5.203:8485] client.QuorumJournalManager
>>>> (IPCLoggerChannel.java:call(357)) - Remote journal 10.120.5.203:8485
>>>> failed to write txns 9635-9635. Will try to write to this JN again after
>>>> the next log roll.
>>>> org.apache.hadoop.ipc.RemoteException(java.io.IOException): IPC's epoch
>>>> 5 is not the current writer epoch  0
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.server.Journal.checkWriteRequest(Journal.java:430)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.server.Journal.journal(Journal.java:331)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.journal(JournalNodeRpcServer.java:142)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolServerSideTranslatorPB.journal(QJournalProtocolServerSideTranslatorPB.java:132)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos$QJournalProtocolService$2.callBlockingMethod(QJournalProtocolProtos.java:14018)
>>>>  at
>>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453)
>>>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:898)
>>>>  at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1693)
>>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1689)
>>>>  at java.security.AccessController.doPrivileged(Native Method)
>>>> at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>  at
>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1332)
>>>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1687)
>>>>
>>>> at org.apache.hadoop.ipc.Client.call(Client.java:1224)
>>>> at
>>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202)
>>>>  at com.sun.proxy.$Proxy9.journal(Unknown Source)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolTranslatorPB.journal(QJournalProtocolTranslatorPB.java:156)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.client.IPCLoggerChannel$7.call(IPCLoggerChannel.java:354)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.client.IPCLoggerChannel$7.call(IPCLoggerChannel.java:347)
>>>>  at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>>> at
>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>>  at
>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>> at java.lang.Thread.run(Thread.java:745)
>>>> 2014-07-31 20:59:33,975 WARN  [Logger channel to rhel2.local/
>>>> 10.120.5.25:8485] client.QuorumJournalManager
>>>> (IPCLoggerChannel.java:call(357)) - Remote journal 10.120.5.25:8485
>>>> failed to write txns 9635-9635. Will try to write to this JN again after
>>>> the next log roll.
>>>> org.apache.hadoop.ipc.RemoteException(java.io.IOException): IPC's epoch
>>>> 5 is not the current writer epoch  0
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.server.Journal.checkWriteRequest(Journal.java:430)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.server.Journal.journal(Journal.java:331)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.journal(JournalNodeRpcServer.java:142)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolServerSideTranslatorPB.journal(QJournalProtocolServerSideTranslatorPB.java:132)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos$QJournalProtocolService$2.callBlockingMethod(QJournalProtocolProtos.java:14018)
>>>>  at
>>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453)
>>>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:898)
>>>>  at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1693)
>>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1689)
>>>>  at java.security.AccessController.doPrivileged(Native Method)
>>>> at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>  at
>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1332)
>>>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1687)
>>>>
>>>> at org.apache.hadoop.ipc.Client.call(Client.java:1224)
>>>> at
>>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202)
>>>>  at com.sun.proxy.$Proxy9.journal(Unknown Source)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolTranslatorPB.journal(QJournalProtocolTranslatorPB.java:156)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.client.IPCLoggerChannel$7.call(IPCLoggerChannel.java:354)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.client.IPCLoggerChannel$7.call(IPCLoggerChannel.java:347)
>>>>  at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>>> at
>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>>  at
>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>> at java.lang.Thread.run(Thread.java:745)
>>>> 2014-07-31 20:59:33,976 FATAL [IPC Server handler 5 on 8020]
>>>> namenode.FSEditLog (JournalSet.java:mapJournalsAndReportErrors(355)) -
>>>> Error: flush failed for required journal (JournalAndStream(mgr=QJM to [
>>>> 10.120.5.203:8485, 10.120.5.247:8485, 10.120.5.25:8485],
>>>> stream=QuorumOutputStream starting at txid 9634))
>>>> org.apache.hadoop.hdfs.qjournal.client.QuorumException: Got too many
>>>> exceptions to achieve quorum size 2/3. 3 exceptions thrown:
>>>> 10.120.5.25:8485: IPC's epoch 5 is not the current writer epoch  0
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.server.Journal.checkWriteRequest(Journal.java:430)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.server.Journal.journal(Journal.java:331)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.journal(JournalNodeRpcServer.java:142)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolServerSideTranslatorPB.journal(QJournalProtocolServerSideTranslatorPB.java:132)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos$QJournalProtocolService$2.callBlockingMethod(QJournalProtocolProtos.java:14018)
>>>> at
>>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453)
>>>>  at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:898)
>>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1693)
>>>>  at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1689)
>>>> at java.security.AccessController.doPrivileged(Native Method)
>>>>  at javax.security.auth.Subject.doAs(Subject.java:415)
>>>> at
>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1332)
>>>>  at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1687)
>>>>
>>>> 10.120.5.203:8485: IPC's epoch 5 is not the current writer epoch  0
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.server.Journal.checkWriteRequest(Journal.java:430)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.server.Journal.journal(Journal.java:331)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.journal(JournalNodeRpcServer.java:142)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolServerSideTranslatorPB.journal(QJournalProtocolServerSideTranslatorPB.java:132)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos$QJournalProtocolService$2.callBlockingMethod(QJournalProtocolProtos.java:14018)
>>>> at
>>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453)
>>>>  at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:898)
>>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1693)
>>>>  at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1689)
>>>> at java.security.AccessController.doPrivileged(Native Method)
>>>>  at javax.security.auth.Subject.doAs(Subject.java:415)
>>>> at
>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1332)
>>>>  at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1687)
>>>>
>>>> 10.120.5.247:8485: IPC's epoch 5 is not the current writer epoch  0
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.server.Journal.checkWriteRequest(Journal.java:430)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.server.Journal.journal(Journal.java:331)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.journal(JournalNodeRpcServer.java:142)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolServerSideTranslatorPB.journal(QJournalProtocolServerSideTranslatorPB.java:132)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos$QJournalProtocolService$2.callBlockingMethod(QJournalProtocolProtos.java:14018)
>>>> at
>>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453)
>>>>  at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:898)
>>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1693)
>>>>  at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1689)
>>>> at java.security.AccessController.doPrivileged(Native Method)
>>>>  at javax.security.auth.Subject.doAs(Subject.java:415)
>>>> at
>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1332)
>>>>  at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1687)
>>>>
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.client.QuorumException.create(QuorumException.java:81)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.client.QuorumCall.rethrowException(QuorumCall.java:213)
>>>> at
>>>> org.apache.hadoop.hdfs.qjournal.client.AsyncLoggerSet.waitForWriteQuorum(AsyncLoggerSet.java:142)
>>>>  at
>>>> org.apache.hadoop.hdfs.qjournal.client.QuorumOutputStream.flushAndSync(QuorumOutputStream.java:107)
>>>> at
>>>> org.apache.hadoop.hdfs.server.namenode.EditLogOutputStream.flush(EditLogOutputStream.java:113)
>>>>  at
>>>> org.apache.hadoop.hdfs.server.namenode.EditLogOutputStream.flush(EditLogOutputStream.java:107)
>>>> at
>>>> org.apache.hadoop.hdfs.server.namenode.JournalSet$JournalSetOutputStream$8.apply(JournalSet.java:490)
>>>>  at
>>>> org.apache.hadoop.hdfs.server.namenode.JournalSet.mapJournalsAndReportErrors(JournalSet.java:350)
>>>> at
>>>> org.apache.hadoop.hdfs.server.namenode.JournalSet.access$100(JournalSet.java:55)
>>>>  at
>>>> org.apache.hadoop.hdfs.server.namenode.JournalSet$JournalSetOutputStream.flush(JournalSet.java:486)
>>>> at
>>>> org.apache.hadoop.hdfs.server.namenode.FSEditLog.logSync(FSEditLog.java:581)
>>>>  at
>>>> org.apache.hadoop.hdfs.server.namenode.FSEditLog.endCurrentLogSegment(FSEditLog.java:946)
>>>> at
>>>> org.apache.hadoop.hdfs.server.namenode.FSEditLog.rollEditLog(FSEditLog.java:884)
>>>>  at
>>>> org.apache.hadoop.hdfs.server.namenode.FSImage.rollEditLog(FSImage.java:1013)
>>>> at
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.rollEditLog(FSNamesystem.java:4436)
>>>>  at
>>>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.rollEditLog(NameNodeRpcServer.java:734)
>>>> at
>>>> org.apache.hadoop.hdfs.protocolPB.NamenodeProtocolServerSideTranslatorPB.rollEditLog(NamenodeProtocolServerSideTranslatorPB.java:129)
>>>>  at
>>>> org.apache.hadoop.hdfs.protocol.proto.NamenodeProtocolProtos$NamenodeProtocolService$2.callBlockingMethod(NamenodeProtocolProtos.java:8762)
>>>> at
>>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:453)
>>>>  at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:898)
>>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1693)
>>>>  at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1689)
>>>> at java.security.AccessController.doPrivileged(Native Method)
>>>>  at javax.security.auth.Subject.doAs(Subject.java:415)
>>>> at
>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1332)
>>>>  at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1687)
>>>> 2014-07-31 20:59:33,976 WARN  [IPC Server handler 5 on 8020]
>>>> client.QuorumJournalManager (QuorumOutputStream.java:abort(72)) - Aborting
>>>> QuorumOutputStream starting at txid 9634
>>>> 2014-07-31 20:59:33,978 INFO  [IPC Server handler 5 on 8020]
>>>> util.ExitUtil (ExitUtil.java:terminate(87)) - Exiting with status 1
>>>> 2014-07-31 20:59:33,982 INFO  [Thread-0] namenode.NameNode
>>>> (StringUtils.java:run(615)) - SHUTDOWN_MSG:
>>>>
>>>>
>>>>
>>>> On Thu, Jul 31, 2014 at 6:08 PM, Colin Kincaid Williams <discord@uw.edu
>>>> > wrote:
>>>>
>>>>> I tried a third time and it just worked?
>>>>>
>>>>> sudo hdfs zkfc -formatZK
>>>>> 2014-07-31 18:07:51,595 INFO  [main] tools.DFSZKFailoverController
>>>>> (DFSZKFailoverController.java:<init>(140)) - Failover controller
configured
>>>>> for NameNode NameNode at rhel1.local/10.120.5.203:8020
>>>>> 2014-07-31 18:07:51,791 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client
>>>>> environment:zookeeper.version=3.4.3-cdh4.1.3--1, built on 01/27/2013
00:13
>>>>> GMT
>>>>> 2014-07-31 18:07:51,791 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client environment:host.name
>>>>> =rhel1.local
>>>>> 2014-07-31 18:07:51,792 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client environment:java.version=1.7.0_60
>>>>> 2014-07-31 18:07:51,792 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client environment:java.vendor=Oracle
>>>>> Corporation
>>>>> 2014-07-31 18:07:51,792 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client
>>>>> environment:java.home=/usr/java/jdk1.7.0_60/jre
>>>>> 2014-07-31 18:07:51,792 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client
>>>>> environment:java.class.path=/etc/hadoop/conf:/usr/lib/hadoop/lib/jersey-core-1.8.jar:/usr/lib/hadoop/lib/commons-cli-1.2.jar:/usr/lib/hadoop/lib/jaxb-impl-2.2.3-1.jar:/usr/lib/hadoop/lib/jackson-mapper-asl-1.8.8.jar:/usr/lib/hadoop/lib/jsch-0.1.42.jar:/usr/lib/hadoop/lib/asm-3.2.jar:/usr/lib/hadoop/lib/kfs-0.3.jar:/usr/lib/hadoop/lib/jsr305-1.3.9.jar:/usr/lib/hadoop/lib/commons-logging-1.1.1.jar:/usr/lib/hadoop/lib/jets3t-0.6.1.jar:/usr/lib/hadoop/lib/xmlenc-0.52.jar:/usr/lib/hadoop/lib/jetty-6.1.26.cloudera.2.jar:/usr/lib/hadoop/lib/jsp-api-2.1.jar:/usr/lib/hadoop/lib/commons-collections-3.2.1.jar:/usr/lib/hadoop/lib/jackson-core-asl-1.8.8.jar:/usr/lib/hadoop/lib/guava-11.0.2.jar:/usr/lib/hadoop/lib/zookeeper-3.4.3-cdh4.1.3.jar:/usr/lib/hadoop/lib/servlet-api-2.5.jar:/usr/lib/hadoop/lib/log4j-1.2.17.jar:/usr/lib/hadoop/lib/avro-1.7.1.cloudera.2.jar:/usr/lib/hadoop/lib/jersey-server-1.8.jar:/usr/lib/hadoop/lib/jasper-compiler-5.5.23.jar:/usr/lib/hadoop/lib/jackson-jaxrs-1.8.8.jar:/usr/lib/hadoop/lib/jline-0.9.94.jar:/usr/lib/hadoop/lib/commons-el-1.0.jar:/usr/lib/hadoop/lib/commons-io-2.1.jar:/usr/lib/hadoop/lib/commons-configuration-1.6.jar:/usr/lib/hadoop/lib/commons-net-3.1.jar:/usr/lib/hadoop/lib/jackson-xc-1.8.8.jar:/usr/lib/hadoop/lib/jasper-runtime-5.5.23.jar:/usr/lib/hadoop/lib/commons-digester-1.8.jar:/usr/lib/hadoop/lib/slf4j-api-1.6.1.jar:/usr/lib/hadoop/lib/commons-beanutils-core-1.8.0.jar:/usr/lib/hadoop/lib/commons-httpclient-3.1.jar:/usr/lib/hadoop/lib/activation-1.1.jar:/usr/lib/hadoop/lib/jetty-util-6.1.26.cloudera.2.jar:/usr/lib/hadoop/lib/commons-beanutils-1.7.0.jar:/usr/lib/hadoop/lib/commons-lang-2.5.jar:/usr/lib/hadoop/lib/snappy-java-1.0.4.1.jar:/usr/lib/hadoop/lib/junit-4.8.2.jar:/usr/lib/hadoop/lib/stax-api-1.0.1.jar:/usr/lib/hadoop/lib/commons-math-2.1.jar:/usr/lib/hadoop/lib/mockito-all-1.8.5.jar:/usr/lib/hadoop/lib/jettison-1.1.jar:/usr/lib/hadoop/lib/jaxb-api-2.2.2.jar:/usr/lib/hadoop/lib/paranamer-2.3.jar:/usr/lib/hadoop/lib/protobuf-java-2.4.0a.jar:/usr/lib/hadoop/lib/jersey-json-1.8.jar:/usr/lib/hadoop/lib/slf4j-log4j12-1.6.1.jar:/usr/lib/hadoop/lib/commons-codec-1.4.jar:/usr/lib/hadoop/.//hadoop-annotations-2.0.0-cdh4.1.3.jar:/usr/lib/hadoop/.//hadoop-common.jar:/usr/lib/hadoop/.//hadoop-common-2.0.0-cdh4.1.3-tests.jar:/usr/lib/hadoop/.//hadoop-common-2.0.0-cdh4.1.3.jar:/usr/lib/hadoop/.//hadoop-annotations.jar:/usr/lib/hadoop/.//hadoop-auth-2.0.0-cdh4.1.3.jar:/usr/lib/hadoop/.//hadoop-auth.jar:/usr/lib/hadoop-hdfs/./:/usr/lib/hadoop-hdfs/lib/jersey-core-1.8.jar:/usr/lib/hadoop-hdfs/lib/commons-cli-1.2.jar:/usr/lib/hadoop-hdfs/lib/jackson-mapper-asl-1.8.8.jar:/usr/lib/hadoop-hdfs/lib/asm-3.2.jar:/usr/lib/hadoop-hdfs/lib/jsr305-1.3.9.jar:/usr/lib/hadoop-hdfs/lib/commons-logging-1.1.1.jar:/usr/lib/hadoop-hdfs/lib/xmlenc-0.52.jar:/usr/lib/hadoop-hdfs/lib/jetty-6.1.26.cloudera.2.jar:/usr/lib/hadoop-hdfs/lib/jsp-api-2.1.jar:/usr/lib/hadoop-hdfs/lib/jackson-core-asl-1.8.8.jar:/usr/lib/hadoop-hdfs/lib/guava-11.0.2.jar:/usr/lib/hadoop-hdfs/lib/zookeeper-3.4.3-cdh4.1.3.jar:/usr/lib/hadoop-hdfs/lib/servlet-api-2.5.jar:/usr/lib/hadoop-hdfs/lib/log4j-1.2.17.jar:/usr/lib/hadoop-hdfs/lib/jersey-server-1.8.jar:/usr/lib/hadoop-hdfs/lib/jline-0.9.94.jar:/usr/lib/hadoop-hdfs/lib/commons-el-1.0.jar:/usr/lib/hadoop-hdfs/lib/commons-io-2.1.jar:/usr/lib/hadoop-hdfs/lib/commons-daemon-1.0.3.jar:/usr/lib/hadoop-hdfs/lib/jasper-runtime-5.5.23.jar:/usr/lib/hadoop-hdfs/lib/jetty-util-6.1.26.cloudera.2.jar:/usr/lib/hadoop-hdfs/lib/commons-lang-2.5.jar:/usr/lib/hadoop-hdfs/lib/protobuf-java-2.4.0a.jar:/usr/lib/hadoop-hdfs/lib/commons-codec-1.4.jar:/usr/lib/hadoop-hdfs/.//hadoop-hdfs-2.0.0-cdh4.1.3.jar:/usr/lib/hadoop-hdfs/.//hadoop-hdfs-2.0.0-cdh4.1.3-tests.jar:/usr/lib/hadoop-hdfs/.//hadoop-hdfs.jar:/usr/lib/hadoop-yarn/lib/jersey-core-1.8.jar:/usr/lib/hadoop-yarn/lib/jackson-mapper-asl-1.8.8.jar:/usr/lib/hadoop-yarn/lib/asm-3.2.jar:/usr/lib/hadoop-yarn/lib/netty-3.2.4.Final.jar:/usr/lib/hadoop-yarn/lib/jackson-core-asl-1.8.8.jar:/usr/lib/hadoop-yarn/lib/jersey-guice-1.8.jar:/usr/lib/hadoop-yarn/lib/log4j-1.2.17.jar:/usr/lib/hadoop-yarn/lib/avro-1.7.1.cloudera.2.jar:/usr/lib/hadoop-yarn/lib/jersey-server-1.8.jar:/usr/lib/hadoop-yarn/lib/guice-3.0.jar:/usr/lib/hadoop-yarn/lib/commons-io-2.1.jar:/usr/lib/hadoop-yarn/lib/aopalliance-1.0.jar:/usr/lib/hadoop-yarn/lib/javax.inject-1.jar:/usr/lib/hadoop-yarn/lib/snappy-java-1.0.4.1.jar:/usr/lib/hadoop-yarn/lib/guice-servlet-3.0.jar:/usr/lib/hadoop-yarn/lib/paranamer-2.3.jar:/usr/lib/hadoop-yarn/lib/protobuf-java-2.4.0a.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-applications-distributedshell.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-api-2.0.0-cdh4.1.3.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-site-2.0.0-cdh4.1.3.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-common-2.0.0-cdh4.1.3.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-api.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-site.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-server-nodemanager.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-server-tests.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-server-tests-2.0.0-cdh4.1.3.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-server-common.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-applications-distributedshell-2.0.0-cdh4.1.3.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-server-web-proxy-2.0.0-cdh4.1.3.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-server-tests-2.0.0-cdh4.1.3-tests.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-server-resourcemanager.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-server-web-proxy.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-server-resourcemanager-2.0.0-cdh4.1.3.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-server-common-2.0.0-cdh4.1.3.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-server-nodemanager-2.0.0-cdh4.1.3.jar:/usr/lib/hadoop-yarn/.//hadoop-yarn-common.jar:/usr/lib/hadoop-0.20-mapreduce/./:/usr/lib/hadoop-0.20-mapreduce/lib/jersey-core-1.8.jar:/usr/lib/hadoop-0.20-mapreduce/lib/ant-contrib-1.0b3.jar:/usr/lib/hadoop-0.20-mapreduce/lib/commons-cli-1.2.jar:/usr/lib/hadoop-0.20-mapreduce/lib/aspectjrt-1.6.5.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jaxb-impl-2.2.3-1.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jackson-mapper-asl-1.8.8.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jsch-0.1.42.jar:/usr/lib/hadoop-0.20-mapreduce/lib/asm-3.2.jar:/usr/lib/hadoop-0.20-mapreduce/lib/kfs-0.3.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jsr305-1.3.9.jar:/usr/lib/hadoop-0.20-mapreduce/lib/commons-logging-1.1.1.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jets3t-0.6.1.jar:/usr/lib/hadoop-0.20-mapreduce/lib/xmlenc-0.52.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jetty-6.1.26.cloudera.2.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jsp-api-2.1.jar:/usr/lib/hadoop-0.20-mapreduce/lib/commons-collections-3.2.1.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jackson-core-asl-1.8.8.jar:/usr/lib/hadoop-0.20-mapreduce/lib/guava-11.0.2.jar:/usr/lib/hadoop-0.20-mapreduce/lib/servlet-api-2.5.jar:/usr/lib/hadoop-0.20-mapreduce/lib/log4j-1.2.17.jar:/usr/lib/hadoop-0.20-mapreduce/lib/avro-1.7.1.cloudera.2.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jersey-server-1.8.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jasper-compiler-5.5.23.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jackson-jaxrs-1.8.8.jar:/usr/lib/hadoop-0.20-mapreduce/lib/commons-el-1.0.jar:/usr/lib/hadoop-0.20-mapreduce/lib/commons-io-2.1.jar:/usr/lib/hadoop-0.20-mapreduce/lib/commons-configuration-1.6.jar:/usr/lib/hadoop-0.20-mapreduce/lib/aspectjtools-1.6.5.jar:/usr/lib/hadoop-0.20-mapreduce/lib/commons-net-3.1.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jackson-xc-1.8.8.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jasper-runtime-5.5.23.jar:/usr/lib/hadoop-0.20-mapreduce/lib/commons-digester-1.8.jar:/usr/lib/hadoop-0.20-mapreduce/lib/slf4j-api-1.6.1.jar:/usr/lib/hadoop-0.20-mapreduce/lib/commons-beanutils-core-1.8.0.jar:/usr/lib/hadoop-0.20-mapreduce/lib/commons-httpclient-3.1.jar:/usr/lib/hadoop-0.20-mapreduce/lib/kfs-0.2.2.jar:/usr/lib/hadoop-0.20-mapreduce/lib/activation-1.1.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jetty-util-6.1.26.cloudera.2.jar:/usr/lib/hadoop-0.20-mapreduce/lib/commons-beanutils-1.7.0.jar:/usr/lib/hadoop-0.20-mapreduce/lib/commons-lang-2.5.jar:/usr/lib/hadoop-0.20-mapreduce/lib/avro-compiler-1.7.1.cloudera.2.jar:/usr/lib/hadoop-0.20-mapreduce/lib/snappy-java-1.0.4.1.jar:/usr/lib/hadoop-0.20-mapreduce/lib/junit-4.8.2.jar:/usr/lib/hadoop-0.20-mapreduce/lib/stax-api-1.0.1.jar:/usr/lib/hadoop-0.20-mapreduce/lib/commons-math-2.1.jar:/usr/lib/hadoop-0.20-mapreduce/lib/mockito-all-1.8.5.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jettison-1.1.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jaxb-api-2.2.2.jar:/usr/lib/hadoop-0.20-mapreduce/lib/hsqldb-1.8.0.10.jar:/usr/lib/hadoop-0.20-mapreduce/lib/paranamer-2.3.jar:/usr/lib/hadoop-0.20-mapreduce/lib/protobuf-java-2.4.0a.jar:/usr/lib/hadoop-0.20-mapreduce/lib/jersey-json-1.8.jar:/usr/lib/hadoop-0.20-mapreduce/lib/commons-codec-1.4.jar:/usr/lib/hadoop-0.20-mapreduce/lib/hadoop-fairscheduler-2.0.0-mr1-cdh4.1.3.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-examples-2.0.0-mr1-cdh4.1.3.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-2.0.0-mr1-cdh4.1.3-ant.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-ant-2.0.0-mr1-cdh4.1.3.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-ant.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-2.0.0-mr1-cdh4.1.3-core.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-test.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-tools.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-core-2.0.0-mr1-cdh4.1.3.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-2.0.0-mr1-cdh4.1.3-examples.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-test-2.0.0-mr1-cdh4.1.3.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-tools-2.0.0-mr1-cdh4.1.3.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-examples.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-2.0.0-mr1-cdh4.1.3-tools.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-2.0.0-mr1-cdh4.1.3-test.jar:/usr/lib/hadoop-0.20-mapreduce/.//hadoop-core.jar
>>>>> 2014-07-31 18:07:51,793 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client
>>>>> environment:java.library.path=//usr/lib/hadoop/lib/native
>>>>> 2014-07-31 18:07:51,801 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client environment:java.io.tmpdir=/tmp
>>>>> 2014-07-31 18:07:51,801 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client environment:java.compiler=<NA>
>>>>> 2014-07-31 18:07:51,801 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client environment:os.name=Linux
>>>>> 2014-07-31 18:07:51,802 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client environment:os.arch=amd64
>>>>> 2014-07-31 18:07:51,802 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client
>>>>> environment:os.version=2.6.32-358.el6.x86_64
>>>>> 2014-07-31 18:07:51,802 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client environment:user.name=root
>>>>> 2014-07-31 18:07:51,802 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client environment:user.home=/root
>>>>> 2014-07-31 18:07:51,802 INFO  [main] zookeeper.ZooKeeper
>>>>> (Environment.java:logEnv(100)) - Client
>>>>> environment:user.dir=/etc/hbase/conf.golden_apple
>>>>> 2014-07-31 18:07:51,813 INFO  [main] zookeeper.ZooKeeper
>>>>> (ZooKeeper.java:<init>(433)) - Initiating client connection,
>>>>> connectString=rhel1.local:2181,rhel6.local:2181,rhel2.local:2181
>>>>> sessionTimeout=5000 watcher=null
>>>>> 2014-07-31 18:07:51,833 INFO  [main-SendThread(rhel1.local:2181)]
>>>>> zookeeper.ClientCnxn (ClientCnxn.java:logStartConnect(958)) - Opening
>>>>> socket connection to server rhel1.local/10.120.5.203:2181. Will not
>>>>> attempt to authenticate using SASL (unknown error)
>>>>> 2014-07-31 18:07:51,844 INFO  [main-SendThread(rhel1.local:2181)]
>>>>> zookeeper.ClientCnxn (ClientCnxn.java:primeConnection(850)) - Socket
>>>>> connection established to rhel1.local/10.120.5.203:2181, initiating
>>>>> session
>>>>> 2014-07-31 18:07:51,852 INFO  [main-SendThread(rhel1.local:2181)]
>>>>> zookeeper.ClientCnxn (ClientCnxn.java:onConnected(1187)) - Session
>>>>> establishment complete on server rhel1.local/10.120.5.203:2181,
>>>>> sessionid = 0x1478902fddc000a, negotiated timeout = 5000
>>>>>  ===============================================
>>>>> The configured parent znode /hadoop-ha/golden-apple already exists.
>>>>> Are you sure you want to clear all failover information from
>>>>> ZooKeeper?
>>>>> WARNING: Before proceeding, ensure that all HDFS services and
>>>>> failover controllers are stopped!
>>>>> ===============================================
>>>>> Proceed formatting /hadoop-ha/golden-apple? (Y or N) 2014-07-31
>>>>> 18:07:51,858 INFO  [main-EventThread] ha.ActiveStandbyElector
>>>>> (ActiveStandbyElector.java:processWatchEvent(538)) - Session connected.
>>>>> Y
>>>>> 2014-07-31 18:08:00,439 INFO  [main] ha.ActiveStandbyElector
>>>>> (ActiveStandbyElector.java:clearParentZNode(314)) - Recursively deleting
>>>>> /hadoop-ha/golden-apple from ZK...
>>>>> 2014-07-31 18:08:00,506 INFO  [main] ha.ActiveStandbyElector
>>>>> (ActiveStandbyElector.java:clearParentZNode(327)) - Successfully deleted
>>>>> /hadoop-ha/golden-apple from ZK.
>>>>> 2014-07-31 18:08:00,541 INFO  [main] ha.ActiveStandbyElector
>>>>> (ActiveStandbyElector.java:ensureParentZNode(299)) - Successfully created
>>>>> /hadoop-ha/golden-apple in ZK.
>>>>> 2014-07-31 18:08:00,545 INFO  [main-EventThread] zookeeper.ClientCnxn
>>>>> (ClientCnxn.java:run(511)) - EventThread shut down
>>>>> 2014-07-31 18:08:00,545 INFO  [main] zookeeper.ZooKeeper
>>>>> (ZooKeeper.java:close(679)) - Session: 0x1478902fddc000a closed
>>>>>
>>>>>
>>>>>
>>>>> On Thu, Jul 31, 2014 at 2:51 PM, Alex Newman <posix4e@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Cheers. That's rough. We don't have that problem here at WanDISCO.
>>>>>>
>>>>>> On Thu, Jul 31, 2014 at 12:46 PM, Colin Kincaid Williams <
>>>>>> discord@uw.edu> wrote:
>>>>>> > Hi this is drocsid / discord from #hbase. Thanks for the help
>>>>>> earlier today.
>>>>>> > Just thought I'd forward this info regarding swapping out the
>>>>>> NameNode in a
>>>>>> > QJM / HA configuration. See you around on #hbase. If you visit
>>>>>> Seattle, feel
>>>>>> > free to give me a shout out.
>>>>>> >
>>>>>> > ---------- Forwarded message ----------
>>>>>> > From: Colin Kincaid Williams <discord@uw.edu>
>>>>>> > Date: Thu, Jul 31, 2014 at 12:35 PM
>>>>>> > Subject: Re: Juggling or swaping out the standby NameNode in
a QJM
>>>>>> / HA
>>>>>> > configuration
>>>>>> > To: user@hadoop.apache.org
>>>>>> >
>>>>>> >
>>>>>> > Hi Jing,
>>>>>> >
>>>>>> > Thanks for the response. I will try this out, and file an Apache
>>>>>> jira.
>>>>>> >
>>>>>> > Best,
>>>>>> >
>>>>>> > Colin Williams
>>>>>> >
>>>>>> >
>>>>>> > On Thu, Jul 31, 2014 at 11:19 AM, Jing Zhao <jing@hortonworks.com>
>>>>>> wrote:
>>>>>> >>
>>>>>> >> Hi Colin,
>>>>>> >>
>>>>>> >>     I guess currently we may have to restart almost all
the
>>>>>> >> daemons/services in order to swap out a standby NameNode
(SBN):
>>>>>> >>
>>>>>> >> 1. The current active NameNode (ANN) needs to know the new
SBN
>>>>>> since in
>>>>>> >> the current implementation the SBN tries to send rollEditLog
RPC
>>>>>> request to
>>>>>> >> ANN periodically (thus if a NN failover happens later, the
>>>>>> original ANN
>>>>>> >> needs to send this RPC to the correct NN).
>>>>>> >> 2. Looks like the DataNode currently cannot do real refreshment
>>>>>> for NN.
>>>>>> >> Look at the code in BPOfferService:
>>>>>> >>
>>>>>> >>   void refreshNNList(ArrayList<InetSocketAddress>
addrs) throws
>>>>>> >> IOException {
>>>>>> >>     Set<InetSocketAddress> oldAddrs = Sets.newHashSet();
>>>>>> >>     for (BPServiceActor actor : bpServices) {
>>>>>> >>       oldAddrs.add(actor.getNNSocketAddress());
>>>>>> >>     }
>>>>>> >>     Set<InetSocketAddress> newAddrs = Sets.newHashSet(addrs);
>>>>>> >>
>>>>>> >>     if (!Sets.symmetricDifference(oldAddrs, newAddrs).isEmpty())
{
>>>>>> >>       // Keep things simple for now -- we can implement
this at a
>>>>>> later
>>>>>> >> date.
>>>>>> >>       throw new IOException(
>>>>>> >>           "HA does not currently support adding a new standby
to a
>>>>>> running
>>>>>> >> DN. " +
>>>>>> >>           "Please do a rolling restart of DNs to reconfigure
the
>>>>>> list of
>>>>>> >> NNs.");
>>>>>> >>     }
>>>>>> >>   }
>>>>>> >>
>>>>>> >> 3. If you're using automatic failover, you also need to
update the
>>>>>> >> configuration of the ZKFC on the current ANN machine, since
ZKFC
>>>>>> will do
>>>>>> >> gracefully fencing by sending RPC to the other NN.
>>>>>> >> 4. Looks like we do not need to restart JournalNodes for
the new
>>>>>> SBN but I
>>>>>> >> have not tried before.
>>>>>> >>
>>>>>> >>     Thus in general we may still have to restart all the
services
>>>>>> (except
>>>>>> >> JNs) and update their configurations. But this may be a
rolling
>>>>>> restart
>>>>>> >> process I guess:
>>>>>> >> 1. Shutdown the old SBN, bootstrap the new SBN, and start
the new
>>>>>> SBN.
>>>>>> >> 2. Keep the ANN and its corresponding ZKFC running, do a
rolling
>>>>>> restart
>>>>>> >> of all the DN to update their configurations
>>>>>> >> 3. After restarting all the DN, stop ANN and the ZKFC, and
update
>>>>>> their
>>>>>> >> configuration. The new SBN should become active.
>>>>>> >>
>>>>>> >>     I have not tried the upper steps, thus please let me
know if
>>>>>> this
>>>>>> >> works or not. And I think we should also document the correct
>>>>>> steps in
>>>>>> >> Apache. Could you please file an Apache jira?
>>>>>> >>
>>>>>> >> Thanks,
>>>>>> >> -Jing
>>>>>> >>
>>>>>> >>
>>>>>> >>
>>>>>> >> On Thu, Jul 31, 2014 at 9:37 AM, Colin Kincaid Williams
<
>>>>>> discord@uw.edu>
>>>>>> >> wrote:
>>>>>> >>>
>>>>>> >>> Hello,
>>>>>> >>>
>>>>>> >>> I'm trying to swap out a standby NameNode in a QJM /
HA
>>>>>> configuration. I
>>>>>> >>> believe the steps to achieve this would be something
similar to:
>>>>>> >>>
>>>>>> >>> Use the Bootstrap standby command to prep the replacment
standby.
>>>>>> Or
>>>>>> >>> rsync if the command fails.
>>>>>> >>>
>>>>>> >>> Somehow update the datanodes, so they push the heartbeat
/
>>>>>> journal to the
>>>>>> >>> new standby
>>>>>> >>>
>>>>>> >>> Update the xml configuration on all nodes to reflect
the
>>>>>> replacment
>>>>>> >>> standby.
>>>>>> >>>
>>>>>> >>> Start the replacment standby
>>>>>> >>>
>>>>>> >>> Use some hadoop command to refresh the datanodes to
the new
>>>>>> NameNode
>>>>>> >>> configuration.
>>>>>> >>>
>>>>>> >>> I am not sure how to deal with the Journal switch, or
if I am
>>>>>> going about
>>>>>> >>> this the right way. Can anybody give me some suggestions
here?
>>>>>> >>>
>>>>>> >>>
>>>>>> >>> Regards,
>>>>>> >>>
>>>>>> >>> Colin Williams
>>>>>> >>>
>>>>>> >>
>>>>>> >>
>>>>>> >> CONFIDENTIALITY NOTICE
>>>>>> >> NOTICE: This message is intended for the use of the individual
or
>>>>>> entity
>>>>>> >> to which it is addressed and may contain information that
is
>>>>>> confidential,
>>>>>> >> privileged and exempt from disclosure under applicable law.
If the
>>>>>> reader of
>>>>>> >> this message is not the intended recipient, you are hereby
>>>>>> notified that any
>>>>>> >> printing, copying, dissemination, distribution, disclosure
or
>>>>>> forwarding of
>>>>>> >> this communication is strictly prohibited. If you have received
>>>>>> this
>>>>>> >> communication in error, please contact the sender immediately
and
>>>>>> delete it
>>>>>> >> from your system. Thank You.
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>>
>>>>>
>>>>>
>>>>
>>>

Mime
View raw message