hadoop-mapreduce-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jitendra Yadav <jeetuyadav200...@gmail.com>
Subject Re: Namenode is failing with expception to join
Date Wed, 07 Aug 2013 08:30:09 GMT
Hi,

Did you configured your Name Node to store multiple copies of its metadata?.

You can recover your name node in that situation.

#hadoop namenode -recover

it will ask you whether you want to continue or not, Please follow the
instructions.

Thanks
On Wed, Aug 7, 2013 at 1:44 PM, Manish Bhoge <manishbhoge@rocketmail.com>wrote:

>  I am not using HA here. All I am trying here is to make a 2 node
> cluster. But before that i wanted to make sure that i am setting up
> everything right and make the HDFS up on Pseudo distributed mode. However,
> I am suspecting a mistake in my /etc/hosts file. As, I have rename the
> local host to myhost-1
>
> Please suggest.
>
>   ------------------------------
> *From:* Azuryy Yu <azuryyyu@gmail.com>
> *To:* user@hadoop.apache.org; Manish Bhoge <manishbhoge@rocketmail.com>
> *Sent:* Wednesday, 7 August 2013 1:08 PM
> *Subject:* Re: Namenode is failing with expception to join
>
>  Manish,
>
> you stop HDFS then start HDFS on the standby name node right?
>
> please looked at https://issues.apache.org/jira/browse/HDFS-5058
>
> there are two solutions:
> 1) start HDFS on the active name node, nor SBN
> 2) copy {namenode.name.dir}/* to the SBN
>
> I advice #1.
>
>
>
>
> On Wed, Aug 7, 2013 at 3:00 PM, Manish Bhoge <manishbhoge@rocketmail.com>wrote:
>
>  I have all configuration fine. But whenever i start namenode it fails
> with a below exception. No clue where to fix this?
>
> "2013-08-07 02:56:22,754 FATAL
> org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join"
>
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of
files = 1
> 2013-08-07 02:56:22,751 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Number of
files under construction = 0
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Image file
of size 115 loaded in 0 seconds.
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Loaded image
for txid 0 from /data/1/dfs/nn/current/fsimage_0000000000000000000
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Reading
org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@5f18223d expecting start
txid #1
> 2013-08-07 02:56:22,752 INFO org.apache.hadoop.hdfs.server.namenode.EditLogInputStream:
Fast-forwarding stream '/data/1/dfs/nn/current/edits_0000000000000515247-0000000000000515255'
to transaction ID 1
> 2013-08-07 02:56:22,753 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping
NameNode metrics system...
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode
metrics system stopped.
> 2013-08-07 02:56:22,754 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode
metrics system shutdown complete.2013-08-07 02:56:22,754 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode:
Exception in namenode join
> java.io.IOException: There appears to be a gap in the edit log.  We expected txid 1,
but got txid 515247.
>
>
>
>
>

Mime
View raw message