hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Azuryy Yu <azury...@gmail.com>
Subject Re: I am about to lose all my data please help
Date Tue, 18 Mar 2014 06:06:15 GMT
I don't think this is the case, because there is;
  <property>
    <name>hadoop.tmp.dir</name>
    <value>/home/hadoop/project/hadoop-data</value>
  </property>


On Tue, Mar 18, 2014 at 1:55 PM, Stanley Shi <sshi@gopivotal.com> wrote:

> one possible reason is that you didn't set the namenode working directory,
> by default it's in "/tmp" folder; and the "/tmp" folder might get deleted
> by the OS without any notification. If this is the case, I am afraid you
> have lost all your namenode data.
>
> *<property>
>   <name>dfs.name.dir</name>
>   <value>${hadoop.tmp.dir}/dfs/name</value>
>   <description>Determines where on the local filesystem the DFS name node
>       should store the name table(fsimage).  If this is a comma-delimited list
>       of directories then the name table is replicated in all of the
>       directories, for redundancy. </description>
> </property>*
>
>
> Regards,
> *Stanley Shi,*
>
>
>
> On Sun, Mar 16, 2014 at 5:29 PM, Mirko Kämpf <mirko.kaempf@gmail.com>wrote:
>
>> Hi,
>>
>> what is the location of the namenodes fsimage and editlogs?
>> And how much memory has the NameNode.
>>
>> Did you work with a Secondary NameNode or a Standby NameNode for
>> checkpointing?
>>
>> Where are your HDFS blocks located, are those still safe?
>>
>> With this information at hand, one might be able to fix your setup, but
>> do not format the old namenode before
>> all is working with a fresh one.
>>
>> Grab a copy of the maintainance guide:
>> http://shop.oreilly.com/product/0636920025085.do?sortby=publicationDate
>> which helps solving such type of problems as well.
>>
>> Best wishes
>> Mirko
>>
>>
>> 2014-03-16 9:07 GMT+00:00 Fatih Haltas <fatih.haltas@nyu.edu>:
>>
>> Dear All,
>>>
>>> I have just restarted machines of my hadoop clusters. Now, I am trying
>>> to restart hadoop clusters again, but getting error on namenode restart. I
>>> am afraid of loosing my data as it was properly running for more than 3
>>> months. Currently, I believe if I do namenode formatting, it will work
>>> again, however, data will be lost. Is there anyway to solve this without
>>> losing the data.
>>>
>>> I will really appreciate any help.
>>>
>>> Thanks.
>>>
>>>
>>> =====================
>>> Here is the logs;
>>> ====================
>>> 2014-02-26 16:02:39,698 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:
>>> /************************************************************
>>> STARTUP_MSG: Starting NameNode
>>> STARTUP_MSG:   host = ADUAE042-LAP-V/127.0.0.1
>>> STARTUP_MSG:   args = []
>>> STARTUP_MSG:   version = 1.0.4
>>> STARTUP_MSG:   build =
>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
>>> 1393290; compiled by 'hortonfo' on Wed Oct  3 05:13:58 UTC 2012
>>> ************************************************************/
>>> 2014-02-26 16:02:40,005 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
>>> hadoop-metrics2.properties
>>> 2014-02-26 16:02:40,019 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>> MetricsSystem,sub=Stats registered.
>>> 2014-02-26 16:02:40,021 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
>>> period at 10 second(s).
>>> 2014-02-26 16:02:40,021 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system
>>> started
>>> 2014-02-26 16:02:40,169 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
>>> registered.
>>> 2014-02-26 16:02:40,193 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source jvm
>>> registered.
>>> 2014-02-26 16:02:40,194 INFO
>>> org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
>>> NameNode registered.
>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet: VM type
>>>     = 64-bit
>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet: 2% max
>>> memory = 17.77875 MB
>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet: capacity
>>>      = 2^21 = 2097152 entries
>>> 2014-02-26 16:02:40,242 INFO org.apache.hadoop.hdfs.util.GSet:
>>> recommended=2097152, actual=2097152
>>> 2014-02-26 16:02:40,273 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner=hadoop
>>> 2014-02-26 16:02:40,273 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup
>>> 2014-02-26 16:02:40,274 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>> isPermissionEnabled=true
>>> 2014-02-26 16:02:40,279 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>> dfs.block.invalidate.limit=100
>>> 2014-02-26 16:02:40,279 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>> isAccessTokenEnabled=false accessKeyUpdateInterval=0 min(s),
>>> accessTokenLifetime=0 min(s)
>>> 2014-02-26 16:02:40,724 INFO
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
>>> FSNamesystemStateMBean and NameNodeMXBean
>>> 2014-02-26 16:02:40,749 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: Caching file names
>>> occuring more than 10 times
>>> 2014-02-26 16:02:40,780 ERROR
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: FSNamesystem
>>> initialization failed.
>>> java.io.IOException: NameNode is not formatted.
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:388)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:362)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:276)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:496)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1279)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1288)
>>> 2014-02-26 16:02:40,781 ERROR
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: java.io.IOException:
>>> NameNode is not formatted.
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:388)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:362)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:276)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:496)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1279)
>>>         at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1288)
>>>
>>> 2014-02-26 16:02:40,781 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>>> /************************************************************
>>> SHUTDOWN_MSG: Shutting down NameNode at ADUAE042-LAP-V/127.0.0.1
>>> ************************************************************/
>>>
>>> ===========================
>>> Here is the core-site.xml
>>> ===========================
>>>  <?xml version="1.0"?>
>>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>>>
>>> <!-- Put site-specific property overrides in this file. -->
>>>
>>> <configuration>
>>> <property>
>>>     <name>fs.default.name</name>
>>>     <value>-BLANKED</value>
>>>   </property>
>>>   <property>
>>>     <name>hadoop.tmp.dir</name>
>>>     <value>/home/hadoop/project/hadoop-data</value>
>>>   </property>
>>> </configuration>
>>>
>>>
>>>
>>>
>>
>

Mime
View raw message