hadoop-common-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Grady Laksmono (JIRA)" <j...@apache.org>
Subject [jira] Commented: (HADOOP-5342) DataNodes do not start up because InconsistentFSStateException on just part of the disks in use
Date Mon, 09 Aug 2010 04:40:22 GMT

    [ https://issues.apache.org/jira/browse/HADOOP-5342?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12896450#action_12896450
] 

Grady Laksmono commented on HADOOP-5342:
----------------------------------------

I'm also experiencing this issue, is there a quick solution for now? This happened to me once
before, then I reformat the HDFS and it works fine, but this happened again, but this time
I have files on my HDFS..


10/08/08 21:33:01 INFO common.Storage: Storage directory /tmp/hadoop-grady/dfs/name does not
exist.
10/08/08 21:33:01 ERROR namenode.FSNamesystem: FSNamesystem initialization failed.
org.apache.hadoop.hdfs.server.common.InconsistentFSStateException: Directory /tmp/hadoop-grady/dfs/name
is in an inconsistent state: storage directory does not exist or is not accessible.
	at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:290)
	at org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:87)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:311)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:292)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:201)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:279)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:956)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:965)
10/08/08 21:33:01 INFO ipc.Server: Stopping server on 9000
10/08/08 21:33:01 ERROR namenode.NameNode: org.apache.hadoop.hdfs.server.common.InconsistentFSStateException:
Directory /tmp/hadoop-grady/dfs/name is in an inconsistent state: storage directory does not
exist or is not accessible.
	at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:290)
	at org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:87)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:311)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:292)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:201)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:279)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:956)
	at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:965)


> DataNodes do not start up because InconsistentFSStateException on just part of the disks
in use
> -----------------------------------------------------------------------------------------------
>
>                 Key: HADOOP-5342
>                 URL: https://issues.apache.org/jira/browse/HADOOP-5342
>             Project: Hadoop Common
>          Issue Type: Bug
>    Affects Versions: 0.18.2
>            Reporter: Christian Kunz
>            Assignee: Hairong Kuang
>            Priority: Critical
>
> After restarting a cluster (including rebooting) the dfs got corrupted because many DataNodes
did not start up, running into the following exception:
> 2009-02-26 22:33:53,774 ERROR org.apache.hadoop.dfs.DataNode: org.apache.hadoop.dfs.InconsistentFSStateException:
Directory xxx  is in an inconsistent state: version file in current directory is missing.
> 	at org.apache.hadoop.dfs.Storage$StorageDirectory.analyzeStorage(Storage.java:326)
> 	at org.apache.hadoop.dfs.DataStorage.recoverTransitionRead(DataStorage.java:105)
> 	at org.apache.hadoop.dfs.DataNode.startDataNode(DataNode.java:306)
> 	at org.apache.hadoop.dfs.DataNode.<init>(DataNode.java:223)
> 	at org.apache.hadoop.dfs.DataNode.makeInstance(DataNode.java:3030)
> 	at org.apache.hadoop.dfs.DataNode.instantiateDataNode(DataNode.java:2985)
> 	at org.apache.hadoop.dfs.DataNode.createDataNode(DataNode.java:2993)
> 	at org.apache.hadoop.dfs.DataNode.main(DataNode.java:3115)
> This happens when using multiple disks with at least one previously marked as read-only,
such that the storage version became out-dated, but after reboot it was mounted read-write,
resulting in the DataNode not starting because of out-dated version.
> This is a big headache. If a DataNode has multiple disks of which at least one has the
correct storage version then out-dated versions should not bring down the DataNode.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


Mime
View raw message