hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Harsh J <ha...@cloudera.com>
Subject Re: ERROR namenode.NameNode: java.io.IOException: Cannot remove current directory: /var/lib/hadoop-0.20/cache/hadoop/dfs/name/current
Date Wed, 01 Feb 2012 20:28:50 GMT
Vijay,

[Moving to cdh-user@cloudera.org |
https://groups.google.com/a/cloudera.org/group/cdh-user/topics since
this is CDH3 specific]

You need to run that command as the 'hdfs' user, since the specific
dirs are writable only by the group 'hadoop':
$ sudo -u hdfs hadoop namenode -format

On Thu, Feb 2, 2012 at 1:27 AM, Vijayakumar Ramdoss
<nellaivijay@gmail.com> wrote:
> Hi All,
> I am trying to start the Namenode from my machine, Its throwing the error
> message,
>                                 *ERROR namenode.NameNode:
> java.io.IOException: Cannot remove current directory:
> /var/lib/hadoop-0.20/cache/hadoop/dfs/name/current*
>
> Please refer the log information from here,
> vijayram@ubuntu:/etc$ hadoop namenode -format
> 12/02/01 14:07:48 INFO namenode.NameNode: STARTUP_MSG:
> /************************************************************
> STARTUP_MSG: Starting NameNode
> STARTUP_MSG:   host = ubuntu/127.0.1.1
> STARTUP_MSG:   args = [-format]
> STARTUP_MSG:   version = 0.20.2-cdh3u3
> STARTUP_MSG:   build =
> file:///data/1/tmp/nightly_2012-01-26_09-40-25_3/hadoop-0.20-0.20.2+923.194-1~squeeze
> -r 03b655719d13929bd68bb2c2f9cee615b389cea9; compiled by 'root' on Thu Jan
> 26 11:54:44 PST 2012
> ************************************************************/
> Re-format filesystem in /var/lib/hadoop-0.20/cache/hadoop/dfs/name ? (Y or
> N) Y
> 12/02/01 14:08:10 INFO util.GSet: VM type       = 64-bit
> 12/02/01 14:08:10 INFO util.GSet: 2% max memory = 17.77875 MB
> 12/02/01 14:08:10 INFO util.GSet: capacity      = 2^21 = 2097152 entries
> 12/02/01 14:08:10 INFO util.GSet: recommended=2097152, actual=2097152
> 12/02/01 14:08:10 INFO security.UserGroupInformation: JAAS Configuration
> already set up for Hadoop, not re-installing.
> 12/02/01 14:08:10 INFO namenode.FSNamesystem: fsOwner=vijayram (auth:SIMPLE)
> 12/02/01 14:08:10 INFO namenode.FSNamesystem: supergroup=supergroup
> 12/02/01 14:08:10 INFO namenode.FSNamesystem: isPermissionEnabled=false
> 12/02/01 14:08:10 INFO namenode.FSNamesystem:
> dfs.block.invalidate.limit=1000
> 12/02/01 14:08:10 INFO namenode.FSNamesystem: isAccessTokenEnabled=false
> accessKeyUpdateInterval=0 min(s), accessTokenLifetime=0 min(s)
> 12/02/01 14:08:10 ERROR namenode.NameNode: java.io.IOException: Cannot
> remove current directory: /var/lib/hadoop-0.20/cache/hadoop/dfs/name/current
>    at
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:292)
>    at
> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1246)
>    at
> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1265)
>    at
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1127)
>    at
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1244)
>    at
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1260)
>
>
> Thanks and Regards
> Vijay
>
> nellaivijay@gmail.com



-- 
Harsh J
Customer Ops. Engineer
Cloudera | http://tiny.cloudera.com/about

Mime
View raw message