hadoop-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Balaji Narayanan (பாலாஜி நாராயணன்) <li...@balajin.net>
Subject Re: Namenode shutting down while creating cluster
Date Sat, 20 Oct 2012 06:12:38 GMT
Seems like an issue with resolution of sk.r252.0. Can you ensure that it
resolves?

On Friday, October 19, 2012, Sundeep Kambhmapati wrote:

> Hi Users,
> My name node is shutting down soon after it is started.
> Here the log. Can some one please help me.
>
> 2012-10-19 23:20:42,143 INFO
> org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:
> /************************************************************
> STARTUP_MSG: Starting NameNode
> STARTUP_MSG:   host = sk.r252.0/10.0.2.15
> STARTUP_MSG:   args = []
> STARTUP_MSG:   version = 0.20.2
> STARTUP_MSG:   build =
> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r
> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010
> ************************************************************/
> 2012-10-19 23:20:42,732 INFO org.apache.hadoop.ipc.metrics.RpcMetrics:
> Initializing RPC Metrics with hostName=NameNode, port=54310
> 2012-10-19 23:20:42,741 INFO
> org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at: sk.r252.0/
> 10.0.2.15:54310
> 2012-10-19 23:20:42,745 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
> Initializing JVM Metrics with processName=NameNode, sessionId=null
> 2012-10-19 23:20:42,747 INFO
> org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics:
> Initializing NameNodeMeterics using context
> object:org.apache.hadoop.metrics.spi.NullContext
> 2012-10-19 23:20:43,074 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
> fsOwner=root,root,bin,daemon,sys,adm,disk,wheel
> 2012-10-19 23:20:43,077 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup
> 2012-10-19 23:20:43,077 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
> isPermissionEnabled=true
> 2012-10-19 23:20:43,231 INFO
> org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics:
> Initializing FSNamesystemMetrics using context
> object:org.apache.hadoop.metrics.spi.NullContext
> 2012-10-19 23:20:43,239 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
> FSNamesystemStatusMBean
> 2012-10-19 23:20:43,359 INFO org.apache.hadoop.hdfs.server.common.Storage:
> Number of files = 1
> 2012-10-19 23:20:43,379 INFO org.apache.hadoop.hdfs.server.common.Storage:
> Number of files under construction = 0
> 2012-10-19 23:20:43,379 INFO org.apache.hadoop.hdfs.server.common.Storage:
> Image file of size 94 loaded in 0 seconds.
> 2012-10-19 23:20:43,380 INFO org.apache.hadoop.hdfs.server.common.Storage:
> Edits file /app/hadoop/tmp/dfs/name/current/edits of size 4 edits # 0
> loaded in 0 seconds.
> 2012-10-19 23:20:43,415 INFO org.apache.hadoop.hdfs.server.common.Storage:
> Image file of size 94 saved in 0 seconds.
> 2012-10-19 23:20:43,612 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished loading
> FSImage in 758 msecs
> 2012-10-19 23:20:43,615 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number of blocks
> = 0
> 2012-10-19 23:20:43,615 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of invalid
> blocks = 0
> 2012-10-19 23:20:43,615 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
> under-replicated blocks = 0
> 2012-10-19 23:20:43,615 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
>  over-replicated blocks = 0
> 2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.StateChange: STATE*
> Leaving safe mode after 0 secs.
> 2012-10-19 23:20:43,616 INFO org.apache.hadoop.hdfs.StateChange: STATE*
> Network topology has 0 racks and 0 datanodes
> 2012-10-19 23:20:43,616 INFO org.apache.hadoop.hdfs.StateChange: STATE*
> UnderReplicatedBlocks has 0 blocks
> 2012-10-19 23:20:44,450 INFO org.mortbay.log: Logging to
> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
> org.mortbay.log.Slf4jLog
> 2012-10-19 23:20:44,711 INFO org.apache.hadoop.http.HttpServer: Port
> returned by webServer.getConnectors()[0].getLocalPort() before open() is
> -1. Opening the listener on 50070
> 2012-10-19 23:20:44,715 INFO org.apache.hadoop.http.HttpServer:
> listener.getLocalPort() returned 50070
> webServer.getConnectors()[0].getLocalPort() returned 50070
> 2012-10-19 23:20:44,715 INFO org.apache.hadoop.http.HttpServer: Jetty
> bound to port 50070
> 2012-10-19 23:20:44,715 INFO org.mortbay.log: jetty-6.1.14
> 2012-10-19 23:20:47,021 INFO org.mortbay.log: Started
> SelectChannelConnector@0.0.0.0:50070
> 2012-10-19 23:20:47,022 INFO
> org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at:
> 0.0.0.0:50070
> 2012-10-19 23:20:47,022 INFO
> org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at:
> 0.0.0.0:50070
> 2012-10-19 23:20:47,067 INFO org.apache.hadoop.ipc.Server: IPC Server
> listener on 54310: starting
> 2012-10-19 23:20:47,086 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 0 on 54310: starting
> 2012-10-19 23:20:47,089 INFO org.apache.hadoop.ipc.Server: IPC Server
> Responder: starting
> 2012-10-19 23:20:47,106 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 1 on 54310: starting
> 2012-10-19 23:20:47,130 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 2 on 54310: starting
> 2012-10-19 23:20:47,148 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 3 on 54310: starting
> 2012-10-19 23:20:47,165 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 4 on 54310: starting
> 2012-10-19 23:20:47,183 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 5 on 54310: starting
> 2012-10-19 23:20:47,200 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 6 on 54310: starting
> 2012-10-19 23:20:47,803 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 9 on 54310: starting
> 2012-10-19 23:20:47,804 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 7 on 54310: starting
> 2012-10-19 23:20:47,806 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 8 on 54310: starting
> 2012-10-19 23:20:48,685 WARN
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: ReplicationMonitor
> thread received InterruptedException.java.lang.InterruptedException: sleep
> interrupted
> 2012-10-19 23:20:48,691 INFO
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
> transactions: 0 Total time for transactions(ms): 0Number of transactions
> batched in Syncs: 0 Number of syncs: 0 SyncTimes(ms): 0
> 2012-10-19 23:20:48,690 INFO
> org.apache.hadoop.hdfs.server.namenode.DecommissionManager: Interrupted
> Monitor
> java.lang.InterruptedException: sleep interrupted
>         at java.lang.Thread.sleep(Native Method)
>         at
> org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor.run(DecommissionManager.java:65)
>         at java.lang.Thread.run(Thread.java:636)
> 2012-10-19 23:20:48,771 INFO org.apache.hadoop.ipc.Server: Stopping server
> on 54310
> 2012-10-19 23:20:48,775 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 0 on 54310: exiting
> 2012-10-19 23:20:48,780 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 1 on 54310: exiting
> 2012-10-19 23:20:48,781 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 2 on 54310: exiting
> 2012-10-19 23:20:48,782 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 3 on 54310: exiting
> 2012-10-19 23:20:48,783 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 4 on 54310: exiting
> 2012-10-19 23:20:48,784 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 5 on 54310: exiting
> 2012-10-19 23:20:48,784 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 6 on 54310: exiting
> 2012-10-19 23:20:48,785 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 7 on 54310: exiting
> 2012-10-19 23:20:48,785 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 8 on 54310: exiting
> 2012-10-19 23:20:48,786 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 9 on 54310: exiting
> 2012-10-19 23:20:48,786 INFO org.apache.hadoop.ipc.Server: Stopping IPC
> Server listener on 54310
> 2012-10-19 23:20:48,788 INFO org.apache.hadoop.ipc.Server: Stopping IPC
> Server Responder
> 2012-10-19 23:20:48,790 ERROR
> org.apache.hadoop.hdfs.server.namenode.NameNode: java.io.IOException:
> Incomplete HDFS URI, no host: hdfs://sk.r252.0:54310
>         at
> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:78)
>         at
> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378)
>         at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
>         at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390)
>         at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196)
>         at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:95)
>          at org.apache.hadoop.fs.Trash.<init>(Trash.java:62)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.startTrashEmptier(NameNode.java:208)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:204)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:279)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:956)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:965)
>
> 2012-10-19 23:20:48,995 INFO
> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
> /************************************************************
> SHUTDOWN_MSG: Shutting down NameNode at sk.r252.0/10.0.2.15
>
> ***core-site.xml***
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
> <property>
>   <name>hadoop.tmp.dir</name>
>   <value>/app/hadoop/tmp</value>
>   <description>A base for other temporary directories.</description>
> </property>
>
> <property>
>   <name>fs.default.name</name>
>   <value>hdfs://sk.r252.0:54310</value>
>   <description>The name of the default file system.  A URI whose
>   scheme and authority determine the FileSystem implementation.  The
>   uri's scheme determines the config property (fs.SCHEME.impl) naming
>   the FileSystem implementation class.  The uri's authority is used to
>   determine the host, port, etc. for a filesystem.</description>
> </property>
> </configuration>
>
> ***mapred-site.xml***
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
> <property>
>   <name>mapred.job.tracker</name>
>   <value>sk.r252.0:54311</value>
>   <description>The host and port that the MapReduce job tracker runs
>   at.  If "local", then jobs are run in-process as a single map
>   and reduce task.
>   </description>
> </property>
> </configuration>
>
> ***hdfs-site.xml***
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
> <property>
>   <name>dfs.replication</name>
>   <value>2</value>
>   <description>Default block replication.
>   The actual number of replications can be specified when the file is
> created.
>   The default is used if replication is not specified in create time.
>   </description>
> </property>
> <property>
>     <name>dfs.http.address</name>
>     <value>0.0.0.0:50070</value>
>   </property>
> </configuration>
>
> Can some one please help me.
>
> Regards
> Sundeep
>
>

-- 
Thanks
-balaji

--
http://balajin.net/blog/
http://flic.kr/balajijegan

Mime
View raw message