hadoop-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Sundeep Kambhmapati <ksundeepsa...@yahoo.co.in>
Subject Re: Namenode shutting down while creating cluster
Date Sat, 20 Oct 2012 13:15:56 GMT
Thank  You Balaji,
I checked gethostbyname(sk.r252.0) it gives 10.0.2.15. This is ipaddress i am getting in ifconfig
also.
ssh sk.r252.0 is sshing to 10.0.2.15
ping sk.r252.0 is pinging to 10.0.2.15.

Can you please help me with the issue?

Regards
Sundeep



________________________________
 From: Balaji Narayanan (பாலாஜி நாராயணன்) <lists@balajin.net>
To: "user@hadoop.apache.org" <user@hadoop.apache.org>; Sundeep Kambhmapati <ksundeepsatya@yahoo.co.in>

Sent: Saturday, 20 October 2012 2:12 AM
Subject: Re: Namenode shutting down while creating cluster
 

Seems like an issue with resolution of sk.r252.0. Can you ensure that it resolves?

On Friday, October 19, 2012, Sundeep Kambhmapati  wrote:

Hi Users,
>My name node is shutting down soon after it is started. 
>Here the log. Can some one please help me.
>
>
>2012-10-19 23:20:42,143 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:
>/************************************************************
>STARTUP_MSG: Starting NameNode
>STARTUP_MSG:   host = sk.r252.0/10.0.2.15
>STARTUP_MSG:   args = []
>STARTUP_MSG:   version = 0.20.2
>STARTUP_MSG:   build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20
-r 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010
>************************************************************/
>2012-10-19 23:20:42,732 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: Initializing RPC
Metrics with hostName=NameNode, port=54310
>2012-10-19 23:20:42,741 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode
up at: sk.r252.0/10.0.2.15:54310
>2012-10-19 23:20:42,745 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: Initializing JVM
Metrics with processName=NameNode, sessionId=null
>2012-10-19 23:20:42,747 INFO org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics:
Initializing NameNodeMeterics using context object:org.apache.hadoop.metrics.spi.NullContext
>2012-10-19 23:20:43,074 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner=root,root,bin,daemon,sys,adm,disk,wheel
>2012-10-19 23:20:43,077 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup
>2012-10-19 23:20:43,077 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: isPermissionEnabled=true
>2012-10-19 23:20:43,231 INFO org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics:
Initializing FSNamesystemMetrics using context object:org.apache.hadoop.metrics.spi.NullContext
>2012-10-19 23:20:43,239 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
FSNamesystemStatusMBean
>2012-10-19 23:20:43,359 INFO org.apache.hadoop.hdfs.server.common.Storage: Number of files
= 1
>2012-10-19 23:20:43,379 INFO org.apache.hadoop.hdfs.server.common.Storage: Number of files
under construction = 0
>2012-10-19 23:20:43,379 INFO org.apache.hadoop.hdfs.server.common.Storage: Image file
of size 94 loaded in 0 seconds.
>2012-10-19 23:20:43,380 INFO org.apache.hadoop.hdfs.server.common.Storage: Edits file
/app/hadoop/tmp/dfs/name/current/edits of size 4 edits # 0 loaded in 0 seconds.
>2012-10-19 23:20:43,415 INFO org.apache.hadoop.hdfs.server.common.Storage: Image file
of size 94 saved in 0 seconds.
>2012-10-19 23:20:43,612 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished
loading FSImage in 758 msecs
>2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total
number of blocks = 0
>2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number
of invalid blocks = 0
>2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number
of under-replicated blocks = 0
>2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number
of  over-replicated blocks = 0
>2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.StateChange: STATE* Leaving safe mode
after 0 secs.
>2012-10-19 23:20:43,616 INFO org.apache.hadoop.hdfs.StateChange: STATE* Network topology
has 0 racks and 0 datanodes
>2012-10-19 23:20:43,616 INFO org.apache.hadoop.hdfs.StateChange: STATE* UnderReplicatedBlocks
has 0 blocks
>2012-10-19 23:20:44,450 INFO org.mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log)
via org.mortbay.log.Slf4jLog
>2012-10-19 23:20:44,711 INFO org.apache.hadoop.http.HttpServer: Port returned by webServer.getConnectors()[0].getLocalPort()
before open() is -1. Opening the listener on 50070
>2012-10-19 23:20:44,715 INFO org.apache.hadoop.http.HttpServer: listener.getLocalPort()
returned 50070 webServer.getConnectors()[0].getLocalPort() returned 50070
>2012-10-19 23:20:44,715 INFO org.apache.hadoop.http.HttpServer: Jetty bound to port 50070
>2012-10-19 23:20:44,715 INFO org.mortbay.log: jetty-6.1.14
>2012-10-19 23:20:47,021 INFO org.mortbay.log: Started SelectChannelConnector@0.0.0.0:50070
>2012-10-19 23:20:47,022 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server
up at: 0.0.0.0:50070
>2012-10-19 23:20:47,022 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server
up at: 0.0.0.0:50070
>2012-10-19 23:20:47,067 INFO org.apache.hadoop.ipc.Server: IPC Server listener on 54310:
starting
>2012-10-19 23:20:47,086 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 on 54310:
starting
>2012-10-19 23:20:47,089 INFO org.apache.hadoop.ipc.Server: IPC Server Responder: starting
>2012-10-19 23:20:47,106 INFO org.apache.hadoop.ipc.Server: IPC Server handler 1 on 54310:
starting
>2012-10-19 23:20:47,130 INFO org.apache.hadoop.ipc.Server: IPC Server handler 2 on 54310:
starting
>2012-10-19 23:20:47,148 INFO org.apache.hadoop.ipc.Server: IPC Server handler 3 on 54310:
starting
>2012-10-19 23:20:47,165 INFO org.apache.hadoop.ipc.Server: IPC Server handler 4 on 54310:
starting
>2012-10-19 23:20:47,183 INFO org.apache.hadoop.ipc.Server: IPC Server handler 5 on 54310:
starting
>2012-10-19 23:20:47,200 INFO org.apache.hadoop.ipc.Server: IPC Server handler 6 on 54310:
starting
>2012-10-19 23:20:47,803 INFO org.apache.hadoop.ipc.Server: IPC Server handler 9 on 54310:
starting
>2012-10-19 23:20:47,804 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 on 54310:
starting
>2012-10-19 23:20:47,806 INFO org.apache.hadoop.ipc.Server: IPC Server handler 8 on 54310:
starting
>2012-10-19 23:20:48,685WARN org.apache.hadoop.hdfs.server.namenode.FSNamesystem: ReplicationMonitor
thread received InterruptedException.java.lang.InterruptedException: sleep interrupted
>2012-10-19 23:20:48,691 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number
of transactions: 0 Total time for transactions(ms): 0Number of transactions batched in Syncs:
0 Number of syncs: 0 SyncTimes(ms): 0
>2012-10-19 23:20:48,690 INFO org.apache.hadoop.hdfs.server.namenode.DecommissionManager:
Interrupted Monitor
>java.lang.InterruptedException: sleep interrupted
>        at java.lang.Thread.sleep(Native Method)
>        at org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor.run(DecommissionManager.java:65)
>        at java.lang.Thread.run(Thread.java:636)
>2012-10-19 23:20:48,771 INFO org.apache.hadoop.ipc.Server: Stopping server on 54310
>2012-10-19 23:20:48,775 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 on 54310:
exiting
>2012-10-19 23:20:48,780 INFO org.apache.hadoop.ipc.Server: IPC Server handler 1 on 54310:
exiting
>2012-10-19 23:20:48,781 INFO org.apache.hadoop.ipc.Server: IPC Server handler 2 on 54310:
exiting
>2012-10-19 23:20:48,782 INFO org.apache.hadoop.ipc.Server: IPC Server handler 3 on 54310:
exiting
>2012-10-19 23:20:48,783 INFO org.apache.hadoop.ipc.Server: IPC Server handler 4 on 54310:
exiting
>2012-10-19 23:20:48,784 INFO org.apache.hadoop.ipc.Server: IPC Server handler 5 on 54310:
exiting
>2012-10-19 23:20:48,784 INFO org.apache.hadoop.ipc.Server: IPC Server handler 6 on 54310:
exiting
>2012-10-19 23:20:48,785 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 on 54310:
exiting
>2012-10-19 23:20:48,785 INFO org.apache.hadoop.ipc.Server: IPC Server handler 8 on 54310:
exiting
>2012-10-19 23:20:48,786 INFO org.apache.hadoop.ipc.Server: IPC Server handler 9 on 54310:
exiting
>2012-10-19 23:20:48,786 INFO org.apache.hadoop.ipc.Server: Stopping IPC Server listener
on 54310
>2012-10-19 23:20:48,788 INFO org.apache.hadoop.ipc.Server: Stopping IPC Server Responder
>2012-10-19 23:20:48,790 ERROR org.apache.hadoop.hdfs.server.namenode.NameNode: java.io.IOException:
Incomplete HDFS URI, no host: hdfs://sk.r252.0:54310
>        at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:78)
>        at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378)
>        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
>        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390)
>        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196)
>        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:95)
>         at org.apache.hadoop.fs.Trash.<init>(Trash.java:62)
>        at org.apache.hadoop.hdfs.server.namenode.NameNode.startTrashEmptier(NameNode.java:208)
>        at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:204)
>        at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:279)
>        at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:956)
>        at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:965)
>
>
>2012-10-19 23:20:48,995 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>/************************************************************
>SHUTDOWN_MSG: Shutting down NameNode at sk.r252.0/10.0.2.15
>
>
>***core-site.xml***
><?xml version="1.0"?>
><?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
>
><!-- Put site-specific property overrides in this file. -->
>
>
><configuration>
><property>
>  <name>hadoop.tmp.dir</name>
>  <value>/app/hadoop/tmp</value>
>  <description>A base for other temporary directories.</description>
></property>
>
>
><property>
>  <name>fs.default.name</name>
>  <value>hdfs://sk.r252.0:54310</value>
>  <description>The name of the default file system.  A URI whose
>  scheme and authority determine the FileSystem implementation.  The
>  uri's scheme determines the config property (fs.SCHEME.impl) naming
>  the FileSystem implementation class.  The uri's authority is used to
>  determine the host, port, etc. for a filesystem.</description>
></property>
></configuration>
>
>
>***mapred-site.xml***
><?xml version="1.0"?>
><?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
>
><!-- Put site-specific property overrides in this file. -->
>
>
><configuration>
><property>
>  <name>mapred.job.tracker</name>
>  <value>sk.r252.0:54311</value>
>  <description>The host and port that the MapReduce job tracker runs
>  at.  If "local", then jobs are run in-process as a single map
>  and reduce task.
>  </description>
></property>
></configuration>
>
>
>***hdfs-site.xml***
><?xml version="1.0"?>
><?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
>
><!-- Put site-specific property overrides in this file. -->
>
>
><configuration>
><property>
>  <name>dfs.replication</name>
>  <value>2</value>
>  <description>Default block replication.
>  The actual number of replications can be specified when the file is created.
>  The default is used if replication is not specified in create time.
>  </description>
></property>
><property>
>    <name>dfs.http.address</name>
>    <value>0.0.0.0:50070</value>
>  </property>
></configuration>
>
>
>Can some one please help me.
>
>
>Regards 
>Sundeep
>
>

-- 
Thanks
-balaji
--
http://balajin.net/blog/
http://flic.kr/balajijegan
Mime
View raw message