hadoop-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Sundeep Kambhmapati <ksundeepsa...@yahoo.co.in>
Subject Namenode shutting down while creating cluster
Date Sat, 20 Oct 2012 03:33:51 GMT
Hi Users,
My name node is shutting down soon after it is started. 
Here the log. Can some one please help me.

2012-10-19 23:20:42,143 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting NameNode
STARTUP_MSG:   host = sk.r252.0/10.0.2.15
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 0.20.2
STARTUP_MSG:   build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20
-r 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010
************************************************************/
2012-10-19 23:20:42,732 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: Initializing RPC Metrics
with hostName=NameNode, port=54310
2012-10-19 23:20:42,741 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up
at: sk.r252.0/10.0.2.15:54310
2012-10-19 23:20:42,745 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: Initializing JVM Metrics
with processName=NameNode, sessionId=null
2012-10-19 23:20:42,747 INFO org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics:
Initializing NameNodeMeterics using context object:org.apache.hadoop.metrics.spi.NullContext
2012-10-19 23:20:43,074 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner=root,root,bin,daemon,sys,adm,disk,wheel
2012-10-19 23:20:43,077 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup
2012-10-19 23:20:43,077 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: isPermissionEnabled=true
2012-10-19 23:20:43,231 INFO org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics:
Initializing FSNamesystemMetrics using context object:org.apache.hadoop.metrics.spi.NullContext
2012-10-19 23:20:43,239 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
FSNamesystemStatusMBean
2012-10-19 23:20:43,359 INFO org.apache.hadoop.hdfs.server.common.Storage: Number of files
= 1
2012-10-19 23:20:43,379 INFO org.apache.hadoop.hdfs.server.common.Storage: Number of files
under construction = 0
2012-10-19 23:20:43,379 INFO org.apache.hadoop.hdfs.server.common.Storage: Image file of size
94 loaded in 0 seconds.
2012-10-19 23:20:43,380 INFO org.apache.hadoop.hdfs.server.common.Storage: Edits file /app/hadoop/tmp/dfs/name/current/edits
of size 4 edits # 0 loaded in 0 seconds.
2012-10-19 23:20:43,415 INFO org.apache.hadoop.hdfs.server.common.Storage: Image file of size
94 saved in 0 seconds.
2012-10-19 23:20:43,612 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished
loading FSImage in 758 msecs
2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number
of blocks = 0
2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
invalid blocks = 0
2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
under-replicated blocks = 0
2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
 over-replicated blocks = 0
2012-10-19 23:20:43,615 INFO org.apache.hadoop.hdfs.StateChange: STATE* Leaving safe mode
after 0 secs.
2012-10-19 23:20:43,616 INFO org.apache.hadoop.hdfs.StateChange: STATE* Network topology has
0 racks and 0 datanodes
2012-10-19 23:20:43,616 INFO org.apache.hadoop.hdfs.StateChange: STATE* UnderReplicatedBlocks
has 0 blocks
2012-10-19 23:20:44,450 INFO org.mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log)
via org.mortbay.log.Slf4jLog
2012-10-19 23:20:44,711 INFO org.apache.hadoop.http.HttpServer: Port returned by webServer.getConnectors()[0].getLocalPort()
before open() is -1. Opening the listener on 50070
2012-10-19 23:20:44,715 INFO org.apache.hadoop.http.HttpServer: listener.getLocalPort() returned
50070 webServer.getConnectors()[0].getLocalPort() returned 50070
2012-10-19 23:20:44,715 INFO org.apache.hadoop.http.HttpServer: Jetty bound to port 50070
2012-10-19 23:20:44,715 INFO org.mortbay.log: jetty-6.1.14
2012-10-19 23:20:47,021 INFO org.mortbay.log: Started SelectChannelConnector@0.0.0.0:50070
2012-10-19 23:20:47,022 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up
at: 0.0.0.0:50070
2012-10-19 23:20:47,022 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up
at: 0.0.0.0:50070
2012-10-19 23:20:47,067 INFO org.apache.hadoop.ipc.Server: IPC Server listener on 54310: starting
2012-10-19 23:20:47,086 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 on 54310:
starting
2012-10-19 23:20:47,089 INFO org.apache.hadoop.ipc.Server: IPC Server Responder: starting
2012-10-19 23:20:47,106 INFO org.apache.hadoop.ipc.Server: IPC Server handler 1 on 54310:
starting
2012-10-19 23:20:47,130 INFO org.apache.hadoop.ipc.Server: IPC Server handler 2 on 54310:
starting
2012-10-19 23:20:47,148 INFO org.apache.hadoop.ipc.Server: IPC Server handler 3 on 54310:
starting
2012-10-19 23:20:47,165 INFO org.apache.hadoop.ipc.Server: IPC Server handler 4 on 54310:
starting
2012-10-19 23:20:47,183 INFO org.apache.hadoop.ipc.Server: IPC Server handler 5 on 54310:
starting
2012-10-19 23:20:47,200 INFO org.apache.hadoop.ipc.Server: IPC Server handler 6 on 54310:
starting
2012-10-19 23:20:47,803 INFO org.apache.hadoop.ipc.Server: IPC Server handler 9 on 54310:
starting
2012-10-19 23:20:47,804 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 on 54310:
starting
2012-10-19 23:20:47,806 INFO org.apache.hadoop.ipc.Server: IPC Server handler 8 on 54310:
starting
2012-10-19 23:20:48,685WARN org.apache.hadoop.hdfs.server.namenode.FSNamesystem: ReplicationMonitor
thread received InterruptedException.java.lang.InterruptedException: sleep interrupted
2012-10-19 23:20:48,691 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
transactions: 0 Total time for transactions(ms): 0Number of transactions batched in Syncs:
0 Number of syncs: 0 SyncTimes(ms): 0
2012-10-19 23:20:48,690 INFO org.apache.hadoop.hdfs.server.namenode.DecommissionManager: Interrupted
Monitor
java.lang.InterruptedException: sleep interrupted
        at java.lang.Thread.sleep(Native Method)
        at org.apache.hadoop.hdfs.server.namenode.DecommissionManager$Monitor.run(DecommissionManager.java:65)
        at java.lang.Thread.run(Thread.java:636)
2012-10-19 23:20:48,771 INFO org.apache.hadoop.ipc.Server: Stopping server on 54310
2012-10-19 23:20:48,775 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 on 54310:
exiting
2012-10-19 23:20:48,780 INFO org.apache.hadoop.ipc.Server: IPC Server handler 1 on 54310:
exiting
2012-10-19 23:20:48,781 INFO org.apache.hadoop.ipc.Server: IPC Server handler 2 on 54310:
exiting
2012-10-19 23:20:48,782 INFO org.apache.hadoop.ipc.Server: IPC Server handler 3 on 54310:
exiting
2012-10-19 23:20:48,783 INFO org.apache.hadoop.ipc.Server: IPC Server handler 4 on 54310:
exiting
2012-10-19 23:20:48,784 INFO org.apache.hadoop.ipc.Server: IPC Server handler 5 on 54310:
exiting
2012-10-19 23:20:48,784 INFO org.apache.hadoop.ipc.Server: IPC Server handler 6 on 54310:
exiting
2012-10-19 23:20:48,785 INFO org.apache.hadoop.ipc.Server: IPC Server handler 7 on 54310:
exiting
2012-10-19 23:20:48,785 INFO org.apache.hadoop.ipc.Server: IPC Server handler 8 on 54310:
exiting
2012-10-19 23:20:48,786 INFO org.apache.hadoop.ipc.Server: IPC Server handler 9 on 54310:
exiting
2012-10-19 23:20:48,786 INFO org.apache.hadoop.ipc.Server: Stopping IPC Server listener on
54310
2012-10-19 23:20:48,788 INFO org.apache.hadoop.ipc.Server: Stopping IPC Server Responder
2012-10-19 23:20:48,790 ERROR org.apache.hadoop.hdfs.server.namenode.NameNode: java.io.IOException:
Incomplete HDFS URI, no host: hdfs://sk.r252.0:54310
        at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:78)
        at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378)
        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390)
        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196)
        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:95)
         at org.apache.hadoop.fs.Trash.<init>(Trash.java:62)
        at org.apache.hadoop.hdfs.server.namenode.NameNode.startTrashEmptier(NameNode.java:208)
        at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:204)
        at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:279)
        at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:956)
        at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:965)

2012-10-19 23:20:48,995 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at sk.r252.0/10.0.2.15

***core-site.xml***
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>
<property>
  <name>hadoop.tmp.dir</name>
  <value>/app/hadoop/tmp</value>
  <description>A base for other temporary directories.</description>
</property>

<property>
  <name>fs.default.name</name>
  <value>hdfs://sk.r252.0:54310</value>
  <description>The name of the default file system.  A URI whose
  scheme and authority determine the FileSystem implementation.  The
  uri's scheme determines the config property (fs.SCHEME.impl) naming
  the FileSystem implementation class.  The uri's authority is used to
  determine the host, port, etc. for a filesystem.</description>
</property>
</configuration>

***mapred-site.xml***
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>
<property>
  <name>mapred.job.tracker</name>
  <value>sk.r252.0:54311</value>
  <description>The host and port that the MapReduce job tracker runs
  at.  If "local", then jobs are run in-process as a single map
  and reduce task.
  </description>
</property>
</configuration>

***hdfs-site.xml***
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<!-- Put site-specific property overrides in this file. -->

<configuration>
<property>
  <name>dfs.replication</name>
  <value>2</value>
  <description>Default block replication.
  The actual number of replications can be specified when the file is created.
  The default is used if replication is not specified in create time.
  </description>
</property>
<property>
    <name>dfs.http.address</name>
    <value>0.0.0.0:50070</value>
  </property>
</configuration>

Can some one please help me.

Regards 
Sundeep

Mime
View raw message