hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Ian jonhson <jonhson....@gmail.com>
Subject Re: Datanodes fail to start
Date Tue, 16 Jun 2009 06:57:39 GMT
If you rebuilt the hadoop, following the wikipage of HowToRelease may
reduce the trouble occurred.


On Sat, May 16, 2009 at 7:20 AM, Pankil Doshi<forpankil@gmail.com> wrote:
> I got the solution..
>
> Namespace IDs where some how incompatible.So I had to clean data dir and
> temp dir ,format the cluster and make a fresh start
>
> Pankil
>
> On Fri, May 15, 2009 at 2:25 AM, jason hadoop <jason.hadoop@gmail.com>wrote:
>
>> There should be a few more lines at the end.
>> We only want the part from last the STARTUP_MSG to the end
>>
>> On one of mine a successfull start looks like this:
>> STARTUP_MSG: Starting DataNode
>> STARTUP_MSG:   host = at/192.168.1.119
>> STARTUP_MSG:   args = []
>> STARTUP_MSG:   version = 0.19.1-dev
>> STARTUP_MSG:   build =  -r ; compiled by 'jason' on Tue Mar 17 04:03:57 PDT
>> 2009
>> ************************************************************/
>> 2009-03-17 03:08:11,884 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Registered
>> FSDatasetStatusMBean
>> 2009-03-17 03:08:11,886 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Opened info server at
>> 50010
>> 2009-03-17 03:08:11,889 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is
>> 1048576 bytes/s
>> 2009-03-17 03:08:12,142 INFO org.mortbay.http.HttpServer: Version
>> Jetty/5.1.4
>> 2009-03-17 03:08:12,155 INFO org.mortbay.util.Credential: Checking Resource
>> aliases
>> 2009-03-17 03:08:12,518 INFO org.mortbay.util.Container: Started
>> org.mortbay.jetty.servlet.WebApplicationHandler@1e184cb
>> 2009-03-17 03:08:12,578 INFO org.mortbay.util.Container: Started
>> WebApplicationContext[/static,/static]
>> 2009-03-17 03:08:12,721 INFO org.mortbay.util.Container: Started
>> org.mortbay.jetty.servlet.WebApplicationHandler@1d9e282
>> 2009-03-17 03:08:12,722 INFO org.mortbay.util.Container: Started
>> WebApplicationContext[/logs,/logs]
>> 2009-03-17 03:08:12,878 INFO org.mortbay.util.Container: Started
>> org.mortbay.jetty.servlet.WebApplicationHandler@14a75bb
>> 2009-03-17 03:08:12,884 INFO org.mortbay.util.Container: Started
>> WebApplicationContext[/,/]
>> 2009-03-17 03:08:12,951 INFO org.mortbay.http.SocketListener: Started
>> SocketListener on 0.0.0.0:50075
>> 2009-03-17 03:08:12,951 INFO org.mortbay.util.Container: Started
>> org.mortbay.jetty.Server@1358f03
>> 2009-03-17 03:08:12,957 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
>> Initializing JVM Metrics with processName=DataNode, sessionId=null
>> 2009-03-17 03:08:13,242 INFO org.apache.hadoop.ipc.metrics.RpcMetrics:
>> Initializing RPC Metrics with hostName=DataNode, port=50020
>> 2009-03-17 03:08:13,264 INFO org.apache.hadoop.ipc.Server: IPC Server
>> Responder: starting
>> 2009-03-17 03:08:13,304 INFO org.apache.hadoop.ipc.Server: IPC Server
>> listener on 50020: starting
>> 2009-03-17 03:08:13,343 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 0 on 50020: starting
>> 2009-03-17 03:08:13,343 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: dnRegistration =
>> DatanodeRegistration(192.168.1.119:50010,
>> storageID=DS-540597485-192.168.1.119-50010-1237022386925, infoPort=50075,
>> ipcPort=50020)
>> 2009-03-17 03:08:13,344 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 1 on 50020: starting
>> 2009-03-17 03:08:13,344 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 2 on 50020: starting
>> 2009-03-17 03:08:13,351 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
>> 192.168.1.119:50010,
>> storageID=DS-540597485-192.168.1.119-50010-1237022386925, infoPort=50075,
>> ipcPort=50020)In DataNode.run, data =
>> FSDataset{dirpath='/tmp/hadoop-0.19.0-jason/dfs/data/current'}
>> 2009-03-17 03:08:13,352 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: using BLOCKREPORT_INTERVAL
>> of 3600000msec Initial delay: 0msec
>> 2009-03-17 03:08:13,391 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 14 blocks
>> got processed in 27 msecs
>> 2009-03-17 03:08:13,392 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Starting Periodic block
>> scanner.
>>
>>
>>
>> On Thu, May 14, 2009 at 9:51 PM, Pankil Doshi <forpankil@gmail.com> wrote:
>>
>> > This is log from datanode.
>> >
>> >
>> > 2009-05-14 00:36:14,559 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 12 msecs
>> > 2009-05-14 01:36:15,768 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 8 msecs
>> > 2009-05-14 02:36:13,975 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 9 msecs
>> > 2009-05-14 03:36:15,189 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 12 msecs
>> > 2009-05-14 04:36:13,384 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 9 msecs
>> > 2009-05-14 05:36:14,592 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 9 msecs
>> > 2009-05-14 06:36:15,806 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 12 msecs
>> > 2009-05-14 07:36:14,008 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 12 msecs
>> > 2009-05-14 08:36:15,204 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 9 msecs
>> > 2009-05-14 09:36:13,430 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 12 msecs
>> > 2009-05-14 10:36:14,642 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 12 msecs
>> > 2009-05-14 11:36:15,850 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 9 msecs
>> > 2009-05-14 12:36:14,193 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 11 msecs
>> > 2009-05-14 13:36:15,454 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 12 msecs
>> > 2009-05-14 14:36:13,662 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 9 msecs
>> > 2009-05-14 15:36:14,930 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 13 msecs
>> > 2009-05-14 16:36:16,151 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 12 msecs
>> > 2009-05-14 17:36:14,407 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 9 msecs
>> > 2009-05-14 18:36:15,659 INFO org.apache.hadoop.dfs.DataNode: BlockReport
>> of
>> > 82 blocks got processed in 10 msecs
>> > 2009-05-14 19:27:02,188 WARN org.apache.hadoop.dfs.DataNode:
>> > java.io.IOException: Call to
>> > hadoopmaster.utdallas.edu/10.110.95.61:9000failed on local except$
>> >        at org.apache.hadoop.ipc.Client.wrapException(Client.java:751)
>> >        at org.apache.hadoop.ipc.Client.call(Client.java:719)
>> >        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216)
>> >        at org.apache.hadoop.dfs.$Proxy4.sendHeartbeat(Unknown Source)
>> >        at org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:690)
>> >        at org.apache.hadoop.dfs.DataNode.run(DataNode.java:2967)
>> >        at java.lang.Thread.run(Thread.java:619)
>> > Caused by: java.io.EOFException
>> >        at java.io.DataInputStream.readInt(DataInputStream.java:375)
>> >        at
>> > org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:500)
>> >        at org.apache.hadoop.ipc.Client$Connection.run(Client.java:442)
>> >
>> > 2009-05-14 19:27:06,198 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect
>> > to server: hadoopmaster.utdallas.edu/10.110.95.61:9000. Already tried 0
>> > time(s).
>> > 2009-05-14 19:27:06,436 INFO org.apache.hadoop.dfs.DataNode:
>> SHUTDOWN_MSG:
>> > /************************************************************
>> > SHUTDOWN_MSG: Shutting down DataNode at Slave1/127.0.1.1
>> > ************************************************************/
>> > 2009-05-14 19:27:21,737 INFO org.apache.hadoop.dfs.DataNode: STARTUP_MSG:
>> > /************************************************************
>> > STARTUP_MSG: Starting DataNode
>> > STARTUP_MSG:   host = Slave1/127.0.1.1
>> >
>> >
>> > On Thu, May 14, 2009 at 11:43 PM, jason hadoop <jason.hadoop@gmail.com
>> > >wrote:
>> >
>> > > The data node logs are on the datanode machines in the log directory.
>> > > You may wish to buy my book and read chapter 4 on hdfs management.
>> > >
>> > > On Thu, May 14, 2009 at 9:39 PM, Pankil Doshi <forpankil@gmail.com>
>> > wrote:
>> > >
>> > > > Can u guide me where can I find datanode log files? As I cannot find
>> it
>> > > in
>> > > > $hadoop/logs and so.
>> > > >
>> > > > I can only find  following files in logs folder :-
>> > > >
>> > > > hadoop-hadoop-namenode-hadoopmaster.log
>> > > >    hadoop-hadoop-namenode-hadoopmaster.out
>> > > >     hadoop-hadoop-namenode-hadoopmaster.out.1
>> > > >   hadoop-hadoop-secondarynamenode-hadoopmaster.log
>> > > >     hadoop-hadoop-secondarynamenode-hadoopmaster.out
>> > > >     hadoop-hadoop-secondarynamenode-hadoopmaster.out.1
>> > > >    history
>> > > >
>> > > >
>> > > > Thanks
>> > > > Pankil
>> > > >
>> > > > On Thu, May 14, 2009 at 11:27 PM, jason hadoop <
>> jason.hadoop@gmail.com
>> > > > >wrote:
>> > > >
>> > > > > You have to examine the datanode log files
>> > > > > the namenode does not start the datanodes, the start script does.
>> > > > > The name node passively waits for the datanodes to connect to
it.
>> > > > >
>> > > > > On Thu, May 14, 2009 at 6:43 PM, Pankil Doshi <forpankil@gmail.com
>> >
>> > > > wrote:
>> > > > >
>> > > > > > Hello Everyone,
>> > > > > >
>> > > > > > Actually I had a cluster which was up.
>> > > > > >
>> > > > > > But i stopped the cluster as i  wanted to format it.But
cant
>> start
>> > it
>> > > > > back.
>> > > > > >
>> > > > > > 1)when i give "start-dfs.sh" I get following on screen
>> > > > > >
>> > > > > > starting namenode, logging to
>> > > > > >
>> > > >
>> > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-namenode-hadoopmaster.out
>> > > > > > slave1.local: starting datanode, logging to
>> > > > > >
>> /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave1.out
>> > > > > > slave3.local: starting datanode, logging to
>> > > > > >
>> /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave3.out
>> > > > > > slave4.local: starting datanode, logging to
>> > > > > >
>> /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave4.out
>> > > > > > slave2.local: starting datanode, logging to
>> > > > > >
>> /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave2.out
>> > > > > > slave5.local: starting datanode, logging to
>> > > > > >
>> /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave5.out
>> > > > > > slave6.local: starting datanode, logging to
>> > > > > >
>> /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave6.out
>> > > > > > slave9.local: starting datanode, logging to
>> > > > > >
>> /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave9.out
>> > > > > > slave8.local: starting datanode, logging to
>> > > > > >
>> /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave8.out
>> > > > > > slave7.local: starting datanode, logging to
>> > > > > >
>> /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave7.out
>> > > > > > slave10.local: starting datanode, logging to
>> > > > > >
>> > /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-datanode-Slave10.out
>> > > > > > hadoopmaster.local: starting secondarynamenode, logging
to
>> > > > > >
>> > > > > >
>> > > > >
>> > > >
>> > >
>> >
>> /Hadoop/hadoop-0.18.3/bin/../logs/hadoop-hadoop-secondarynamenode-hadoopmaster.out
>> > > > > >
>> > > > > >
>> > > > > > 2) from log file named "hadoop-hadoop-namenode-hadoopmaster.log"
>> I
>> > > get
>> > > > > > following
>> > > > > >
>> > > > > >
>> > > > > >
>> > > > > > 2009-05-14 20:28:23,515 INFO org.apache.hadoop.dfs.NameNode:
>> > > > STARTUP_MSG:
>> > > > > > /************************************************************
>> > > > > > STARTUP_MSG: Starting NameNode
>> > > > > > STARTUP_MSG:   host = hadoopmaster/127.0.0.1
>> > > > > > STARTUP_MSG:   args = []
>> > > > > > STARTUP_MSG:   version = 0.18.3
>> > > > > > STARTUP_MSG:   build =
>> > > > > >
>> https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.18-r
>> > > > > > 736250;
>> > > > > > compiled by 'ndaley' on Thu Jan 22 23:12:08 UTC 2009
>> > > > > > ************************************************************/
>> > > > > > 2009-05-14 20:28:23,717 INFO
>> > > org.apache.hadoop.ipc.metrics.RpcMetrics:
>> > > > > > Initializing RPC Metrics with hostName=NameNode, port=9000
>> > > > > > 2009-05-14 20:28:23,728 INFO org.apache.hadoop.dfs.NameNode:
>> > Namenode
>> > > > up
>> > > > > > at:
>> > > > > > hadoopmaster.local/192.168.0.1:9000
>> > > > > > 2009-05-14 20:28:23,733 INFO
>> > > org.apache.hadoop.metrics.jvm.JvmMetrics:
>> > > > > > Initializing JVM Metrics with processName=NameNode,
>> sessionId=null
>> > > > > > 2009-05-14 20:28:23,743 INFO
>> org.apache.hadoop.dfs.NameNodeMetrics:
>> > > > > > Initializing NameNodeMeterics using context
>> > > > > > object:org.apache.hadoop.metrics.spi.NullContext
>> > > > > > 2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem:
>> > > > > >
>> > > > > >
>> > > > >
>> > > >
>> > >
>> >
>> fsOwner=hadoop,hadoop,adm,dialout,fax,cdrom,floppy,tape,audio,dip,video,plugdev,fuse,lpadmin,admin,sambashare
>> > > > > > 2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem:
>> > > > > > supergroup=supergroup
>> > > > > > 2009-05-14 20:28:23,856 INFO org.apache.hadoop.fs.FSNamesystem:
>> > > > > > isPermissionEnabled=true
>> > > > > > 2009-05-14 20:28:23,883 INFO
>> > > org.apache.hadoop.dfs.FSNamesystemMetrics:
>> > > > > > Initializing FSNamesystemMeterics using context
>> > > > > > object:org.apache.hadoop.metrics.spi.NullContext
>> > > > > > 2009-05-14 20:28:23,885 INFO org.apache.hadoop.fs.FSNamesystem:
>> > > > > Registered
>> > > > > > FSNamesystemStatusMBean
>> > > > > > 2009-05-14 20:28:23,964 INFO org.apache.hadoop.dfs.Storage:
>> Number
>> > of
>> > > > > files
>> > > > > > = 1
>> > > > > > 2009-05-14 20:28:23,971 INFO org.apache.hadoop.dfs.Storage:
>> Number
>> > of
>> > > > > files
>> > > > > > under construction = 0
>> > > > > > 2009-05-14 20:28:23,971 INFO org.apache.hadoop.dfs.Storage:
Image
>> > > file
>> > > > of
>> > > > > > size 80 loaded in 0 seconds.
>> > > > > > 2009-05-14 20:28:23,972 INFO org.apache.hadoop.dfs.Storage:
Edits
>> > > file
>> > > > > > edits
>> > > > > > of size 4 edits # 0 loaded in 0 seconds.
>> > > > > > 2009-05-14 20:28:23,974 INFO org.apache.hadoop.fs.FSNamesystem:
>> > > > Finished
>> > > > > > loading FSImage in 155 msecs
>> > > > > > 2009-05-14 20:28:23,976 INFO org.apache.hadoop.fs.FSNamesystem:
>> > Total
>> > > > > > number
>> > > > > > of blocks = 0
>> > > > > > 2009-05-14 20:28:23,988 INFO org.apache.hadoop.fs.FSNamesystem:
>> > > Number
>> > > > of
>> > > > > > invalid blocks = 0
>> > > > > > 2009-05-14 20:28:23,988 INFO org.apache.hadoop.fs.FSNamesystem:
>> > > Number
>> > > > of
>> > > > > > under-replicated blocks = 0
>> > > > > > 2009-05-14 20:28:23,988 INFO org.apache.hadoop.fs.FSNamesystem:
>> > > Number
>> > > > of
>> > > > > > over-replicated blocks = 0
>> > > > > > 2009-05-14 20:28:23,988 INFO org.apache.hadoop.dfs.StateChange:
>> > > STATE*
>> > > > > > Leaving safe mode after 0 secs.
>> > > > > > *2009-05-14 20:28:23,989 INFO org.apache.hadoop.dfs.StateChange:
>> > > STATE*
>> > > > > > Network topology has 0 racks and 0 datanodes*
>> > > > > > 2009-05-14 20:28:23,989 INFO org.apache.hadoop.dfs.StateChange:
>> > > STATE*
>> > > > > > UnderReplicatedBlocks has 0 blocks
>> > > > > > 2009-05-14 20:28:29,128 INFO org.mortbay.util.Credential:
>> Checking
>> > > > > Resource
>> > > > > > aliases
>> > > > > > 2009-05-14 20:28:29,243 INFO org.mortbay.http.HttpServer:
Version
>> > > > > > Jetty/5.1.4
>> > > > > > 2009-05-14 20:28:29,244 INFO org.mortbay.util.Container:
Started
>> > > > > > HttpContext[/static,/static]
>> > > > > > 2009-05-14 20:28:29,245 INFO org.mortbay.util.Container:
Started
>> > > > > > HttpContext[/logs,/logs]
>> > > > > > 2009-05-14 20:28:29,750 INFO org.mortbay.util.Container:
Started
>> > > > > > org.mortbay.jetty.servlet.WebApplicationHandler@7fcebc9f
>> > > > > > 2009-05-14 20:28:29,838 INFO org.mortbay.util.Container:
Started
>> > > > > > WebApplicationContext[/,/]
>> > > > > > 2009-05-14 20:28:29,843 INFO org.mortbay.http.SocketListener:
>> > Started
>> > > > > > SocketListener on 0.0.0.0:50070
>> > > > > > 2009-05-14 20:28:29,843 INFO org.mortbay.util.Container:
Started
>> > > > > > org.mortbay.jetty.Server@61acfa31
>> > > > > > 2009-05-14 20:28:29,843 INFO org.apache.hadoop.fs.FSNamesystem:
>> > > > > Web-server
>> > > > > > up at: 0.0.0.0:50070
>> > > > > > 2009-05-14 20:28:29,843 INFO org.apache.hadoop.ipc.Server:
IPC
>> > Server
>> > > > > > Responder: starting
>> > > > > > 2009-05-14 20:28:29,844 INFO org.apache.hadoop.ipc.Server:
IPC
>> > Server
>> > > > > > listener on 9000: starting
>> > > > > > 2009-05-14 20:28:29,865 INFO org.apache.hadoop.ipc.Server:
IPC
>> > Server
>> > > > > > handler 0 on 9000: starting
>> > > > > > 2009-05-14 20:28:29,876 INFO org.apache.hadoop.ipc.Server:
IPC
>> > Server
>> > > > > > handler 1 on 9000: starting
>> > > > > > 2009-05-14 20:28:29,877 INFO org.apache.hadoop.ipc.Server:
IPC
>> > Server
>> > > > > > handler 2 on 9000: starting
>> > > > > > 2009-05-14 20:28:29,877 INFO org.apache.hadoop.ipc.Server:
IPC
>> > Server
>> > > > > > handler 3 on 9000: starting
>> > > > > > 2009-05-14 20:28:29,878 INFO org.apache.hadoop.ipc.Server:
IPC
>> > Server
>> > > > > > handler 4 on 9000: starting
>> > > > > > 2009-05-14 20:28:29,879 INFO org.apache.hadoop.ipc.Server:
IPC
>> > Server
>> > > > > > handler 5 on 9000: starting
>> > > > > > 2009-05-14 20:28:29,879 INFO org.apache.hadoop.ipc.Server:
IPC
>> > Server
>> > > > > > handler 6 on 9000: starting
>> > > > > > 2009-05-14 20:28:29,881 INFO org.apache.hadoop.ipc.Server:
IPC
>> > Server
>> > > > > > handler 7 on 9000: starting
>> > > > > > 2009-05-14 20:28:29,881 INFO org.apache.hadoop.ipc.Server:
IPC
>> > Server
>> > > > > > handler 8 on 9000: starting
>> > > > > > 2009-05-14 20:28:29,882 INFO org.apache.hadoop.ipc.Server:
IPC
>> > Server
>> > > > > > handler 9 on 9000: starting
>> > > > > > 2009-05-14 20:33:35,774 INFO org.apache.hadoop.fs.FSNamesystem:
>> > Roll
>> > > > Edit
>> > > > > > Log from 192.168.0.1
>> > > > > > 2009-05-14 20:33:35,775 INFO org.apache.hadoop.fs.FSNamesystem:
>> > > Number
>> > > > of
>> > > > > > transactions: 0 Total time for transactions(ms): 0 Number
of
>> syncs:
>> > 0
>> > > > > > SyncTimes(ms): 0
>> > > > > > 2009-05-14 20:33:36,310 INFO org.apache.hadoop.fs.FSNamesystem:
>> > Roll
>> > > > > > FSImage
>> > > > > > from 192.168.0.1
>> > > > > > 2009-05-14 20:33:36,311 INFO org.apache.hadoop.fs.FSNamesystem:
>> > > Number
>> > > > of
>> > > > > > transactions: 0 Total time for transactions(ms): 0 Number
of
>> syncs:
>> > 0
>> > > > > > SyncTimes(ms): 0
>> > > > > >
>> > > > > >
>> > > > > >
>> > > > > > 3) my hadoop-site.xml for refrence
>> > > > > >
>> > > > > > <?xml version="1.0"?>
>> > > > > > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>> > > > > >
>> > > > > > <!-- Put site-specific property overrides in this file.
-->
>> > > > > > <configuration>
>> > > > > >  <property>
>> > > > > >    <name>fs.default.name</name>
>> > > > > >    <value>hdfs://hadoopmaster.local:9000</value>
>> > > > > >  </property>
>> > > > > >  <property>
>> > > > > >    <name>mapred.job.tracker</name>
>> > > > > >    <value>hadoopmaster.local:9001</value>
>> > > > > >  </property>
>> > > > > >  <property>
>> > > > > >    <name>dfs.replication</name>
>> > > > > >    <value>3</value>
>> > > > > >  </property>
>> > > > > > <property>
>> > > > > >  <name>mapred.child.java.opts</name>
>> > > > > >  <value>-Xmx512m</value>
>> > > > > > </property>
>> > > > > > <property>
>> > > > > >  <name>hadoop.tmp.dir</name>
>> > > > > >  <value>/Hadoop/Temp</value>
>> > > > > >  <description>A base for other temporary
>> directories.</description>
>> > > > > > </property>
>> > > > > > <property>
>> > > > > >  <name>dfs.data.dir</name>
>> > > > > >  <value>/Hadoop/Data,/data/Hadoop</value>
>> > > > > >  <description>Determines where on the local filesystem
an DFS
>> data
>> > > node
>> > > > > >  should store its blocks.  If this is a comma-delimited
>> > > > > >  list of directories, then data will be stored in all named
>> > > > > >  directories, typically on different devices.
>> > > > > >  Directories that do not exist are ignored.
>> > > > > >  </description>
>> > > > > > </property>
>> > > > > > </configuration>
>> > > > > >
>> > > > > >
>> > > > > > the main thing i find in log it says "*2009-05-14 20:28:23,989
>> INFO
>> > > > > > org.apache.hadoop.dfs.StateChange: STATE* Network topology
has 0
>> > > racks
>> > > > > and
>> > > > > > 0
>> > > > > > datanodes" *which means it cannot start datanodes*.*but
why is it
>> > so?
>> > > I
>> > > > > > have
>> > > > > > all my datanodes in my slave.xml and it detects that on
start up
>> > > screen
>> > > > > > when
>> > > > > > i give start-dfs.sh.
>> > > > > >
>> > > > > >
>> > > > > > Can anyone throw some lights on my problem.
>> > > > > >
>> > > > > > Thanks
>> > > > > > Pankil
>> > > > > >
>> > > > >
>> > > > >
>> > > > >
>> > > > > --
>> > > > > Alpha Chapters of my book on Hadoop are available
>> > > > > http://www.apress.com/book/view/9781430219422
>> > > > > www.prohadoopbook.com a community for Hadoop Professionals
>> > > > >
>> > > >
>> > >
>> > >
>> > >
>> > > --
>> > > Alpha Chapters of my book on Hadoop are available
>> > > http://www.apress.com/book/view/9781430219422
>> > > www.prohadoopbook.com a community for Hadoop Professionals
>> > >
>> >
>>
>>
>>
>> --
>> Alpha Chapters of my book on Hadoop are available
>> http://www.apress.com/book/view/9781430219422
>> www.prohadoopbook.com a community for Hadoop Professionals
>>
>

Mime
View raw message