hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Brian Wolf <brw...@gmail.com>
Subject Re: hadoop under cygwin issue
Date Sat, 13 Mar 2010 01:20:01 GMT
Hi Alex,

I am back on this problem.  Seems it works, but I have this issue with 
connecting to server.
I can connect 'ssh localhost' ok.

Thanks
Brian

$ bin/hadoop jar hadoop-*-examples.jar pi 2 2
Number of Maps  = 2
Samples per Map = 2
10/03/12 17:16:17 INFO ipc.Client: Retrying connect to server: 
localhost/127.0.0.1:9000. Already tried 0 time(s).
10/03/12 17:16:19 INFO ipc.Client: Retrying connect to server: 
localhost/127.0.0.1:9000. Already tried 1 time(s).



Alex Kozlov wrote:
> Can you endeavor a simpler job (just to make sure your setup works):
>
> $ hadoop jar $HADOOP_INSTALL/hadoop-*-examples.jar pi 2 2
>
> Alex K
>
> On Wed, Feb 3, 2010 at 8:26 PM, Brian Wolf <brw314@gmail.com> wrote:
>
>   
>> Alex, thanks for the help,  it seems to start now, however
>>
>>
>> $ bin/hadoop jar hadoop-*-examples.jar grep -fs local input output
>> 'dfs[a-z.]+'
>> 10/02/03 20:02:41 WARN fs.FileSystem: "local" is a deprecated filesystem
>> name. Use "file:///" instead.
>> 10/02/03 20:02:43 INFO mapred.FileInputFormat: Total input paths to process
>> : 3
>> 10/02/03 20:02:44 INFO mapred.JobClient: Running job: job_201002031354_0013
>> 10/02/03 20:02:45 INFO mapred.JobClient:  map 0% reduce 0%
>>
>>
>>
>> it hangs here (is pseudo cluster  supposed to work?)
>>
>>
>> these are bottom of various log files
>>
>> conf log file
>>
>>
>> <property><name>fs.s3.impl</name><value>org.apache.hadoop.fs.s3.S3FileSystem</value></property>
>>
>> <property><name>mapred.input.dir</name><value>file:/C:/OpenSSH/usr/local/hadoop-0.19.2/input</value></property>
>> <property><name>mapred.job.tracker.http.address</name><value>0.0.0.0:50030
>> </value></property>
>> <property><name>io.file.buffer.size</name><value>4096</value></property>
>>
>> <property><name>mapred.jobtracker.restart.recover</name><value>false</value></property>
>>
>> <property><name>io.serializations</name><value>org.apache.hadoop.io.serializer.WritableSerialization</value></property>
>>
>> <property><name>dfs.datanode.handler.count</name><value>3</value></property>
>>
>> <property><name>mapred.reduce.copy.backoff</name><value>300</value></property>
>> <property><name>mapred.task.profile</name><value>false</value></property>
>>
>> <property><name>dfs.replication.considerLoad</name><value>true</value></property>
>>
>> <property><name>jobclient.output.filter</name><value>FAILED</value></property>
>>
>> <property><name>mapred.tasktracker.map.tasks.maximum</name><value>2</value></property>
>>
>> <property><name>io.compression.codecs</name><value>org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.BZip2Codec</value></property>
>> <property><name>fs.checkpoint.size</name><value>67108864</value></property>
>>
>>
>> bottom
>> namenode log
>>
>> added to blk_6520091160827873550_1036 size 570
>> 2010-02-03 20:02:43,826 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=brian,None,Administrators,Users    ip=/127.0.0.1    cmd=create
>>  src=/cygwin/tmp/hadoop-brian/mapred/system/job_201002031354_0013/job.xml
>>  dst=null    perm=brian:supergroup:rw-r--r--
>> 2010-02-03 20:02:43,866 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=brian,None,Administrators,Users    ip=/127.0.0.1    cmd=setPermission
>>    src=/cygwin/tmp/hadoop-brian/mapred/system/job_201002031354_0013/job.xml
>>    dst=null    perm=brian:supergroup:rw-r--r--
>> 2010-02-03 20:02:44,026 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
>> NameSystem.allocateBlock:
>> /cygwin/tmp/hadoop-brian/mapred/system/job_201002031354_0013/job.xml.
>> blk_517844159758473296_1037
>> 2010-02-03 20:02:44,076 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
>> NameSystem.addStoredBlock: blockMap updated: 127.0.0.1:50010 is added to
>> blk_517844159758473296_1037 size 16238
>> 2010-02-03 20:02:44,257 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=brian,None,Administrators,Users    ip=/127.0.0.1    cmd=open
>>  src=/cygwin/tmp/hadoop-brian/mapred/system/job_201002031354_0013/job.xml
>>  dst=null    perm=null
>> 2010-02-03 20:02:44,527 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=brian,None,Administrators,Users    ip=/127.0.0.1    cmd=open
>>  src=/cygwin/tmp/hadoop-brian/mapred/system/job_201002031354_0013/job.jar
>>  dst=null    perm=null
>> 2010-02-03 20:02:45,258 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=brian,None,Administrators,Users    ip=/127.0.0.1    cmd=open
>>  src=/cygwin/tmp/hadoop-brian/mapred/system/job_201002031354_0013/job.split
>>    dst=null    perm=null
>>
>>
>> bottom
>> datanode log
>>
>> 2010-02-03 20:02:44,046 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Receiving block
>> blk_517844159758473296_1037 src: /127.0.0.1:4069 dest: /127.0.0.1:50010
>> 2010-02-03 20:02:44,076 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /
>> 127.0.0.1:4069, dest: /127.0.0.1:50010, bytes: 16238, op: HDFS_WRITE,
>> cliID: DFSClient_-1424524646, srvID:
>> DS-1812377383-192.168.1.5-50010-1265088397104, blockid:
>> blk_517844159758473296_1037
>> 2010-02-03 20:02:44,086 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder 0 for block
>> blk_517844159758473296_1037 terminating
>> 2010-02-03 20:02:44,457 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /
>> 127.0.0.1:50010, dest: /127.0.0.1:4075, bytes: 16366, op: HDFS_READ,
>> cliID: DFSClient_-548531246, srvID:
>> DS-1812377383-192.168.1.5-50010-1265088397104, blockid:
>> blk_517844159758473296_1037
>> 2010-02-03 20:02:44,677 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /
>> 127.0.0.1:50010, dest: /127.0.0.1:4076, bytes: 135168, op: HDFS_READ,
>> cliID: DFSClient_-548531246, srvID:
>> DS-1812377383-192.168.1.5-50010-1265088397104, blockid:
>> blk_-2806977820057440405_1035
>> 2010-02-03 20:02:45,278 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode.clienttrace: src: /
>> 127.0.0.1:50010, dest: /127.0.0.1:4077, bytes: 578, op: HDFS_READ, cliID:
>> DFSClient_-548531246, srvID: DS-1812377383-192.168.1.5-50010-1265088397104,
>> blockid: blk_6520091160827873550_1036
>> 2010-02-03 20:04:10,451 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataBlockScanner: Verification
>> succeeded for blk_3301977249866081256_1031
>> 2010-02-03 20:09:35,658 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataBlockScanner: Verification
>> succeeded for blk_9116729021606317943_1025
>> 2010-02-03 20:09:44,671 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataBlockScanner: Verification
>> succeeded for blk_8602436668984954947_1026
>>
>>
>>
>>
>>
>> jobtracker log
>>
>> Input size for job job_201002031354_0012 = 53060
>> 2010-02-03 19:48:37,599 INFO org.apache.hadoop.mapred.JobInProgress: Split
>> info for job:job_201002031354_0012
>> 2010-02-03 19:48:37,649 INFO org.apache.hadoop.net.NetworkTopology: Adding
>> a new node: /default-rack/localhost
>> 2010-02-03 19:48:37,659 INFO org.apache.hadoop.mapred.JobInProgress:
>> tip:task_201002031354_0012_m_000000 has split on
>> node:/default-rack/localhost
>> 2010-02-03 19:48:37,659 INFO org.apache.hadoop.mapred.JobInProgress:
>> tip:task_201002031354_0012_m_000001 has split on
>> node:/default-rack/localhost
>> 2010-02-03 19:48:37,659 INFO org.apache.hadoop.mapred.JobInProgress:
>> tip:task_201002031354_0012_m_000002 has split on
>> node:/default-rack/localhost
>> 2010-02-03 19:48:37,659 INFO org.apache.hadoop.mapred.JobInProgress:
>> tip:task_201002031354_0012_m_000003 has split on
>> node:/default-rack/localhost
>> 2010-02-03 20:02:45,278 INFO org.apache.hadoop.mapred.JobInProgress: Input
>> size for job job_201002031354_0013 = 53060
>> 2010-02-03 20:02:45,278 INFO org.apache.hadoop.mapred.JobInProgress: Split
>> info for job:job_201002031354_0013
>> 2010-02-03 20:02:45,278 INFO org.apache.hadoop.mapred.JobInProgress:
>> tip:task_201002031354_0013_m_000000 has split on
>> node:/default-rack/localhost
>> 2010-02-03 20:02:45,278 INFO org.apache.hadoop.mapred.JobInProgress:
>> tip:task_201002031354_0013_m_000001 has split on
>> node:/default-rack/localhost
>> 2010-02-03 20:02:45,278 INFO org.apache.hadoop.mapred.JobInProgress:
>> tip:task_201002031354_0013_m_000002 has split on
>> node:/default-rack/localhost
>> 2010-02-03 20:02:45,278 INFO org.apache.hadoop.mapred.JobInProgress:
>> tip:task_201002031354_0013_m_000003 has split on
>> node:/default-rack/localhost
>>
>>
>>
>> Thanks
>> Brian
>>
>>
>>
>>
>>
>> Alex Kozlov wrote:
>>
>>     
>>> Try
>>>
>>> $ bin/hadoop jar hadoop-*-examples.jar grep
>>> file:///usr/local/hadoop-0.19.2/input output 'dfs[a-z.]+'
>>>
>>> file:/// is a magical prefix to force hadoop to look for the file in the
>>> local FS
>>>
>>> You can also force it to look into local FS by giving '-fs local' or '-fs
>>> file:///' option to the hadoop executable
>>>
>>> These options basically overwrite the *fs.default.name* configuration
>>> setting, which should be in your core-site.xml file
>>>
>>> You can also copy the content of the input directory to HDFS by executing
>>>
>>> $ bin/hadoop fs -mkdir input
>>> $ bin/hadoop fs -copyFromLocal input/* input
>>>
>>> Hope this helps
>>>
>>> Alex K
>>>
>>> On Wed, Feb 3, 2010 at 2:17 PM, Brian Wolf <brw314@gmail.com> wrote:
>>>
>>>
>>>
>>>       
>>>> Alex Kozlov wrote:
>>>>
>>>>
>>>>
>>>>         
>>>>> Live Nodes <http://localhost:50070/dfshealth.jsp#LiveNodes>   
 :
>>>>> 0
>>>>>
>>>>> You datanode is dead.  Look at the logs in the $HADOOP_HOME/logs
>>>>> directory
>>>>> (or where your logs are) and check the errors.
>>>>>
>>>>> Alex K
>>>>>
>>>>> On Mon, Feb 1, 2010 at 1:59 PM, Brian Wolf <brw314@gmail.com> wrote:
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>           
>>>> Thanks for your help, Alex,
>>>>
>>>> I managed to get past that problem, now I have this problem:
>>>>
>>>> However, when I try to run this example as stated on the quickstart
>>>> webpage:
>>>>
>>>>
>>>> bin/hadoop jar hadoop-*-examples.jar grep input  output 'dfs[a-z.]+'
>>>>
>>>> I get this error;
>>>> =============================================================
>>>> java.io.IOException:       Not a file:
>>>> hdfs://localhost:9000/user/brian/input/conf
>>>> =========================================================
>>>> so it seems to default to my home directory looking for "input" it
>>>> apparently  needs an absolute filepath, however, when I  run that way:
>>>>
>>>> $ bin/hadoop jar hadoop-*-examples.jar grep
>>>> /usr/local/hadoop-0.19.2/input
>>>>  output 'dfs[a-z.]+'
>>>>
>>>> ==============================================================
>>>> org.apache.hadoop.mapred.InvalidInputException: Input path does not
>>>> exist:
>>>> hdfs://localhost:9000/usr/local/hadoop-0.19.2/input
>>>> ==============================================================
>>>> It still isn't happy although this part -> /usr/local/hadoop-0.19.2/input
>>>>  <-  does exist
>>>>
>>>>  Aaron,
>>>>
>>>>
>>>>         
>>>>> Thanks or your help. I  carefully went through the steps again a couple
>>>>>           
>>>>>> times , and ran
>>>>>>
>>>>>> after this
>>>>>> bin/hadoop namenode -format
>>>>>>
>>>>>> (by the way, it asks if I want to reformat, I've tried it both ways)
>>>>>>
>>>>>>
>>>>>> then
>>>>>>
>>>>>>
>>>>>> bin/start-dfs.sh
>>>>>>
>>>>>> and
>>>>>>
>>>>>> bin/start-all.sh
>>>>>>
>>>>>>
>>>>>> and then
>>>>>> bin/hadoop fs -put conf input
>>>>>>
>>>>>> now the return for this seemed cryptic:
>>>>>>
>>>>>>
>>>>>> put: Target input/conf is a directory
>>>>>>
>>>>>> (??)
>>>>>>
>>>>>>  and when I tried
>>>>>>
>>>>>> bin/hadoop jar hadoop-*-examples.jar grep input output 'dfs[a-z.]+'
>>>>>>
>>>>>> It says something about 0 nodes
>>>>>>
>>>>>> (from log file)
>>>>>>
>>>>>> 2010-02-01 13:26:29,874 INFO
>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>>>>> ugi=brian,None,Administrators,Users    ip=/127.0.0.1    cmd=create
>>>>>>
>>>>>>
>>>>>>  src=/cygwin/tmp/hadoop-SYSTEM/mapred/system/job_201002011323_0001/job.jar
>>>>>>  dst=null    perm=brian:supergroup:rw-r--r--
>>>>>> 2010-02-01 13:26:30,045 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>>>> handler 3 on 9000, call
>>>>>>
>>>>>>
>>>>>> addBlock(/cygwin/tmp/hadoop-SYSTEM/mapred/system/job_201002011323_0001/job.jar,
>>>>>> DFSClient_725490811) from 127.0.0.1:3003: error: java.io.IOException:
>>>>>> File
>>>>>> /cygwin/tmp/hadoop-SYSTEM/mapred/system/job_201002011323_0001/job.jar
>>>>>> could
>>>>>> only be replicated to 0 nodes, instead of 1
>>>>>> java.io.IOException: File
>>>>>> /cygwin/tmp/hadoop-SYSTEM/mapred/system/job_201002011323_0001/job.jar
>>>>>> could
>>>>>> only be replicated to 0 nodes, instead of 1
>>>>>>  at
>>>>>>
>>>>>>
>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1287)
>>>>>>  at
>>>>>>
>>>>>>
>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:351)
>>>>>>  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>>>  at
>>>>>>
>>>>>>
>>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>>>>>>  at
>>>>>>
>>>>>>
>>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> To maybe rule out something regarding ports or ssh , when I run
>>>>>> netstat:
>>>>>>
>>>>>>  TCP    127.0.0.1:9000         0.0.0.0:0              LISTENING
>>>>>>  TCP    127.0.0.1:9001         0.0.0.0:0              LISTENING
>>>>>>
>>>>>>
>>>>>> and when I browse to http://localhost:50070/
>>>>>>
>>>>>>
>>>>>>   Cluster Summary
>>>>>>
>>>>>> * * * 21 files and directories, 0 blocks = 21 total. Heap Size is
8.01
>>>>>> MB
>>>>>> /
>>>>>> 992.31 MB (0%)
>>>>>> *
>>>>>> Configured Capacity     :       0 KB
>>>>>> DFS Used        :       0 KB
>>>>>> Non DFS Used    :       0 KB
>>>>>> DFS Remaining   :       0 KB
>>>>>> DFS Used%       :       100 %
>>>>>> DFS Remaining%  :       0 %
>>>>>> Live Nodes <http://localhost:50070/dfshealth.jsp#LiveNodes>
    :
>>>>>> 0
>>>>>> Dead Nodes <http://localhost:50070/dfshealth.jsp#DeadNodes>
    :
>>>>>> 0
>>>>>>
>>>>>>
>>>>>> so I'm a bit still in the dark, I guess.
>>>>>>
>>>>>> Thanks
>>>>>> Brian
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> Aaron Kimball wrote:
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>             
>>>>>>> Brian, it looks like you missed a step in the instructions. You'll
>>>>>>> need
>>>>>>> to
>>>>>>> format the hdfs filesystem instance before starting the NameNode
>>>>>>> server:
>>>>>>>
>>>>>>> You need to run:
>>>>>>>
>>>>>>> $ bin/hadoop namenode -format
>>>>>>>
>>>>>>> .. then you can do bin/start-dfs.sh
>>>>>>> Hope this helps,
>>>>>>> - Aaron
>>>>>>>
>>>>>>>
>>>>>>> On Sat, Jan 30, 2010 at 12:27 AM, Brian Wolf <brw314@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>               
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> I am trying to run Hadoop 0.19.2 under cygwin as per directions
on
>>>>>>>> the
>>>>>>>> hadoop "quickstart" web page.
>>>>>>>>
>>>>>>>> I know sshd is running and I can "ssh localhost" without
a password.
>>>>>>>>
>>>>>>>> This is from my hadoop-site.xml
>>>>>>>>
>>>>>>>> <configuration>
>>>>>>>> <property>
>>>>>>>> <name>hadoop.tmp.dir</name>
>>>>>>>> <value>/cygwin/tmp/hadoop-${user.name}</value>
>>>>>>>> </property>
>>>>>>>> <property>
>>>>>>>> <name>fs.default.name</name>
>>>>>>>> <value>hdfs://localhost:9000</value>
>>>>>>>> </property>
>>>>>>>> <property>
>>>>>>>> <name>mapred.job.tracker</name>
>>>>>>>> <value>localhost:9001</value>
>>>>>>>> </property>
>>>>>>>> <property>
>>>>>>>> <name>mapred.job.reuse.jvm.num.tasks</name>
>>>>>>>> <value>-1</value>
>>>>>>>> </property>
>>>>>>>> <property>
>>>>>>>> <name>dfs.replication</name>
>>>>>>>> <value>1</value>
>>>>>>>> </property>
>>>>>>>> <property>
>>>>>>>> <name>dfs.permissions</name>
>>>>>>>> <value>false</value>
>>>>>>>> </property>
>>>>>>>> <property>
>>>>>>>> <name>webinterface.private.actions</name>
>>>>>>>> <value>true</value>
>>>>>>>> </property>
>>>>>>>> </configuration>
>>>>>>>>
>>>>>>>> These are errors from my log files:
>>>>>>>>
>>>>>>>>
>>>>>>>> 2010-01-30 00:03:33,091 INFO
>>>>>>>> org.apache.hadoop.ipc.metrics.RpcMetrics:
>>>>>>>> Initializing RPC Metrics with hostName=NameNode, port=9000
>>>>>>>> 2010-01-30 00:03:33,121 INFO
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode
up at:
>>>>>>>> localhost/
>>>>>>>> 127.0.0.1:9000
>>>>>>>> 2010-01-30 00:03:33,161 INFO
>>>>>>>> org.apache.hadoop.metrics.jvm.JvmMetrics:
>>>>>>>> Initializing JVM Metrics with processName=NameNode, sessionId=null
>>>>>>>> 2010-01-30 00:03:33,181 INFO
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics:
>>>>>>>> Initializing
>>>>>>>> NameNodeMeterics using context
>>>>>>>> object:org.apache.hadoop.metrics.spi.NullContext
>>>>>>>> 2010-01-30 00:03:34,603 INFO
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>>>>>> fsOwner=brian,None,Administrators,Users
>>>>>>>> 2010-01-30 00:03:34,603 INFO
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>>>>>> supergroup=supergroup
>>>>>>>> 2010-01-30 00:03:34,603 INFO
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>>>>>> isPermissionEnabled=false
>>>>>>>> 2010-01-30 00:03:34,653 INFO
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics:
>>>>>>>> Initializing FSNamesystemMetrics using context
>>>>>>>> object:org.apache.hadoop.metrics.spi.NullContext
>>>>>>>> 2010-01-30 00:03:34,653 INFO
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
>>>>>>>> FSNamesystemStatusMBean
>>>>>>>> 2010-01-30 00:03:34,803 INFO
>>>>>>>> org.apache.hadoop.hdfs.server.common.Storage:
>>>>>>>> Storage directory C:\cygwin\tmp\hadoop-brian\dfs\name does
not exist.
>>>>>>>> 2010-01-30 00:03:34,813 ERROR
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: FSNamesystem
>>>>>>>> initialization failed.
>>>>>>>> org.apache.hadoop.hdfs.server.common.InconsistentFSStateException:
>>>>>>>> Directory C:\cygwin\tmp\hadoop-brian\dfs\name is in an inconsistent
>>>>>>>> state:
>>>>>>>> storage directory does not exist or is not accessible.
>>>>>>>>  at
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:278)
>>>>>>>>  at
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:87)
>>>>>>>>  at
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:309)
>>>>>>>>  at
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:288)
>>>>>>>>  at
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:163)
>>>>>>>>  at
>>>>>>>>
>>>>>>>>
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:208)
>>>>>>>>  at
>>>>>>>>
>>>>>>>>
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:194)
>>>>>>>>  at
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:859)
>>>>>>>>  at
>>>>>>>>
>>>>>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:868)
>>>>>>>> 2010-01-30 00:03:34,823 INFO org.apache.hadoop.ipc.Server:
Stopping
>>>>>>>> server
>>>>>>>> on 9000
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> =========================================================
>>>>>>>>
>>>>>>>> 2010-01-29 15:13:30,270 INFO org.apache.hadoop.ipc.Client:
Retrying
>>>>>>>> connect
>>>>>>>> to server: localhost/127.0.0.1:9000. Already tried 9 time(s).
>>>>>>>> problem cleaning system directory: null
>>>>>>>> java.net.ConnectException: Call to localhost/127.0.0.1:9000
failed
>>>>>>>> on
>>>>>>>> connection exception: java.net.ConnectException: Connection
refused:
>>>>>>>> no
>>>>>>>> further information
>>>>>>>>  at org.apache.hadoop.ipc.Client.wrapException(Client.java:724)
>>>>>>>>  at org.apache.hadoop.ipc.Client.call(Client.java:700)
>>>>>>>>  at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216)
>>>>>>>>  at $Proxy4.getProtocolVersion(Unknown Source)
>>>>>>>>  at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:348)
>>>>>>>>  at
>>>>>>>>
>>>>>>>> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:104)
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Thanks
>>>>>>>> Brian
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>                 
>>>>>>>
>>>>>>>               
>>>>>>             
>>>>>
>>>>>           
>>>>         
>>>
>>>       
>>     
>
>   


Mime
View raw message