hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jander g <jande...@gmail.com>
Subject Re: Some connection error when running HADOOP
Date Sat, 16 Oct 2010 14:19:37 GMT
Hi,
Didn't you format the HDFS ? or maybe the HDFS was broken.

On Sat, Oct 16, 2010 at 9:01 PM, 冯超 <hsmyy198712@gmail.com> wrote:

> Dear All,
>    I met a problem that when I start the hadoop ,I could not find the slave
> nodes , my OS system is CentOS, when run the mapreduce program, it alarms
> me
> that:
>
>
> 10/10/16 20:32:41 INFO ipc.Client: Retrying connect to server: zawc1/
> 192.168.1.101:9000. Already tried 0 time(s).
> 10/10/16 20:32:42 INFO ipc.Client: Retrying connect to server: zawc1/
> 192.168.1.101:9000. Already tried 1 time(s).
> 10/10/16 20:32:43 INFO ipc.Client: Retrying connect to server: zawc1/
> 192.168.1.101:9000. Already tried 2 time(s).
> 10/10/16 20:32:44 INFO ipc.Client: Retrying connect to server: zawc1/
> 192.168.1.101:9000. Already tried 3 time(s).
> 10/10/16 20:32:45 INFO ipc.Client: Retrying connect to server: zawc1/
> 192.168.1.101:9000. Already tried 4 time(s).
> 10/10/16 20:32:46 INFO ipc.Client: Retrying connect to server: zawc1/
> 192.168.1.101:9000. Already tried 5 time(s).
> 10/10/16 20:32:47 INFO ipc.Client: Retrying connect to server: zawc1/
> 192.168.1.101:9000. Already tried 6 time(s).
> 10/10/16 20:32:48 INFO ipc.Client: Retrying connect to server: zawc1/
> 192.168.1.101:9000. Already tried 7 time(s).
> 10/10/16 20:32:49 INFO ipc.Client: Retrying connect to server: zawc1/
> 192.168.1.101:9000. Already tried 8 time(s).
> 10/10/16 20:32:50 INFO ipc.Client: Retrying connect to server: zawc1/
> 192.168.1.101:9000. Already tried 9 time(s).
> java.net.ConnectException: Call to zawc1/192.168.1.101:9000 failed on
> connection exception: java.net.ConnectException: Connection refused
>        at org.apache.hadoop.ipc.Client.wrapException(Client.java:724)
>        at org.apache.hadoop.ipc.Client.call(Client.java:700)
>        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216)
>        at $Proxy0.getProtocolVersion(Unknown Source)
>        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:348)
>        at
> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:104)
>        at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:176)
>        at
>
> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:75)
>        at
> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1367)
>        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:56)
>        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1379)
>        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:215)
>        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:120)
>        at
> org.apache.hadoop.examples.PiEstimator.launch(PiEstimator.java:187)
>        at org.apache.hadoop.examples.PiEstimator.run(PiEstimator.java:245)
>        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>        at org.apache.hadoop.examples.PiEstimator.main(PiEstimator.java:252)
>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>        at
>
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>        at
>
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>        at java.lang.reflect.Method.invoke(Method.java:597)
>        at
>
> org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:68)
>        at
> org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:141)
>        at
> org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:61)
>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>        at
>
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>        at
>
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>        at java.lang.reflect.Method.invoke(Method.java:597)
>        at org.apache.hadoop.util.RunJar.main(RunJar.java:165)
>        at org.apache.hadoop.mapred.JobShell.run(JobShell.java:54)
>        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
>        at org.apache.hadoop.mapred.JobShell.main(JobShell.java:68)
> Caused by: java.net.ConnectException: Connection refused
>        at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>        at
> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:574)
>        at sun.nio.ch.SocketAdaptor.connect(SocketAdaptor.java:100)
>        at
> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:300)
>        at
> org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:177)
>        at org.apache.hadoop.ipc.Client.getConnection(Client.java:801)
>        at org.apache.hadoop.ipc.Client.call(Client.java:686)
>        ... 31 more
>
>    Who can tell me what is the reason? Thanks for any help...
>



-- 
Thanks,
Jander

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message