hadoop-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Namikaze Minato <lloydsen...@gmail.com>
Subject Re: Problem running example (wrong IP address)
Date Fri, 25 Sep 2015 17:12:52 GMT
http://sprunge.us/EAJJ
http://sprunge.us/hDSX
http://sprunge.us/CKag

On 25 September 2015 at 18:53, Brahma Reddy Battula
<brahmareddy.battula@hotmail.com> wrote:
> sorry,I am not able to access the logs, could please post in paste bin or
> attach the 192.168.51.6( as your query is why different IP) DN logs and
> namenode logs here..?
>
>
>
>
> Thanks And Regards
> Brahma Reddy Battula
>
>
> ________________________________
> Date: Fri, 25 Sep 2015 11:16:55 -0500
> Subject: Re: Problem running example (wrong IP address)
> From: dwmaillist@gmail.com
> To: user@hadoop.apache.org
>
>
> Brahma,
>
> Thanks for the reply. I'll keep this conversation here in the user list. The
> /etc/hosts file is identical on all three nodes
>
> hadoop@hadoop-data1:~$ cat /etc/hosts
> 127.0.0.1 localhost
> 192.168.51.4 hadoop-master
> 192.168.52.4 hadoop-data1
> 192.168.52.6 hadoop-data2
>
> hadoop@hadoop-data2:~$ cat /etc/hosts
> 127.0.0.1 localhost
> 192.168.51.4 hadoop-master
> 192.168.52.4 hadoop-data1
> 192.168.52.6 hadoop-data2
>
> hadoop@hadoop-master:~$ cat /etc/hosts
> 127.0.0.1 localhost
> 192.168.51.4 hadoop-master
> 192.168.52.4 hadoop-data1
> 192.168.52.6 hadoop-data2
>
> Here are the startup logs for all three nodes:
> https://gist.github.com/dwatrous/7241bb804a9be8f9303f
> https://gist.github.com/dwatrous/bcd85cda23d6eca3a68b
> https://gist.github.com/dwatrous/922c4f773aded0137fa3
>
> Thanks for your help.
>
>
> On Fri, Sep 25, 2015 at 10:33 AM, Brahma Reddy Battula
> <brahmareddy.battula@huawei.com> wrote:
>
> Seems DN started in three machines and failed in
> hadoop-data1(192.168.52.4)..
>
>
> 192.168.51.6 : giving IP as 192.168.51.1...can you please check /etc/hosts
> file of 192.168.51.6 (might be 192.168.51.1 is configured in /etc/hosts)
>
> 192.168.52.4 : datanode startup might be failed ( you can check this node
> logs)
>
> 192.168.51.4 :  Datanode starup is success..which is in master node..
>
>
>
> Thanks & Regards
>  Brahma Reddy Battula
>
>
>
> ________________________________
> From: Daniel Watrous [dwmaillist@gmail.com]
> Sent: Friday, September 25, 2015 8:41 PM
> To: user@hadoop.apache.org
> Subject: Re: Problem running example (wrong IP address)
>
> I'm still stuck on this and posted it to stackoverflow:
> http://stackoverflow.com/questions/32785256/hadoop-datanode-binds-wrong-ip-address
>
> Thanks,
> Daniel
>
> On Fri, Sep 25, 2015 at 8:28 AM, Daniel Watrous <dwmaillist@gmail.com>
> wrote:
>
> I could really use some help here. As you can see from the output below, the
> two attached datanodes are identified with a non-existent IP address. Can
> someone tell me how that gets selected or how to explicitly set it. Also,
> why are both datanodes shown under the same name/IP?
>
> hadoop@hadoop-master:~$ hdfs dfsadmin -report
> Configured Capacity: 84482326528 (78.68 GB)
> Present Capacity: 75745546240 (70.54 GB)
> DFS Remaining: 75744862208 (70.54 GB)
> DFS Used: 684032 (668 KB)
> DFS Used%: 0.00%
> Under replicated blocks: 0
> Blocks with corrupt replicas: 0
> Missing blocks: 0
> Missing blocks (with replication factor 1): 0
>
> -------------------------------------------------
> Live datanodes (2):
>
> Name: 192.168.51.1:50010 (192.168.51.1)
> Hostname: hadoop-data1
> Decommission Status : Normal
> Configured Capacity: 42241163264 (39.34 GB)
> DFS Used: 303104 (296 KB)
> Non DFS Used: 4302479360 (4.01 GB)
> DFS Remaining: 37938380800 (35.33 GB)
> DFS Used%: 0.00%
> DFS Remaining%: 89.81%
> Configured Cache Capacity: 0 (0 B)
> Cache Used: 0 (0 B)
> Cache Remaining: 0 (0 B)
> Cache Used%: 100.00%
> Cache Remaining%: 0.00%
> Xceivers: 1
> Last contact: Fri Sep 25 13:25:37 UTC 2015
>
>
> Name: 192.168.51.4:50010 (hadoop-master)
> Hostname: hadoop-master
> Decommission Status : Normal
> Configured Capacity: 42241163264 (39.34 GB)
> DFS Used: 380928 (372 KB)
> Non DFS Used: 4434300928 (4.13 GB)
> DFS Remaining: 37806481408 (35.21 GB)
> DFS Used%: 0.00%
> DFS Remaining%: 89.50%
> Configured Cache Capacity: 0 (0 B)
> Cache Used: 0 (0 B)
> Cache Remaining: 0 (0 B)
> Cache Used%: 100.00%
> Cache Remaining%: 0.00%
> Xceivers: 1
> Last contact: Fri Sep 25 13:25:38 UTC 2015
>
>
>
> On Thu, Sep 24, 2015 at 5:05 PM, Daniel Watrous <dwmaillist@gmail.com>
> wrote:
>
> The IP address is clearly wrong, but I'm not sure how it gets set. Can
> someone tell me how to configure it to choose a valid IP address?
>
> On Thu, Sep 24, 2015 at 3:26 PM, Daniel Watrous <dwmaillist@gmail.com>
> wrote:
>
> I just noticed that both datanodes appear to have chosen that IP address and
> bound that port for HDFS communication.
>
> http://screencast.com/t/OQNbrWFF
>
> Any idea why this would be? Is there some way to specify which IP/hostname
> should be used for that?
>
> On Thu, Sep 24, 2015 at 3:11 PM, Daniel Watrous <dwmaillist@gmail.com>
> wrote:
>
> When I try to run a map reduce example, I get the following error:
>
> hadoop@hadoop-master:~$ hadoop jar
> /usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.7.1.jar
> pi 10 30
> Number of Maps  = 10
> Samples per Map = 30
> 15/09/24 20:04:28 INFO hdfs.DFSClient: Exception in createBlockOutputStream
> java.io.IOException: Got error, status message , ack with firstBadLink as
> 192.168.51.1:50010
>         at
> org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:140)
>         at
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:1334)
>         at
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1237)
>         at
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:449)
> 15/09/24 20:04:28 INFO hdfs.DFSClient: Abandoning
> BP-852923283-127.0.1.1-1443119668806:blk_1073741825_1001
> 15/09/24 20:04:28 INFO hdfs.DFSClient: Excluding datanode
> DatanodeInfoWithStorage[192.168.51.1:50010,DS-45f6e06d-752e-41e8-ac25-ca88bce80d00,DISK]
> 15/09/24 20:04:28 WARN hdfs.DFSClient: Slow waitForAckedSeqno took 65357ms
> (threshold=30000ms)
> Wrote input for Map #0
>
> I'm not sure why it's trying to access 192.168.51.1:50010, which isn't even
> a valid IP address in my setup.
>
> Daniel
>
>
>
>
>
>

Mime
View raw message