hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Yi Liu (JIRA)" <j...@apache.org>
Subject [jira] [Comment Edited] (HDFS-7392) org.apache.hadoop.hdfs.DistributedFileSystem open invalid URI forever
Date Thu, 20 Nov 2014 02:09:33 GMT

    [ https://issues.apache.org/jira/browse/HDFS-7392?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14218889#comment-14218889
] 

Yi Liu edited comment on HDFS-7392 at 11/20/14 2:08 AM:
--------------------------------------------------------

[~vacekf], I get the issue from your description now. I think it is an issue from the code
logic, and now I try to reproduce it in my local env and go to confirm it.

I repeat the issue here: the issue (client open the URL in dead loop) occurs when you use
client to access HDFS, but you fill a wrong filesystem URL and there is no namenode in the
host corresponding to that URL, and the FQDN of that host can be resolved to two IP addresses.
Is that right?


was (Author: hitliuyi):
[~vacekf], I get the issue from your description now. I think it is an issue from the code
logic, and now I try to reproduce it in my local env and go to confirm it.

I repeat the issue here: the issue (client will open in dead loop) occurs when you use client
to access HDFS, but you fill a wrong filesystem URL and there is no namenode in the host corresponding
to that URL, and the FQDN of that host can be resolved to two IP addresses. Is that right?

> org.apache.hadoop.hdfs.DistributedFileSystem open invalid URI forever
> ---------------------------------------------------------------------
>
>                 Key: HDFS-7392
>                 URL: https://issues.apache.org/jira/browse/HDFS-7392
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: hdfs-client
>            Reporter: Frantisek Vacek
>            Priority: Critical
>         Attachments: 1.png, 2.png
>
>
> In some specific circumstances, org.apache.hadoop.hdfs.DistributedFileSystem.open(invalid
URI) never timeouts and last forever.
> What are specific circumstances:
> 1) HDFS URI (hdfs://share.example.com:8020/someDir/someFile.txt) should point to valid
IP address but without name node service running on it.
> 2) There should be at least 2 IP addresses for such a URI. See output below:
> {quote}
> [~/proj/quickbox]$ nslookup share.example.com
> Server:         127.0.1.1
> Address:        127.0.1.1#53
> share.example.com canonical name = internal-realm-share-example-com-1234.us-east-1.elb.amazonaws.com.
> Name:   internal-realm-share-example-com-1234.us-east-1.elb.amazonaws.com
> Address: 192.168.1.223
> Name:   internal-realm-share-example-com-1234.us-east-1.elb.amazonaws.com
> Address: 192.168.1.65
> {quote}
> In such a case the org.apache.hadoop.ipc.Client.Connection.updateAddress() returns sometimes
true (even if address didn't actually changed see img. 1) and the timeoutFailures counter
is set to 0 (see img. 2). The maxRetriesOnSocketTimeouts (45) is never reached and connection
attempt is repeated forever.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Mime
View raw message