hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Daryn Sharp (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
Date Thu, 25 Apr 2013 14:22:15 GMT

    [ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13641848#comment-13641848

Daryn Sharp commented on HDFS-3934:

Quick review just by eyeballing the patch:

It'd be nice to continue to use the {{HostsFileReader}} and post-process the result.  Otherwise
it's a consistency/maintenance to copy-n-paste any new parsing functionality.

Why does the reader need to instantiate dummy {{DatanodeID}}?  It appears to be for repeatedly
making the somewhat fragile assumption that xferAddr is ipAddr+port?  If that relationship
changes, we've got a problem...

Patch appears to have dropped supported for the node's registration name.  [~eli] wanted me
to maintain that feature in HDFS-3990.  If we need to keep it, doing a lookup and a canonical
lookup (can trigger another dns lookup) isn't compatible with supporting the reg name.

Doing a lookup followed by {{getCanonicalName}} is a bad idea.  It does 2 more lookups: hostname
-> PTR -> A so it can resolve CNAMES to IP to hostname.  With this change I think it
will cause 3 lookups per host.

Question about "// If no transfer port was specified, we take a guess".  Why needed, and what
are the ramifications for getting this wrong?  Just a display issue?

It _feels_ like de-dupping the display should be a bit easier to do w/o changing core node
registration logic?
> duplicative dfs_hosts entries handled wrong
> -------------------------------------------
>                 Key: HDFS-3934
>                 URL: https://issues.apache.org/jira/browse/HDFS-3934
>             Project: Hadoop HDFS
>          Issue Type: Bug
>    Affects Versions: 2.0.1-alpha
>            Reporter: Andy Isaacson
>            Assignee: Colin Patrick McCabe
>            Priority: Minor
>         Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch,
HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch
> A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname
ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts
because {{getDatanodeListForReport}} does not handle such a "pseudo-duplicate" correctly:
> # the "Remove any nodes we know about from the map" loop no longer has the knowledge
to remove the spurious entries
> # the "The remaining nodes are ones that are referenced by the hosts files" loop does
not do hostname lookups, so does not know that the IP and hostname refer to the same host.
> Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP
output:  The *Node* column shows ":50010" as the nodename, with HTML markup {{<a href="http://:50075/browseDirectory.jsp?namenodeInfoPort=50070&amp;dir=%2F&amp;nnaddr="

This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

View raw message