hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Hadoop QA (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HDFS-4448) HA NN does not start with wildcard address configured for other NN when security is enabled
Date Thu, 23 Apr 2015 04:00:40 GMT

    [ https://issues.apache.org/jira/browse/HDFS-4448?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14508413#comment-14508413
] 

Hadoop QA commented on HDFS-4448:
---------------------------------

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  14m 33s | Pre-patch trunk compilation is healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any @author tags.
|
| {color:red}-1{color} | tests included |   0m  0s | The patch doesn't appear to include any
new or modified tests.  Please justify why no new tests are needed for this patch. Also please
list what manual steps were performed to verify this patch. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that end in whitespace.
|
| {color:green}+1{color} | javac |   7m 22s | There were no new javac warning messages. |
| {color:green}+1{color} | javadoc |   9m 40s | There were no new javadoc warning messages.
|
| {color:green}+1{color} | release audit |   0m 22s | The applied patch does not increase
the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   5m 28s | There were no new checkstyle issues. |
| {color:green}+1{color} | install |   1m 34s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 32s | The patch built with eclipse:eclipse.
|
| {color:green}+1{color} | findbugs |   3m  5s | The patch does not introduce any new Findbugs
(version 2.0.3) warnings. |
| {color:green}+1{color} | native |   3m 12s | Pre-build of native portion |
| {color:green}+1{color} | hdfs tests | 163m 56s | Tests passed in hadoop-hdfs. |
| | | 209m 48s | |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | http://issues.apache.org/jira/secure/attachment/12727482/HDFS-4448.2.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / 0ebe84d |
| hadoop-hdfs test log | https://builds.apache.org/job/PreCommit-HDFS-Build/10354/artifact/patchprocess/testrun_hadoop-hdfs.txt
|
| Test Results | https://builds.apache.org/job/PreCommit-HDFS-Build/10354/testReport/ |
| Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/10354//console |


This message was automatically generated.

> HA NN does not start with wildcard address configured for other NN when security is enabled
> -------------------------------------------------------------------------------------------
>
>                 Key: HDFS-4448
>                 URL: https://issues.apache.org/jira/browse/HDFS-4448
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: ha, namenode, security
>    Affects Versions: 2.0.3-alpha
>            Reporter: Aaron T. Myers
>            Assignee: Aaron T. Myers
>             Fix For: 2.8.0
>
>         Attachments: HDFS-4448.2.patch, HDFS-4448.patch, HDFS-4448.patch
>
>
> Currently if one tries to configure HA NNs use the wildcard HTTP address when security
is enabled, the NN will fail to start with an error like the following:
> {code}
> java.lang.IllegalArgumentException: java.io.IOException: Cannot use a wildcard address
with security. Must explicitly set bind address for Kerberos
> {code}
> This is the case even if one configures an actual address for the other NN's HTTP address.
There's no good reason for this, since we now check for the local address being set to 0.0.0.0
and determine the canonical hostname for Kerberos purposes using {{InetAddress.getLocalHost().getCanonicalHostName()}},
so we should remove the restriction.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Mime
View raw message