Return-Path: Delivered-To: apmail-hadoop-core-dev-archive@www.apache.org Received: (qmail 70961 invoked from network); 26 May 2009 19:02:59 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (140.211.11.3) by minotaur.apache.org with SMTP; 26 May 2009 19:02:59 -0000 Received: (qmail 52807 invoked by uid 500); 26 May 2009 19:03:10 -0000 Delivered-To: apmail-hadoop-core-dev-archive@hadoop.apache.org Received: (qmail 52753 invoked by uid 500); 26 May 2009 19:03:10 -0000 Mailing-List: contact core-dev-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: core-dev@hadoop.apache.org Delivered-To: mailing list core-dev@hadoop.apache.org Received: (qmail 52669 invoked by uid 99); 26 May 2009 19:03:10 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 26 May 2009 19:03:10 +0000 X-ASF-Spam-Status: No, hits=-1998.5 required=10.0 tests=ALL_TRUSTED,WEIRD_PORT X-Spam-Check-By: apache.org Received: from [140.211.11.140] (HELO brutus.apache.org) (140.211.11.140) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 26 May 2009 19:03:07 +0000 Received: from brutus (localhost [127.0.0.1]) by brutus.apache.org (Postfix) with ESMTP id D9F3E234C1E9 for ; Tue, 26 May 2009 12:02:45 -0700 (PDT) Message-ID: <756439471.1243364565891.JavaMail.jira@brutus> Date: Tue, 26 May 2009 12:02:45 -0700 (PDT) From: "Jakob Homan (JIRA)" To: core-dev@hadoop.apache.org Subject: [jira] Commented: (HADOOP-5626) SecondaryNamenode may report incorrect info host name In-Reply-To: <1404622165.1238955912965.JavaMail.jira@brutus> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 X-Virus-Checked: Checked by ClamAV on apache.org [ https://issues.apache.org/jira/browse/HADOOP-5626?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12713190#action_12713190 ] Jakob Homan commented on HADOOP-5626: ------------------------------------- In BackupNode::getHttpServerAddress this issue also arises and is handled by using the rpcAddress, which has already been resolved to a real address. It would be good to move your solution to NetUtils and access it from both the SecondaryNameNode and the BackupNode, to avoid code duplication. > SecondaryNamenode may report incorrect info host name > ----------------------------------------------------- > > Key: HADOOP-5626 > URL: https://issues.apache.org/jira/browse/HADOOP-5626 > Project: Hadoop Core > Issue Type: Bug > Components: dfs > Reporter: Carlos Valiente > Assignee: Todd Lipcon > Priority: Minor > Fix For: 0.21.0 > > Attachments: HADOOP-5626.patch, hadoop-5626.txt > > > I have set up {{dfs.secondary.http.address}} like this: > {code} > > dfs.secondary.http.address > secondary.example.com:50090 > > {code} > In my setup {{secondary.example.com}} resolves to an IP address (say, 192.168.0.10) which is not the same as the host's name (as returned by {{InetAddress.getLocalHost().getHostAddress()}}, say 192.168.0.1). > In this situation, edit log related transfers fail. From the namenode log: > {code} > 2009-04-05 13:32:39,128 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Roll Edit Log from 192.168.0.10 > 2009-04-05 13:32:39,168 WARN org.mortbay.log: /getimage: java.io.IOException: GetImage failed. java.net.ConnectException: Connection refused > at java.net.PlainSocketImpl.socketConnect(Native Method) > at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:333) > at java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:195) > at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:182) > at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366) > at java.net.Socket.connect(Socket.java:519) > at java.net.Socket.connect(Socket.java:469) > at sun.net.NetworkClient.doConnect(NetworkClient.java:163) > at sun.net.www.http.HttpClient.openServer(HttpClient.java:394) > at sun.net.www.http.HttpClient.openServer(HttpClient.java:529) > at sun.net.www.http.HttpClient.(HttpClient.java:233) > at sun.net.www.http.HttpClient.New(HttpClient.java:306) > at sun.net.www.http.HttpClient.New(HttpClient.java:323) > at sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:837) > at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:778) > at sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:703) > at sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1026) > at org.apache.hadoop.hdfs.server.namenode.TransferFsImage.getFileClient(TransferFsImage.java:151) > ... > {code} > From the secondary namenode log: > {code} > 2009-04-05 13:42:39,238 ERROR org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: Exception in doCheckpoint: > 2009-04-05 13:42:39,238 ERROR org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: java.io.FileNotFoundException: http://nn.example.com:50070/getimage?putimage=1&port=50090&machine= > 192.168.0.1&token=-19:1243068779:0:1238929357000:1238929031783 > at sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1288) > at org.apache.hadoop.hdfs.server.namenode.TransferFsImage.getFileClient(TransferFsImage.java:151) > at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.putFSImage(SecondaryNameNode.java:294) > at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doCheckpoint(SecondaryNameNode.java:333) > at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.run(SecondaryNameNode.java:239) > at java.lang.Thread.run(Thread.java:619) > {code} -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.