Return-Path: X-Original-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 3368CEE92 for ; Fri, 15 Feb 2013 13:11:55 +0000 (UTC) Received: (qmail 68734 invoked by uid 500); 15 Feb 2013 13:11:50 -0000 Delivered-To: apmail-hadoop-mapreduce-user-archive@hadoop.apache.org Received: (qmail 68313 invoked by uid 500); 15 Feb 2013 13:11:49 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 68281 invoked by uid 99); 15 Feb 2013 13:11:48 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 15 Feb 2013 13:11:48 +0000 X-ASF-Spam-Status: No, hits=3.2 required=5.0 tests=FREEMAIL_REPLY,HTML_MESSAGE,NORMAL_HTTP_TO_IP,RCVD_IN_DNSWL_NONE,SPF_PASS,WEIRD_PORT X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of vijayjtuk@hotmail.com designates 157.55.1.146 as permitted sender) Received: from [157.55.1.146] (HELO dub0-omc2-s7.dub0.hotmail.com) (157.55.1.146) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 15 Feb 2013 13:11:40 +0000 Received: from DUB118-DS9 ([157.55.1.136]) by dub0-omc2-s7.dub0.hotmail.com with Microsoft SMTPSVC(6.0.3790.4675); Fri, 15 Feb 2013 05:11:19 -0800 X-EIP: [EX97A4hIj3TJR+OR22NZpKPz0ujKbMJAeb9uHvUXv5w=] X-Originating-Email: [vijayjtuk@hotmail.com] Message-ID: From: Vijay Thakorlal To: References: In-Reply-To: Subject: RE: getimage failed in Name Node Log Date: Fri, 15 Feb 2013 13:11:26 +0000 MIME-Version: 1.0 Content-Type: multipart/alternative; boundary="----=_NextPart_000_0049_01CE0B7D.F67D08D0" X-Mailer: Microsoft Office Outlook 12.0 thread-index: Ac4LeTBSzmIzZTEVRWaj2znXV8tURQAAlrlw Content-Language: en-gb X-OriginalArrivalTime: 15 Feb 2013 13:11:19.0874 (UTC) FILETIME=[F1F89220:01CE0B7D] X-Virus-Checked: Checked by ClamAV on apache.org ------=_NextPart_000_0049_01CE0B7D.F67D08D0 Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Hi Janesh, I think your SNN may be starting up with the wrong IP, I'm sure the machine parameter should say 192.168.0.101? http://namenode:50070/getimage?putimage=1 &port=50090&machine=0.0.0.0&token=-32:1989419481:0:1360849430000:13608491228 45 Are you able to retrieve the fsimage from the SNN from the command line? Using curl or wget: wget 'http://192.168.0.105:50070/getimage?getimage=1' -O fsimage.dmp If this actually retrieves an error page then the NN is reachable from the SNN and the port is definitely open. Otherwise double check that this is not due to the OS firewall blocking the connection, assuming it is on? That said the PrivilegedActionException in the error may actually mean it's Vijay From: janesh mishra [mailto:janeshmishra@gmail.com] Sent: 15 February 2013 12:27 To: user@hadoop.apache.org Subject: getimage failed in Name Node Log Hi, I am new in Hadoop and i set the hadoop cluster with the help of Michell Noll Multi-Node setup (http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi- node-cluster/). When i setup the single Node Hadoop then every things works fine. But in Multi Node setup i found that my fsimage and editlogs files are not updated on SNN, roll back of edit is done i have edit.new on NN Logs Form NN: 2013-02-14 19:13:52,468 ERROR org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException as:hduser cause:java.net.ConnectException: Connection refused 2013-02-14 19:13:52,468 ERROR org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException as:hduser cause:java.net.ConnectException: Connection refused 2013-02-14 19:13:52,477 WARN org.mortbay.log: /getimage: java.io.IOException: GetImage failed. java.net.ConnectException: Connection refused Logs From SNN: -------------- 2013-02-14 19:13:52,350 INFO org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: Posted URL namenode:50070putimage=1&port=50090&machine=0.0.0.0&token=32:1989419481:0:13 60849430000:1360849122845 2013-02-14 19:13:52,374 ERROR org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: Exception in doCheckpoint: 2013-02-14 19:13:52,375 ERROR org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: java.io.FileNotFoundException: http://namenode:50070/getimage?putimage=1 &port=50090&machine=0.0.0.0&token=-32:1989419481:0:1360849430000:13608491228 45 at sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection .java:1613) atorg.apache.hadoop.hdfs.server.namenode.TransferFsImage.getFileClient(Trans ferFsImage.java:160) at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.putFSImage(Secondar yNameNode.java:377) at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doCheckpoint(Second aryNameNode.java:418) at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doWork(SecondaryNam eNode.java:312) at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.run(SecondaryNameNo de.java:275) at java.lang.Thread.run(Thread.java:722) My setup includes Version : hadoop-1.0.4 1. Name Node (192.168.0.105) 2. Secondary Name Node (192.168.0.101) 3. Data Node (192.168.0.100) Name Node also works as Data Node. Conf File For Name Node: core-hdfs.xml ------------- hadoop.tmp.dir /app/hadoop/tmp A base for other temporary directories. fs.default.name hdfs://namenode:54310 The name of the default file system. A URI whose scheme and authority determine the FileSystem implementation. The uri's scheme determines the config property (fs.SCHEME.impl) naming the FileSystem implementation class. The uri's authority is used to determine the host, port, etc. for a filesystem. fs.checkpoint.period 300 The number of seconds between two periodic checkpoints. hdfs-site.xml ------------- dfs.replication 2 Default block replication. The actual number of replications can be specified when the file is created. The default is used if replication is not specified in create time. dfs.hosts /usr/local/hadoop/includehosts ips that works as datanode dfs.namenode.secondary.http-address secondarynamenode:50090 The address and the base port on which the dfs NameNode Web UI will listen. If the port is 0, the server will start on a free port. dfs.http.address namenode:50070 The address and the base port on which the dfs NameNode Web UI will listen. If the port is 0, the server will start on a free port. I sync these file to all my nodes. (I read somewhere in Cloud Era doc that all nodes should have same conf files). Please help me out. Thanks Janesh ------=_NextPart_000_0049_01CE0B7D.F67D08D0 Content-Type: text/html; charset="us-ascii" Content-Transfer-Encoding: quoted-printable

Hi Janesh,

 

I think your SNN may be starting up with the wrong IP, I’m sure = the machine parameter should say 192.168.0.101?

 

http://namenode:50070/getimage?putimage=3D1&port=3D50090&machine=3D0.0.0.0&token=3D-32:1989419481:0= :1360849430000:1360849122845

 

Are you able to retrieve the fsimage from the SNN from the command = line? Using curl or wget:

 

wget  'http://192.168.0.105:50070/getimage?getimage=3D1' -O = fsimage.dmp

 

If this actually retrieves an error page then the NN is reachable = from the SNN and the port is definitely open. Otherwise double check = that this is not due to the OS firewall blocking the connection, = assuming it is on?

 

That said the PrivilegedActionException in the error may actually = mean it’s

 

Vijay

 

From:= = janesh mishra [mailto:janeshmishra@gmail.com]
Sent: 15 = February 2013 12:27
To: = user@hadoop.apache.org
Subject: getimage failed in Name Node = Log

 

Hi,

I am new in Hadoop and = i set the hadoop cluster with the help of Michell Noll Multi-Node setup = (http://www.michael-noll.com/tutorials/running-ha= doop-on-ubuntu-linux-multi-node-cluster/). When i setup the single = Node Hadoop then every things works fine.

But in Multi Node = setup i found that my fsimage and editlogs files are not updated = on SNN, roll back of edit is done i have edit.new on NN =

Logs Form NN: =

2013-02-14 = 19:13:52,468 ERROR = org.apache.hadoop.security.UserGroupInformation: = PriviledgedActionException as:hduser cause:java.net.ConnectException: = Connection refused

2013-02-14 = 19:13:52,468 ERROR = org.apache.hadoop.security.UserGroupInformation: = PriviledgedActionException as:hduser cause:java.net.ConnectException: = Connection refused

2013-02-14 = 19:13:52,477 WARN org.mortbay.log: /getimage: java.io.IOException: = GetImage failed. java.net.ConnectException: Connection refused =

Logs From SNN: =

-------------- =

2013-02-14 = 19:13:52,350 INFO = org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: Posted URL = namenode:50070putimage=3D1&port=3D50090&machine=3D0.0.0.0&tok= en=3D32:1989419481:0:1360849430000:1360849122845

2013-02-14 = 19:13:52,374 ERROR = org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: Exception in = doCheckpoint:

2013-02-14 = 19:13:52,375 ERROR = org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: = java.io.FileNotFoundException:

http://namenode:50070/getimage?putimage=3D1&port=3D50090&machi= ne=3D0.0.0.0&token=3D-32:1989419481:0:1360849430000:1360849122845=

at = sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnect= ion.java:1613)

atorg.apache.hadoop.hdf= s.server.namenode.TransferFsImage.getFileClient(TransferFsImage.java:160)= at = org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.putFSImage(Secon= daryNameNode.java:377) at = org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doCheckpoint(Sec= ondaryNameNode.java:418) at = org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doWork(Secondary= NameNode.java:312) at = org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.run(SecondaryNam= eNode.java:275) at java.lang.Thread.run(Thread.java:722) =

My = setup includes

Version : hadoop-1.0.4 =

   1. Name = Node (192.168.0.105)

   2. = Secondary Name Node (192.168.0.101)

   3. Data = Node (192.168.0.100)

Name Node also works = as Data Node.

Conf File For Name = Node:

core-hdfs.xml =

------------- =

<?xml = version=3D"1.0"?>

<?xml-stylesheet = type=3D"text/xsl" href=3D"configuration.xsl"?> =

<!-- Put = site-specific property overrides in this file. -->

<configuration> =

<property> =

<name>hadoop.tmp.= dir</name>

<value>/app/hadoo= p/tmp</value>

<description>A = base for other temporary directories.</description> =

</property> =

<property> =

<name>fs.default.name</name> =

<value>hdfs://nam= enode:54310</value>

<description>The = name of the default file system. A URI whose

scheme and authority = determine the FileSystem implementation. The

uri's scheme = determines the config property (fs.SCHEME.impl) naming

the FileSystem = implementation class. The uri's authority is used to

determine the host, = port, etc. for a filesystem.</description>

</property> =

<property> =

<name>fs.checkpoi= nt.period</name>

<value>300</va= lue>

<description>The = number of seconds between two periodic checkpoints.

</description> =

</property> =

</configuration> =

hdfs-site.xml =

------------- =

<?xml = version=3D"1.0"?>

<?xml-stylesheet = type=3D"text/xsl" href=3D"configuration.xsl"?> =

<!-- Put = site-specific property overrides in this file. -->

<configuration> =

<property> =

<name>dfs.replica= tion</name>

<value>2</valu= e>

<description>Defa= ult block replication.

The actual number of = replications can be specified when the file is created. =

The = default is used if replication is not specified in create time. =

</description> =

</property> =

<property> =

<name>dfs.hosts&l= t;/name>

<value>/usr/local= /hadoop/includehosts</value>

<description>ips = that works as datanode</description>

</property> =

<property> =

<name>dfs.namenod= e.secondary.http-address</name>

<value>secondaryn= amenode:50090</value>

<description> =

The = address and the base port on which the dfs NameNode Web UI will listen. =

If = the port is 0, the server will start on a free port.

</description> =

</property> =

<property> =

<name>dfs.http.ad= dress</name>

<value>namenode:5= 0070</value>

<description> =

The = address and the base port on which the dfs NameNode Web UI will listen. =

If = the port is 0, the server will start on a free port.

</description> =

</property> =

</configuration> =

I = sync these file to all my nodes. (I read somewhere in Cloud Era doc that = all nodes should have same conf files).

Please help me out. =

Thanks =

Janesh =

------=_NextPart_000_0049_01CE0B7D.F67D08D0--