hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From praveenesh kumar <praveen...@gmail.com>
Subject Re: SecondaryNameNode not connecting to NameNode : PriviledgedActionException
Date Mon, 04 Jun 2012 12:01:38 GMT
Try giving value to dfs.secondary.http.address in hdfs-site.xml on your SNN.
In your logs, its starting SNN webserver at 0.0.0.0:50090. Its better if we
provide which IP it should start at.
Also I am assuming you are not having any firewalls enable between these 2
machines right ?

Regards,
Praveenesh

On Mon, Jun 4, 2012 at 5:05 PM, <ramon.pin@accenture.com> wrote:

> I configured dfs.http.address on SNN's hdfs-site.xml but still gets:
>
> /************************************************************
> STARTUP_MSG: Starting SecondaryNameNode
> STARTUP_MSG:   host = hadoop01/192.168.0.11
> STARTUP_MSG:   args = [-checkpoint, force]
> STARTUP_MSG:   version = 1.0.3
> STARTUP_MSG:   build =
> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
> 1335192; compiled by 'hortonfo' on Tue May  8 20:31:25 UTC 2012
> ************************************************************/
> 12/06/04 13:34:24 INFO namenode.SecondaryNameNode: Starting web server as:
> hadoop
> 12/06/04 13:34:24 INFO mortbay.log: Logging to
> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
> org.mortbay.log.Slf4jLog
> 12/06/04 13:34:24 INFO http.HttpServer: Added global filtersafety
> (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
> 12/06/04 13:34:24 INFO http.HttpServer: Port returned by
> webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening
> the listener on 50090
> 12/06/04 13:34:24 INFO http.HttpServer: listener.getLocalPort() returned
> 50090 webServer.getConnectors()[0].getLocalPort() returned 50090
> 12/06/04 13:34:24 INFO http.HttpServer: Jetty bound to port 50090
> 12/06/04 13:34:24 INFO mortbay.log: jetty-6.1.26
> 12/06/04 13:34:25 INFO mortbay.log: Started
> SelectChannelConnector@0.0.0.0:50090
> 12/06/04 13:34:25 INFO namenode.SecondaryNameNode: Web server init done
> 12/06/04 13:34:25 INFO namenode.SecondaryNameNode: Secondary Web-server up
> at: 0.0.0.0:50090
> 12/06/04 13:34:25 INFO namenode.SecondaryNameNode: Secondary image servlet
> up at: 0.0.0.0:50090
> 12/06/04 13:34:25 WARN namenode.SecondaryNameNode: Checkpoint Period
> :3600 secs (60 min)
> 12/06/04 13:34:25 WARN namenode.SecondaryNameNode: Log Size Trigger
>  :67108864 bytes (65536 KB)
> 12/06/04 13:34:25 ERROR security.UserGroupInformation:
> PriviledgedActionException as:hadoop cause:java.net.ConnectException:
> Connection refused
> 12/06/04 13:34:25 ERROR namenode.SecondaryNameNode: checkpoint: Connection
> refused
> 12/06/04 13:34:25 INFO namenode.SecondaryNameNode: SHUTDOWN_MSG:
> /************************************************************
> SHUTDOWN_MSG: Shutting down SecondaryNameNode at hadoop01/192.168.0.11
> ************************************************************/
>
> -----Original Message-----
> From: praveenesh kumar [mailto:praveenesh@gmail.com]
> Sent: lunes, 04 de junio de 2012 13:15
> To: common-user@hadoop.apache.org
> Subject: Re: SecondaryNameNode not connecting to NameNode :
> PriviledgedActionException
>
> I am not sure what could be the exact issue but when configuring secondary
> NN to NN, you need to tell your SNN where the actual NN resides.
> Try adding - dfs.http.address on your secondary namenode machine having
> value as <NN:port> on hdfs-site.xml Port should be on which your NN url is
> opening - means your NN web browser http port.
>
> Regards,
> Praveenesh
> On Mon, Jun 4, 2012 at 4:37 PM, <ramon.pin@accenture.com> wrote:
>
> > Hello. I'm facing a issue when trying to configure my
> > SecondaryNameNode on a different machine than my NameNode. When both
> > are on the same machine everything works fine but after moving the
> secondary to a new machine I get:
> >
> > 2012-05-28 09:57:36,832 ERROR
> > org.apache.hadoop.security.UserGroupInformation:
> > PriviledgedActionException as:hadoop cause:java.net.ConnectException:
> > Connection refused
> > 2012-05-28 09:57:36,832 ERROR
> > org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: Exception in
> > doCheckpoint:
> > 2012-05-28 09:57:36,834 ERROR
> > org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode:
> > java.net.ConnectException: Connection refused
> >        at java.net.PlainSocketImpl.socketConnect(Native Method)
> >        at
> >
> java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:327)
> >        at
> >
> java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:191)
> >        at
> >
> java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:180)
> >        at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:384)
> >        at java.net.Socket.connect(Socket.java:546)
> >        at java.net.Socket.connect(Socket.java:495)
> >        at sun.net.NetworkClient.doConnect(NetworkClient.java:178)
> >        at sun.net.www.http.HttpClient.openServer(HttpClient.java:409)
> >        at sun.net.www.http.HttpClient.openServer(HttpClient.java:530)
> >        at sun.net.www.http.HttpClient.<init>(HttpClient.java:240)
> >        at sun.net.www.http.HttpClient.New(HttpClient.java:321)
> >        at sun.net.www.http.HttpClient.New(HttpClient.java:338)
> >        at
> >
> sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:935)
> >        at
> >
> sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:876)
> >        at
> > sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.
> > java:801)
> >
> > Is there any configuration I'm missing? At this point my
> > mapred-site.xml is very simple just:
> >
> > <?xml version="1.0"?>
> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> > <configuration>  <property>
> >    <name>mapred.job.tracker</name>
> >    <value>hadoop00:9001</value>
> >  </property>
> >  <property>
> >    <name>mapred.system.dir</name>
> >    <value>/home/hadoop/mapred/system</value>
> >  </property>
> >  <property>
> >    <name>mapred.local.dir</name>
> >    <value>/home/hadoop/mapred/local</value>
> >  </property>
> >  <property>
> >    <name>mapred.jobtracker.taskScheduler</name>
> >    <value>org.apache.hadoop.mapred.FairScheduler</value>
> >  </property>
> >  <property>
> >    <name>mapred.fairscheduler.allocation.file</name>
> >    <value>/home/hadoop/hadoop/conf/fairscheduler.xml</value>
> >  </property>
> > </configuration>
> >
> >
> >
> > ________________________________
> > Subject to local law, communications with Accenture and its affiliates
> > including telephone calls and emails (including content), may be
> > monitored by our systems for the purposes of security and the
> > assessment of internal compliance with Accenture policy.
> >
> > ______________________________________________________________________
> > ________________
> >
> > www.accenture.com
> >
>
> ________________________________
> Subject to local law, communications with Accenture and its affiliates
> including telephone calls and emails (including content), may be monitored
> by our systems for the purposes of security and the assessment of internal
> compliance with Accenture policy.
>
> ______________________________________________________________________________________
>
> www.accenture.com
>
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message