hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From ashish pareek <pareek...@gmail.com>
Subject Problem in viewing WEB UI
Date Wed, 17 Jun 2009 05:42:38 GMT
Hi,

      When I run command *bin/hadoop dfsadmin -report *it shows that 2
datanodes are alive but when i try to http://hadoopmster:50070/ but the
problem is that it opens doesnot opne
http://hadoopmaster:50070/dfshealth.jsp page and throws *error HTTP: 404 .
So why is't happening like this?
*
Regards,
Ashish Pareek


 On Wed, Jun 17, 2009 at 10:06 AM, Sugandha Neaolekar <
sugandha.n87@gmail.com> wrote:

> Well, You just have to specify the address in the URL address bar as::
> http://hadoopmaster:50070 U'll be able to see the web UI..!
>
>
> On Tue, Jun 16, 2009 at 7:17 PM, ashish pareek <pareekash@gmail.com>wrote:
>
>> HI Sugandha,
>>                    Hmmm........ your suggestion helped and Now I am able
>> to run two datanode ....one on the same machine as name node and other on
>> the different machine ........Thanks a lot :)
>>
>>                  But the problem is now I am not able to see web UI .....
>> for  both datanode and as well as name node....
>> should I have to consider some more things in the site.xml ? if so please
>> help...
>>
>> Thanking you again,
>> regards,
>> Ashish Pareek.
>>
>> On Tue, Jun 16, 2009 at 3:10 PM, Sugandha Naolekar <
>> sugandha.n87@gmail.com> wrote:
>>
>>> hi,,!
>>>
>>>
>>> First of all, get your concepts clear of hadoop.
>>> You can refer to the following
>>>
>>> site::
>>> http://www.google.co.in/url?sa=t&source=web&ct=res&cd=1&url=http%3A%2F%2Fwww.michael-noll.com%2Fwiki%2FRunning_Hadoop_On_Ubuntu_Linux_(Single-Node_Cluster)&ei=lGU3Spv2FZbLjAe19KmiDQ&usg=AFQjCNFbmVGsoChOSMzCB3tRhoV0ylHOzA&sig2=t2AJ_nf24SFtveN4PHS_TA<http://www.google.co.in/url?sa=t&source=web&ct=res&cd=1&url=http%3A%2F%2Fwww.michael-noll.com%2Fwiki%2FRunning_Hadoop_On_Ubuntu_Linux_%28Single-Node_Cluster%29&ei=lGU3Spv2FZbLjAe19KmiDQ&usg=AFQjCNFbmVGsoChOSMzCB3tRhoV0ylHOzA&sig2=t2AJ_nf24SFtveN4PHS_TA>
>>>
>>>
>>> I have small doubt whether in the mater.xml and slave.xml we can have
>>> same port numbers to both of them like
>>>
>>>
>>> for slave :::::
>>>
>>> <property>
>>>     <name>fs.default.name</name>
>>>     <value>hdfs://hadoopslave:
>>>>
>>>> 9000</value>
>>>>   </property>
>>>>
>>>>
>>>>  for master:::
>>>>
>>>> <property>
>>>>     <name>fs.default.name</name>
>>>>     <value>hdfs://hadoopmaster:9000</value>
>>>>   </property>
>>>>
>>>>
>>>
>>> Well, any  two daemons or services can run on the same port unless, they
>>> are not run on the same machine.If you wish to run DN and NN on the same
>>> machine, their port numbers have to be different.
>>>
>>>
>>>
>>>
>>> On Tue, Jun 16, 2009 at 2:55 PM, ashish pareek <pareekash@gmail.com>wrote:
>>>
>>>> HI sugandha,
>>>>
>>>>
>>>>
>>>> and one more thing can we have in slave:::
>>>>
>>>> <property>
>>>>   <name>dfs.datanode.address</
>>>>>
>>>>> name>
>>>>>   <value>hadoopmaster:9000</value>
>>>>>     <value>hadoopslave:9001</value>
>>>>>   </property>
>>>>>
>>>>
>>>
>>> Also, fs,default.name is the tag which specifies the default filesystem.
>>> And generaLLY, it is run on namenode. So, it;s value has to be a namenode's
>>> address only and not slave's.
>>>
>>>
>>>>
>>>> Else if you have complete procedure for installing and running Hadoop in
>>>> cluster can you please send it to me ...... I need to step up hadoop with
in
>>>> two days and show it to my guide.Currently I am doing my masters.
>>>>
>>>> Thanks for your spending time
>>>
>>>
>>> Try for the above, and this should work!
>>>
>>>>
>>>>
>>>> regards,
>>>> Ashish Pareek
>>>>
>>>>
>>>> On Tue, Jun 16, 2009 at 2:33 PM, Sugandha Naolekar <
>>>> sugandha.n87@gmail.com> wrote:
>>>>
>>>>> Following changes are to be done::
>>>>>
>>>>> Under master folder::
>>>>>
>>>>> -> put slaves address as well under the values of
>>>>> tag(dfs.datanode.address)
>>>>>
>>>>> -> You want to make namenode as datanode as well. As per your config
>>>>> file, you have specified hadoopmaster  in your slave file. If you don't
want
>>>>> that, remove ti from slaves file.
>>>>>
>>>>> UNder slave folder::
>>>>>
>>>>> -> put only slave's (the m/c where you intend to run your datanode)'s
>>>>> address.under datanode.address tag. Else
>>>>> it should go as such::
>>>>>
>>>>> <property>
>>>>>   <name>dfs.datanode.address</name>
>>>>>   <value>hadoopmaster:9000</value>
>>>>>     <value>hadoopslave:9001</value>
>>>>>   </property>
>>>>>
>>>>> Also, your port numbers hould be different. the daemons NN,DN,JT,TT
>>>>> should run independently on different ports.
>>>>>
>>>>>
>>>>> On Tue, Jun 16, 2009 at 2:05 PM, Sugandha Naolekar <
>>>>> sugandha.n87@gmail.com> wrote:
>>>>>
>>>>>>
>>>>>>
>>>>>> ---------- Forwarded message ----------
>>>>>> From: ashish pareek <pareekash@gmail.com>
>>>>>> Date: Tue, Jun 16, 2009 at 2:00 PM
>>>>>> Subject: Re: org.apache.hadoop.ipc.client : trying connect to server
>>>>>> failed
>>>>>> To: Sugandha Naolekar <sugandha.n87@gmail.com>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, Jun 16, 2009 at 1:58 PM, ashish pareek <pareekash@gmail.com>wrote:
>>>>>>
>>>>>>> HI ,
>>>>>>>      I am sending .tar.gz format containing both master and datanode
>>>>>>> config files .......
>>>>>>>
>>>>>>> Regards,
>>>>>>> Ashish Pareek
>>>>>>>
>>>>>>>
>>>>>>> On Tue, Jun 16, 2009 at 1:47 PM, Sugandha Naolekar <
>>>>>>> sugandha.n87@gmail.com> wrote:
>>>>>>>
>>>>>>>> can u pls send me a zip or a tar file? I don't have windows
systems
>>>>>>>> but, linux
>>>>>>>>
>>>>>>>>
>>>>>>>> On Tue, Jun 16, 2009 at 1:19 PM, ashish pareek <pareekash@gmail.com
>>>>>>>> > wrote:
>>>>>>>>
>>>>>>>>> HI Sungandha ,
>>>>>>>>>                       Thanks for your reply .... I am
sending you
>>>>>>>>> master and slave configuration files if you can go through
it and tell me
>>>>>>>>> where I am going wrong it would be helpful.
>>>>>>>>>
>>>>>>>>>                         Hope to get a reply soon ...........
Thanks
>>>>>>>>> again!
>>>>>>>>>
>>>>>>>>> Regards,
>>>>>>>>> Ashish Pareek
>>>>>>>>>
>>>>>>>>> On Tue, Jun 16, 2009 at 11:12 AM, Sugandha Naolekar <
>>>>>>>>> sugandha.n87@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hi Ashish!
>>>>>>>>>>
>>>>>>>>>> Try for the following things::
>>>>>>>>>>
>>>>>>>>>> -> Check the config file(hadoop-site.xml) of namenode.
>>>>>>>>>> -> Make sure, the tag(dfs.datanode.addres)'s value
you have given
>>>>>>>>>> correctly
>>>>>>>>>> it's IP,and the name of that machine.
>>>>>>>>>> -> Also, check for the name added in /etc/hosts
file.
>>>>>>>>>> -> Check for the ssh keys of datanodes present
in namenode's
>>>>>>>>>> known_hosts
>>>>>>>>>> file
>>>>>>>>>> -> check for the value of dfs.datanode.addres
on datanode's config
>>>>>>>>>> file.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Tue, Jun 16, 2009 at 10:58 AM, ashish pareek <
>>>>>>>>>> pareekash@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>> > HI ,
>>>>>>>>>> >     I am trying to step up a hadoop cluster
on 3GB machine and
>>>>>>>>>> using hadoop
>>>>>>>>>> > 0.18.3 and  have followed procedure given in
 apache hadoop site
>>>>>>>>>> for hadoop
>>>>>>>>>> > cluster.
>>>>>>>>>> >     In conf/slaves I have added two datanode
i.e including the
>>>>>>>>>> namenode
>>>>>>>>>> > vitrual machine and other machine virtual machine
(datanode)
>>>>>>>>>>  ..... and
>>>>>>>>>> > have
>>>>>>>>>> > set up passwordless ssh between both virtual
machines ..... But
>>>>>>>>>> now problem
>>>>>>>>>> > is when I run command :
>>>>>>>>>> >
>>>>>>>>>> > bin/hadoop start-all.sh
>>>>>>>>>> >
>>>>>>>>>> > It start only one datanode on the same namenode
vitrual machine
>>>>>>>>>> but it
>>>>>>>>>> > doesn't start the datanode on other machine.....
>>>>>>>>>> >
>>>>>>>>>> > in logs/hadoop-datanode.log  i get message
>>>>>>>>>> >
>>>>>>>>>> >
>>>>>>>>>> >  INFO org.apache.hadoop.ipc.Client: Retrying
>>>>>>>>>> >  connect to server: hadoop1/192.168.1.28:9000.
Already
>>>>>>>>>> >
>>>>>>>>>> >  tried 1 time(s).
>>>>>>>>>> >
>>>>>>>>>> >  2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client:
>>>>>>>>>> Retrying
>>>>>>>>>> >  connect to server: hadoop1/192.168.1.28:9000.
Already tried 2
>>>>>>>>>> time(s).
>>>>>>>>>> >
>>>>>>>>>> >  2009-05-09 18:35:14,266 INFO org.apache.hadoop.ipc.Client:
>>>>>>>>>> Retrying
>>>>>>>>>> >  connect to server: hadoop1/192.168.1.28:9000.
Already tried 3
>>>>>>>>>> time(s).
>>>>>>>>>> >
>>>>>>>>>> >
>>>>>>>>>> > .
>>>>>>>>>> > .
>>>>>>>>>> > .
>>>>>>>>>> > .
>>>>>>>>>> > .
>>>>>>>>>> > .
>>>>>>>>>> > .
>>>>>>>>>> > .
>>>>>>>>>> > .
>>>>>>>>>> >
>>>>>>>>>> > .
>>>>>>>>>> > .
>>>>>>>>>> >
>>>>>>>>>> > .
>>>>>>>>>> >
>>>>>>>>>> > I have tried formatting and start the cluster
again .....but
>>>>>>>>>> still I
>>>>>>>>>> > get the same error.
>>>>>>>>>> >
>>>>>>>>>> > So can any one help in solving this problem.
:)
>>>>>>>>>> >
>>>>>>>>>> > Thanks
>>>>>>>>>> >
>>>>>>>>>> > Regards
>>>>>>>>>> >
>>>>>>>>>> > Ashish Pareek
>>>>>>>>>> >
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Regards!
>>>>>>>>>> Sugandha
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Regards!
>>>>>>>> Sugandha
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Regards!
>>>>>> Sugandha
>>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Regards!
>>>>> Sugandha
>>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> Regards!
>>> Sugandha
>>>
>>
>>
>
>
> --
> Regards!
> Sugandha
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message