hadoop-general mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From C J <xine....@googlemail.com>
Subject Re: My secondary namenode seem not be running, and may be the reason of my problem!!!
Date Fri, 03 Jul 2009 12:17:20 GMT
*Hallo Tim, and everyone,
I have  fixed the passphraseless, now each of the three machines can ssh the
others without a password. thanks to you I guess I moved one step forward
now the output of my  shell is the following:*

*My current status

**when starting the dfs*
alshain:~/Desktop/hadoop-0.18.3 # bin/start-dfs.sh
starting namenode, logging to
/root/Desktop/hadoop-0.18.3/logs/hadoop-root-namenode-alshain.out

134.130.222.17: starting datanode, logging to
/root/Desktop/hadoop-0.18.3/logs/hadoop-root-datanode-adhil.out

134.130.222.20: starting secondarynamenode, logging to
/root/Desktop/hadoop-0.18.3/logs/hadoop-root-secondarynamenode-alshain.out

134.130.222.18: starting secondarynamenode, logging to
/root/Desktop/hadoop-0.18.3/logs/hadoop-root-secondarynamenode-albali.out

*when starting the mapred*
alshain:~/Desktop/hadoop-0.18.3 # bin/start-mapred.sh
starting jobtracker, logging to
/root/Desktop/hadoop-0.18.3/logs/hadoop-root-jobtracker-alshain.out
1
34.130.222.17: starting tasktracker, logging to
/root/Desktop/hadoop-0.18.3/logs/hadoop-root-tasktracker-adhil.out

*Problem-1*

*when trying to list the content of the input folder*
alshain:~/Desktop/hadoop-0.18.3 # bin/hadoop dfs -ls
usr/tina/wordcount/input
09/07/03 14:00:21 INFO ipc.Client: Retrying connect to server:
alshain.mobnets.rwth-aachen.de/127.0.0.2:14000. Already tried 0 time(s).
09/07/03 14:00:22 INFO ipc.Client: Retrying connect to server:
alshain.mobnets.rwth-aachen.de/127.0.0.2:14000. Already tried 1 time(s).
09/07/03 14:00:23 INFO ipc.Client: Retrying connect to server:
alshain.mobnets.rwth-aachen.de/127.0.0.2:14000. Already tried 2 time(s).
09/07/03 14:00:24 INFO ipc.Client: Retrying connect to server:
alshain.mobnets.rwth-aachen.de/127.0.0.2:14000. Already tried 3 time(s).
09/07/03 14:00:25 INFO ipc.Client: Retrying connect to server:
alshain.mobnets.rwth-aachen.de/127.0.0.2:14000. Already tried 4 time(s).


*Problem 2
**
*alshain:~/Desktop/hadoop-0.18.3 # bin/stop-dfs.sh
no namenode to stop
134.130.222.17: stopping datanode
134.130.222.20: stopping secondarynamenode
134.130.222.18: stopping secondarynamenode

alshain:~/Desktop/hadoop-0.18.3 # bin/stop-mapred.sh
no jobtracker to stop
134.130.222.17: stopping tasktracker

This looks as if the namenode and the jobtracker were never started? any
idea why


*What I have checked

tried to ping the namenode using its name, it is successfull from any
machine

*alshain:~/Desktop/hadoop-0.18.3 # ping alshain.mobnets.rwth-aachen.de
PING alshain.mobnets.rwth-aachen.de (127.0.0.2) 56(84) bytes of data.
64 bytes from alshain.mobnets.rwth-aachen.de (127.0.0.2): icmp_seq=1 ttl=64
time=0.011 ms
64 bytes from alshain.mobnets.rwth-aachen.de (127.0.0.2): icmp_seq=2 ttl=64
time=0.010 ms
64 bytes from alshain.mobnets.rwth-aachen.de (127.0.0.2): icmp_seq=3 ttl=64
time=0.009 ms
--- alshain.mobnets.rwth-aachen.de ping statistics ---
3 packets transmitted, 3 received, 0% packet loss, time 1998ms
rtt min/avg/max/mdev = 0.009/0.010/0.011/0.000 ms

*tried to ping the namenode using its ip address, it is successful from any
machine

*alshain:~/Desktop/hadoop-0.18.3 # ping 134.130.222.20
PING 134.130.222.20 (134.130.222.20) 56(84) bytes of data.
64 bytes from 134.130.222.20: icmp_seq=1 ttl=64 time=0.038 ms
64 bytes from 134.130.222.20: icmp_seq=2 ttl=64 time=0.009 ms
64 bytes from 134.130.222.20: icmp_seq=3 ttl=64 time=0.008 ms
64 bytes from 134.130.222.20: icmp_seq=4 ttl=64 time=0.009 ms
64 bytes from 134.130.222.20: icmp_seq=5 ttl=64 time=0.010 ms

--- 134.130.222.20 ping statistics ---
5 packets transmitted, 5 received, 0% packet loss, time 3997ms
rtt min/avg/max/mdev = 0.008/0.014/0.038/0.012 ms



*Can anyone tell me why the namenode and the jobtracker were never started?
I appreciate your help,
CJ
*

On Fri, Jul 3, 2009 at 11:26 AM, tim robertson <timrobertson100@gmail.com>wrote:

> "at this stage the shell is blocked unless I press the enter key" - my
> cluster did this until I realised passphraseless ssh was not working
> between the master and the nodes and the master to itself.
> http://fak3r.com/2006/08/10/howto-passwordless-ssh-logins/
>
> Cheers,
>
> Tim
>
>
>
> On Fri, Jul 3, 2009 at 11:21 AM, C J<xine.jar@googlemail.com> wrote:
> > Hallo everyone,
> > I have installed the hadoop 0.18.3 on three linux machines, I am trying
> to
> > run the
> > example of WordCountv1.0 on a cluster. But I guess I have a problem
> > somewhere.
> > *
> > Problem*
> >
> > *After formating the name node:*
> > I am getting several STARTUP_MSG and at the end a "SHUTDOWN_MSG: shutting
> > down the namenode..."
> > is this normal?
> >
> > *Afterwards I try starting the dfs:**
> > *I get a message "starting namenode..."* *afterwards I get another
> message
> > "starting secondary namenode"
> > at this stage the shell is blocked unless I press the enter key. Then the
> > system tries to start another secondary
> > namenode and the shell is then not blocked. What is going on?
> > *
> > Then I proceed an try starting the mapred:*
> > I get the two messages "starting jobtracker....." and "starting
> tasktracker"
> >
> > *Following the tutorial for runnin WordCode v1.0, If I try to list the
> files
> > in the input folder I have created
> > *I get the famous error "Retrying connect to server:134.130.222.20:9000"
> > .what am I doing something wrong?
> >
> >
> > *Steps I have already verified*
> >
> > *I have already checked the iptables of the three machines and they
> look**:
> > **
> > *Chain INPUT (policy ACCEPT)
> > target         prot opt source destiantion
> >
> > Chain FORWARD (policy ACCEPT)
> > target         prot opt source destination
> >
> > Chain OUTPUT (policy ACCEPT)
> > target        prot opt source destination
> >
> > *
> > My hadoop-site.xml file looks like this
> > * configuration>
> >    <property>
> >        <name>fs.default.name</name>
> >        <value>134.130.222.20:9000/</value>
> >    </property>
> >    <property>
> >        <name>mapred.job.tracker</name>
> >        <value>134.130.222.18:9001</value>
> >    </property>
> >    ........
> > </configuration>
> >
> > *Can someone help me out?*
> > *Thank you, CJ*
> >
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message