hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jagadesh_Doddi <Jagadesh_Do...@satyam.com>
Subject RE: Reducer hangs at 16%
Date Tue, 24 Feb 2009 09:28:20 GMT
I have opened the ports 50010, 50030, 50060, 50070, 50075 and 50090
It works fine now. Thanks Matie.

Thanks
Jagadesh Doddi


-----Original Message-----
From: Matei Zaharia [mailto:matei@cloudera.com]
Sent: Monday, February 23, 2009 10:06 PM
To: core-user@hadoop.apache.org
Subject: Re: Reducer hangs at 16%

The fact that it works with one slave node doesn't mean much, because when
the slave is alone, it's copying map outputs from itself and thus not going
through the firewall. It sounds like the slaves can't open a connection to
each other, which could well mean a firewall problem. Can you look at the
output of the reduce task (by clicking it in the "running tasks" column in
the web UI and going on to see the last 8k of output)? I imagine it will
have fetched data from one slave and will be failing to connect to the other
one.

On Mon, Feb 23, 2009 at 5:03 AM, Jagadesh_Doddi
<Jagadesh_Doddi>wrote:

> It works as longs as I use any one of the slave nodes.
> The moment I add both the slave nodes to conf/slaves, It fails.
> So there is no issue with firewall or /etc/hosts entries.
>
> Thanks and Regards
>
> Jagadesh Doddi
>
>
>
> -----Original Message-----
> From: Amar Kamat [mailto:amarrk@yahoo-inc.com]
> Sent: Monday, February 23, 2009 6:26 PM
> To: core-user@hadoop.apache.org
> Subject: Re: Reducer hangs at 16%
>
> Looks like the reducer is able to fetch map output files from the local
> box but fails to fetch it from the remote box. Can you check if there is
> no firewall issue or /etc/hosts entries are correct?
> Amar
> Jagadesh_Doddi wrote:
> > Hi
> >
> > I have changed the configuration to run Name node and job tracker on the
> same system.
> > The job is started with bin/start-all.sh on NN
> > With a single slave node, the job completes in 12 seconds, and the
> console output is shown below:
> >
> > [root@Fedora1 hadoop-0.18.3]# bin/hadoop jar samples/wordcount.jar
> org.myorg.WordCount input output1
> > 09/02/23 17:19:30 WARN mapred.JobClient: Use GenericOptionsParser for
> parsing the arguments. Applications should implement Tool for the same.
> > 09/02/23 17:19:30 INFO mapred.FileInputFormat: Total input paths to
> process : 1
> > 09/02/23 17:19:30 INFO mapred.FileInputFormat: Total input paths to
> process : 1
> > 09/02/23 17:19:30 INFO mapred.JobClient: Running job:
> job_200902231717_0001
> > 09/02/23 17:19:31 INFO mapred.JobClient:  map 0% reduce 0%
> > 09/02/23 17:19:37 INFO mapred.JobClient:  map 100% reduce 0%
> > 09/02/23 17:19:42 INFO mapred.JobClient: Job complete:
> job_200902231717_0001
> > 09/02/23 17:19:42 INFO mapred.JobClient: Counters: 16
> > 09/02/23 17:19:42 INFO mapred.JobClient:   Job Counters
> > 09/02/23 17:19:42 INFO mapred.JobClient:     Data-local map tasks=2
> > 09/02/23 17:19:42 INFO mapred.JobClient:     Launched reduce tasks=1
> > 09/02/23 17:19:42 INFO mapred.JobClient:     Launched map tasks=2
> > 09/02/23 17:19:42 INFO mapred.JobClient:   Map-Reduce Framework
> > 09/02/23 17:19:42 INFO mapred.JobClient:     Map output records=25
> > 09/02/23 17:19:42 INFO mapred.JobClient:     Reduce input records=23
> > 09/02/23 17:19:42 INFO mapred.JobClient:     Map output bytes=238
> > 09/02/23 17:19:42 INFO mapred.JobClient:     Map input records=5
> > 09/02/23 17:19:42 INFO mapred.JobClient:     Combine output records=46
> > 09/02/23 17:19:42 INFO mapred.JobClient:     Map input bytes=138
> > 09/02/23 17:19:42 INFO mapred.JobClient:     Combine input records=48
> > 09/02/23 17:19:42 INFO mapred.JobClient:     Reduce input groups=23
> > 09/02/23 17:19:42 INFO mapred.JobClient:     Reduce output records=23
> > 09/02/23 17:19:42 INFO mapred.JobClient:   File Systems
> > 09/02/23 17:19:42 INFO mapred.JobClient:     HDFS bytes written=175
> > 09/02/23 17:19:42 INFO mapred.JobClient:     Local bytes written=648
> > 09/02/23 17:19:42 INFO mapred.JobClient:     HDFS bytes read=208
> > 09/02/23 17:19:42 INFO mapred.JobClient:     Local bytes read=281
> >
> > With two slave nodes, the job completes in 13 minutes, and the console
> output is shown below:
> >
> > [root@Fedora1 hadoop-0.18.3]# bin/hadoop jar samples/wordcount.jar
> org.myorg.WordCount input output2
> > 09/02/23 17:25:38 WARN mapred.JobClient: Use GenericOptionsParser for
> parsing the arguments. Applications should implement Tool for the same.
> > 09/02/23 17:25:38 INFO mapred.FileInputFormat: Total input paths to
> process : 1
> > 09/02/23 17:25:38 INFO mapred.FileInputFormat: Total input paths to
> process : 1
> > 09/02/23 17:25:39 INFO mapred.JobClient: Running job:
> job_200902231722_0001
> > 09/02/23 17:25:40 INFO mapred.JobClient:  map 0% reduce 0%
> > 09/02/23 17:25:42 INFO mapred.JobClient:  map 50% reduce 0%
> > 09/02/23 17:25:43 INFO mapred.JobClient:  map 100% reduce 0%
> > 09/02/23 17:25:58 INFO mapred.JobClient:  map 100% reduce 16%
> > 09/02/23 17:38:31 INFO mapred.JobClient: Task Id :
> attempt_200902231722_0001_m_000000_0, Status : FAILED
> > Too many fetch-failures
> > 09/02/23 17:38:31 WARN mapred.JobClient: Error reading task outputNo
> route to host
> > 09/02/23 17:38:31 WARN mapred.JobClient: Error reading task outputNo
> route to host
> > 09/02/23 17:38:43 INFO mapred.JobClient: Job complete:
> job_200902231722_0001
> > 09/02/23 17:38:43 INFO mapred.JobClient: Counters: 16
> > 09/02/23 17:38:43 INFO mapred.JobClient:   Job Counters
> > 09/02/23 17:38:43 INFO mapred.JobClient:     Data-local map tasks=3
> > 09/02/23 17:38:43 INFO mapred.JobClient:     Launched reduce tasks=1
> > 09/02/23 17:38:43 INFO mapred.JobClient:     Launched map tasks=3
> > 09/02/23 17:38:43 INFO mapred.JobClient:   Map-Reduce Framework
> > 09/02/23 17:38:43 INFO mapred.JobClient:     Map output records=25
> > 09/02/23 17:38:43 INFO mapred.JobClient:     Reduce input records=23
> > 09/02/23 17:38:43 INFO mapred.JobClient:     Map output bytes=238
> > 09/02/23 17:38:43 INFO mapred.JobClient:     Map input records=5
> > 09/02/23 17:38:43 INFO mapred.JobClient:     Combine output records=46
> > 09/02/23 17:38:43 INFO mapred.JobClient:     Map input bytes=138
> > 09/02/23 17:38:43 INFO mapred.JobClient:     Combine input records=48
> > 09/02/23 17:38:43 INFO mapred.JobClient:     Reduce input groups=23
> > 09/02/23 17:38:43 INFO mapred.JobClient:     Reduce output records=23
> > 09/02/23 17:38:43 INFO mapred.JobClient:   File Systems
> > 09/02/23 17:38:43 INFO mapred.JobClient:     HDFS bytes written=175
> > 09/02/23 17:38:43 INFO mapred.JobClient:     Local bytes written=648
> > 09/02/23 17:38:43 INFO mapred.JobClient:     HDFS bytes read=208
> > 09/02/23 17:38:43 INFO mapred.JobClient:     Local bytes read=281
> >
> > Thanks
> >
> > Jagadesh
> >
> >
> >
> > -----Original Message-----
> > From: Jothi Padmanabhan [mailto:jothipn@yahoo-inc.com]
> > Sent: Monday, February 23, 2009 4:57 PM
> > To: core-user@hadoop.apache.org
> > Subject: Re: Reducer hangs at 16%
> >
> > OK. I am guessing that your problem arises from having two entries for
> > master. The master should be the node where the JT is run (for
> > start-mapred.sh) and NN is run (for start-dfs.sh). This might need a bit
> > more effort to set up. To start with, you might want to try out having
> both
> > the JT and NN in the same machine (the node designated as master) and
> then
> > try start-all.sh. You need to configure you hadoop-site.xml correctly as
> > well.
> >
> > Jothi
> >
> >
> >
> >
> > On 2/23/09 4:36 PM, "Jagadesh_Doddi" <Jagadesh_Doddi@satyam.com> wrote:
> >
> >
> >> Hi
> >>
> >> I have setup as per the documentation in hadoop site.
> >> On namenode, I am running bin/start-dfs.sh and on job tracker, I am
> running
> >> bin\start-mapred.sh
> >>
> >> Thanks and Regards
> >>
> >> Jagadesh Doddi
> >>
> >>
> >>
> >> -----Original Message-----
> >> From: Jothi Padmanabhan [mailto:jothipn@yahoo-inc.com]
> >> Sent: Monday, February 23, 2009 4:00 PM
> >> To: core-user@hadoop.apache.org
> >> Subject: Re: Reducer hangs at 16%
> >>
> >> Hi,
> >>
> >> This looks like a set up issue. See
> >>
> http://hadoop.apache.org/core/docs/current/cluster_setup.html#Configuration+
> >> Files
> >> On how to set this up correctly.
> >>
> >> As an aside, how are you bringing up the hadoop daemons (JobTracker,
> >> Namenode, TT and Datanodes)?  Are you manually bringing them up or are
> you
> >> using bin/start-all.sh?
> >>
> >> Jothi
> >>
> >>
> >> On 2/23/09 3:14 PM, "Jagadesh_Doddi" <Jagadesh_Doddi@satyam.com> wrote:
> >>
> >>
> >>> I have setup a distributed environment on Fedora OS to run Hadoop.
> >>> System Fedora1 is the name node, Fedora2 is Job tracker, Fedora3 and
> Fedora4
> >>> are task trackers.
> >>> Conf/masters contains the entries Fedora1, Fedors2, and conf/slaves
> contains
> >>> the entries Fedora3, Fedora4.
> >>> When I run the sample wordcount example with single task tracker
> (either
> >>> Fedora3 or Fedora4), it works fine and the job completes in a few
> seconds.
> >>> However, when I add the other task tracker in conf/slaves, the reducer
> stop
> >>> at
> >>> 16% and the job completes after 13 minutes.
> >>> The same problem exists in versions 16.4, 17.2.1 and 18.3. The output
> on the
> >>> namenode console is shown below:
> >>>
> >>> [root@Fedora1 hadoop-0.17.2.1Cluster]# bin/hadoop jar
> samples/wordcount.jar
> >>> org.myorg.WordCount input output
> >>> 09/02/19 17:43:18 INFO mapred.FileInputFormat: Total input paths to
> process :
> >>> 1
> >>> 09/02/19 17:43:19 INFO mapred.JobClient: Running job:
> job_200902191741_0001
> >>> 09/02/19 17:43:20 INFO mapred.JobClient:  map 0% reduce 0%
> >>> 09/02/19 17:43:26 INFO mapred.JobClient:  map 50% reduce 0%
> >>> 09/02/19 17:43:27 INFO mapred.JobClient:  map 100% reduce 0%
> >>> 09/02/19 17:43:35 INFO mapred.JobClient:  map 100% reduce 16%
> >>> 09/02/19 17:56:15 INFO mapred.JobClient: Task Id :
> >>> task_200902191741_0001_m_000001_0, Status : FAILED
> >>> Too many fetch-failures
> >>> 09/02/19 17:56:15 WARN mapred.JobClient: Error reading task outputNo
> route to
> >>> host
> >>> 09/02/19 17:56:18 WARN mapred.JobClient: Error reading task outputNo
> route to
> >>> host
> >>> 09/02/19 17:56:25 INFO mapred.JobClient:  map 100% reduce 81%
> >>> 09/02/19 17:56:26 INFO mapred.JobClient:  map 100% reduce 100%
> >>> 09/02/19 17:56:27 INFO mapred.JobClient: Job complete:
> job_200902191741_0001
> >>> 09/02/19 17:56:27 INFO mapred.JobClient: Counters: 16
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:   Job Counters
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     Launched map tasks=3
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     Launched reduce tasks=1
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     Data-local map tasks=3
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:   Map-Reduce Framework
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     Map input records=5
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     Map output records=25
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     Map input bytes=138
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     Map output bytes=238
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     Combine input records=25
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     Combine output records=23
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     Reduce input groups=23
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     Reduce input records=23
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     Reduce output records=23
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:   File Systems
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     Local bytes read=522
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     Local bytes written=1177
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     HDFS bytes read=208
> >>> 09/02/19 17:56:27 INFO mapred.JobClient:     HDFS bytes written=175
> >>>
> >>> Appreciate any help on this.
> >>>
> >>> Thanks
> >>>
> >>> Jagadesh
> >>>
> >>> DISCLAIMER:
> >>> This email (including any attachments) is intended for the sole use of
> the
> >>> intended recipient/s and may contain material that is CONFIDENTIAL AND
> >>> PRIVATE
> >>> COMPANY INFORMATION. Any review or reliance by others or copying or
> >>> distribution or forwarding of any or all of the contents in this
> message is
> >>> STRICTLY PROHIBITED. If you are not the intended recipient, please
> contact
> >>> the
> >>> sender by email and delete all copies; your cooperation in this regard
> is
> >>> appreciated.
> >>>
> >>
> >> DISCLAIMER:
> >> This email (including any attachments) is intended for the sole use of
> the
> >> intended recipient/s and may contain material that is CONFIDENTIAL AND
> PRIVATE
> >> COMPANY INFORMATION. Any review or reliance by others or copying or
> >> distribution or forwarding of any or all of the contents in this message
> is
> >> STRICTLY PROHIBITED. If you are not the intended recipient, please
> contact the
> >> sender by email and delete all copies; your cooperation in this regard
> is
> >> appreciated.
> >>
> >
> >
> >
> > DISCLAIMER:
> > This email (including any attachments) is intended for the sole use of
> the intended recipient/s and may contain material that is CONFIDENTIAL AND
> PRIVATE COMPANY INFORMATION. Any review or reliance by others or copying or
> distribution or forwarding of any or all of the contents in this message is
> STRICTLY PROHIBITED. If you are not the intended recipient, please contact
> the sender by email and delete all copies; your cooperation in this regard
> is appreciated.
> >
>
>
>
> DISCLAIMER:
> This email (including any attachments) is intended for the sole use of the
> intended recipient/s and may contain material that is CONFIDENTIAL AND
> PRIVATE COMPANY INFORMATION. Any review or reliance by others or copying or
> distribution or forwarding of any or all of the contents in this message is
> STRICTLY PROHIBITED. If you are not the intended recipient, please contact
> the sender by email and delete all copies; your cooperation in this regard
> is appreciated.
>

DISCLAIMER:
This email (including any attachments) is intended for the sole use of the intended recipient/s
and may contain material that is CONFIDENTIAL AND PRIVATE COMPANY INFORMATION. Any review
or reliance by others or copying or distribution or forwarding of any or all of the contents
in this message is STRICTLY PROHIBITED. If you are not the intended recipient, please contact
the sender by email and delete all copies; your cooperation in this regard is appreciated.

Mime
View raw message