Return-Path: X-Original-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 4E170974C for ; Mon, 7 May 2012 16:24:36 +0000 (UTC) Received: (qmail 31997 invoked by uid 500); 7 May 2012 16:24:34 -0000 Delivered-To: apmail-hadoop-mapreduce-user-archive@hadoop.apache.org Received: (qmail 31934 invoked by uid 500); 7 May 2012 16:24:34 -0000 Mailing-List: contact mapreduce-user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: mapreduce-user@hadoop.apache.org Delivered-To: mailing list mapreduce-user@hadoop.apache.org Received: (qmail 31925 invoked by uid 99); 7 May 2012 16:24:34 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 07 May 2012 16:24:34 +0000 X-ASF-Spam-Status: No, hits=1.7 required=5.0 tests=FREEMAIL_ENVFROM_END_DIGIT,HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of lordjoe2000@gmail.com designates 209.85.160.176 as permitted sender) Received: from [209.85.160.176] (HELO mail-gy0-f176.google.com) (209.85.160.176) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 07 May 2012 16:24:29 +0000 Received: by ghbz10 with SMTP id z10so3598850ghb.35 for ; Mon, 07 May 2012 09:24:08 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type; bh=KVibrxHTwHVKg26u78i74wpp5S/uPqFvsqKKWY+EdQo=; b=hEL9QaeQfadoA5pYdHkeWdALT3P2/RHiEFFMVp2dyMEFSX21BIcR59QZOoLOKSOCZm 2/i59W/DmrO6K/y3cFep3k5iMvvvVv7yjLX/CDHSMZjBldtl6sz/tSDCmIWU4osum3PQ nMJJapqIUmte6asTbnB6Oj9pfiHmiWmQiz6aa5D7QhUmaO6E49SgH4hkeEO3bmORPuLk 8qO1MteUt3bEFcbb1OFk4DOanvSueLuaeDe4+QBMizzEmWtZZB3AiYy6rHBz+fq34DeJ OweUENbBqmOxBsgrf9SAD0J7tT+u6Ocj4vYJpE5M4/Mc9wSKYH+MwPZdOsiD0Y3VfFh+ 2pfw== MIME-Version: 1.0 Received: by 10.60.169.174 with SMTP id af14mr22194320oec.13.1336407845694; Mon, 07 May 2012 09:24:05 -0700 (PDT) Received: by 10.182.169.39 with HTTP; Mon, 7 May 2012 09:24:05 -0700 (PDT) In-Reply-To: References: Date: Mon, 7 May 2012 09:24:05 -0700 Message-ID: Subject: Re: Ant Colony Optimization for Travelling Salesman Problem in Hadoop From: Steve Lewis To: mapreduce-user@hadoop.apache.org Content-Type: multipart/alternative; boundary=bcaec54c5392eaed1804bf74b29f X-Virus-Checked: Checked by ClamAV on apache.org --bcaec54c5392eaed1804bf74b29f Content-Type: text/plain; charset=ISO-8859-1 Fair enough - I write a lot of InputFormats since for most of my problems a line of text is not the proper unit - I read fasta files - read lines intil you hit a line starting with > and xml fragments - read until you hit a closing tag On Mon, May 7, 2012 at 9:03 AM, GUOJUN Zhu wrote: > > The default FileInputformat split the file according to the size. If you > use line text data, the TextFileInputformat respects the line structure for > input. We got splits as small as a few KBs. The file split is a tricky > business, especially when you want it to respect your logical boundary. It > is better to use the existing battle-test code than invent your own wheel. > > Zhu, Guojun > Modeling Sr Graduate > 571-3824370 > guojun_zhu@freddiemac.com > Financial Engineering > Freddie Mac > > > *Steve Lewis * > > 05/07/2012 11:17 AM > Please respond to > mapreduce-user@hadoop.apache.org > > To > mapreduce-user@hadoop.apache.org > cc > Subject > Re: Ant Colony Optimization for Travelling Salesman Problem in Hadoop > > > > > Yes but it is the job of the InputFormat code to implement the behavior - > it is not necessary to do so and in other cases I choose to create more > mappers when the mapper has a lot of work > > On Mon, May 7, 2012 at 7:54 AM, GUOJUN Zhu <*guojun_zhu@freddiemac.com*> > wrote: > > We are using old API of 0.20. I think when you set "mapred.reduce.tasks" > with certain number N and use fileinputformat, the default behavior is that > any file will be split into that number, N, each split smaller than the > default block size. Of course, other restriction, such as > "mapred.min.split.size" cannot be set too large (default is as small as > possible I think). > > Zhu, Guojun > Modeling Sr Graduate* > **571-3824370* <571-3824370>* > **guojun_zhu@freddiemac.com* > Financial Engineering > Freddie Mac > > *sharat attupurath <**sharat_a@hotmail.com* *>* > > 05/05/2012 11:37 AM > > > Please respond to* > **mapreduce-user@hadoop.apache.org* > > To > <*mapreduce-user@hadoop.apache.org* > > cc > Subject > RE: Ant Colony Optimization for Travelling Salesman Problem in Hadoop > > > > > > > Since the input files are very small, the default input formats in Hadoop > all generate just a single InputSplit, so only a single map task is > executed, and we wont have much parallelism. > > I was thinking of writing an InputFormat that would read the whole file as > an InputSplit and replicate this input split n times (where n would be the > number of ants in a single stage) so that we'll have n mappers. > Also I want the input format to return the value as the adjacency matrix > of the graph (calculating it from the coordinates in the input file). > > But I can't find a way to do this. Is it possible? Or is it better to just > have the input as Text and create the adjacency matrix in the mappers? > > ------------------------------ > Date: Sat, 5 May 2012 08:20:34 -0700 > Subject: Re: Ant Colony Optimization for Travelling Salesman Problem in > Hadoop > From: *lordjoe2000@gmail.com* > To: *mapreduce-user@hadoop.apache.org* > > yes - if you know how you can put it in distributed cache or if it is > small put in as a String in the config or have all InputFormats read it > from somewhere > > On Sat, May 5, 2012 at 8:08 AM, sharat attupurath <*sharat_a@hotmail.com*> > wrote: > I looked at both the files. in AbstractNShotInputFormat it is mentioned > that this input format does not read from files. My input is in a text > file. I want the whole file as a single record. So is it enough if i just > copy the contents of the file and return it as a string from > getValueFromIndex() ? > > ------------------------------ > Date: Fri, 4 May 2012 13:15:46 -0700 > > Subject: Re: Ant Colony Optimization for Travelling Salesman Problem in > Hadoop > From: *lordjoe2000@gmail.com* > To: *mapreduce-user@hadoop.apache.org* > > Look at NShotInputFormat > call setNumberKeys to set the number of ants then > change the method getValueFromIndex to return the text representing the > original problem to each mapper - > what happens in the second and third jobs is an exercise but the saved > test needs to have pheremones and the original problem > > > On Fri, May 4, 2012 at 9:54 AM, sharat attupurath <*sharat_a@hotmail.com*> > wrote: > Hi, > > Thanks a lot for the quick reply sir! We are new to Apache Hadoop and > haven't yet understood it properly yet. Can you please elaborate on how we > can have multiple stages of mapreduce jobs for combining the trails as you > have mentioned? > > We have been trying to find out how to write a custom splitter and almost > all online resources tell to subclass the FileInputFormat and write only a > custom RecordReader? Will it be possible to generate our splits in that way? > > Regards > > Sharat > > ------------------------------ > Date: Fri, 4 May 2012 09:22:51 -0700 > Subject: Re: Ant Colony Optimization for Travelling Salesman Problem in > Hadoop > From: *lordjoe2000@gmail.com* > To: *mapreduce-user@hadoop.apache.org* > > > > > On Fri, May 4, 2012 at 9:06 AM, sharat attupurath <*sharat_a@hotmail.com*> > wrote: > Hi, > > We are trying to parallelize the ant colony optimization algorithm for TSP > over hadoop and are facing some issues. We are using TSPLIB as input files. > The input is a text file containing eucledian coordinates of the cities - > first column is city number and the next two columns contain the x and y > coordinates respectively. > > What we intend to do is to take input from this single file, send copies > of it to multiple mappers (each mapper acts like the ant in the algorithm), > each mapper works on its input to find its own TSP solution that it outputs > and finally the reducer outputs the smallest tour found by the mappers. > Hope we are in the right track. Here are the issues: > > 1) Since the input file is small, we need to force hadoop to fire up > multiple map tasks by replicating the input. How can we make an InputSplit > of the whole file and replicate it so that the input can be sent to > multiple mappers? > Write a custom splitter sending the same data to all mappers - the only > critical criteria is the number of "Ants" > > 2) the algorithm uses a shared pheromone array and each mapper needs to > read and write data from this. How can we share the pheromone data across > the mappers. > You cannot share data across mappers and should not attempt to do so - > Better to use the reducer(s) to combine trails combine first pass trails > and > then pass the combined trails to another mapreduce job - this with the > original problem plus the current pheremone trails > > Hope the questions are clear enough. Any help would be greatly > appreciated. > > Thank you > > Regards > > Sharat > > > > -- > Steven M. Lewis PhD > 4221 105th Ave NE > Kirkland, WA 98033 * > **206-384-1340* <206-384-1340> (cell) > Skype lordjoe_com > > > > > > -- > Steven M. Lewis PhD > 4221 105th Ave NE > Kirkland, WA 98033 * > **206-384-1340* <206-384-1340> (cell) > Skype lordjoe_com > > > > > > -- > Steven M. Lewis PhD > 4221 105th Ave NE > Kirkland, WA 98033 * > **206-384-1340* <206-384-1340> (cell) > Skype lordjoe_com > > > > > > -- > Steven M. Lewis PhD > 4221 105th Ave NE > Kirkland, WA 98033 > 206-384-1340 (cell) > Skype lordjoe_com > > > -- Steven M. Lewis PhD 4221 105th Ave NE Kirkland, WA 98033 206-384-1340 (cell) Skype lordjoe_com --bcaec54c5392eaed1804bf74b29f Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable Fair enough - I write a lot of InputFormats since for most of my problems a= line of text is not the proper unit -
I read fasta files - read lines = intil you hit a line starting with > and xml fragments - read until you = hit a closing tag


On Mon, May 7, 2012 at 9:03 AM, GUO= JUN Zhu <guojun_zhu@freddiemac.com> wrote:

The default FileInputformat split the file according to the size. =A0If you use line text data, the TextFileInput= format respects the line structure for input. =A0 We got splits as small as a few KBs. =A0The file split is a tricky business, especially when you want it to respect your logical boundary. It is better to use the existing battle-test code than invent your own wheel.

Zhu, Guojun
Modeling Sr Graduate
571-38= 24370
guojun_zhu@f= reddiemac.com
Financial Engineering
Freddie Mac



=A0 =A0Steve Lewi= s <lordjoe2000@= gmail.com>

=A0 =A005/07/2012 11:17 AM

=A0 =A0Please re= spond to
mapre= duce-user@hadoop.apache.org

To
mapreduce-user@hadoop.apache.org
cc
Subject
Re: Ant Colony Optimization f= or Travelling Salesman Problem in Hadoop





Yes but it is the job of the InputFormat code to imple= ment the behavior - it is not necessary to do so and in other cases I choose to create more mappers when the mapper has a lot of work

On Mon, May 7, 2012 at 7:54 AM, GUOJUN Zhu <= guojun_zhu@freddiemac.com> wrote:

We are using old API of 0.20. =A0I think when you set
"mapred.reduce.tasks" with certain number N and use fileinputformat, the default behavior is that any file will be split into that number, N, each split smaller than the default block size. Of course, other restriction, such as "mapred.= min.split.size" cannot be set too large (default is as small as possible I think). <= font size=3D"3" face=3D"sans-serif">

Zhu, Guojun
Modeling Sr Graduate
571-3824370
<= font size=3D"3" color=3D"blue" face=3D"sans-serif">guojun_zhu@freddiemac= .com
Financial Engineering
Freddie Mac


=A0 =A0sharat att= upurath <sharat_a@hotmail.co= m>

=A0 =A005/05/2012 11:37 AM


=A0 =A0Please re= spond to
mapreduce-user= @hadoop.apache.org

To
<mapreduce-user@hadoop.apache.= org>
cc
Subject
RE: Ant Colony Optimization f= or Travelling Salesman Problem in Hadoop







Since the input files are very small, the default input formats in Hadoop all generate just a single InputSplit, so only a single map task is execute= d, and we wont have much parallelism.

I was thinking of writing an InputFormat that would read the whole file as an InputSplit and replicate this input split n times (where n would be the number of ants in a single stage) so that we'll have n mappers.<= br> Also I want the input format to return the value as the adjacency matrix of the graph (calculating it from the coordinates in the input file).

But I can't find a way to do this. Is it possible? Or is it better to j= ust have the input as Text and create the adjacency matrix in the mappers?


Date: Sat, 5 May 2012 08:20:34 -0700 Subject: Re: Ant Colony Optimization for Travelling Salesman Problem in Hadoop
From:
lordjoe2000@gmail.com
To:
mapreduce-user@had= oop.apache.org

yes - if you know how you can put it in distributed cache or if it is small put in as a String in the config or have all InputFormats read it from somewhere


On Sat, May 5, 2012 at 8:08 AM, sharat attupurath <
sharat_a@hotmail.com> wrote:
I looked at both the files. in AbstractNShotInputFormat it is mentioned that this input format does not read from files. My input is in a text file. I want the whole file as a single record. So is it enough if i just copy the contents of the file and return it as a string from getValueFromIn= dex() ?



Date: Fri, 4 May 2012 13:15:46 -0700

Subject: Re: Ant Colony Optimization for Travelling Salesman Problem in Hadoop
From:
lordjoe2000@gmail.com
To:
mapreduce-user@had= oop.apache.org

Look at NShotInputFormat
call setNumberKeys to set the number of ants then
<= /font>
change the method getValueFromIndex to return the text representing the original problem to each mapper -

what happens in the second and third jobs is an exercise but the saved test needs to have pheremones and the original problem


On Fri, May 4, 2012 at 9:54 AM, sharat attupurath <
sharat_a@hotmail.com> wrote:
Hi,

Thanks a lot for the quick reply sir! We are new to Apache Hadoop and haven= 't yet understood it properly yet. Can you please elaborate on how we can have multiple stages of mapreduce jobs for combining the trails as you have mentioned?

We have been trying to find out how to write a custom splitter and almost all online resources tell to subclass the FileInputFormat and write only a custom RecordReader? Will it be possible to generate our splits in that way?

Regards

Sharat



Date: Fri, 4 May 2012 09:22:51 -0700 Subject: Re: Ant Colony Optimization for Travelling Salesman Problem in Hadoop
From:
lordjoe2000@gmail.com
To:
mapreduce-user@had= oop.apache.org




On Fri, May 4, 2012 at 9:06 AM, sharat attupurath <
sharat_a@hotmail.com> wrote:
Hi,

We are trying to parallelize the ant colony optimization algorithm for TSP over hadoop and are facing some issues. We are using TSPLIB as input files. The input is a text file containing eucledian coordinates of the cities - first column is city number and the next two columns contain the x and y coordinates respectively.

What we intend to do is to take input from this single file, send copies of it to multiple mappers (each mapper acts like the ant in the algorithm), each mapper works on its input to find its own TSP solution that it outputs and finally the reducer outputs the smallest tour found by the mappers. Hope we are in the right track. Here are the issues:

1) Since the input file is small, we need to force hadoop to fire up multip= le map tasks by replicating the input. How can we make an InputSplit of the whole file and replicate it so that the input can be sent to multiple mappe= rs?

Write a custom splitter sending the same data to all mappers - the only critical criteria is the number of "Ants"

2) the algorithm uses a shared pheromone array and each mapper needs to read and write data from this. How can we share the pheromone data across the mappers.

You cannot share data across mappers and should not attempt to do so -
Better to use the reducer(s) to combine trails =A0combine first pass trails and
=
then pass the combined trails to another mapreduce job - this with the original problem plus the current pheremone trails


Hope the questions are clear enough. Any help would be greatly appreciated.

Thank you

Regards


Sharat




--
Steven M. Lewis PhD

4221 105th Ave NE

Kirkland, WA 98033

206-384-1340 (cell)
Skype lordjoe_com






--
Steven M. Lewis PhD

4221 105th Ave NE

Kirkland, WA 98033

206-384-1340 (cell)
Skype lordjoe_com






--
Steven M. Lewis PhD

4221 105th Ave NE

Kirkland, WA 98033

206-384-1340 (cell)
Skype lordjoe_com






--
Steven M. Lewis PhD

4221 105th Ave NE
Kirkland, WA 98033
206-384-1340 (cell)
Skype lordjoe_com





--
S= teven M. Lewis PhD
4221 105th Ave NE
Kirkland, WA 98033
=
206-384-1340 (cell)
Skype lordjoe_com


--bcaec54c5392eaed1804bf74b29f--