hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Raja Nagendra Kumar <Nagendra.R...@tejasoft.com>
Subject Re: Hadoop project - help needed
Date Sun, 17 Jul 2011 02:02:48 GMT

Haoop is meant for Distributed Data Storage as the pre-requisite and Data
Analysis later.
In your case this seem to be parallel downloads and store the images in
different place.
Yes, you could just do what you want to do with flicker api in mapper with
empty one liner junk data.. however you would not use this data with in

Raja Nagendra Kumar,

parismav wrote:
> Hello dear forum, 
> i am working on a project on apache Hadoop, i am totally new to this
> software and i need some help understanding the basic features!
> To sum up, for my project i have configured hadoop so that it runs 3
> datanodes on one machine.
> The project's main goal is, to use both Flickr API (flickr.com) libraries
> and hadoop libraries on Java, so that each one of the 3 datanodes, chooses
> a Flickr group and returns photos' info from that group.
> In order to do that, i have 3 flickr accounts, each one with a different
> api key. 
> I dont need any help on the flickr side of the code, ofcourse. But what i
> dont understand, is how to use the Mapper and Reducer part of the code. 
> What input do i have to give the Map() function? 
> do i have to contain this whole "info downloading" process in the map()
> function? 
> In a few words, how do i convert my code so that it runs distributedly on
> hadoop? 
> thank u!

View this message in context: http://old.nabble.com/Hadoop-project---help-needed-tp31741968p32076097.html
Sent from the Hadoop core-user mailing list archive at Nabble.com.

View raw message