hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Vetle Roeim" <vet...@opera.com>
Subject Re: JAR packaging
Date Mon, 30 Oct 2006 11:29:53 GMT
On Sat, 28 Oct 2006 22:13:35 +0200, Albert Chern <albert.chern@gmail.com>  
wrote:

> I'm not sure if the first option works.  If it does let me know.  One of  
> the
> developers taught me to use option 2 by creating a jar with your
> dependencies in lib/.  The tasktrackers will automatically include
> everything in lib/ on their classpaths.

Yeah, I ended up using this method as well, after getting  
ClassNotFoundException on some instances. Haven't tried the first method  
in a while, though.


> On 10/28/06, Grant Ingersoll <gsingers@apache.org> wrote:
>>
>> I'm not sure I am understanding this correctly and I don't see
>> anything on this in the Getting Started section, so...
>>
>> It seems that when I want to run my application in distributed mode,
>> I should invoke the <hadoop_home>/bin/hadoop jar <jar> (or bin/hadoop
>> <main-class>) and it will copy my JAR onto the DFS and then
>> distribute the other nodes in the cluster can access it and run it.
>>
>> Classpath wise, there seems to be two options:
>>
>> 1. Have all the appropriate dependencies available so they are read
>> in by the start up commands and included in the classpath.  Does this
>> means they all need to be on each node at startup time?
>>
>> 2. Create a single JAR made up of the contents of all the dependencies
>>
>> Also, the paths must be exactly the same on all the nodes, right?
>>
>> Is this correct or am I missing something?
>>
>> Thanks,
>> Grant
>>



-- 
Vetle Roeim
Team Manager, Information Systems
Opera Software ASA <URL: http://www.opera.com/ >

Mime
View raw message