Return-Path: Delivered-To: apmail-lucene-hadoop-commits-archive@locus.apache.org Received: (qmail 24673 invoked from network); 5 Sep 2006 08:42:36 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (209.237.227.199) by minotaur.apache.org with SMTP; 5 Sep 2006 08:42:36 -0000 Received: (qmail 74524 invoked by uid 500); 5 Sep 2006 08:42:36 -0000 Delivered-To: apmail-lucene-hadoop-commits-archive@lucene.apache.org Received: (qmail 74454 invoked by uid 500); 5 Sep 2006 08:42:36 -0000 Mailing-List: contact hadoop-commits-help@lucene.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: hadoop-dev@lucene.apache.org Delivered-To: mailing list hadoop-commits@lucene.apache.org Received: (qmail 74445 invoked by uid 99); 5 Sep 2006 08:42:36 -0000 Received: from asf.osuosl.org (HELO asf.osuosl.org) (140.211.166.49) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 05 Sep 2006 01:42:36 -0700 X-ASF-Spam-Status: No, hits=-0.0 required=10.0 tests=SPF_HELO_PASS X-Spam-Check-By: apache.org Received: from [192.87.106.226] (HELO ajax.apache.org) (192.87.106.226) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 05 Sep 2006 01:42:35 -0700 Received: from ajax.apache.org (localhost [127.0.0.1]) by ajax.apache.org (Postfix) with ESMTP id 635B4D495D for ; Tue, 5 Sep 2006 09:42:14 +0100 (BST) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: Apache Wiki To: hadoop-commits@lucene.apache.org Date: Tue, 05 Sep 2006 08:42:14 -0000 Message-ID: <20060905084214.21271.40358@ajax.apache.org> Subject: [Lucene-hadoop Wiki] Trivial Update of "GettingStartedWithHadoop" by SameerParanjpye X-Virus-Checked: Checked by ClamAV on apache.org X-Spam-Rating: minotaur.apache.org 1.6.2 0/1000/N Dear Wiki user, You have subscribed to a wiki page or wiki category on "Lucene-hadoop Wiki" for change notification. The following page has been changed by SameerParanjpye: http://wiki.apache.org/lucene-hadoop/GettingStartedWithHadoop ------------------------------------------------------------------------------ = Downloading and installing Hadoop = - Hadoop can be downloaded from [http://www.apache.org/dyn/closer.cgi/lucene/hadoop/ here]. You may also download a nightly build from [http://cvs.apache.org/dist/lucene/hadoop/nightly/ here] or check out the code from [http://lucene.apache.org/hadoop/version_control.html subversion] and build it with [http://ant.apache.org Ant]. To install Hadoop untar the tar file in your install directory. So the directory structure would like installdir/hadoop-[version]/. All the scripts to run Hadoop are in hadoop-[version]/bin. I will refer to this directory as hadoop/bin from now on. + Hadoop can be downloaded from [http://www.apache.org/dyn/closer.cgi/lucene/hadoop/ here]. You may also download a nightly build from [http://cvs.apache.org/dist/lucene/hadoop/nightly/ here] or check out the code from [http://lucene.apache.org/hadoop/version_control.html subversion] and build it with [http://ant.apache.org Ant]. Select a directory to install Hadoop under (let's call it ) and untar the tarball in that directory. This will create a directory called hadoop- under . All scripts and tools needed to run Hadoop are present in the directory hadoop-/bin. This directory will subsequently be referred to as "hadoop/bin" in this document. = Starting Hadoop using Hadoop scripts = This section explains how to set up a Hadoop cluster running Hadoop DFS and Hadoop Mapreduce. The startup scripts are in hadoop/bin. The file that contains all the slave nodes that would join the DFS and map reduce cluster is the slaves file in hadoop/conf. Edit the slaves file to add nodes to your cluster. You need to edit the slaves file only on the machines you plan to run the Jobtracker and Namenode on. In case you want to run a single node cluster you do not have to edit the slaves file. Next edit the file hadoop-env.sh in the hadoop/conf directory. Make sure JAVA_HOME is set correctly. You can change the other environment variables as per your requirements. HADOOP_HOME is automatically determined depending on where you run your hadoop scripts from.