hadoop-common-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Wiki <wikidi...@apache.org>
Subject [Hadoop Wiki] Update of "GettingStartedWithHadoop" by DavidBiesack
Date Tue, 15 Feb 2011 14:47:55 GMT
Dear Wiki user,

You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change notification.

The "GettingStartedWithHadoop" page has been changed by DavidBiesack.
The comment on this change is: add note that this page is not current; link to config doc.
http://wiki.apache.org/hadoop/GettingStartedWithHadoop?action=diff&rev1=30&rev2=31

--------------------------------------------------

  ## page was renamed from GettingStartedWithHadoop
  = Downloading and installing Hadoop =
  
- Hadoop can be downloaded from one of the [[http://www.apache.org/dyn/closer.cgi/hadoop/|Apache
download mirrors]]. You may also download a [[http://people.apache.org/builds/lucene/hadoop/nightly/|nightly
build]] or check out the code from [[http://lucene.apache.org/hadoop/version_control.html|subversion]]
and build it with [[http://ant.apache.org|Ant]]. Select a directory to install Hadoop under
(let's say {{{/foo/bar/hadoop-install}}}) and untar the tarball in that directory. A directory
corresponding to the version of Hadoop downloaded will be created under the {{{/foo/bar/hadoop-install}}}
directory. For instance, if version 0.6.0 of Hadoop was downloaded untarring as described
above will create the directory {{{/foo/bar/hadoop-install/hadoop-0.6.0}}}. The examples in
this document assume the existence of an environment variable {{{$HADOOP_INSTALL}}} that represents
the path to all versions of Hadoop installed. In the above instance {{{HADOOP_INSTALL=/foo/bar/hadoop-install}}}.
They further assume the existence of a symlink named {{{hadoop}}} in {{{$HADOOP_INSTALL}}}
that points to the version of Hadoop being used. For instance, if version 0.6.0 is being used
then {{{$HADOOP_INSTALL/hadoop -> hadoop-0.6.0}}}. All tools used to run Hadoop will be
present in the directory {{{$HADOOP_INSTALL/hadoop/bin}}}. All configuration files for Hadoop
will be present in the directory {{{$HADOOP_INSTALL/hadoop/conf}}}.
+ Hadoop can be downloaded from one of the [[http://www.apache.org/dyn/closer.cgi/hadoop/|Apache
download mirrors]]. You may also download a [[http://people.apache.org/builds/lucene/hadoop/nightly/|nightly
build]] or check out the code from [[http://lucene.apache.org/hadoop/version_control.html|subversion]]
and build it with [[http://ant.apache.org|Ant]]. Select a directory to install Hadoop under
(let's say {{{/foo/bar/hadoop-install}}}) and untar the tarball in that directory. A directory
corresponding to the version of Hadoop downloaded will be created under the {{{/foo/bar/hadoop-install}}}
directory. For instance, if version 0.21.0 of Hadoop was downloaded untarring as described
above will create the directory {{{/foo/bar/hadoop-install/hadoop-0.21.0}}}. The examples
in this document assume the existence of an environment variable {{{$HADOOP_INSTALL}}} that
represents the path to all versions of Hadoop installed. In the above instance {{{HADOOP_INSTALL=/foo/bar/hadoop-install}}}.
They further assume the existence of a symlink named {{{hadoop}}} in {{{$HADOOP_INSTALL}}}
that points to the version of Hadoop being used. For instance, if version 0.21.0 is being
used then {{{$HADOOP_INSTALL/hadoop -> hadoop-0.21.0}}}. All tools used to run Hadoop will
be present in the directory {{{$HADOOP_INSTALL/hadoop/bin}}}. All configuration files for
Hadoop will be present in the directory {{{$HADOOP_INSTALL/hadoop/conf}}}.
  
  == Startup scripts ==
  
  The {{{$HADOOP_INSTALL/hadoop/bin}}} directory contains some scripts used to launch Hadoop
DFS and Hadoop Map/Reduce daemons. These are:
  
-  * {{{start-all.sh}}} - Starts all Hadoop daemons, the namenode, datanodes, the jobtracker
and tasktrackers.
+  * {{{start-dfs.sh}}} - Starts the Hadoop DFS daemons, the namenode and datanodes. Use this
before  {{{start-mapred.sh}}} 
-  * {{{stop-all.sh}}} - Stops all Hadoop daemons.
+  * {{{stop-dfs.sh}}} - Stops the Hadoop DFS daemons.
   * {{{start-mapred.sh}}} - Starts the Hadoop Map/Reduce daemons, the jobtracker and tasktrackers.
   * {{{stop-mapred.sh}}} - Stops the Hadoop Map/Reduce daemons.
-  * {{{start-dfs.sh}}} - Starts the Hadoop DFS daemons, the namenode and datanodes.
-  * {{{stop-dfs.sh}}} - Stops the Hadoop DFS daemons.
+  * {{{start-all.sh}}} - Starts all Hadoop daemons, the namenode, datanodes, the jobtracker
and tasktrackers. Deprecated; use {{{start-dfs.sh}}} then {{{start-mapred.sh}}} 
+  * {{{stop-all.sh}}} - Stops all Hadoop daemons. Deprecated; use {{{stop-mapred.sh}}} then
{{{stop-dfs.sh}}}
  
  It is also possible to run the Hadoop daemons as Windows Services using the [[http://wrapper.tanukisoftware.org/|Java
Service Wrapper]] (download this separately). This still requires Cygwin to be installed as
Hadoop requires its df command. See the following JIRA issues for details:
  
@@ -23, +23 @@

   * https://issues.apache.org/jira/browse/HADOOP-1526
  
  == Configuration files ==
+ 
+ [[http://hadoop.apache.org/common/docs/current/cluster_setup.html#Configurationml | Hadoop
Cluster Setup/Configuration]] contains a description of Hadoop configuration for 0.21.0. The
information on this wiki page is not current. See also QuickStart which is current for 0.21.0.
  
  The {{{$HADOOP_INSTALL/hadoop/conf}}} directory contains some configuration files for Hadoop.
These are:
  

Mime
View raw message