hadoop-common-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Wiki <wikidi...@apache.org>
Subject [Hadoop Wiki] Update of "QuickStart" by SteveKallestad
Date Thu, 20 Feb 2014 02:49:49 GMT
Dear Wiki user,

You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change notification.

The "QuickStart" page has been changed by SteveKallestad:
https://wiki.apache.org/hadoop/QuickStart?action=diff&rev1=35&rev2=36

Comment:
updated broken links, clouderas download page is now singular for all the distributions. 


  == Get up and running fast ==
  
  The fastest way may be to just install a pre-configured virtual Hadoop environment.  Two
such environments are:
-  * The [[http://www.cloudera.com/hadoop-training-virtual-machine|Cloudera Training Virtual
Machine]]. This image runs within the free VMWare player and has Hadoop, Hive, Pig and examples
pre-loaded. Video lectures and screencasts walk you through everything.
-  * The [[http://opensolaris.org/os/project/livehadoop/|OpenSolaris Hadoop Live CD]].  This
virtual Hadoop cluster runs entirely off the CD, and does not require you to install any new
software on your system.
+  * The [[http://www.cloudera.com/content/cloudera-content/cloudera-docs/DemoVMs/Cloudera-QuickStart-VM/cloudera_quickstart_vm.html|Cloudera
QuickStart Virtual Machine]]. This image runs within the free VMWare player, VirtualBox, or
KVM and has Hadoop, Hive, Pig and examples pre-loaded. Video lectures and screencasts walk
you through everything.
+  * The [[http://hortonworks.com/products/hortonworks-sandbox/|Hortonworks Sandbox]].  The
sandbox is a pre-configured virtual machine that comes with a dozen interactive Hadoop tutorials.
  
  
+ Cloudera also provides their [[http://www.cloudera.com/content/cloudera/en/products-and-services/cdh.html|distribution
for Hadoop]] (Apache 2.0 Licensed), including support for Hive and Pig and configuration management
for [[http://www.cloudera.com/content/cloudera-content/cloudera-docs/CDH4/latest/CDH-Version-and-Packaging-Information/cdhvd_topic_2.html|various
operating systems]].
- Cloudera also provides their [[http://www.cloudera.com/hadoop|distribution for Hadoop]]
(Apache 2.0 Licensed), including support for Hive and Pig and configuration management, in
the following formats:
-  * [[http://www.cloudera.com/hadoop-rpm|RPMs for Redhat based systems]] (Centos, Fedora,
RHEL, etc)
-  * [[http://www.cloudera.com/hadoop-deb|Debian Packages for Debian based systems]] (Debian,
Ubuntu, etc)
-  * [[http://www.cloudera.com/hadoop-ec2|AMI for Amazon EC2]]
  
  If you want to work exclusively with Hadoop code directly from Apache, the following articles
from the website will be most useful:
-  * [[http://hadoop.apache.org/docs/stable/single_node_setup.html|Single-Node Setup]]
+  * [[http://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-common/SingleNodeSetup.html|Single-Node
Setup]]
-  * [[http://hadoop.apache.org/docs/stable/cluster_setup.html|Cluster Setup]]
+  * [[http://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-common/ClusterSetup.html|Cluster
Setup]]
  
  Note for the above Apache links, if you're having trouble getting "ssh localhost" to work
on the following OS's:
  
@@ -31, +28 @@

   1. mapred.map.tasks and mapred.reduce.tasks in conf/mapred-site.xml. As a rule of thumb,
use 10x the number of slave processors for mapred.map.tasks, and 2x the number of slave processors
for mapred.reduce.tasks.
   1. Finally, list all slave hostnames or IP addresses in your conf/slaves file, one per
line. Then format your filesystem and start your cluster on your master node, as above.
  
- See [[http://hadoop.apache.org/common/docs/stable/cluster_setup.html#Configurationml | Hadoop
Cluster Setup/Configuration]] for details.
+ See [[http://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-common/ClusterSetup.html#Configuring_the_Hadoop_Daemons_in_Non-Secure_Mode
| Hadoop Cluster Setup/Configuration]] for details.
  

Mime
View raw message