spark-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From pwend...@apache.org
Subject [01/50] git commit: File rename
Date Tue, 10 Sep 2013 19:40:15 GMT
Updated Branches:
  refs/heads/branch-0.8 fc6fbfe7d -> 5667faafd


File rename


Project: http://git-wip-us.apache.org/repos/asf/incubator-spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-spark/commit/22b982d2
Tree: http://git-wip-us.apache.org/repos/asf/incubator-spark/tree/22b982d2
Diff: http://git-wip-us.apache.org/repos/asf/incubator-spark/diff/22b982d2

Branch: refs/heads/branch-0.8
Commit: 22b982d2bc76197c85ed50558a0fc676dffcc5ef
Parents: 61c4762
Author: Patrick Wendell <pwendell@gmail.com>
Authored: Sat Sep 7 14:37:54 2013 -0700
Committer: Patrick Wendell <pwendell@gmail.com>
Committed: Sat Sep 7 14:38:54 2013 -0700

----------------------------------------------------------------------
 docs/_layouts/global.html                |  2 +-
 docs/cdh-hdp.md                          | 78 ---------------------------
 docs/hadoop-third-party-distributions.md | 76 ++++++++++++++++++++++++++
 3 files changed, 77 insertions(+), 79 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-spark/blob/22b982d2/docs/_layouts/global.html
----------------------------------------------------------------------
diff --git a/docs/_layouts/global.html b/docs/_layouts/global.html
index 3a3b8dc..2f6bdca 100755
--- a/docs/_layouts/global.html
+++ b/docs/_layouts/global.html
@@ -98,7 +98,7 @@
                             <ul class="dropdown-menu">
                                 <li><a href="configuration.html">Configuration</a></li>
                                 <li><a href="tuning.html">Tuning Guide</a></li>
-                                <li><a href="cdh-hdp.html">Running with CDH/HDP</a></li>
+                                <li><a href="hadoop-third-party-distributions.html">Running
with CDH/HDP</a></li>
                                 <li><a href="hardware-provisioning.html">Hardware
Provisioning</a></li>
                                 <li><a href="building-with-maven.html">Building
Spark with Maven</a></li>
                                 <li><a href="contributing-to-spark.html">Contributing
to Spark</a></li>

http://git-wip-us.apache.org/repos/asf/incubator-spark/blob/22b982d2/docs/cdh-hdp.md
----------------------------------------------------------------------
diff --git a/docs/cdh-hdp.md b/docs/cdh-hdp.md
deleted file mode 100644
index 679fb11..0000000
--- a/docs/cdh-hdp.md
+++ /dev/null
@@ -1,78 +0,0 @@
----
-layout: global
-title: Running with Cloudera and HortonWorks Distributions
----
-
-Spark can run against all versions of Cloudera's Distribution Including Hadoop (CDH) and
-the Hortonworks Data Platform (HDP). There are a few things to keep in mind when using Spark
with
-these distributions:
-
-# Compile-time Hadoop Version
-When compiling Spark, you'll need to 
-[set the HADOOP_VERSION flag](http://localhost:4000/index.html#a-note-about-hadoop-versions):
-
-    HADOOP_VERSION=1.0.4 sbt/sbt assembly
-
-The table below lists the corresponding HADOOP_VERSION for each CDH/HDP release. Note that
-some Hadoop releases are binary compatible across client versions. This means the pre-built
Spark
-distribution may "just work" without you needing to compile. That said, we recommend compiling
with 
-the _exact_ Hadoop version you are running to avoid any compatibility errors.
-
-<table>
-  <tr valign="top">
-    <td>
-      <h3>CDH Releases</h3>
-      <table class="table" style="width:350px;">
-        <tr><th>Version</th><th>HADOOP_VERSION</th></tr>
-        <tr><td>CDH 4.X.X (YARN mode)</td><td>2.0.0-chd4.X.X</td></tr>
-        <tr><td>CDH 4.X.X</td><td>2.0.0-mr1-chd4.X.X</td></tr>
-        <tr><td>CDH 3u6</td><td>0.20.2-cdh3u6</td></tr>
-        <tr><td>CDH 3u5</td><td>0.20.2-cdh3u5</td></tr>
-        <tr><td>CDH 3u4</td><td>0.20.2-cdh3u4</td></tr>
-      </table>
-    </td>
-    <td>
-      <h3>HDP Releases</h3>
-      <table class="table" style="width:350px;">
-        <tr><th>Version</th><th>HADOOP_VERSION</th></tr>
-        <tr><td>HDP 1.3</td><td>1.2.0</td></tr>
-        <tr><td>HDP 1.2</td><td>1.1.2</td></tr>
-        <tr><td>HDP 1.1</td><td>1.0.3</td></tr>
-        <tr><td>HDP 1.0</td><td>1.0.3</td></tr>
-      </table>
-    </td>
-  </tr>
-</table>
-
-# Where to Run Spark
-As described in the [Hardware Provisioning](hardware-provisioning.html#storage-systems) guide,
-Spark can run in a variety of deployment modes:
-
-* Using dedicated set of Spark nodes in your cluster. These nodes should be co-located with
your
-  Hadoop installation.
-* Running on the same nodes as an existing Hadoop installation, with a fixed amount memory
and 
-  cores dedicated to Spark on each node.
-* Run Spark alongside Hadoop using a cluster resource manager, such as YARN or Mesos.
-
-These options are identical for those using CDH and HDP. Note that if you have a YARN cluster,
-but still prefer to run Spark on a dedicated set of nodes rather than scheduling through
YARN, 
-use `mr1` versions of HADOOP_HOME when compiling.
-
-# Inheriting Cluster Configuration
-If you plan to read and write from HDFS using Spark, there are two Hadoop configuration files
that
-should be included on Spark's classpath:
-
-* `hdfs-site.xml`, which provides default behaviors for the HDFS client.
-* `core-site.xml`, which sets the default filesystem name.
-
-The location of these configuration files varies across CDH and HDP versions, but
-a common location is inside of `/etc/hadoop/conf`. Some tools, such as Cloudera Manager,
create
-configurations on-the-fly, but offer a mechanisms to download copies of them.
-
-There are a few ways to make these files visible to Spark:
-
-* You can copy these files into `$SPARK_HOME/conf` and they will be included in Spark's
-classpath automatically.
-* If you are running Spark on the same nodes as Hadoop _and_ your distribution includes both
-`hdfs-site.xml` and `core-site.xml` in the same directory, you can set `HADOOP_CONF_DIR`

-in `$SPARK_HOME/spark-env.sh` to that directory.

http://git-wip-us.apache.org/repos/asf/incubator-spark/blob/22b982d2/docs/hadoop-third-party-distributions.md
----------------------------------------------------------------------
diff --git a/docs/hadoop-third-party-distributions.md b/docs/hadoop-third-party-distributions.md
new file mode 100644
index 0000000..9f4f354
--- /dev/null
+++ b/docs/hadoop-third-party-distributions.md
@@ -0,0 +1,76 @@
+---
+layout: global
+title: Running with Cloudera and HortonWorks Distributions
+---
+
+Spark can run against all versions of Cloudera's Distribution Including Hadoop (CDH) and
+the Hortonworks Data Platform (HDP). There are a few things to keep in mind when using Spark
with
+these distributions:
+
+# Compile-time Hadoop Version
+When compiling Spark, you'll need to 
+[set the HADOOP_VERSION flag](http://localhost:4000/index.html#a-note-about-hadoop-versions):
+
+    HADOOP_VERSION=1.0.4 sbt/sbt assembly
+
+The table below lists the corresponding HADOOP_VERSION for each CDH/HDP release. Note that
+some Hadoop releases are binary compatible across client versions. This means the pre-built
Spark
+distribution may "just work" without you needing to compile. That said, we recommend compiling
with 
+the _exact_ Hadoop version you are running to avoid any compatibility errors.
+
+<table>
+  <tr valign="top">
+    <td>
+      <h3>CDH Releases</h3>
+      <table class="table" style="width:350px;">
+        <tr><th>Version</th><th>HADOOP_VERSION</th></tr>
+        <tr><td>CDH 4.X.X (YARN mode)</td><td>2.0.0-chd4.X.X</td></tr>
+        <tr><td>CDH 4.X.X</td><td>2.0.0-mr1-chd4.X.X</td></tr>
+        <tr><td>CDH 3u6</td><td>0.20.2-cdh3u6</td></tr>
+        <tr><td>CDH 3u5</td><td>0.20.2-cdh3u5</td></tr>
+        <tr><td>CDH 3u4</td><td>0.20.2-cdh3u4</td></tr>
+      </table>
+    </td>
+    <td>
+      <h3>HDP Releases</h3>
+      <table class="table" style="width:350px;">
+        <tr><th>Version</th><th>HADOOP_VERSION</th></tr>
+        <tr><td>HDP 1.3</td><td>1.2.0</td></tr>
+        <tr><td>HDP 1.2</td><td>1.1.2</td></tr>
+        <tr><td>HDP 1.1</td><td>1.0.3</td></tr>
+        <tr><td>HDP 1.0</td><td>1.0.3</td></tr>
+      </table>
+    </td>
+  </tr>
+</table>
+
+# Where to Run Spark
+As described in the [Hardware Provisioning](hardware-provisioning.html#storage-systems) guide,
+Spark can run in a variety of deployment modes:
+
+* Using dedicated set of Spark nodes in your cluster. These nodes should be co-located with
your
+  Hadoop installation.
+* Running on the same nodes as an existing Hadoop installation, with a fixed amount memory
and 
+  cores dedicated to Spark on each node.
+* Run Spark alongside Hadoop using a cluster resource manager, such as YARN or Mesos.
+
+These options are identical for those using CDH and HDP. 
+
+# Inheriting Cluster Configuration
+If you plan to read and write from HDFS using Spark, there are two Hadoop configuration files
that
+should be included on Spark's classpath:
+
+* `hdfs-site.xml`, which provides default behaviors for the HDFS client.
+* `core-site.xml`, which sets the default filesystem name.
+
+The location of these configuration files varies across CDH and HDP versions, but
+a common location is inside of `/etc/hadoop/conf`. Some tools, such as Cloudera Manager,
create
+configurations on-the-fly, but offer a mechanisms to download copies of them.
+
+There are a few ways to make these files visible to Spark:
+
+* You can copy these files into `$SPARK_HOME/conf` and they will be included in Spark's
+classpath automatically.
+* If you are running Spark on the same nodes as Hadoop _and_ your distribution includes both
+`hdfs-site.xml` and `core-site.xml` in the same directory, you can set `HADOOP_CONF_DIR`

+in `$SPARK_HOME/spark-env.sh` to that directory.


Mime
View raw message