spark-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From pwend...@apache.org
Subject git commit: Revert "[SPARK-1150] fix repo location in create script"
Date Sun, 02 Mar 2014 01:16:04 GMT
Repository: spark
Updated Branches:
  refs/heads/master 9aa095711 -> ec992e182


Revert "[SPARK-1150] fix repo location in create script"

This reverts commit 9aa095711858ce8670e51488f66a3d7c1a821c30.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/ec992e18
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/ec992e18
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/ec992e18

Branch: refs/heads/master
Commit: ec992e182231da7313d85d10b3d5fd5975c44c8b
Parents: 9aa0957
Author: Patrick Wendell <pwendell@gmail.com>
Authored: Sat Mar 1 17:15:38 2014 -0800
Committer: Patrick Wendell <pwendell@gmail.com>
Committed: Sat Mar 1 17:15:38 2014 -0800

----------------------------------------------------------------------
 conf/spark-env.sh.template                                |  2 +-
 .../org/apache/spark/deploy/worker/WorkerArguments.scala  | 10 ++--------
 docs/spark-standalone.md                                  |  4 ++--
 3 files changed, 5 insertions(+), 11 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/spark/blob/ec992e18/conf/spark-env.sh.template
----------------------------------------------------------------------
diff --git a/conf/spark-env.sh.template b/conf/spark-env.sh.template
index 619fc27..6432a56 100755
--- a/conf/spark-env.sh.template
+++ b/conf/spark-env.sh.template
@@ -15,7 +15,7 @@
 # - SPARK_MASTER_IP, to bind the master to a different IP address or hostname
 # - SPARK_MASTER_PORT / SPARK_MASTER_WEBUI_PORT, to use non-default ports
 # - SPARK_WORKER_CORES, to set the number of cores to use on this machine
-# - SPARK_WORKER_MEM, to set how much memory to use (e.g. 1000m, 2g)
+# - SPARK_WORKER_MEMORY, to set how much memory to use (e.g. 1000m, 2g)
 # - SPARK_WORKER_PORT / SPARK_WORKER_WEBUI_PORT
 # - SPARK_WORKER_INSTANCES, to set the number of worker processes per node
 # - SPARK_WORKER_DIR, to set the working directory of worker processes

http://git-wip-us.apache.org/repos/asf/spark/blob/ec992e18/core/src/main/scala/org/apache/spark/deploy/worker/WorkerArguments.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/deploy/worker/WorkerArguments.scala b/core/src/main/scala/org/apache/spark/deploy/worker/WorkerArguments.scala
index 52c4419..d35d5be 100644
--- a/core/src/main/scala/org/apache/spark/deploy/worker/WorkerArguments.scala
+++ b/core/src/main/scala/org/apache/spark/deploy/worker/WorkerArguments.scala
@@ -18,15 +18,13 @@
 package org.apache.spark.deploy.worker
 
 import java.lang.management.ManagementFactory
-import org.apache.spark.Logging
 
 import org.apache.spark.util.{IntParam, MemoryParam, Utils}
 
 /**
  * Command-line parser for the master.
  */
-private[spark] class WorkerArguments(args: Array[String]) extends Logging {
-  initLogging()
+private[spark] class WorkerArguments(args: Array[String]) {
   var host = Utils.localHostName()
   var port = 0
   var webUiPort = 8081
@@ -42,13 +40,9 @@ private[spark] class WorkerArguments(args: Array[String]) extends Logging
{
   if (System.getenv("SPARK_WORKER_CORES") != null) {
     cores = System.getenv("SPARK_WORKER_CORES").toInt
   }
-  if (System.getenv("SPARK_WORKER_MEM") != null) {
-    memory = Utils.memoryStringToMb(System.getenv("SPARK_WORKER_MEM"))
-  } else if (System.getenv("SPARK_WORKER_MEMORY") != null) {
-    logWarning("SPARK_WORKER_MEMORY is deprecated. Please use SPARK_WORKER_MEM instead")
+  if (System.getenv("SPARK_WORKER_MEMORY") != null) {
     memory = Utils.memoryStringToMb(System.getenv("SPARK_WORKER_MEMORY"))
   }
-
   if (System.getenv("SPARK_WORKER_WEBUI_PORT") != null) {
     webUiPort = System.getenv("SPARK_WORKER_WEBUI_PORT").toInt
   }

http://git-wip-us.apache.org/repos/asf/spark/blob/ec992e18/docs/spark-standalone.md
----------------------------------------------------------------------
diff --git a/docs/spark-standalone.md b/docs/spark-standalone.md
index a2dec86..51fb3a4 100644
--- a/docs/spark-standalone.md
+++ b/docs/spark-standalone.md
@@ -104,8 +104,8 @@ You can optionally configure the cluster further by setting environment
variable
     <td>Total number of cores to allow Spark applications to use on the machine (default:
all available cores).</td>
   </tr>
   <tr>
-    <td><code>SPARK_WORKER_MEM</code></td>
-    <td>Total amount of memory to allow Spark applications to use on the machine, e.g.
<code>1000m</code>, <code>2g</code> (default: total memory minus 1
GB); note that each application's <i>individual</i> memory is configured using
its <code>spark.executor.memory</code> property. The old variable </code>SPARK_WORKER_MEMORY</code>
has been deprecated.</td>
+    <td><code>SPARK_WORKER_MEMORY</code></td>
+    <td>Total amount of memory to allow Spark applications to use on the machine, e.g.
<code>1000m</code>, <code>2g</code> (default: total memory minus 1
GB); note that each application's <i>individual</i> memory is configured using
its <code>spark.executor.memory</code> property.</td>
   </tr>
   <tr>
     <td><code>SPARK_WORKER_WEBUI_PORT</code></td>


Mime
View raw message