spark-reviews mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From zjffdu <>
Subject [GitHub] spark pull request: [SPARK-12994][CORE] It is not necessary to cre...
Date Wed, 24 Feb 2016 02:56:01 GMT
Github user zjffdu commented on a diff in the pull request:
    --- Diff: core/src/main/scala/org/apache/spark/SparkContext.scala ---
    @@ -243,7 +243,7 @@ class SparkContext(config: SparkConf) extends Logging with ExecutorAllocationCli
       private[spark] def eventLogDir: Option[URI] = _eventLogDir
       private[spark] def eventLogCodec: Option[String] = _eventLogCodec
    -  def isLocal: Boolean = (master == "local" || master.startsWith("local["))
    +  def isLocal: Boolean = Utils.isLocal(_conf)
    --- End diff --
    Keep using def, because val will cause test fails (val will be evaluated when declared
while def will be evaluated when invoked ) Although I can use config instead of _conf, but
considering all the variables around isLocal is using _conf, I don't want to make it inconsistent
and involve any potential issue. 

If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at or file a JIRA ticket
with INFRA.

To unsubscribe, e-mail:
For additional commands, e-mail:

View raw message