spark-reviews mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From sitalkedia <>
Subject [GitHub] spark pull request: [SPARK-14056] Appends s3 specific configuratio...
Date Thu, 24 Mar 2016 16:29:44 GMT
Github user sitalkedia commented on a diff in the pull request:
    --- Diff: core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala ---
    @@ -74,13 +74,12 @@ class SparkHadoopUtil extends Logging {
    -  /**
    -   * Return an appropriate (subclass) of Configuration. Creating config can initializes
some Hadoop
    -   * subsystems.
    -   */
    -  def newConfiguration(conf: SparkConf): Configuration = {
    -    val hadoopConf = new Configuration()
    +  /**
    +    * Appends S3-specific, spark.hadoop.*, and spark.spark.buffer.size configurations
to a Hadoop
    --- End diff --
    Thanks, removed extra "spark." . 
    You are right, `HiveConf` is being initialized in a separate code path which Spark isn't
initializing properly. I am not very familiar with hive side of things to comment on why it
was done that way. But the `TODO` in `TableReader.scala` suggests that it is the right place
to initialize the `HiveConf`. 

If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at or file a JIRA ticket
with INFRA.

To unsubscribe, e-mail:
For additional commands, e-mail:

View raw message