hive-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From ga...@apache.org
Subject [11/50] [abbrv] hive git commit: HIVE-10527: NPE in SparkUtilities::isDedicatedCluster [Spark Branch] (Rui reviewed by Jimmy)
Date Fri, 05 Jun 2015 23:16:24 GMT
HIVE-10527: NPE in SparkUtilities::isDedicatedCluster [Spark Branch] (Rui reviewed by Jimmy)


Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/81e8e665
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/81e8e665
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/81e8e665

Branch: refs/heads/hbase-metastore
Commit: 81e8e665247f8bff927cb206b5bb178455047d73
Parents: cee1183
Author: Rui Li <rui.li@intel.com>
Authored: Thu Apr 30 10:15:34 2015 +0800
Committer: Xuefu Zhang <xzhang@Cloudera.com>
Committed: Mon Jun 1 14:03:30 2015 -0700

----------------------------------------------------------------------
 .../apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java    | 1 +
 1 file changed, 1 insertion(+)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hive/blob/81e8e665/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java
b/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java
index 058d91d..21398d8 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/HiveSparkClientFactory.java
@@ -112,6 +112,7 @@ public class HiveSparkClientFactory {
     String sparkMaster = hiveConf.get("spark.master");
     if (sparkMaster == null) {
       sparkMaster = sparkConf.get("spark.master");
+      hiveConf.set("spark.master", sparkMaster);
     }
     if (sparkMaster.equals("yarn-cluster")) {
       sparkConf.put("spark.yarn.maxAppAttempts", "1");


Mime
View raw message