ambari-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From nc...@apache.org
Subject [04/50] [abbrv] ambari git commit: Revert "AMBARI-20400 Yarn should not copy Tez and Slider tar ball if Tez and Sliders are not installed on the cluster (dili)"
Date Wed, 29 Mar 2017 22:40:22 GMT
Revert "AMBARI-20400 Yarn should not copy Tez and Slider tar ball if Tez and Sliders are not
installed on the cluster (dili)"

This reverts commit 3ab4c8d2ae0a7f28bc73d8438f79d2f18f96cc01.


Project: http://git-wip-us.apache.org/repos/asf/ambari/repo
Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/742fdff6
Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/742fdff6
Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/742fdff6

Branch: refs/heads/branch-feature-AMBARI-12556
Commit: 742fdff6b74d6324c2cc29d700ce96cbaa42786c
Parents: fe02991
Author: Di Li <dili@apache.org>
Authored: Tue Mar 28 15:13:40 2017 -0400
Committer: Di Li <dili@apache.org>
Committed: Tue Mar 28 15:13:40 2017 -0400

----------------------------------------------------------------------
 .../libraries/functions/copy_tarball.py         | 24 +------------------
 .../python/stacks/2.0.6/configs/default.json    | 25 --------------------
 2 files changed, 1 insertion(+), 48 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/ambari/blob/742fdff6/ambari-common/src/main/python/resource_management/libraries/functions/copy_tarball.py
----------------------------------------------------------------------
diff --git a/ambari-common/src/main/python/resource_management/libraries/functions/copy_tarball.py
b/ambari-common/src/main/python/resource_management/libraries/functions/copy_tarball.py
index 63b6926..31a9be4 100644
--- a/ambari-common/src/main/python/resource_management/libraries/functions/copy_tarball.py
+++ b/ambari-common/src/main/python/resource_management/libraries/functions/copy_tarball.py
@@ -64,18 +64,6 @@ TARBALL_MAP = {
              "/{0}/apps/{1}/spark2/spark2-{0}-yarn-archive.tar.gz".format(STACK_NAME_PATTERN,
STACK_VERSION_PATTERN))
 }
 
-SERVICE_MAP = {
-  "slider": "SLIDER",
-  "tez": "TEZ_CLIENT",
-  "pig": "PIG",
-  "sqoop": "SQOOP_CLIENT",
-  "hive": "HIVE_CLIENT",
-  "mapreduce": "HADOOP_CLIENT",
-  "hadoop_streaming": "MAPREDUCE2_CLIENT",
-  "tez_hive2": "HIVE_CLIENT",
-  "spark": "SPARK_CLIENT",
-  "spark2": "SPARK2_CLIENT"
-}
 
 def get_sysprep_skip_copy_tarballs_hdfs():
   import params
@@ -211,7 +199,7 @@ def _get_single_version_from_stack_select():
 
 
 def copy_to_hdfs(name, user_group, owner, file_mode=0444, custom_source_file=None, custom_dest_file=None,
force_execute=False,
-                 use_upgrading_version_during_upgrade=True, replace_existing_files=False,
skip=False, skip_component_check=False):
+                 use_upgrading_version_during_upgrade=True, replace_existing_files=False,
skip=False):
   """
   :param name: Tarball name, e.g., tez, hive, pig, sqoop.
   :param user_group: Group to own the directory.
@@ -222,8 +210,6 @@ def copy_to_hdfs(name, user_group, owner, file_mode=0444, custom_source_file=Non
   :param force_execute: If true, will execute the HDFS commands immediately, otherwise, will
defer to the calling function.
   :param use_upgrading_version_during_upgrade: If true, will use the version going to during
upgrade. Otherwise, use the CURRENT (source) version.
   :param skip: If true, tarballs will not be copied as the cluster deployment uses prepped
VMs.
-  :param skip_component_check: If true, will skip checking if a given component is installed
on the node for a file under its dir to be copied.
-                               This is in case the file is not mapped to a component but
rather to a specific location (JDK jar, Ambari jar, etc).
   :return: Will return True if successful, otherwise, False.
   """
   import params
@@ -240,14 +226,6 @@ def copy_to_hdfs(name, user_group, owner, file_mode=0444, custom_source_file=Non
     Logger.warning("Skipping copying {0} to {1} for {2} as it is a sys prepped host.".format(str(source_file),
str(dest_file), str(name)))
     return True
 
-  if not skip_component_check:
-    #Use components installed on the node to check if a file can be copied into HDFS
-    local_components = default("/localComponents", [])
-    component = SERVICE_MAP.get(name)
-    if component not in local_components:
-      Logger.info("{0} is not installed on the host. Skip copying {1}".format(component,
source_file))
-      return False
-
   Logger.info("Source file: {0} , Dest file in HDFS: {1}".format(source_file, dest_file))
 
   if not os.path.exists(source_file):

http://git-wip-us.apache.org/repos/asf/ambari/blob/742fdff6/ambari-server/src/test/python/stacks/2.0.6/configs/default.json
----------------------------------------------------------------------
diff --git a/ambari-server/src/test/python/stacks/2.0.6/configs/default.json b/ambari-server/src/test/python/stacks/2.0.6/configs/default.json
index ceb0ca0..fa7419f 100644
--- a/ambari-server/src/test/python/stacks/2.0.6/configs/default.json
+++ b/ambari-server/src/test/python/stacks/2.0.6/configs/default.json
@@ -1,26 +1,4 @@
 {
-    "localComponents": [
-        "NAMENODE",
-        "SECONDARY_NAMENODE",
-        "ZOOKEEPER_SERVER",
-        "DATANODE",
-        "HDFS_CLIENT",
-        "ZOOKEEPER_CLIENT",
-        "RESOURCEMANAGER",
-        "HISTORYSERVER",
-        "NODEMANAGER",
-        "YARN_CLIENT",
-        "MAPREDUCE2_CLIENT",
-        "SLIDER",
-        "PIG",
-        "SQOOP_CLIENT",
-        "HIVE_CLIENT",
-        "TEZ_CLIENT",
-        "HADOOP_CLIENT",
-        "HIVE_CLIENT",
-        "SPARK_CLIENT",
-        "SPARK2_CLIENT"
-    ],
     "roleCommand": "SERVICE_CHECK",
     "clusterName": "c1",
     "hostname": "c6401.ambari.apache.org",
@@ -566,9 +544,6 @@
             "tez.staging-dir": "/tmp/${user.name}/staging",
             "tez.am.am-rm.heartbeat.interval-ms.max": "250"
         },
-        "slider-env": {
-            "content": "export JAVA_HOME={{java64_home}}\nexport HADOOP_CONF_DIR={{hadoop_conf_dir}}"
-        },
         "yarn-env": {
             "yarn_pid_dir_prefix": "/var/run/hadoop-yarn",
             "apptimelineserver_heapsize": "1024",


Mime
View raw message