hawq-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From w...@apache.org
Subject [2/2] incubator-hawq git commit: HAWQ-186. Remove some useless files
Date Thu, 17 Dec 2015 04:48:15 GMT
HAWQ-186. Remove some useless files


Project: http://git-wip-us.apache.org/repos/asf/incubator-hawq/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-hawq/commit/10909db4
Tree: http://git-wip-us.apache.org/repos/asf/incubator-hawq/tree/10909db4
Diff: http://git-wip-us.apache.org/repos/asf/incubator-hawq/diff/10909db4

Branch: refs/heads/master
Commit: 10909db4a2f0a20121f14e666eb0e253547b9ecb
Parents: d9e1578
Author: Wen Lin <wlin@pivotal.io>
Authored: Thu Dec 17 12:55:30 2015 +0800
Committer: Wen Lin <wlin@pivotal.io>
Committed: Thu Dec 17 12:55:30 2015 +0800

----------------------------------------------------------------------
 depends/libyarn/releng/Makefile                 | 235 --------
 depends/libyarn/releng/bin/cleanup-phd.sh       |  43 --
 depends/libyarn/releng/bin/gphdinst.sh          | 536 -------------------
 depends/libyarn/releng/bin/ha_failover.sh       |  36 --
 depends/libyarn/releng/bin/install-phd.sh       |  71 ---
 depends/libyarn/releng/bin/setup_gphdinst.sh    | 175 ------
 .../bin/templates/.core-site-secure-ha.xml.swp  | Bin 12288 -> 0 bytes
 .../releng/bin/templates/core-site-ha.xml       |  28 -
 .../bin/templates/core-site-secure-ha.xml       |  39 --
 .../releng/bin/templates/core-site-secure.xml   |  34 --
 .../libyarn/releng/bin/templates/core-site.xml  |  24 -
 .../releng/bin/templates/download-config        |   5 -
 .../releng/bin/templates/hadoop-env-ha.sh       |  72 ---
 .../releng/bin/templates/hadoop-env-secure.sh   |  73 ---
 .../libyarn/releng/bin/templates/hadoop-env.sh  |  72 ---
 .../templates/hdfs-site-datanode-secure-ha.xml  | 240 ---------
 .../bin/templates/hdfs-site-datanode-secure.xml | 193 -------
 .../releng/bin/templates/hdfs-site-ha.xml       | 172 ------
 .../templates/hdfs-site-namenode-secure-ha.xml  | 224 --------
 .../bin/templates/hdfs-site-namenode-secure.xml | 181 -------
 .../bin/templates/hdfs-site-secure-ha.xml       | 240 ---------
 .../releng/bin/templates/hdfs-site-secure.xml   | 193 -------
 .../libyarn/releng/bin/templates/hdfs-site.xml  | 123 -----
 .../releng/bin/templates/log4j-ha.properties    |  75 ---
 .../libyarn/releng/bin/templates/yarn-env.sh    |  87 ---
 .../libyarn/releng/bin/templates/yarn-site.xml  | 112 ----
 depends/libyarn/releng/bin/templates/zoo.cfg    |   8 -
 depends/libyarn/releng/etc/ivy.xml              |  38 --
 depends/libyarn/releng/etc/local/.gitignore     |   1 -
 .../releng/etc/local/hdfs-site.xml.simple       |  65 ---
 .../libyarn/releng/make/dependencies/.gitignore |   1 -
 .../libyarn/releng/make/dependencies/build.xml  |  84 ---
 .../libyarn/releng/make/dependencies/cacerts    | Bin 83678 -> 0 bytes
 .../releng/make/dependencies/ivy-report.css     | 279 ----------
 depends/libyarn/releng/make/dependencies/ivy.sh |  70 ---
 .../libyarn/releng/make/dependencies/ivy.xml    |  42 --
 .../releng/make/dependencies/ivy_util.sh        |  60 ---
 .../releng/make/dependencies/ivysettings.xml    |  68 ---
 depends/libyarn/releng/publish/ivy.xml          |  36 --
 depends/libyarn/releng/publish/ivysettings.xml  |  56 --
 depends/libyarn/releng/set_bld_arch.sh          |  63 ---
 src/port/win32.ico                              | Bin 22486 -> 0 bytes
 tools/bin/ext/yaml/__init__.pyc                 | Bin 11360 -> 0 bytes
 tools/bin/ext/yaml/composer.pyc                 | Bin 4204 -> 0 bytes
 tools/bin/ext/yaml/constructor.pyc              | Bin 22333 -> 0 bytes
 tools/bin/ext/yaml/cyaml.pyc                    | Bin 3846 -> 0 bytes
 tools/bin/ext/yaml/dumper.pyc                   | Bin 2614 -> 0 bytes
 tools/bin/ext/yaml/emitter.pyc                  | Bin 32743 -> 0 bytes
 tools/bin/ext/yaml/error.pyc                    | Bin 3040 -> 0 bytes
 tools/bin/ext/yaml/events.pyc                   | Bin 5094 -> 0 bytes
 tools/bin/ext/yaml/loader.pyc                   | Bin 1925 -> 0 bytes
 tools/bin/ext/yaml/nodes.pyc                    | Bin 2236 -> 0 bytes
 tools/bin/ext/yaml/parser.pyc                   | Bin 14891 -> 0 bytes
 tools/bin/ext/yaml/reader.pyc                   | Bin 6802 -> 0 bytes
 tools/bin/ext/yaml/representer.pyc              | Bin 15117 -> 0 bytes
 tools/bin/ext/yaml/resolver.pyc                 | Bin 6782 -> 0 bytes
 tools/bin/ext/yaml/scanner.pyc                  | Bin 34012 -> 0 bytes
 tools/bin/ext/yaml/serializer.pyc               | Bin 4430 -> 0 bytes
 tools/bin/ext/yaml/tokens.pyc                   | Bin 6678 -> 0 bytes
 59 files changed, 4154 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/Makefile
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/Makefile b/depends/libyarn/releng/Makefile
deleted file mode 100644
index 851f819..0000000
--- a/depends/libyarn/releng/Makefile
+++ /dev/null
@@ -1,235 +0,0 @@
-## ======================================================================
-## 
-## ======================================================================
-
-BLD_TOP:=$(shell sh -c pwd)
-SHELL=/bin/bash
-
-##-------------------------------------------------------------------------------------
-## dependent modules
-##
-## NOTE: Dependent project module version is kept in $(BLD_TOP)/make/dependencies/ivy.xml
-##-------------------------------------------------------------------------------------
-
-GREP_SED_VAR = $(BLD_TOP)/make/dependencies/ivy.xml | sed -e 's|\(.*\)rev="\(.*\)"[ ]*conf\(.*\)|\2|'
-CMAKE_VER  = $(shell grep "\"cmake\"" $(GREP_SED_VAR))
-
-TARGET_VERSION = ${LIBHDFS3_VERSION}
-
-ifeq "$(LIBHDFS3_VERSION)" ""
-TARGET_VERSION = `cat ${BLD_TOP}/../build/version`
-endif
-
-ifeq "$(LIBHDFS3_NOW)" ""
-TARGET_TIME = `date +%Y%m%d%H%M%S`
-endif
-
-ifeq "$(LIBHDFS3_TEST_URL)" ""
-LIBHDFS3_TEST_URL = localhost:9000
-endif
-
-ifneq "$(LIBYARN_PUBLISH_OVERWRITE)" ""
-LIBYARN_PUBLISH_OVERWRITE = -overwrite
-endif
-
-##-------------------------------------------------------------------------------------
-
-# Determine the default target architecture on this machine, if override not set
-ifeq "$(BLD_ARCH)" ""
-export BLD_ARCH:=$(shell $(BLD_TOP)/set_bld_arch.sh)
-endif
-
-export PATH:=${BLD_TOP}/../ext/${BLD_ARCH}/bin:${BLD_TOP}/../ext/${BLD_ARCH}/cmake-${CMAKE_VER}/bin:${PATH}
-
-# construct test directory prefix on hdfs.
-CUR_DATE=$(shell date +%s)
-TEST_HDFS_PREFIX="${BLD_ARCH}/${CUR_DATE}/"
-
-# prepare HDFS instance
-ifeq "${HDFS_CONFIG}" ""
-export HDFS_CONFIG:=local
-endif
-
-export HDFS_CONFIG_PATH:=${BLD_TOP}/etc/${HDFS_CONFIG}/hdfs-site.xml
-
-ifeq "${HDFS_USER}" ""
-export HDFS_USER:=${USER}
-endif
-
-ifeq "${DATA_PATH}" ""
-export DATA_PATH:=/tmp/libyarn-pulse
-endif
-
-ifneq "${LIBHDFS_HDFS_SUPERUSER}" ""
-export HDFS_SUPERUSER_OPTION:=-DHDFS_SUPERUSER=${LIBHDFS_HDFS_SUPERUSER}
-endif
-
-# include thirdparty infrastructure which depends on WHERE_THE...
-BLD_WHERE_THE_LIBRARY_THINGS_ARE=LD_LIBRARY_PATH
-osx106_x86_WHERE_THE_LIBRARY_THINGS_ARE=DYLD_LIBRARY_PATH
-ifneq "$($(BLD_ARCH)_WHERE_THE_LIBRARY_THINGS_ARE)" ""
-BLD_WHERE_THE_LIBRARY_THINGS_ARE=$($(BLD_ARCH)_WHERE_THE_LIBRARY_THINGS_ARE)
-endif
-
-SED_INLINE_EDIT_BAKUP_FILE=
-
-ifeq "$(BLD_ARCH)" "osx106_x86"
-	SED_INLINE_EDIT_BAKUP_FILE=''
-	export $(BLD_WHERE_THE_LIBRARY_THINGS_ARE):=/opt/gcc-4.4.2-osx106/lib:${BLD_TOP}/../ext/${BLD_ARCH}/lib:$($(BLD_WHERE_THE_LIBRARY_THINGS_ARE))
-else
-	export $(BLD_WHERE_THE_LIBRARY_THINGS_ARE):=/opt/gcc-4.4.2/lib64:${BLD_TOP}/../ext/${BLD_ARCH}/lib:$($(BLD_WHERE_THE_LIBRARY_THINGS_ARE))
-endif
-
-##-------------------
-## set compiler
-##-------------------
-
-CCOMPILER=${CC}
-CXXCOMPILER=${CXX}
-
-ifeq "${CCOMPILER}" ""
-CCOMPILER=gcc
-endif
-
-ifeq "${CXXCOMPILER}" ""
-CXXCOMPILER=g++
-endif
-
-## ----------------------------------------------------------------------
-## libhdfs targets
-## ----------------------------------------------------------------------
-
-all: libhdfs-sync_tools libhdfs-clean libhdfs-configure libhdfs-build libhdfs-unittest
-
-${BLD_TOP}/../build:
-	mkdir ${BLD_TOP}/../build
-
-libhdfs-sync_tools:
-	make sync_tools
-
-libhdfs-clean:
-	@rm -rf ${BLD_TOP}/../dist ${BLD_TOP}/../build
-
-libhdfs-configure: ${BLD_TOP}/../build
-	(cd ${BLD_TOP}/../build && cmake ${BLD_TOP}/.. -DTEST_HDFS_PREFIX=${TEST_HDFS_PREFIX} $(LIBHDFS3_CONF_FLAGS) \
-													-DCMAKE_C_COMPILER=${CCOMPILER} -DCMAKE_CXX_COMPILER=${CXXCOMPILER} \
-	                                                -DCMAKE_INSTALL_PREFIX=${BLD_TOP}/../dist/${BLD_ARCH} \
-                                                    -DCMAKE_PREFIX_PATH=${BLD_TOP}/../ext/${BLD_ARCH} \
-                                                    ${HDFS_SUPERUSER_OPTION})
-
-libhdfs-build:
-	(cd ${BLD_TOP}/../build && make)
-
-libhdfs-install:
-	rm -rf ${BLD_TOP}/../dist/${BLD_ARCH}
-	(cd ${BLD_TOP}/../build && make install)
-
-libhdfs-test:
-	(cd ${BLD_TOP}/../build && make test)
-	
-libhdfs-showcov:
-	(cd ${BLD_TOP}/../build && make ShowCoverage)
-
-libhdfs-functiontest:
-	(export GTEST_OUTPUT="xml:functiontest.xml"; cd ${BLD_TOP}/../build && make functiontest)
-
-libhdfs-securetest:
-	(export GTEST_OUTPUT="xml:securetest.xml"; cd ${BLD_TOP}/../build && make securetest)
-
-libhdfs-unittest:
-	(export GTEST_OUTPUT="xml:unittest.xml"; cd ${BLD_TOP}/../build && make unittest)
-
-libhdfs-update-function-test.xml:
-	(cd ${BLD_TOP}/../test/data && chmod a+w function-test.xml && sed -e 's|localhost:9000|${LIBHDFS3_TEST_URL}|' -i ${SED_INLINE_EDIT_BAKUP_FILE} function-test.xml)
-	(cd ${BLD_TOP}/../test/data && chmod a+w function-secure.xml && sed -e 's|localhost:9000|${LIBHDFS3_TEST_URL}|' -i ${SED_INLINE_EDIT_BAKUP_FILE} function-secure.xml)
-
-libhdfs-package: libhdfs-install
-	(cd ${BLD_TOP}/../dist/${BLD_ARCH} && find . -type f -or -type l | xargs ${MD5} > /tmp/checksums.libyarn && mv /tmp/checksums.libyarn ${BLD_TOP}/../dist/${BLD_ARCH} && \
-	 cd ${BLD_TOP}/../dist/ && tar -czvf libyarn-${BLD_ARCH}-${TARGET_VERSION}.targz ${BLD_ARCH} )
-
-libhdfs-update-ivys:
-	(cd ${BLD_TOP}/../dist/ && mkdir -p ${TARGET_VERSION}/ivys && cp ${BLD_TOP}/etc/ivy.xml ${TARGET_VERSION}/ivys/ivy-${TARGET_VERSION}.xml \
-		&& sed -e 's|<version>|'${TARGET_VERSION}'|' -i ${SED_INLINE_EDIT_BAKUP_FILE} ${TARGET_VERSION}/ivys/ivy-${TARGET_VERSION}.xml \
-		&& sed -e 's|<time>|'${TARGET_TIME}'|' -i ${SED_INLINE_EDIT_BAKUP_FILE} ${TARGET_VERSION}/ivys/ivy-${TARGET_VERSION}.xml)
-
-libhdfs-ivy-package: libhdfs-package libhdfs-update-ivys
-	(cd ${BLD_TOP}/../dist/ && mkdir -p ${TARGET_VERSION}/targzs && cp libyarn-${BLD_ARCH}-${TARGET_VERSION}.targz ${TARGET_VERSION}/targzs/ && tar -cjvf libyarn-${BLD_ARCH}-${TARGET_VERSION}.tar.bz2 ${TARGET_VERSION})
-
-libhdfs-update-hdfs-conf:
-	(cd ${BLD_TOP}/etc/${HDFS_CONFIG} && cp -f hdfs-site.xml.simple hdfs-site.xml && sed -e 's|@user@|${HDFS_USER}|' -i ${SED_INLINE_EDIT_BAKUP_FILE} hdfs-site.xml \
-		&& mkdir -p ${DATA_PATH} && sed -e 's|@data_path@|${DATA_PATH}|' -i ${SED_INLINE_EDIT_BAKUP_FILE} hdfs-site.xml)
-
-libhdfs-cleanup-phd:
-	(${BLD_TOP}/bin/cleanup-phd.sh)
-
-libhdfs-install-phd: libhdfs-update-hdfs-conf
-	(${BLD_TOP}/bin/install-phd.sh)
-
-## ----------------------------------------------------------------------
-## Sync/Clean tools
-## ----------------------------------------------------------------------
-## Populate/clean up dependent releng supported tools.  The projects are
-## downloaded and installed into /opt/releng/...
-##
-## Tool dependencies and platform config mappings are defined in:
-##   * Apache Ivy dependency definition file
-##       releng/make/dependencies/ivy.xml
-## ----------------------------------------------------------------------
-
-opt_write_test:
-	@if [ ! -w /opt ]; then \
-	    echo ""; \
-	    echo "======================================================================"; \
-	    echo "ERROR: /opt is not writable."; \
-	    echo "----------------------------------------------------------------------"; \
-	    echo "  Supporting tools are stored in /opt.  Please ensure you have"; \
-	    echo "  write access to /opt"; \
-	    echo "======================================================================"; \
-	    echo ""; \
-	    exit 1; \
-	fi
-
-/opt/releng/apache-ant: 
-	${MAKE} opt_write_test
-	echo "Sync Ivy project dependency management framework ..."
-	curl --silent http://build-prod.dh.greenplum.com/tools/apache-ant.1.8.1.tar.gz -o /tmp/apache-ant.1.8.1.tar.gz
-	( umask 002; [ ! -d /opt/releng ] && mkdir -p /opt/releng; \
-	   cd /opt/releng; \
-	   gunzip -qc /tmp/apache-ant.1.8.1.tar.gz | tar xf -; \
-	   rm -f /tmp/apache-ant.1.8.1.tar.gz; \
-	   chmod -R a+w /opt/releng/apache-ant )
-
-sync_tools: opt_write_test /opt/releng/apache-ant
-	@cd make/dependencies; \
-	 (umask 002; /opt/releng/apache-ant/bin/ant -DBLD_ARCH=$(BLD_ARCH) resolve); \
-	 echo "Resolve finished"
-
-clean_tools: opt_write_test
-	@cd releng/make/dependencies; \
-	/opt/releng/apache-ant/bin/ant clean; \
-	rm -rf /opt/releng/apache-ant; \
-
-# Publish Libhdfs
-local_publish: /opt/releng/apache-ant
-	@if [ -f publish/ivy.xml ]; then \
-		java -Xmx512M -jar /opt/releng/apache-ant/lib/ivy-2.2.0.jar \
-			-settings publish/ivysettings.xml \
-			-ivy publish/ivy.xml \
-			-publish local \
-			-publishpattern "../dist/[artifact]-[revision].[ext]" \
-			-revision ${TARGET_VERSION} \
-			-status release \
-			-overwrite; \
-			fi; 
-public_publish: /opt/releng/apache-ant
-	@if [ -f publish/ivy.xml ]; then \
-		java -Xmx512M -Djavax.net.ssl.trustStore=$(BLD_TOP)/make/dependencies/cacerts \
-			-jar /opt/releng/apache-ant/lib/ivy-2.2.0.jar \
-			-settings publish/ivysettings.xml \
-			-ivy publish/ivy.xml \
-			-publish public \
-			-publishpattern "../dist/[artifact]-[revision].[ext]" \
-			-revision ${TARGET_VERSION} \
-			-status release \
-			$(LIBYARN_PUBLISH_OVERWRITE); \
-	fi;

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/cleanup-phd.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/cleanup-phd.sh b/depends/libyarn/releng/bin/cleanup-phd.sh
deleted file mode 100644
index 7a9c2ff..0000000
--- a/depends/libyarn/releng/bin/cleanup-phd.sh
+++ /dev/null
@@ -1,43 +0,0 @@
-#!/bin/sh
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-# 
-#   http://www.apache.org/licenses/LICENSE-2.0
-# 
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied.  See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-if [ "x${DATA_PATH}" != "x" ]; then
-	if [ -e ${DATA_PATH} ]; then
-		echo "dump namenode log"
-		cat ${DATA_PATH}/hadoop*/logs/*namenode*log*
-		echo "dump datanode log"
-		cat ${DATA_PATH}/hadoop*/logs/*datanode*log*
-	fi 
-
-	if [ "x${KEEP_LOG}" != "xtrue" ]; then
-		rm -rf ${DATA_PATH}/*
-	fi
-
-	rm -rf ${DATA_PATH}/fs
-	rm -rf ${DATA_PATH}/data
-fi
-
-NAMENODE_PIDS=`ps aux | grep java | grep NameNode | awk '{print $2}'`
-if [ "x${NAMENODE_PIDS}" != "x" ]; then
-	kill -9 ${NAMENODE_PIDS}
-fi
-
-DATANODE_PIDS=`ps aux | grep java | grep DataNode | awk '{print $2}'`
-if [ "x${DATANODE_PIDS}" != "x" ]; then
-	kill -9 ${DATANODE_PIDS}
-fi

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/gphdinst.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/gphdinst.sh b/depends/libyarn/releng/bin/gphdinst.sh
deleted file mode 100644
index 99b7085..0000000
--- a/depends/libyarn/releng/bin/gphdinst.sh
+++ /dev/null
@@ -1,536 +0,0 @@
-#!/bin/bash
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-# 
-#   http://www.apache.org/licenses/LICENSE-2.0
-# 
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied.  See the License for the
-# specific language governing permissions and limitations
-# under the License.
-## ======================================================================
-## TODO
-## * make script work for single and muli-node installs
-## * should we be able to identify the usage of the hdfs instance
-## * need to determine how to setup/start httpd
-## * how to pass in server name values
-## ======================================================================
-
-# Do not run on gpdb9.rel.dh.greenplum.com as this script will delete its shared HDFS cluster
-if [[ $(hostname) =~ "gpdb9.rel" ]]; then
-    echo "ABORT! Do not run this script on gpdb9.rel.dh.greenplum.com as this script will delete its shared cluster!"
-    exit 1
-fi
-
-# Make sure only RHEL6.2 machines can run our script
-cat /etc/redhat-release | grep "6." > /dev/null
-if [ $? -ne 0 ]; then
-    echo "This script must be run on a RHEL6.2 cluster"
-    exit 1
-fi
-
-# Make sure only root can run our script
-if [[ $EUID -ne 0 ]]; then
-   echo "This script must be run as root" 1>&2
-   exit 1
-fi
-
-# Make sure if using HDFS HA that there are at least 3 datanodes not including a namenode
-if [ "${ENABLE_QUORUM}" = "true" ] && [ ! $( echo ${DATANODES} | wc -w ) -ge 3 ]; then
-    echo "You must have at least three datanodes to have as the zkservers"
-    exit 1
-fi
-
-##
-## Courtesy - http://stevemorin.blogspot.com/2007/10/bash-get-self-directory-trick.html
-## Updated to run on rhel3 (kite12)
-##
-
-script_path="$(cd $(dirname $0); pwd -P)/$(basename $0)"
-
-[[ ! -f "$script_path" ]] && script_path="$(cd $(/usr/bin/dirname "$0"); pwd -P)/$(basename "$0")"
-[[ ! -f "$script_path" ]] && script_path="" && echo 'No full path to running script found!' && exit 1
-
-script_dir="${script_path%/*}"
-
-##
-## Global variables
-##
-
-DCA_SETUP=${DCA_SETUP:=false}
-DOWNLOAD_DIR=${DOWNLOAD_DIR:=~/downloads}
-ARCHIVE_URL=${ARCHIVE_URL:=http://hdsh132.lss.emc.com:8080/view/Ongoing-sprint-release-build/job/HudsonHD2_0_1_0ReleaseBuild_ongoing_sprint_release/lastSuccessfulBuild/artifact/*zip*/archive.zip}
-ARCHIVE=$( basename ${ARCHIVE_URL} )
-HADOOP_CONF_DIR=/etc/gphd/hadoop/conf
-ZK_CONF_DIR=/etc/gphd/zookeeper/conf
-NAMENODE=${NAMENODE:=smdw}
-SECONDARY_NAMENODE=${SECONDARY_NAMENODE:=mdw}
-DATANODES=${DATANODES:=sdw1 sdw2}
-RESOURCEMANAGER=${RESOURCEMANAGER:=smdw}
-NODEMANAGERS=${DATANODES:=sdw1 sdw2}
-HDFS_DATA_DIRECTORY_ROOT=${HDFS_DATA_DIRECTORY_ROOT:=/data}
-JOURNAL_DATA_ROOT=${JOURNAL_DATA_ROOT:=/data/journal}
-export HADOOP_HOME=${HADOOP_HOME:=/usr/lib/gphd/hadoop}
-export HADOOP_YARN_HOME=${HADOOP_YARN_HOME:=/usr/lib/gphd/hadoop-yarn}
-export ZK_HOME=${ZK_HOME:=/usr/lib/gphd/zookeeper}
-export ZOO_LOG_DIR=${ZOO_LOG_DIR:=/tmp/zookeeper/logs}
-export JAVA_HOME=${JAVA_HOME:=/opt/jdk1.7.0_15}
-
-## ======================================================================
-## Functions
-## ======================================================================
-
-## ----------------------------------------------------------------------
-## gphd_stop_remove()
-##   o Stop any existing namenode and datanode processes
-##   o Display running java processes with "jps" command
-##   o Remove hadoop*, bigtop* and zookeeper rpms
-##   o Remove /usr/lib/gphd
-## ----------------------------------------------------------------------
-
-display_jps(){
-    echo "======================================================================"
-    for node in ${NAMENODE} ${SECONDARY_NAMENODE} ${DATANODES/%${SECONDARY_NAMENODE}/}; do
-        echo $node
-        ssh $node "${JAVA_HOME}/bin/jps"
-        echo
-    done
-    echo "======================================================================"
-}
-
-gphd_stop_remove(){
-
-    ## Sanity Stop request for NN, 2NN, and DN only (Actual stop is when run finishes in sys_mgmt_utils.sh)
-    ssh ${NAMENODE}           "[ -f ${HADOOP_HOME}/sbin/hadoop-daemon.sh ] && sudo -u hdfs ${HADOOP_HOME}/sbin/hadoop-daemon.sh stop namenode"
-    ssh ${SECONDARY_NAMENODE} "[ -f ${HADOOP_HOME}/sbin/hadoop-daemon.sh ] && sudo -u hdfs ${HADOOP_HOME}/sbin/hadoop-daemon.sh stop secondarynamenode"
-
-    for node in ${DATANODES}; do
-	ssh $node "[ -f ${HADOOP_HOME}/sbin/hadoop-daemon.sh ] && sudo -u hdfs ${HADOOP_HOME}/sbin/hadoop-daemon.sh stop datanode"
-    done
-
-    display_jps
-
-    for node in ${NAMENODE} ${SECONDARY_NAMENODE} ${DATANODES/%${SECONDARY_NAMENODE}/}; do
-	    ssh $node "rm -rf /data*/hdfs /tmp/hadoop-*.pid /tmp/hsperfdata_*;"
-    done
-
-    # Remove hadoop, bigtop, and zookeeper rpms in each node
-    for node in ${NAMENODE} ${SECONDARY_NAMENODE} ${DATANODES/%${SECONDARY_NAMENODE}/}; do
-        ssh $node '\
-        hostname; \
-        for rpm in hadoop bigtop zookeeper; do \
-            echo "----------------------------------------------------------------------"; \
-            echo "Processing $rpm"; \
-            rpm -qa | grep $rpm; \
-            if [ $? = 0 ]; then \
-                yum remove -y $( rpm -qa | grep $rpm ); \
-                rpm -qa | grep $rpm; \
-            fi; \
-            rm -rf /etc/gphd/hadoop* /etc/gphd/zookeeper* /usr/lib/gphd/zookeeper* /var/lib/gphd/hadoop* /var/log/gphd/hadoop* /var/lib/gphd/zookeeper*; \
-            echo "----------------------------------------------------------------------"; \
-        done'
-    done
-}
-
-retrieve_hdfs_archive(){
-    echo "======================================================================"
-    
-    echo "Retrieving ${ARCHIVE_URL}"
-    
-    if [ ! -d "${DOWNLOAD_DIR}" ]; then
-        mkdir ${DOWNLOAD_DIR}
-    fi
-    
-    rm -rf ${DOWNLOAD_DIR}/*
-    
-    wget --no-verbose ${ARCHIVE_URL} -O ${DOWNLOAD_DIR}/${ARCHIVE}
-    if [ $? != 0 ]; then
-        echo "FATAL: retrieval failed (${ARCHIVE_URL})."
-        exit 2
-    fi
-    
-    echo "======================================================================"
-}
-
-## ######################################################################
-## MAIN Script
-## ######################################################################
-
-##
-## If necessary, stop hadoop services and remove previously installed
-## rpms (and files).
-##
-
-gphd_stop_remove
-
-##
-## Retrieve latest hdfs archive from jenkins server
-##
-
-retrieve_hdfs_archive
-
-##
-## Process retrieved archive
-##
-
-cd ${DOWNLOAD_DIR}
-
-if [ ! -f "${ARCHIVE}" ]; then
-    echo "FATAL: archive does note exist (${ARCHIVE})"
-    exit 2
-fi
-
-ls | grep archive.zip >> /dev/null
-if [ $? == 0 ]; then
-    unzip ${ARCHIVE}
-    if [ $? != 0 ]; then
-	echo "FATAL: unzip of archive failed (${ARCHIVE})."
-	exit 2
-    fi
-
-    cd archive/dist
-fi
-
-GPHD_TARBALL=$( ls )
-if [ ! -f "${GPHD_TARBALL}" ]; then
-    echo "FATAL: gphd tarball does not exist."
-    exit 2
-fi
-    
-GPHD_DIR=$( basename * .tar.gz )
-
-tar xf ${GPHD_TARBALL}
-if [ $? != 0 ]; then
-    echo "FATAL: tar expansion failed (${GPHD_TARBALL})."
-    exit 2
-fi
-
-##
-## Prepare file based yum repository
-##
-
-echo "======================================================================"
-
-cd ${GPHD_DIR}
-REPO_DIR=$( pwd )
-createrepo ${REPO_DIR}
-
-echo "======================================================================"
-
-rm -f /etc/yum.repos.d/gphd.repo
-
-cat > /etc/yum.repos.d/gphd.repo <<-EOF
-	[gphd]
-	name=Greenplum HD repository
-	baseurl=http://${SECONDARY_NAMENODE}/yum/base
-	enabled=1
-	gpgcheck=0
-EOF
-
-if [ ! -d /var/www/html/yum ] ; then
-    mkdir -p /var/www/html/yum
-    chmod 777 /var/www/html/yum
-fi
-rm -rf /var/www/html/yum/base
-
-rsync -auv ${REPO_DIR}/* /var/www/html/yum/base
-chmod -R 777 /var/www/html/yum/base
-cd ${DOWNLOAD_DIR}
-rm -rf ${DOWNLOAD_DIR}/*
-
-service httpd start
-
-##
-## Configure hdfs installation
-##
-
-if [ "${ENABLE_QUORUM}" = "true" ] && [ -n "${KDC}" ] ; then
-    CORE_SITE=core-site-secure-ha.xml
-    HADOOP_ENV=hadoop-env-secure.sh
-    HDFS_SITE=hdfs-site-secure-ha.xml
-elif [ "${ENABLE_QUORUM}" = "true" ] ; then
-    CORE_SITE=core-site-ha.xml
-    HADOOP_ENV=hadoop-env-ha.sh
-    HDFS_SITE=hdfs-site-ha.xml
-elif [ -n "${KDC}" ] ; then
-    CORE_SITE=core-site-secure.xml
-    HADOOP_ENV=hadoop-env-secure.sh
-    HDFS_SITE=hdfs-site-secure.xml
-else
-    CORE_SITE=core-site.xml
-    HADOOP_ENV=hadoop-env.sh
-    HDFS_SITE=hdfs-site.xml
-fi
-
-
-if [ "${ENABLE_QUORUM}" = "true" ] ; then
-    ZOO_LOG=log4j-ha.properties
-    ZOO_CFG=zoo.cfg
-
-    # Deploy zookeeper and jounalnode depend on datanodes number.
-    datanode_num=0
-
-    for node in ${DATANODES/%${SECONDARY_NAMENODE}/}; do 
-       datanode_num=`expr $datanode_num + 1` 
-    done
-
-    for num in {1..3}; do
-        declare ZKSERVER_$num=$( echo $DATANODES | cut -d" " -f ${num} )
-
-        if [ $datanode_num -ge 6 ]; then
-            num_jn=`expr $num + 3`
-        elif [ $datanode_num -gt 3 ]; then
-            num_jn=`expr $datanode_num - 3 + $num` 
-        else
-            num_jn=$num
-        fi 
-
-        declare JOURNALNODE_$num=$( echo $DATANODES | cut -d " " -f ${num_jn} )
-    done
-fi
-
-sed -e "s|%HDFS_HOST%|${NAMENODE}:9000|" --in-place=.orig ${script_dir}/templates/${CORE_SITE}
-
-if [ "${DCA_SETUP}" = true ]; then
-    sed -e "s|%JAVA_HOME%|${JAVA_HOME}|" \
-        -e "s|%NAMENODE_MEMORY%|-Xmx10240m|" \
-        -e "s|%DATANODE_MEMORY%|-Xmx8192m|" \
-        -e "s|%SECONDARYNAMENODE_MEMORY%|-Xmx10240m|" \
-        -e "s|%HADOOP_HOME%|${HADOOP_HOME}|" \
-        --in-place=.orig ${script_dir}/templates/${HADOOP_ENV}
-elif [ "${DCA_SETUP}" = "half" ]; then
-    sed -e "s|%JAVA_HOME%|${JAVA_HOME}|" \
-        -e "s|%NAMENODE_MEMORY%|-Xmx10240m|" \
-        -e "s|%DATANODE_MEMORY%|-Xmx8192m|" \
-        -e "s|%SECONDARYNAMENODE_MEMORY%|-Xmx10240m|" \
-        -e "s|%HADOOP_HOME%|${HADOOP_HOME}|" \
-        --in-place=.orig ${script_dir}/templates/${HADOOP_ENV}
-else
-    sed -e "s|%JAVA_HOME%|${JAVA_HOME}|" \
-        -e "s|%NAMENODE_MEMORY%|-Xmx4096m|" \
-        -e "s|%DATANODE_MEMORY%|-Xmx2048m|" \
-        -e "s|%SECONDARYNAMENODE_MEMORY%|-Xmx2048m|" \
-        -e "s|%HADOOP_HOME%|${HADOOP_HOME}|" \
-        --in-place=.orig ${script_dir}/templates/${HADOOP_ENV}
-fi
-
-if [ "${TWO_NAMENODE_DIR}" = true ]; then
-    sed -e "s|%DATA_DIR_ROOT%/hdfs/name<|${HDFS_DATA_DIRECTORY_ROOT}/hdfs/name,file:${HDFS_DATA_DIRECTORY_ROOT}/hdfs/name2<|" --in-place=.orig ${script_dir}/templates/${HDFS_SITE}
-fi
-
-sed -e "s|%DATA_DIR_ROOT%|${HDFS_DATA_DIRECTORY_ROOT}|" --in-place=.orig ${script_dir}/templates/${HDFS_SITE}
-
-DATA_FILESYSTEMS=$( ssh $( echo ${DATANODES} | cut -d" " -f 1 ) "df -kh 2> /dev/null | grep data | wc -l" )
-
-if [ $DATA_FILESYSTEMS -ge 2 ] ; then
-    echo "/data1 and /data2 found... using both!"
-    sed -e "s|%HDFS_DATANODE_DIR%|file:${HDFS_DATA_DIRECTORY_ROOT}1/hdfs/data,file:${HDFS_DATA_DIRECTORY_ROOT}2/hdfs/data|" --in-place=.orig ${script_dir}/templates/${HDFS_SITE}
-else
-    sed -e "s|%HDFS_DATANODE_DIR%|${HDFS_DATA_DIRECTORY_ROOT}/hdfs/data|" --in-place=.orig ${script_dir}/templates/${HDFS_SITE}
-fi
-
-if [ "${ENABLE_QUORUM}" = "true" ] ; then
-    sed -e "s|%ZKSERVER1%|${ZKSERVER_1}|" -e "s|%ZKSERVER2%|${ZKSERVER_2}|" -e "s|%ZKSERVER3%|${ZKSERVER_3}|" --in-place=.orig ${script_dir}/templates/${ZOO_CFG}
-    sed -e "s|%ZKSERVER1%|${ZKSERVER_1}|" -e "s|%ZKSERVER2%|${ZKSERVER_2}|" -e "s|%ZKSERVER3%|${ZKSERVER_3}|" --in-place=.orig ${script_dir}/templates/${CORE_SITE}
-
-    sed -e "s|%NAMENODE%|${NAMENODE}|" -e "s|%SECONDARY_NAMENODE%|${SECONDARY_NAMENODE}|" -e "s|%JOURNALNODE1%|${JOURNALNODE_1}|" -e "s|%JOURNALNODE2%|${JOURNALNODE_2}|" \
-        -e "s|%JOURNALNODE3%|${JOURNALNODE_3}|" -e "s|%JOURNAL_DATA_ROOT%|${JOURNAL_DATA_ROOT}|" --in-place=.orig ${script_dir}/templates/${HDFS_SITE}
-
-    sed -e "s|%ZOO_LOG_DIR%|${ZOO_LOG_DIR}|" --in-place=.orig ${script_dir}/templates/${ZOO_LOG}
-fi
-
-if [ "${YARN}" = "true" ] ; then
-    if [ "${DCA_SETUP}" = true ]; then
-       RESOURCE_MEM="8192" 
-    elif [ "${DCA_SETUP}" = "half" ]; then
-       RESOURCE_MEM="8192"
-    else
-       RESOURCE_MEM="2048"
-    fi
-    sed -e "s|%RESOURCE_MEM%|${RESOURCE_MEM}|" --in-place=.orig ${script_dir}/templates/yarn-site.xml
-    sed -e "s|%JAVA_HOME%|${JAVA_HOME}|" --in-place=.orig ${script_dir}/templates/yarn-env.sh
-fi
-
-rm ${script_dir}/templates/slaves
-for datanode in ${DATANODES}; do
-    echo "${datanode}" >> ${script_dir}/templates/slaves
-done
-
-rm ${script_dir}/templates/masters
-echo "${NAMENODE}" >> ${script_dir}/templates/masters
-echo "${SECONDARY_NAMENODE}" >> ${script_dir}/templates/masters
-
-##
-## Install hdfs rpms
-##
-
-for node in ${NAMENODE} ${SECONDARY_NAMENODE} ${DATANODES/%${SECONDARY_NAMENODE}/}; do
-    echo ""
-    echo "Remove old gphd directories in $node"
-    ssh $node "rm -rf /var/log/gphd/* /usr/lib/gphd/*"
-done
-
-for node in ${NAMENODE} ${SECONDARY_NAMENODE} ${DATANODES/%${SECONDARY_NAMENODE}/}; do
-    echo ""
-    echo "Passing YUM repo file to $node"
-    scp /etc/yum.repos.d/gphd.repo $node:/etc/yum.repos.d/gphd.repo
-
-    echo "Installing hdfs in $node"
-    ssh $node 'yum --disablerepo "*" --enablerepo "gphd" list available; \
-               yum install -y hadoop-conf-pseudo; \
-               yum install -y zookeeper-server;'
-
-    ##
-    ## Post rpm installation processing
-    ##
-
-    echo "cp libexec files in $node"
-    ssh $node "cp ${HADOOP_HOME}/libexec/* ${HADOOP_HOME}-hdfs/libexec"
-    ssh $node "cp /usr/lib/gphd/hadoop/libexec/* /usr/lib/gphd/hadoop-yarn/libexec/"
-
-    echo "scp slaves, masters, and core-site.xml file to $node"
-    scp ${script_dir}/templates/slaves $node:${HADOOP_CONF_DIR}/slaves
-    scp ${script_dir}/templates/masters $node:${HADOOP_CONF_DIR}/masters
-    scp ${script_dir}/templates/${CORE_SITE} $node:${HADOOP_CONF_DIR}/core-site.xml
-    scp ${script_dir}/templates/${HADOOP_ENV} $node:${HADOOP_CONF_DIR}/hadoop-env.sh
-
-    if [ "${YARN}" = "true" ] ; then
-        scp ${script_dir}/templates/yarn-env.sh $node:${HADOOP_CONF_DIR}/yarn-env.sh
-        scp ${script_dir}/templates/yarn-site.xml $node:${HADOOP_CONF_DIR}/yarn-site.xml
-
-        ssh $node mkdir ${HADOOP_YARN_HOME}/logs
-        ssh $node chmod 777 ${HADOOP_YARN_HOME}/logs
-        ssh $node chmod 777 /var/log/gphd/hadoop-yarn
-    fi
-
-    if [ "${ENABLE_QUORUM}" = "true" ] ; then
-	ssh $node "mkdir /usr/lib/gphd/zookeeper/etc; \
-                   ln -s /etc/gphd/zookeeper/conf /usr/lib/gphd/zookeeper/etc/zookeeper; \
-                   echo ZOO_LOG_DIR=/tmp/zookeeper/logs/ >> ${ZK_HOME}/bin/zkEnv.sh"
-	scp ${script_dir}/templates/${ZOO_CFG} $node:${ZK_CONF_DIR}/zoo.cfg
-	scp ${script_dir}/templates/${ZOO_LOG} $node:${ZK_CONF_DIR}/log4j.properties
-
-	if [ $node == $ZKSERVER_1 ] ; then
-	    ZK_NUM=1
-	elif [ $node == $ZKSERVER_2 ] ; then
-	    ZK_NUM=2
-	elif [ $node == $ZKSERVER_3 ] ; then
-	    ZK_NUM=3
-	fi
-
-	if [ $node == $ZKSERVER_1 ] || [ $node == $ZKSERVER_2 ] || [ $node == $ZKSERVER_3 ] ; then
-	    ssh $node "mkdir -p /tmp/zookeeper/logs && chown -R zookeeper.hadoop /tmp/zookeeper/ && echo ${ZK_NUM} > /tmp/zookeeper/myid; \
-                       grep ZK_HOME ~/.bash_profile >> /dev/null; \
-                       if [ $? != 0 ] ; then \
-                           echo 'export ZK_HOME=/usr/lib/gphd/zookeeper' >> ~/.bash_profile; \
-                           echo 'export ZOO_LOG_DIR=/tmp/zookeeper/logs' >> ~/.bash_profile; \
-                       fi"
-	fi
-    fi
-
-    scp ${script_dir}/templates/${HDFS_SITE} $node:${HADOOP_CONF_DIR}/hdfs-site.xml
-
-    ssh $node mkdir ${HADOOP_HOME}/logs
-    ssh $node chmod 777 ${HADOOP_HOME}/logs
-    echo ""
-done
-
-if [ "${YARN}" = "true" ] ; then
-    for node in ${RESOURCEMANAGER} ${NODEMANAGERS}; do
-        ssh ${node} cp /etc/gphd/hadoop/conf.empty/capacity-scheduler.xml /usr/lib/gphd/hadoop/etc/hadoop/
-    done
-fi
-
-if [ "${ENABLE_QUORUM}" = "true" ] ; then
-    if [ -n "${KDC}" ] ; then
-        HDFS_DN_SSH_COMMAND=""
-    else
-        HDFS_DN_SSH_COMMAND="sudo -u hdfs"
-    fi
-
-    for zkserver in ${ZKSERVER_1} ${ZKSERVER_2} ${ZKSERVER_3}; do
-	ssh $zkserver "sudo JAVA_HOME=${JAVA_HOME} /etc/init.d/zookeeper-server start"
-    done
-
-    for journalnode in ${JOURNALNODE_1} ${JOURNALNODE_2} ${JOURNALNODE_3}; do
-	ssh $journalnode "rm -rf ${JOURNAL_DATA_ROOT}; mkdir ${JOURNAL_DATA_ROOT}; chown hdfs.hdfs ${JOURNAL_DATA_ROOT};\
-                          JAVA_HOME=${JAVA_HOME} ${HDFS_DN_SSH_COMMAND} ${HADOOP_HOME}/sbin/hadoop-daemon.sh start journalnode"
-    done
-
-    ssh ${NAMENODE} "\
-      JAVA_HOME=${JAVA_HOME} sudo -u hdfs ${HADOOP_HOME}-hdfs/bin/hdfs namenode -format -force 2>&1 | tee ${HADOOP_HOME}/logs/hdfs-namenode-format.out; \
-      grep 'Exiting with status 0' ${HADOOP_HOME}/logs/hdfs-namenode-format.out; \
-      if [ \$? != 0 ]; then \
-        echo 'FATAL: format failed.'; \
-        exit 1; \
-      fi;
-      sudo -u hdfs ${HADOOP_HOME}/sbin/hadoop-daemon.sh start namenode"
-
-    ssh ${SECONDARY_NAMENODE} "\
-      JAVA_HOME=${JAVA_HOME} sudo -u hdfs ${HADOOP_HOME}-hdfs/bin/hdfs namenode -bootstrapStandby 2>&1 | tee ${HADOOP_HOME}/logs/hdfs-namenode-bootstrapStandby.out; \
-      grep 'Exiting with status 0' ${HADOOP_HOME}/logs/hdfs-namenode-bootstrapStandby.out; \
-      if [ \$? != 0 ]; then \
-        echo 'FATAL: bootstrapStandby failed.'; \
-        exit 1; \
-      fi;
-      sudo -u hdfs ${HADOOP_HOME}/sbin/hadoop-daemon.sh start namenode"
-
-    ssh ${NAMENODE} "JAVA_HOME=${JAVA_HOME} sudo -u hdfs ${HADOOP_HOME}-hdfs/bin/hdfs zkfc -formatZK -force"
-    for namenode in ${NAMENODE} ${SECONDARY_NAMENODE}; do
-	ssh $namenode "JAVA_HOME=${JAVA_HOME} sudo -u hdfs ${HADOOP_HOME}/sbin/hadoop-daemon.sh start zkfc"
-    done
-
-    for datanode in ${DATANODES}; do
-	    ssh $datanode ${HDFS_DN_SSH_COMMAND} /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start datanode
-    done
-else
-    ##
-    ## Format hdfs file system
-    ##
-
-    ssh ${NAMENODE} "\
-      sudo -u hdfs ${HADOOP_HOME}-hdfs/bin/hdfs namenode -format -force 2>&1 | tee ${HADOOP_HOME}/logs/hdfs-namenode-format.out; \
-      grep 'Exiting with status 0' ${HADOOP_HOME}/logs/hdfs-namenode-format.out; \
-      if [ \$? != 0 ]; then \
-        echo 'FATAL: format failed.'; \
-        exit 1; \
-      fi"
-
-    ##
-    ## Startup hadoop namenode and datanode services
-    ##
-
-    ssh ${NAMENODE}           sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start namenode
-
-    ssh ${SECONDARY_NAMENODE} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start secondarynamenode
-
-    for datanode in ${DATANODES}; do
-	if [ -n "${KDC}" ] ; then
-	    ssh $datanode     /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start datanode
-	else
-	    ssh $datanode     sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start datanode
-	fi
-    done
-
-    if [ "${YARN}" = "true" ] ; then
-        ssh ${PHD_USER}@${RESOURCEMANAGER} JAVA_HOME=${JAVA_HOME} /usr/lib/gphd/hadoop-yarn/sbin/yarn-daemon.sh start resourcemanager 
-        for node in ${NODEMANAGERS}; do
-            ssh ${PHD_USER}@${node} JAVA_HOME=${JAVA_HOME} /usr/lib/gphd/hadoop-yarn/sbin/yarn-daemon.sh start nodemanager 
-        done
-    fi
-fi
-
-##
-## Before exiting, display running java processes with "jps" command
-##
-
-display_jps
-
-exit 0

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/ha_failover.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/ha_failover.sh b/depends/libyarn/releng/bin/ha_failover.sh
deleted file mode 100644
index f596041..0000000
--- a/depends/libyarn/releng/bin/ha_failover.sh
+++ /dev/null
@@ -1,36 +0,0 @@
-#!/bin/bash
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-# 
-#   http://www.apache.org/licenses/LICENSE-2.0
-# 
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied.  See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-nn1_status=`/usr/bin/hdfs haadmin -getServiceState nn1`
-echo nn1 status is: $nn1_status
-nn2_status=`/usr/bin/hdfs haadmin -getServiceState nn2`
-echo nn2 status is: $nn2_status
-if [ "${nn1_status}" = "active" ]; then
-    echo "hdfs haadmin -failover nn1 nn2"
-    hdfs haadmin -failover nn1 nn2
-elif [ "${nn1_status}" = "standby" ]; then
-    echo "hdfs haadmin -failover nn2 nn1"
-    hdfs haadmin -failover nn2 nn1
-else
-    echo "Can't get valid status of nn1, exit."
-    exit 1
-fi
-nn1_status=`/usr/bin/hdfs haadmin -getServiceState nn1`
-echo nn1 status now is: $nn1_status
-nn2_status=`/usr/bin/hdfs haadmin -getServiceState nn2`
-echo nn2 status now is: $nn2_status

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/install-phd.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/install-phd.sh b/depends/libyarn/releng/bin/install-phd.sh
deleted file mode 100644
index 4d63191..0000000
--- a/depends/libyarn/releng/bin/install-phd.sh
+++ /dev/null
@@ -1,71 +0,0 @@
-#!/bin/sh
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-# 
-#   http://www.apache.org/licenses/LICENSE-2.0
-# 
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied.  See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-if [ -z ${DATA_PATH} ]; then
-	echo "DATA_PATH not set"
-	exit 1 
-fi
-
-if [ -d ${DATA_PATH} ]; then
-	cd ${DATA_PATH}
-else
-	echo "DATA_PATH not exist"
-        exit 1
-fi
-
-HADOOP_TARBALL=`curl --silent --no-buffer "http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/latest/?C=M;O=D" | grep -o 'PHD-2.[0-9].[0-9].[0-9]-bin-[0-9]*.tar.gz' | head -n 1`
-echo "HADOOP tarball: " ${HADOOP_TARBALL}
-
-HADOOP_URL="http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/latest/"${HADOOP_TARBALL}
-echo "Download HADOOP from " ${HADOOP_URL}
-
-curl --silent -o ${DATA_PATH}/${HADOOP_TARBALL} -L ${HADOOP_URL}
-
-tar -xzf ${DATA_PATH}/${HADOOP_TARBALL}
-
-HADOOP_PACKAGE=`echo ${HADOOP_TARBALL} | grep -o 'PHD-[0-9].[0-9].[0-9].[0-9]-bin-[0-9]*'`
-HADOOP_VERSION=`ls ${DATA_PATH}/${HADOOP_PACKAGE}/hadoop/tar/*.tar.gz | grep -o 'hadoop-[0-9].[0-9].[0-9]-[A-Za-z0-9\-]*-[0-9].[0-9].[0-9].[0-9]'`
-echo "HADOOP version: " ${HADOOP_VERSION}
-
-if [ -z ${HADOOP_VERSION} ]; then
-	echo "cannot get HADOOP version"
-	exit 1
-fi
-
-tar -xzf ${HADOOP_PACKAGE}/hadoop/tar/${HADOOP_VERSION}.tar.gz
-
-if [ -z ${HDFS_CONFIG_PATH} ]; then
-        echo "HDFS_CONFIG_PATH not set"
-        exit 1
-fi
-
-if [ -f ${HDFS_CONFIG_PATH} ]; then
-	cp -f ${HDFS_CONFIG_PATH} ${DATA_PATH}/${HADOOP_VERSION}/etc/hadoop/
-else
-	echo "HDFS_CONFIG_PATH not a file"
-	exit 1
-fi
-
-HADOOP_BIN=${DATA_PATH}/${HADOOP_VERSION}/bin
-HADOOP_SBIN=${DATA_PATH}/${HADOOP_VERSION}/sbin
-
-${HADOOP_BIN}/hdfs namenode -format
-${HADOOP_SBIN}/hadoop-daemon.sh start namenode
-${HADOOP_SBIN}/hadoop-daemon.sh start datanode
-${HADOOP_BIN}/hdfs dfs -mkdir hdfs://localhost:9000/user
-${HADOOP_BIN}/hdfs dfs -chmod 777 hdfs://localhost:9000/user

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/setup_gphdinst.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/setup_gphdinst.sh b/depends/libyarn/releng/bin/setup_gphdinst.sh
deleted file mode 100644
index 64a82c1..0000000
--- a/depends/libyarn/releng/bin/setup_gphdinst.sh
+++ /dev/null
@@ -1,175 +0,0 @@
-#!/bin/bash
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-# 
-#   http://www.apache.org/licenses/LICENSE-2.0
-# 
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied.  See the License for the
-# specific language governing permissions and limitations
-# under the License.
-#
-# Usage: ./setup_gphdinst.sh setup / start HK / stop HK 
-# HK means this is for HA/Kerberos
-#
-
-JAVA_HOME=${JAVA_HOME:="/opt/jdk1.7.0_15"}
-HADOOP_HOME="/usr/lib/gphd/hadoop"
-LOCAL_HDFS_RPMINSTALL="true"
-DCA_SETUP=${DCA_SETUP:=false}
-TWO_NAMENODE_DIR=${TWO_NAMENODE_DIR:=false}
-SEGMENT_SYSTEM_LIST=${SEGMENT_SYSTEM_LIST:="sdw1 sdw2 sdw3"}
-MASTERHOST=${MASTERHOST:=mdw}
-STANDBYHOST=${STANDBYHOST:=smdw}
-DATANODES_LIST=${SEGMENT_SYSTEM_LIST}
-USER="gpadmin"
-STEP="$1"
-ENABLE_QUORUM=${ENABLE_QUORUM:=true}
-
-if [ "${ENABLE_QUORUM}" = "true" ]; then 
-    ZKSERVER_LIST=$( echo ${SEGMENT_SYSTEM_LIST} | cut -d' ' -f1-3 )
-
-    # Deploy zookeeper and jounalnode depend on datanodes number.
-    datanode_num=0
-
-    for node in ${SEGMENT_SYSTEM_LIST}; do   
-       datanode_num=`expr $datanode_num + 1` 
-    done 
-
-    if [ $datanode_num -ge 6 ]; then 
-        JOURNALNODE_LIST=$( echo ${SEGMENT_SYSTEM_LIST} | cut -d' ' -f4-6 )
-    elif [ $datanode_num -eq 5 ]; then 
-        JOURNALNODE_LIST=$( echo ${SEGMENT_SYSTEM_LIST} | cut -d' ' -f3-5 )
-    elif [ $datanode_num -eq 4 ]; then 
-        JOURNALNODE_LIST=$( echo ${SEGMENT_SYSTEM_LIST} | cut -d' ' -f2-4 )
-    else 
-        JOURNALNODE_LIST=$( echo ${SEGMENT_SYSTEM_LIST} | cut -d' ' -f1-3 )
-    fi
-    echo "ZKSERVER_LIST is ${ZKSERVER_LIST}"
-    echo "JOURNALNODE_LIST is ${JOURNALNODE_LIST}"
-fi
-
-if [ "${STEP}" = "setup" ]; then
-    PHD_PACKAGE=$( curl --silent --no-buffer "http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/testing/?C=N;O=D" | grep -o 'PHD-[0-9].[0-9].[0-9].[0-9]-[0-9]*.tar.gz' | head -n 1 )
-    HDFS_DOWNLOAD_URL="http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/testing/${PHD_PACKAGE}"
-
-    echo "Hadoop Package Download URL: ${HDFS_DOWNLOAD_URL}"
-    echo "Setting up latest HDFS..."
-
-    sudo DATANODES="${DATANODES_LIST}" \
-         NAMENODE=${STANDBYHOST} \
-         SECONDARY_NAMENODE=${MASTERHOST} \
-         RESOURCEMANAGER=${STANDBYHOST} \
-         NODEMANAGERS="${DATANODES_LIST}" \
-         ARCHIVE_URL=${HDFS_DOWNLOAD_URL} \
-         JAVA_HOME=${JAVA_HOME} \
-         DCA_SETUP=${DCA_SETUP} \
-         TWO_NAMENODE_DIR=${TWO_NAMENODE_DIR} \
-         KDC=${KDC} \
-         ENABLE_QUORUM=${ENABLE_QUORUM} \
-         YARN=${YARN} \
-         PHD_USER=${USER} \
-         bash gphdinst.sh
-fi
-
-if [ -n "${KDC}" ] ; then
-    HDFS_DN_SSH_COMMAND="sudo"
-elif [ "$2" = "HK" ]; then
-    HDFS_DN_SSH_COMMAND="sudo"
-else
-    HDFS_DN_SSH_COMMAND="sudo -u hdfs"
-fi
-
-if [ "${STEP}" = "stop" ]; then
-    if [ -f "${HADOOP_HOME}/bin/stop-all.sh" ]; then 
-        ${HADOOP_HOME}/bin/stop-all.sh
-    fi   
-
-    if [ "${LOCAL_HDFS_RPMINSTALL}" = "true" ]; then
-        if [ "${USE_MASTER_AS_DATANODE}" = "true" ]; then
-            DATANODES_LIST="${SEGMENT_SYSTEM_LIST} ${MASTERHOST}"
-        else
-            DATANODES_LIST=${SEGMENT_SYSTEM_LIST}
-        fi
-
-        if [ "${ENABLE_QUORUM}" = "true" ]; then
-            echo "Stopping local HDFS HA cluster..."
-            for datanode in ${DATANODES_LIST}; do
-                ssh $datanode ${HDFS_DN_SSH_COMMAND} /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop datanode
-            done
-
-            for node in ${JOURNALNODE_LIST}; do
-                ssh $node ${HDFS_DN_SSH_COMMAND} /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop journalnode
-            done
-
-            for node in ${ZKSERVER_LIST}; do
-                ssh $node sudo JAVA_HOME=${JAVA_HOME} /etc/init.d/zookeeper-server stop
-            done
-
-            sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop zkfc
-            ssh ${STANDBYHOST} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop zkfc
-            sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop namenode
-            ssh ${STANDBYHOST} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop namenode
-        else
-            echo "Stopping local HDFS cluster..."
-
-            sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop secondarynamenode
-
-            for datanode in ${DATANODES_LIST}; do
-                ssh $datanode ${HDFS_DN_SSH_COMMAND} /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop datanode
-                if [ "${YARN}" = "true" ] ; then
-                    ssh $datanode "/usr/lib/gphd/hadoop-yarn/sbin/yarn-daemon.sh stop nodemanager"
-                fi
-            done
-
-            ssh ${STANDBYHOST} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh stop namenode
-            if [ "${YARN}" = "true" ] ; then
-                ssh ${STANDBYHOST} /usr/lib/gphd/hadoop-yarn/sbin/yarn-daemon.sh stop resourcemanager
-            fi
-         fi
-    fi
-fi
-
-if [ "${STEP}" = "start" ]; then
-
-                                if [ "${ENABLE_QUORUM}" = "true" ]; then
-                                    echo "Starting up local HDFS HA"
-
-                                    for node in ${JOURNALNODE_LIST}; do
-                                        ssh $node ${HDFS_DN_SSH_COMMAND} /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start journalnode
-                                    done
-
-                                    for node in ${ZKSERVER_LIST}; do
-                                        ssh $node sudo JAVA_HOME=${JAVA_HOME} /etc/init.d/zookeeper-server start
-                                    done
-                    
-                                    ssh ${STANDBYHOST} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start zkfc
-                                    sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start zkfc
-                                    ssh ${STANDBYHOST} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start namenode
-                                    sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start namenode
-
-                                    for datanode in ${DATANODES_LIST}; do
-                                        ssh $datanode ${HDFS_DN_SSH_COMMAND} /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start datanode
-                                    done
-                                else
-                                    echo "Starting up local HDFS"
-                                    ssh ${STANDBYHOST} sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start namenode
-                                    if [ "${YARN}" = "true" ]; then
-                                        ssh ${STANDBYHOST} /usr/lib/gphd/hadoop-yarn/sbin/yarn-daemon.sh start resourcemanager
-                                    fi
-                                    for datanode in ${DATANODES_LIST}; do
-                                        ssh $datanode ${HDFS_DN_SSH_COMMAND} /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start datanode
-                                        if [ "${YARN}" = "true" ]; then
-                                            ssh $datanode /usr/lib/gphd/hadoop-yarn/sbin/yarn-daemon.sh start nodemanager
-                                        fi
-                                    done
-                                    sudo -u hdfs /usr/lib/gphd/hadoop/sbin/hadoop-daemon.sh start secondarynamenode
-                                fi
-fi

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/.core-site-secure-ha.xml.swp
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/.core-site-secure-ha.xml.swp b/depends/libyarn/releng/bin/templates/.core-site-secure-ha.xml.swp
deleted file mode 100644
index 43e78f7..0000000
Binary files a/depends/libyarn/releng/bin/templates/.core-site-secure-ha.xml.swp and /dev/null differ

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/core-site-ha.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/core-site-ha.xml b/depends/libyarn/releng/bin/templates/core-site-ha.xml
deleted file mode 100644
index ef31461..0000000
--- a/depends/libyarn/releng/bin/templates/core-site-ha.xml
+++ /dev/null
@@ -1,28 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-  Licensed under the Apache License, Version 2.0 (the "License");
-  you may not use this file except in compliance with the License.
-  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
-    <property>
-        <name>fs.defaultFS</name>
-        <value>hdfs://gphd-cluster</value>
-    </property>
-    <property>
-        <name>ha.zookeeper.quorum</name>
-        <value>%ZKSERVER1%:2181,%ZKSERVER2%:2181,%ZKSERVER3%:2181</value>
-    </property>
-</configuration>

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/core-site-secure-ha.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/core-site-secure-ha.xml b/depends/libyarn/releng/bin/templates/core-site-secure-ha.xml
deleted file mode 100755
index cf96ce1..0000000
--- a/depends/libyarn/releng/bin/templates/core-site-secure-ha.xml
+++ /dev/null
@@ -1,39 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-  Licensed under the Apache License, Version 2.0 (the "License");
-  you may not use this file except in compliance with the License.
-  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
-    <property>
-        <name>hadoop.security.authentication</name>
-        <value>kerberos</value>
-    </property>
-    
-    <property>
-        <name>hadoop.security.authorization</name>
-        <value>true</value>
-    </property>
-
-    <property>
-        <name>fs.defaultFS</name>
-        <value>hdfs://gphd-cluster</value>
-    </property>
-
-    <property>
-        <name>ha.zookeeper.quorum</name>
-        <value>%ZKSERVER1%:2181,%ZKSERVER2%:2181,%ZKSERVER3%:2181</value>
-    </property>
-</configuration>

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/core-site-secure.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/core-site-secure.xml b/depends/libyarn/releng/bin/templates/core-site-secure.xml
deleted file mode 100755
index 1424ebb..0000000
--- a/depends/libyarn/releng/bin/templates/core-site-secure.xml
+++ /dev/null
@@ -1,34 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-  Licensed under the Apache License, Version 2.0 (the "License");
-  you may not use this file except in compliance with the License.
-  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
-    <property>
-        <name>fs.default.name</name>
-        <value>hdfs://%HDFS_HOST%</value>
-    </property>
-
-    <property>
-        <name>hadoop.security.authentication</name>
-        <value>kerberos</value>
-    </property>
-    
-    <property>
-        <name>hadoop.security.authorization</name>
-        <value>true</value>
-    </property>
-</configuration>

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/core-site.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/core-site.xml b/depends/libyarn/releng/bin/templates/core-site.xml
deleted file mode 100644
index 120e1d4..0000000
--- a/depends/libyarn/releng/bin/templates/core-site.xml
+++ /dev/null
@@ -1,24 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-  Licensed under the Apache License, Version 2.0 (the "License");
-  you may not use this file except in compliance with the License.
-  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
-    <property>
-        <name>fs.default.name</name>
-        <value>hdfs://%HDFS_HOST%</value>
-    </property>
-</configuration>

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/download-config
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/download-config b/depends/libyarn/releng/bin/templates/download-config
deleted file mode 100644
index e517b49..0000000
--- a/depends/libyarn/releng/bin/templates/download-config
+++ /dev/null
@@ -1,5 +0,0 @@
-hawq-1.1.0=http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/stable/PHD1.0.1/PHD-1.0.1.0-19.tar.gz
-hawq-1.1.1=http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/stable/PHD1.0.2/PHD-1.0.2.0-7.tar.gz
-hawq-1.1.2=http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/stable/PHD1.0.3/PHD-1.0.3.0-66.tar.gz
-hawq-1.1.3=http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/stable/PHD1.1.0/PHD-1.1.0.0-76.tar.gz
-hawq-1.1.4=http://hdp4-mdw1.wbe.dh.greenplum.com/dist/PHD/testing/PHD-1.1.1.0-82.tar.gz

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hadoop-env-ha.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hadoop-env-ha.sh b/depends/libyarn/releng/bin/templates/hadoop-env-ha.sh
deleted file mode 100644
index ede6744..0000000
--- a/depends/libyarn/releng/bin/templates/hadoop-env-ha.sh
+++ /dev/null
@@ -1,72 +0,0 @@
-# Copyright 2011 The Apache Software Foundation
-#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# Set Hadoop-specific environment variables here.
-
-# The only required environment variable is JAVA_HOME.  All others are
-# optional.  When running a distributed configuration it is best to
-# set JAVA_HOME in this file, so that it is correctly defined on
-# remote nodes.
-
-# The java implementation to use.
-export JAVA_HOME=%JAVA_HOME%
-export HADOOP_HOME=%HADOOP_HOME%
-export PATH=$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$PATH
-export HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/etc/hadoop"}
-
-# Extra Java CLASSPATH elements.  Automatically insert capacity-scheduler.
-for f in $HADOOP_HOME/contrib/capacity-scheduler/*.jar; do
-  if [ "$HADOOP_CLASSPATH" ]; then
-    export HADOOP_CLASSPATH=$HADOOP_CLASSPATH:$f
-  else
-    export HADOOP_CLASSPATH=$f
-  fi
-done
-
-# Extra Java runtime options.  Empty by default.
-export HADOOP_OPTS="-Djava.net.preferIPv4Stack=true"
-
-# Command specific options appended to HADOOP_OPTS when specified
-export HADOOP_NAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} %NAMENODE_MEMORY% $HADOOP_NAMENODE_OPTS"
-export HADOOP_DATANODE_OPTS="-Dhadoop.security.logger=ERROR,RFAS %DATANODE_MEMORY% -Xss256k $HADOOP_DATANODE_OPTS"
-
-export HADOOP_SECONDARYNAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} %SECONDARYNAMENODE_MEMORY% $HADOOP_SECONDARYNAMENODE_OPTS"
-
-# The following applies to multiple commands (fs, dfs, fsck, distcp etc)
-export HADOOP_CLIENT_OPTS="-Xmx128m $HADOOP_CLIENT_OPTS"
-
-# On secure datanodes, user to run the datanode as after dropping privileges
-export HADOOP_SECURE_DN_USER=${HADOOP_SECURE_DN_USER}
-
-# Where log files are stored.  $HADOOP_HOME/logs by default.
-#export HADOOP_LOG_DIR=${HADOOP_LOG_DIR}/$USER
-
-# Where log files are stored in the secure data environment.
-export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER}
-
-# The directory where pid files are stored. /tmp by default.
-# NOTE: this should be set to a directory that can only be written to by
-#       the user that will run the hadoop daemons.  Otherwise there is the
-#       potential for a symlink attack.
-export HADOOP_PID_DIR=${HADOOP_PID_DIR}
-export HADOOP_SECURE_DN_PID_DIR=${HADOOP_PID_DIR}
-
-# A string representing this instance of hadoop. $USER by default.
-export HADOOP_IDENT_STRING=$USER
-
-umask 022

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hadoop-env-secure.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hadoop-env-secure.sh b/depends/libyarn/releng/bin/templates/hadoop-env-secure.sh
deleted file mode 100755
index 03f9900..0000000
--- a/depends/libyarn/releng/bin/templates/hadoop-env-secure.sh
+++ /dev/null
@@ -1,73 +0,0 @@
-# Copyright 2011 The Apache Software Foundation
-#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# Set Hadoop-specific environment variables here.
-
-# The only required environment variable is JAVA_HOME.  All others are
-# optional.  When running a distributed configuration it is best to
-# set JAVA_HOME in this file, so that it is correctly defined on
-# remote nodes.
-
-# The java implementation to use.
-export JAVA_HOME=%JAVA_HOME%
-export HADOOP_HOME=%HADOOP_HOME%
-export PATH=$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$PATH
-export HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/etc/hadoop"}
-
-# Extra Java CLASSPATH elements.  Automatically insert capacity-scheduler.
-for f in $HADOOP_HOME/contrib/capacity-scheduler/*.jar; do
-  if [ "$HADOOP_CLASSPATH" ]; then
-    export HADOOP_CLASSPATH=$HADOOP_CLASSPATH:$f
-  else
-    export HADOOP_CLASSPATH=$f
-  fi
-done
-
-# Extra Java runtime options.  Empty by default.
-export HADOOP_OPTS="-Djava.net.preferIPv4Stack=true"
-
-# Command specific options appended to HADOOP_OPTS when specified
-export HADOOP_NAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} %NAMENODE_MEMORY% $HADOOP_NAMENODE_OPTS"
-export HADOOP_DATANODE_OPTS="-Dhadoop.security.logger=ERROR,RFAS %DATANODE_MEMORY% -Xss256k $HADOOP_DATANODE_OPTS"
-
-export HADOOP_SECONDARYNAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} %SECONDARYNAMENODE_MEMORY% $HADOOP_SECONDARYNAMENODE_OPTS"
-
-# The following applies to multiple commands (fs, dfs, fsck, distcp etc)
-export HADOOP_CLIENT_OPTS="-Xmx128m $HADOOP_CLIENT_OPTS"
-
-# On secure datanodes, user to run the datanode as after dropping privileges
-export HADOOP_SECURE_DN_USER=hdfs
-export JSVC_HOME=/usr/bin
-
-# Where log files are stored.  $HADOOP_HOME/logs by default.
-#export HADOOP_LOG_DIR=${HADOOP_LOG_DIR}/$USER
-
-# Where log files are stored in the secure data environment.
-export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER}
-
-# The directory where pid files are stored. /tmp by default.
-# NOTE: this should be set to a directory that can only be written to by
-#       the user that will run the hadoop daemons.  Otherwise there is the
-#       potential for a symlink attack.
-export HADOOP_PID_DIR=${HADOOP_PID_DIR}
-export HADOOP_SECURE_DN_PID_DIR=${HADOOP_PID_DIR}
-
-# A string representing this instance of hadoop. $USER by default.
-export HADOOP_IDENT_STRING=$USER
-
-umask 022

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hadoop-env.sh
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hadoop-env.sh b/depends/libyarn/releng/bin/templates/hadoop-env.sh
deleted file mode 100644
index ede6744..0000000
--- a/depends/libyarn/releng/bin/templates/hadoop-env.sh
+++ /dev/null
@@ -1,72 +0,0 @@
-# Copyright 2011 The Apache Software Foundation
-#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# Set Hadoop-specific environment variables here.
-
-# The only required environment variable is JAVA_HOME.  All others are
-# optional.  When running a distributed configuration it is best to
-# set JAVA_HOME in this file, so that it is correctly defined on
-# remote nodes.
-
-# The java implementation to use.
-export JAVA_HOME=%JAVA_HOME%
-export HADOOP_HOME=%HADOOP_HOME%
-export PATH=$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$PATH
-export HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/etc/hadoop"}
-
-# Extra Java CLASSPATH elements.  Automatically insert capacity-scheduler.
-for f in $HADOOP_HOME/contrib/capacity-scheduler/*.jar; do
-  if [ "$HADOOP_CLASSPATH" ]; then
-    export HADOOP_CLASSPATH=$HADOOP_CLASSPATH:$f
-  else
-    export HADOOP_CLASSPATH=$f
-  fi
-done
-
-# Extra Java runtime options.  Empty by default.
-export HADOOP_OPTS="-Djava.net.preferIPv4Stack=true"
-
-# Command specific options appended to HADOOP_OPTS when specified
-export HADOOP_NAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} %NAMENODE_MEMORY% $HADOOP_NAMENODE_OPTS"
-export HADOOP_DATANODE_OPTS="-Dhadoop.security.logger=ERROR,RFAS %DATANODE_MEMORY% -Xss256k $HADOOP_DATANODE_OPTS"
-
-export HADOOP_SECONDARYNAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} %SECONDARYNAMENODE_MEMORY% $HADOOP_SECONDARYNAMENODE_OPTS"
-
-# The following applies to multiple commands (fs, dfs, fsck, distcp etc)
-export HADOOP_CLIENT_OPTS="-Xmx128m $HADOOP_CLIENT_OPTS"
-
-# On secure datanodes, user to run the datanode as after dropping privileges
-export HADOOP_SECURE_DN_USER=${HADOOP_SECURE_DN_USER}
-
-# Where log files are stored.  $HADOOP_HOME/logs by default.
-#export HADOOP_LOG_DIR=${HADOOP_LOG_DIR}/$USER
-
-# Where log files are stored in the secure data environment.
-export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER}
-
-# The directory where pid files are stored. /tmp by default.
-# NOTE: this should be set to a directory that can only be written to by
-#       the user that will run the hadoop daemons.  Otherwise there is the
-#       potential for a symlink attack.
-export HADOOP_PID_DIR=${HADOOP_PID_DIR}
-export HADOOP_SECURE_DN_PID_DIR=${HADOOP_PID_DIR}
-
-# A string representing this instance of hadoop. $USER by default.
-export HADOOP_IDENT_STRING=$USER
-
-umask 022

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure-ha.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure-ha.xml b/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure-ha.xml
deleted file mode 100755
index 966234d..0000000
--- a/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure-ha.xml
+++ /dev/null
@@ -1,240 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-  Licensed under the Apache License, Version 2.0 (the "License");
-  you may not use this file except in compliance with the License.
-  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
-    <property>
-        <name>dfs.encrypt.data.transfer</name>
-        <value>false</value>
-    </property>
-    <property>
-        <name>dfs.encrypt.data.transfer.algorithm</name>
-        <value>rc4</value>
-        <description>may be "rc4" or "3des" - 3des has a significant performance impact</description>
-    </property>
-    <property>
-        <name>dfs.datanode.keytab.file</name>
-        <value>/etc/gphd/hawq-krb5.keytab</value>
-    </property>
-    <property>
-        <name>dfs.datanode.kerberos.principal</name>
-        <value>hdfs/_HOST@HAWQ.PIVOTAL.COM</value>
-    </property>
-    <property>
-      <name>dfs.datanode.kerberos.http.principal</name>
-      <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
-    </property>
-    <property>
-      <name>dfs.datanode.address</name>
-      <value>0.0.0.0:1004</value>
-    </property>
-    <property>
-      <name>dfs.datanode.http.address</name>
-      <value>0.0.0.0:1006</value>
-    </property>
-    <property>
-        <name>dfs.datanode.data.dir.perm</name>
-        <value>755</value>
-    </property>
-    <property>
-        <name>dfs.journalnode.keytab.file</name>
-        <value>/etc/gphd/hawq-krb5.keytab</value>
-    </property>
-    <property>
-        <name>dfs.journalnode.kerberos.principal</name>
-        <value>hdfs/_HOST@HAWQ.PIVOTAL.COM</value>
-    </property>
-    <property>
-      <name>dfs.journalnode.kerberos.internal.spnego.principal</name>
-      <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
-    </property>
-    <property>
-        <name>dfs.namenode.keytab.file</name>
-        <value>/etc/gphd/hawq-krb5.keytab</value>
-    </property>
-    <property>
-        <name>dfs.namenode.kerberos.principal</name>
-        <value>hdfs/_HOST@HAWQ.PIVOTAL.COM</value>
-    </property>
-    <property>
-      <name>dfs.namenode.kerberos.http.principal</name>
-      <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
-    </property>
-    <property>
-      <name>dfs.namenode.kerberos.internal.spnego.principal</name>
-      <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
-    </property>
-
-    <property>
-      <name>dfs.block.access.token.enable</name>
-      <value>true</value>
-    </property>
-
-    <property>
-      <name>dfs.web.authentication.kerberos.principal</name>
-      <value>HTTP/_HOST@HAWQ.PIVOTAL.COM</value>
-    </property>
-    <property>
-      <name>dfs.web.authentication.kerberos.keytab</name>
-      <value>/etc/gphd/hawq-krb5.keytab</value>
-    </property>
-    <property>
-        <name>dfs.web.ugi</name>
-        <value>hdfs,hdfs</value>
-        <description>The user account used by the web interface.</description>
-    </property>
-    <property>
-        <name>dfs.permissions</name>
-        <value>true</value>
-    </property>
-    <property>
-        <name>dfs.webhdfs.enabled</name>
-        <value>true</value>
-    </property>
-    <property>
-        <name>dfs.support.append</name>
-        <value>true</value>
-    </property>
-    <property>
-        <name>dfs.client.read.shortcircuit</name>
-        <value>false</value>
-    </property>
-    <property>
-        <name>dfs.block.local-path-access.user</name>
-        <value>gpadmin</value>
-        <description>
-            Specify the user allowed to do short circuit read
-        </description>
-    </property>
-    <property>
-        <name>dfs.safemode.extension</name>
-        <value>0</value>
-    </property>
-    <property>
-        <name>dfs.safemode.min.datanodes</name>
-        <value>1</value>
-    </property>
-    <property>
-        <name>hadoop.tmp.dir</name>
-        <value>%DATA_DIR_ROOT%</value>
-    </property>
-    <property>
-        <name>dfs.namenode.name.dir</name>
-        <value>file:%DATA_DIR_ROOT%/hdfs/name</value>
-    </property>
-    <property>
-        <name>dfs.namenode.checkpoint.dir</name>
-        <value>file:%DATA_DIR_ROOT%/hdfs/namesecondary</value>
-    </property>
-    <property>
-        <name>dfs.namenode.checkpoint.period</name>
-        <value>3600</value>
-    </property>
-    <property>
-        <name>dfs.datanode.data.dir</name>
-        <value>%HDFS_DATANODE_DIR%</value>
-    </property>
-    <property>
-        <name>dfs.replication</name>
-        <value>3</value>
-    </property>
-    <property>
-        <name>dfs.datanode.max.transfer.threads</name>
-        <value>40960</value>
-    </property>
-    <property>
-        <name>dfs.client.socket-timeout</name>
-        <value>300000000</value>
-    </property>
-    <property>
-        <name>dfs.datanode.handler.count</name>
-        <value>60</value>
-    </property>
-    <property>
-        <name>dfs.namenode.handler.count</name>
-        <value>60</value>
-    </property>
-    <property>
-        <name>ipc.client.connection.maxidletime</name>
-        <value>3600000</value>
-    </property>
-    <property>
-        <name>ipc.server.handler.queue.size</name>
-        <value>3300</value>
-    </property>
-    <property>
-        <name>ipc.client.connection</name>
-        <value>3</value>
-    </property>
-    <property>
-        <name>dfs.namenode.accesstime.precision</name>
-        <value>-1</value>
-    </property>
-
-    <property>
-        <name>dfs.nameservices</name>
-        <value>gphd-cluster</value>
-    </property>
-    <property>
-        <name>dfs.ha.namenodes.gphd-cluster</name>
-        <value>nn1,nn2</value>
-    </property>
-    <property>
-        <name>dfs.namenode.rpc-address.gphd-cluster.nn1</name>
-        <value>%NAMENODE%:9000</value>
-    </property>
-    <property>
-        <name>dfs.namenode.rpc-address.gphd-cluster.nn2</name>
-        <value>%SECONDARY_NAMENODE%:9000</value>
-    </property>
-    <property>
-        <name>dfs.namenode.http-address.gphd-cluster.nn1</name>
-        <value>%NAMENODE%:50070</value>
-    </property>
-    <property>
-        <name>dfs.namenode.http-address.gphd-cluster.nn2</name>
-        <value>%SECONDARY_NAMENODE%:50070</value>
-    </property>
-    <property>
-        <name>dfs.namenode.shared.edits.dir</name>
-        <value>qjournal://%ZKSERVER1%:8485;%ZKSERVER2%:8485;%ZKSERVER3%:8485/gphd-cluster</value>
-    </property>
-    <property>
-        <name>dfs.journalnode.edits.dir</name>
-        <value>%JOURNAL_DATA_ROOT%</value> 
-    </property>
-    <property>
-        <name>dfs.client.failover.proxy.provider.gphd-cluster</name>
-        <value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
-    </property>
-    <property>
-        <name>dfs.ha.fencing.methods</name>
-        <value>shell(/bin/true)</value>
-    </property>
-    <property>
-        <name>dfs.ha.fencing.ssh.connect-timeout</name>
-        <value>20000</value>
-    </property>
-    <property>
-        <name>dfs.ha.automatic-failover.enabled</name>
-        <value>true</value>
-    </property>
-        <property>
-        <name>dfs.namenode.fs-limits.min-block-size</name>
-        <value>1024</value>
-    </property>
-</configuration>

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure.xml b/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure.xml
deleted file mode 100755
index 89b33cc..0000000
--- a/depends/libyarn/releng/bin/templates/hdfs-site-datanode-secure.xml
+++ /dev/null
@@ -1,193 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-  Licensed under the Apache License, Version 2.0 (the "License");
-  you may not use this file except in compliance with the License.
-  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
-    <property>
-        <name>dfs.encrypt.data.transfer</name>
-        <value>false</value>
-    </property>
-    <property>
-        <name>dfs.encrypt.data.transfer.algorithm</name>
-        <value>rc4</value>
-        <description>may be "rc4" or "3des" - 3des has a significant performance impact</description>
-    </property>
-    <property>
-        <name>dfs.datanode.keytab.file</name>
-        <value>/etc/gphd/hawq-krb5.keytab</value>
-    </property>
-    <property>
-        <name>dfs.datanode.kerberos.principal</name>
-        <value>hdfs/%DATANODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
-    </property>
-    <property>
-      <name>dfs.datanode.kerberos.http.principal</name>
-      <value>HTTP/%DATANODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
-    </property>
-    <property>
-      <name>dfs.datanode.address</name>
-      <value>0.0.0.0:1004</value>
-    </property>
-    <property>
-      <name>dfs.datanode.http.address</name>
-      <value>0.0.0.0:1006</value>
-    </property>
-    <property>
-        <name>dfs.datanode.data.dir.perm</name>
-        <value>755</value>
-    </property>
-    <property>
-        <name>dfs.namenode.keytab.file</name>
-        <value>/etc/gphd/hawq-krb5.keytab</value>
-    </property>
-    <property>
-        <name>dfs.namenode.kerberos.principal</name>
-        <value>hdfs/%NAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
-    </property>
-    <property>
-      <name>dfs.namenode.kerberos.http.principal</name>
-      <value>HTTP/%NAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
-    </property>
-    <property>
-      <name>dfs.namenode.kerberos.internal.spnego.principal</name>
-      <value>HTTP/%NAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
-    </property>
-
-    <property>
-      <name>dfs.secondary.namenode.keytab.file</name>
-      <value>/etc/gphd/hawq-krb5.keytab</value>
-    </property>
-    <property>
-      <name>dfs.secondary.namenode.kerberos.principal</name>
-      <value>hdfs/%SECONDARYNAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
-    </property>
-    <property>
-      <name>dfs.secondary.namenode.kerberos.http.principal</name>
-      <value>HTTP/%SECONDARYNAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
-    </property>
-    <property>
-      <name>dfs.secondary.namenode.kerberos.internal.spnego.principal</name>
-      <value>HTTP/%SECONDARYNAMENODE_PRINCIPAL%@HAWQ.PIVOTAL.COM</value>
-    </property>
-
-    <property>
-      <name>dfs.block.access.token.enable</name>
-      <value>true</value>
-    </property>
-
-    <property>
-      <name>dfs.web.authentication.kerberos.principal</name>
-      <value>HTTP/%WEBAUTH_HOST%@HAWQ.PIVOTAL.COM</value>
-    </property>
-    <property>
-      <name>dfs.web.authentication.kerberos.keytab</name>
-      <value>/etc/gphd/hawq-krb5.keytab</value>
-    </property>
-
-    <property>
-        <name>dfs.web.ugi</name>
-        <value>hdfs,hdfs</value>
-        <description>The user account used by the web interface.</description>
-    </property>
-    <property>
-        <name>dfs.permissions</name>
-        <value>true</value>
-    </property>
-    <property>
-        <name>dfs.webhdfs.enabled</name>
-        <value>true</value>
-    </property>
-    <property>
-        <name>dfs.support.append</name>
-        <value>true</value>
-    </property>
-    <property>
-        <name>dfs.client.read.shortcircuit</name>
-        <value>false</value>
-    </property>
-    <property>
-        <name>dfs.block.local-path-access.user</name>
-        <value>gpadmin</value>
-        <description>
-            Specify the user allowed to do short circuit read
-        </description>
-    </property>
-    <property>
-        <name>dfs.safemode.extension</name>
-        <value>0</value>
-    </property>
-    <property>
-        <name>dfs.safemode.min.datanodes</name>
-        <value>1</value>
-    </property>
-    <property>
-        <name>hadoop.tmp.dir</name>
-        <value>%DATA_DIR_ROOT%</value>
-    </property>
-    <property>
-        <name>dfs.namenode.name.dir</name>
-        <value>file:%DATA_DIR_ROOT%/hdfs/name</value>
-    </property>
-    <property>
-        <name>dfs.namenode.checkpoint.dir</name>
-        <value>file:%DATA_DIR_ROOT%/hdfs/namesecondary</value>
-    </property>
-    <property>
-        <name>dfs.namenode.checkpoint.period</name>
-        <value>3600</value>
-    </property>
-    <property>
-        <name>dfs.datanode.data.dir</name>
-        <value>%HDFS_DATANODE_DIR%</value>
-    </property>
-    <property>
-        <name>dfs.replication</name>
-        <value>3</value>
-    </property>
-    <property>
-        <name>dfs.datanode.max.transfer.threads</name>
-        <value>40960</value>
-    </property>
-    <property>
-        <name>dfs.client.socket-timeout</name>
-        <value>300000000</value>
-    </property>
-    <property>
-        <name>dfs.datanode.handler.count</name>
-        <value>60</value>
-    </property>
-    <property>
-        <name>dfs.namenode.handler.count</name>
-        <value>60</value>
-    </property>
-    <property>
-        <name>ipc.client.connection.maxidletime</name>
-        <value>3600000</value>
-    </property>
-    <property>
-        <name>ipc.server.handler.queue.size</name>
-        <value>3300</value>
-    </property>
-    <property>
-        <name>ipc.client.connection</name>
-        <value>3</value>
-    </property>
-    <property>
-        <name>dfs.namenode.accesstime.precision</name>
-        <value>-1</value>
-    </property>
-</configuration>

http://git-wip-us.apache.org/repos/asf/incubator-hawq/blob/10909db4/depends/libyarn/releng/bin/templates/hdfs-site-ha.xml
----------------------------------------------------------------------
diff --git a/depends/libyarn/releng/bin/templates/hdfs-site-ha.xml b/depends/libyarn/releng/bin/templates/hdfs-site-ha.xml
deleted file mode 100644
index 7fb7c1d..0000000
--- a/depends/libyarn/releng/bin/templates/hdfs-site-ha.xml
+++ /dev/null
@@ -1,172 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-  Licensed under the Apache License, Version 2.0 (the "License");
-  you may not use this file except in compliance with the License.
-  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
-    <property>
-        <name>dfs.web.ugi</name>
-        <value>hdfs,hdfs</value>
-        <description>The user account used by the web interface.</description>
-    </property>
-    <property>
-        <name>dfs.permissions</name>
-        <value>true</value>
-    </property>
-    <property>
-        <name>dfs.webhdfs.enabled</name>
-        <value>true</value>
-    </property>
-    <property>
-        <name>dfs.support.append</name>
-        <value>true</value>
-    </property>
-    <property>
-        <name>dfs.client.read.shortcircuit</name>
-        <value>false</value>
-    </property>
-    <property>
-        <name>dfs.block.local-path-access.user</name>
-        <value>gpadmin</value>
-        <description>
-            Specify the user allowed to do short circuit read
-        </description>
-    </property>
-    <property>
-        <name>dfs.safemode.extension</name>
-        <value>0</value>
-    </property>
-    <property>
-        <name>dfs.safemode.min.datanodes</name>
-        <value>1</value>
-    </property>
-    <property>
-        <name>hadoop.tmp.dir</name>
-        <value>%DATA_DIR_ROOT%</value>
-    </property>
-    <property>
-        <name>dfs.namenode.name.dir</name>
-        <value>file:%DATA_DIR_ROOT%/hdfs/name</value>
-    </property>
-    <property>
-        <name>dfs.namenode.checkpoint.dir</name>
-        <value>file:%DATA_DIR_ROOT%/hdfs/namesecondary</value>
-    </property>
-    <property>
-        <name>dfs.namenode.checkpoint.period</name>
-        <value>3600</value>
-    </property>
-    <property>
-        <name>dfs.datanode.data.dir</name>
-        <value>%HDFS_DATANODE_DIR%</value>
-    </property>
-    <property>
-        <name>dfs.datanode.data.dir.perm</name>
-        <value>755</value>
-    </property>
-    <property>
-        <name>dfs.replication</name>
-        <value>3</value>
-    </property>
-    <property>
-        <name>dfs.datanode.max.transfer.threads</name>
-        <value>40960</value>
-    </property>
-    <property>
-        <name>dfs.client.socket-timeout</name>
-        <value>300000000</value>
-    </property>
-    <property>
-        <name>dfs.datanode.handler.count</name>
-        <value>60</value>
-    </property>
-    <property>
-        <name>dfs.namenode.handler.count</name>
-        <value>60</value>
-    </property>
-    <property>
-        <name>ipc.client.connection.maxidletime</name>
-        <value>3600000</value>
-    </property>
-    <property>
-        <name>ipc.server.handler.queue.size</name>
-        <value>3300</value>
-    </property>
-    <property>
-        <name>ipc.client.connection</name>
-        <value>3</value>
-    </property>
-    <property>
-        <name>dfs.namenode.accesstime.precision</name>
-        <value>-1</value>
-    </property>
-    <property>
-        <name>dfs.datanode.http.address</name>
-        <value>0.0.0.0:59075</value>
-        <description>
-            The datanode http server address and port. If the port is 0 then the server will start on a free port.
-        </description>
-    </property>
-
-    <property>
-      <name>dfs.nameservices</name>
-      <value>gphd-cluster</value>
-    </property>
-    <property>
-      <name>dfs.ha.namenodes.gphd-cluster</name>
-      <value>nn1,nn2</value>
-    </property>
-    <property>
-      <name>dfs.namenode.rpc-address.gphd-cluster.nn1</name>
-      <value>%NAMENODE%:9000</value>
-    </property>
-    <property>
-      <name>dfs.namenode.rpc-address.gphd-cluster.nn2</name>
-      <value>%SECONDARY_NAMENODE%:9000</value>
-    </property>
-    <property>
-      <name>dfs.namenode.http-address.gphd-cluster.nn1</name>
-      <value>%NAMENODE%:50070</value>
-    </property>
-    <property>
-      <name>dfs.namenode.http-address.gphd-cluster.nn2</name>
-      <value>%SECONDARY_NAMENODE%:50070</value>
-    </property>
-    <property>
-      <name>dfs.namenode.shared.edits.dir</name>
-      <value>qjournal://%JOURNALNODE1%:8485;%JOURNALNODE2%:8485;%JOURNALNODE3%:8485/gphd-cluster</value>
-    </property>
-    <property>
-      <name>dfs.journalnode.edits.dir</name>
-      <value>%JOURNAL_DATA_ROOT%</value>
-    </property>
-    <property>
-      <name>dfs.client.failover.proxy.provider.gphd-cluster</name>
-      <value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
-    </property>
-    <property>
-      <name>dfs.ha.fencing.methods</name>
-      <value>shell(/bin/true)</value>
-    </property>
-    <property>
-      <name>dfs.ha.fencing.ssh.connect-timeout</name>
-      <value>20000</value>
-    </property>
-    <property>
-      <name>dfs.ha.automatic-failover.enabled</name>
-      <value>true</value>
-    </property>
-</configuration>



Mime
View raw message