bigtop-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From ofle...@apache.org
Subject [bigtop] 01/01: zookeeper-3.4.11 hadoop-3.1.2 maven enforcer
Date Sat, 22 Jun 2019 22:16:03 GMT
This is an automated email from the ASF dual-hosted git repository.

oflebbe pushed a commit to branch bigtop-alpha
in repository https://gitbox.apache.org/repos/asf/bigtop.git

commit fbccab4c92cd41b44a83202ec468da20695b791b
Author: Olaf Flebbe <of@oflebbe.de>
AuthorDate: Thu Dec 28 16:18:23 2017 +0100

    zookeeper-3.4.11
    hadoop-3.1.2
    maven enforcer
---
 .../src/common/hadoop/do-component-build           |  10 +-
 .../src/common/hadoop/install_hadoop.sh            |  36 ++---
 ...tch0-HADOOP-12366.-expose-calculated-paths.diff | 125 ---------------
 .../hadoop/patch0-HADOOP-15122-trunk.002.diff      |  13 ++
 .../src/common/hadoop/patch0-maven-enforcer.diff   |  40 +++++
 .../src/common/hadoop/patch2-HADOOP-14597.diff     | 143 -----------------
 .../src/common/zookeeper/do-component-build        |   6 +-
 .../common/zookeeper/patch0-ZOOKEEPER-1897v3.diff  |  14 --
 .../common/zookeeper/patch1-ZOOKEEPER-1911.diff    |  25 ---
 .../common/zookeeper/patch2-ZOOKEEPER-1926.diff    | 177 ---------------------
 .../common/zookeeper/patch3-ZOOKEEPER-2064.diff    | 121 --------------
 .../src/common/zookeeper/patch4-removeasm.diff     |  17 --
 .../common/zookeeper/patch5-ZOOKEEPER-2594.diff    |  66 --------
 bigtop-packages/src/deb/hadoop/control             |   4 +-
 .../src/deb/hadoop/hadoop-httpfs.install           |   1 -
 .../src/deb/hadoop/hadoop-mapreduce.install        |   1 -
 bigtop-packages/src/deb/hadoop/hadoop.install      |   2 -
 bigtop-packages/src/deb/hadoop/rules               |   4 -
 bigtop-packages/src/rpm/hadoop/SPECS/hadoop.spec   |  17 +-
 bigtop.bom                                         |  14 +-
 20 files changed, 88 insertions(+), 748 deletions(-)

diff --git a/bigtop-packages/src/common/hadoop/do-component-build b/bigtop-packages/src/common/hadoop/do-component-build
index 2a1a634..75e27ac 100644
--- a/bigtop-packages/src/common/hadoop/do-component-build
+++ b/bigtop-packages/src/common/hadoop/do-component-build
@@ -131,18 +131,18 @@ MAVEN_OPTS="$MAVEN_OPTS -DskipTests -DskipTest -DskipITs"
 . $(dirname ${0})/maven_deploy.sh
 
 # Build artifacts
-mvn $ANT_OPTS $BUNDLE_SNAPPY -Pdist -Pnative -Psrc -Dtar ${MAVEN_OPTS} -Dtomcat.version=${BIGTOP_TOMCAT_VERSION} -Dtomcat.download.url="$BIGTOP_TOMCAT_DOWNLOAD_URL" install package ${EXTRA_GOALS} "$@"
-mvn site site:stage ${MAVEN_OPTS} $@
+mvn $ANT_OPTS $BUNDLE_SNAPPY -Pdist -Pnative -Psrc -Dtar ${MAVEN_OPTS}" install package "$@"
+#mvn site site:stage ${MAVEN_OPTS} $@
 
 (cd build ; tar --strip-components=1 -xzvf  ../hadoop-dist/target/hadoop-${HADOOP_VERSION}.tar.gz)
-(cd build/src ; tar --strip-components=1 -xzvf  ../../hadoop-dist/target/hadoop-${HADOOP_VERSION}-src.tar.gz)
+#(cd build/src ; tar --strip-components=1 -xzvf  ../../hadoop-dist/target/hadoop-${HADOOP_VERSION}-src.tar.gz)
 
 # Copy generated docs
 mkdir -p build/share
-cp -r target/staging/hadoop-project build/share/doc
+#cp -r target/staging/hadoop-project build/share/doc
 
 # Create a manifest for hadoop client package
-(cd hadoop-client/target/hadoop-client-*/share/hadoop/client/lib ; ls) > build/hadoop-client.list
+(cd hadoop-client-modules/hadoop-client/target/hadoop-client-*/share/hadoop/client/lib ; ls) > build/hadoop-client.list
 
 # Copy fuse output to the build directory
 cp hadoop-hdfs-project/hadoop-hdfs-native-client/target/main/native/fuse-dfs/fuse_dfs build/bin
diff --git a/bigtop-packages/src/common/hadoop/install_hadoop.sh b/bigtop-packages/src/common/hadoop/install_hadoop.sh
index 9957dcf..fcfbf96 100755
--- a/bigtop-packages/src/common/hadoop/install_hadoop.sh
+++ b/bigtop-packages/src/common/hadoop/install_hadoop.sh
@@ -181,7 +181,7 @@ done
 
 #libexec
 install -d -m 0755 ${SYSTEM_LIBEXEC_DIR}
-cp ${BUILD_DIR}/libexec/* ${SYSTEM_LIBEXEC_DIR}/
+cp -r ${BUILD_DIR}/libexec/* ${SYSTEM_LIBEXEC_DIR}/
 cp ${DISTRO_DIR}/hadoop-layout.sh ${SYSTEM_LIBEXEC_DIR}/
 install -m 0755 ${DISTRO_DIR}/init-hdfs.sh ${SYSTEM_LIBEXEC_DIR}/
 install -m 0755 ${DISTRO_DIR}/init-hcfs.json ${SYSTEM_LIBEXEC_DIR}/
@@ -192,7 +192,7 @@ rm -rf ${SYSTEM_LIBEXEC_DIR}/*.cmd
 install -d -m 0755 ${HADOOP_DIR}
 cp ${BUILD_DIR}/share/hadoop/common/*.jar ${HADOOP_DIR}/
 cp ${BUILD_DIR}/share/hadoop/common/lib/hadoop-auth*.jar ${HADOOP_DIR}/
-cp ${BUILD_DIR}/share/hadoop/mapreduce/lib/hadoop-annotations*.jar ${HADOOP_DIR}/
+cp ${BUILD_DIR}/share/hadoop/common/lib/hadoop-annotations*.jar ${HADOOP_DIR}/
 install -d -m 0755 ${MAPREDUCE_DIR}
 cp ${BUILD_DIR}/share/hadoop/mapreduce/hadoop-mapreduce*.jar ${MAPREDUCE_DIR}
 cp ${BUILD_DIR}/share/hadoop/tools/lib/*.jar ${MAPREDUCE_DIR}
@@ -205,20 +205,18 @@ chmod 644 ${HADOOP_DIR}/*.jar ${MAPREDUCE_DIR}/*.jar ${HDFS_DIR}/*.jar ${YARN_DI
 # lib jars
 install -d -m 0755 ${HADOOP_DIR}/lib
 cp ${BUILD_DIR}/share/hadoop/common/lib/*.jar ${HADOOP_DIR}/lib
-install -d -m 0755 ${MAPREDUCE_DIR}/lib
-cp ${BUILD_DIR}/share/hadoop/mapreduce/lib/*.jar ${MAPREDUCE_DIR}/lib
 install -d -m 0755 ${HDFS_DIR}/lib 
 cp ${BUILD_DIR}/share/hadoop/hdfs/lib/*.jar ${HDFS_DIR}/lib
 install -d -m 0755 ${YARN_DIR}/lib
 cp ${BUILD_DIR}/share/hadoop/yarn/lib/*.jar ${YARN_DIR}/lib
-chmod 644 ${HADOOP_DIR}/lib/*.jar ${MAPREDUCE_DIR}/lib/*.jar ${HDFS_DIR}/lib/*.jar ${YARN_DIR}/lib/*.jar
+chmod 644 ${HADOOP_DIR}/lib/*.jar ${HDFS_DIR}/lib/*.jar ${YARN_DIR}/lib/*.jar
 
 # Install webapps
 cp -ra ${BUILD_DIR}/share/hadoop/hdfs/webapps ${HDFS_DIR}/
 
 # bin
 install -d -m 0755 ${HADOOP_DIR}/bin
-cp -a ${BUILD_DIR}/bin/{hadoop,rcc,fuse_dfs} ${HADOOP_DIR}/bin
+cp -a ${BUILD_DIR}/bin/{hadoop,fuse_dfs} ${HADOOP_DIR}/bin
 install -d -m 0755 ${HDFS_DIR}/bin
 cp -a ${BUILD_DIR}/bin/hdfs ${HDFS_DIR}/bin
 install -d -m 0755 ${YARN_DIR}/bin
@@ -230,7 +228,7 @@ cp -a ${BUILD_DIR}/bin/mapred ${YARN_DIR}/bin
 
 # sbin
 install -d -m 0755 ${HADOOP_DIR}/sbin
-cp -a ${BUILD_DIR}/sbin/{hadoop-daemon,hadoop-daemons,slaves}.sh ${HADOOP_DIR}/sbin
+cp -a ${BUILD_DIR}/sbin/{hadoop-daemon,hadoop-daemons}.sh ${HADOOP_DIR}/sbin
 install -d -m 0755 ${HDFS_DIR}/sbin
 cp -a ${BUILD_DIR}/sbin/{distribute-exclude,refresh-namenodes}.sh ${HDFS_DIR}/sbin
 install -d -m 0755 ${YARN_DIR}/sbin
@@ -244,7 +242,7 @@ install -d -m 0755 ${HADOOP_NATIVE_LIB_DIR}
 for library in libhdfs.so.0.0.0; do
   cp ${BUILD_DIR}/lib/native/${library} ${SYSTEM_LIB_DIR}/
   ldconfig -vlN ${SYSTEM_LIB_DIR}/${library}
-  ln -s ${library} ${SYSTEM_LIB_DIR}/${library/.so.*/}.so
+  ln -sf ${library} ${SYSTEM_LIB_DIR}/${library/.so.*/}.so
 done
 
 install -d -m 0755 ${SYSTEM_INCLUDE_DIR}
@@ -254,7 +252,7 @@ cp ${BUILD_DIR}/lib/native/*.a ${HADOOP_NATIVE_LIB_DIR}/
 for library in `cd ${BUILD_DIR}/lib/native ; ls libsnappy.so.1.* 2>/dev/null` libhadoop.so.1.0.0; do
   cp ${BUILD_DIR}/lib/native/${library} ${HADOOP_NATIVE_LIB_DIR}/
   ldconfig -vlN ${HADOOP_NATIVE_LIB_DIR}/${library}
-  ln -s ${library} ${HADOOP_NATIVE_LIB_DIR}/${library/.so.*/}.so
+  ln -sf ${library} ${HADOOP_NATIVE_LIB_DIR}/${library/.so.*/}.so
 done
 
 # Install fuse wrapper
@@ -304,7 +302,7 @@ cp ${DISTRO_DIR}/conf.empty/mapred-site.xml $HADOOP_ETC_DIR/conf.empty
 # so that it can still be used as example, but doesn't affect anything
 # by default
 sed -i -e '/^[^#]/s,^,#,' ${BUILD_DIR}/etc/hadoop/hadoop-env.sh
-cp ${BUILD_DIR}/etc/hadoop/* $HADOOP_ETC_DIR/conf.empty
+cp -r ${BUILD_DIR}/etc/hadoop/* $HADOOP_ETC_DIR/conf.empty
 rm -rf $HADOOP_ETC_DIR/conf.empty/*.cmd
 
 # docs
@@ -321,7 +319,7 @@ done
 # HTTPFS
 install -d -m 0755 ${HTTPFS_DIR}/sbin
 cp ${BUILD_DIR}/sbin/httpfs.sh ${HTTPFS_DIR}/sbin/
-cp -r ${BUILD_DIR}/share/hadoop/httpfs/tomcat/webapps ${HTTPFS_DIR}/webapps
+#???cp -r ${BUILD_DIR}/share/hadoop/httpfs/tomcat/webapps ${HTTPFS_DIR}/webapps
 install -d -m 0755 ${PREFIX}/var/lib/hadoop-httpfs
 install -d -m 0755 $HTTPFS_ETC_DIR/conf.empty
 
@@ -331,14 +329,14 @@ HTTP_DIRECTORY=$HTTPFS_ETC_DIR/tomcat-conf.dist
 HTTPS_DIRECTORY=$HTTPFS_ETC_DIR/tomcat-conf.https
 
 install -d -m 0755 ${HTTP_DIRECTORY}
-cp -r ${BUILD_DIR}/share/hadoop/httpfs/tomcat/conf ${HTTP_DIRECTORY}
-chmod 644 ${HTTP_DIRECTORY}/conf/*
+#??cp -r ${BUILD_DIR}/share/hadoop/httpfs/tomcat/conf ${HTTP_DIRECTORY}
+#?? chmod 644 ${HTTP_DIRECTORY}/conf/*
 install -d -m 0755 ${HTTP_DIRECTORY}/WEB-INF
-mv ${HTTPFS_DIR}/webapps/webhdfs/WEB-INF/*.xml ${HTTP_DIRECTORY}/WEB-INF/
+#??mv ${HTTPFS_DIR}/webapps/webhdfs/WEB-INF/*.xml ${HTTP_DIRECTORY}/WEB-INF/
 
 cp -r ${HTTP_DIRECTORY} ${HTTPS_DIRECTORY}
-mv ${HTTPS_DIRECTORY}/conf/ssl-server.xml ${HTTPS_DIRECTORY}/conf/server.xml
-rm ${HTTP_DIRECTORY}/conf/ssl-server.xml
+#??mv ${HTTPS_DIRECTORY}/conf/ssl-server.xml ${HTTPS_DIRECTORY}/conf/server.xml
+#?? rm ${HTTP_DIRECTORY}/conf/ssl-server.xml
 
 mv $HADOOP_ETC_DIR/conf.empty/httpfs* $HTTPFS_ETC_DIR/conf.empty
 sed -i -e '/<\/configuration>/i\
@@ -351,7 +349,7 @@ sed -i -e '/<\/configuration>/i\
 for conf in conf.pseudo ; do
   install -d -m 0755 $HADOOP_ETC_DIR/$conf
   # Install the upstream config files
-  cp ${BUILD_DIR}/etc/hadoop/* $HADOOP_ETC_DIR/$conf
+  cp -r ${BUILD_DIR}/etc/hadoop/* $HADOOP_ETC_DIR/$conf
   # Remove the ones that shouldn't be installed
   rm -rf $HADOOP_ETC_DIR/$conf/httpfs*
   rm -rf $HADOOP_ETC_DIR/$conf/*.cmd
@@ -365,9 +363,9 @@ cp ${BUILD_DIR}/etc/hadoop/log4j.properties $HADOOP_ETC_DIR/conf.pseudo
 
 # FIXME: Provide a convenience link for configuration (HADOOP-7939)
 install -d -m 0755 ${HADOOP_DIR}/etc
-ln -s ${HADOOP_ETC_DIR##${PREFIX}}/conf ${HADOOP_DIR}/etc/hadoop
+ln -sf ${HADOOP_ETC_DIR##${PREFIX}}/conf ${HADOOP_DIR}/etc/hadoop
 install -d -m 0755 ${YARN_DIR}/etc
-ln -s ${HADOOP_ETC_DIR##${PREFIX}}/conf ${YARN_DIR}/etc/hadoop
+ln -sf ${HADOOP_ETC_DIR##${PREFIX}}/conf ${YARN_DIR}/etc/hadoop
 
 # Create log, var and lib
 install -d -m 0755 $PREFIX/var/{log,run,lib}/hadoop-hdfs
diff --git a/bigtop-packages/src/common/hadoop/patch0-HADOOP-12366.-expose-calculated-paths.diff b/bigtop-packages/src/common/hadoop/patch0-HADOOP-12366.-expose-calculated-paths.diff
deleted file mode 100644
index 026c01f..0000000
--- a/bigtop-packages/src/common/hadoop/patch0-HADOOP-12366.-expose-calculated-paths.diff
+++ /dev/null
@@ -1,125 +0,0 @@
-From f10ec9f9bec56bcfa5e2015ea87a8a5054d096c5 Mon Sep 17 00:00:00 2001
-From: Anton Chevychalov <pub@mnu.pp.ru>
-Date: Tue, 5 Sep 2017 15:14:19 +0300
-Subject: [PATCH] [PATCH] HADOOP-12366. expose calculated paths
-
-Adapted the patch to apply cleanly on branch-2.8.1
----
- hadoop-common-project/hadoop-common/src/main/bin/hadoop | 10 ++++++++++
- hadoop-hdfs-project/hadoop-hdfs/src/main/bin/hdfs       |  9 +++++++++
- hadoop-mapreduce-project/bin/mapred                     |  9 +++++++++
- hadoop-yarn-project/hadoop-yarn/bin/yarn                |  9 +++++++++
- 4 files changed, 37 insertions(+)
-
-diff --git a/hadoop-common-project/hadoop-common/src/main/bin/hadoop b/hadoop-common-project/hadoop-common/src/main/bin/hadoop
-index 1575996..44be2e3 100755
---- a/hadoop-common-project/hadoop-common/src/main/bin/hadoop
-+++ b/hadoop-common-project/hadoop-common/src/main/bin/hadoop
-@@ -38,6 +38,7 @@ function print_usage(){
-   echo "                             YARN applications, not this command."
-   echo "  checknative [-a|-h]  check native hadoop and compression libraries availability"
-   echo "  distcp <srcurl> <desturl> copy file or directories recursively"
-+  echo "  envvars              display computed Hadoop environment variables"
-   echo "  archive -archiveName NAME -p <parent path> <src>* <dest> create a hadoop archive"
-   echo "  classpath            prints the class path needed to get the"
-   echo "                       Hadoop jar and the required libraries"
-@@ -116,6 +117,15 @@ case $COMMAND in
-       CLASSPATH=${CLASSPATH}:${TOOL_PATH}
-     elif [ "$COMMAND" = "daemonlog" ] ; then
-       CLASS=org.apache.hadoop.log.LogLevel
-+    elif [ "$COMMAND" = "envvars" ] ; then
-+      echo "JAVA_HOME='${JAVA_HOME}'"
-+      echo "HADOOP_COMMON_HOME='${HADOOP_COMMON_HOME}'"
-+      echo "HADOOP_COMMON_DIR='${HADOOP_COMMON_DIR}'"
-+      echo "HADOOP_COMMON_LIB_JARS_DIR='${HADOOP_COMMON_LIB_JARS_DIR}'"
-+      echo "HADOOP_COMMON_LIB_NATIVE_DIR='${HADOOP_COMMON_LIB_NATIVE_DIR}'"
-+      echo "HADOOP_CONF_DIR='${HADOOP_CONF_DIR}'"
-+      echo "HADOOP_TOOLS_PATH='${TOOL_PATH}'"
-+      exit 0
-     elif [ "$COMMAND" = "archive" ] ; then
-       CLASS=org.apache.hadoop.tools.HadoopArchives
-       CLASSPATH=${CLASSPATH}:${TOOL_PATH}
-diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/hdfs b/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/hdfs
-index fbfbaf2..714280e 100755
---- a/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/hdfs
-+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/bin/hdfs
-@@ -47,6 +47,7 @@ function print_usage(){
-   echo "  datanode             run a DFS datanode"
-   echo "  debug                run a Debug Admin to execute HDFS debug commands"
-   echo "  dfsadmin             run a DFS admin client"
-+  echo "  envvars              display computed Hadoop environment variables"
-   echo "  haadmin              run a DFS HA admin client"
-   echo "  fsck                 run a DFS filesystem checking utility"
-   echo "  balancer             run a cluster balancing utility"
-@@ -161,6 +162,14 @@ elif [ "$COMMAND" = "haadmin" ] ; then
-   CLASS=org.apache.hadoop.hdfs.tools.DFSHAAdmin
-   CLASSPATH=${CLASSPATH}:${TOOL_PATH}
-   HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
-+elif [ "$COMMAND" = "envvars" ] ; then
-+  echo "JAVA_HOME='${JAVA_HOME}'"
-+  echo "HADOOP_HDFS_HOME='${HADOOP_HDFS_HOME}'"
-+  echo "HDFS_DIR='${HDFS_DIR}'"
-+  echo "HDFS_LIB_JARS_DIR='${HDFS_LIB_JARS_DIR}'"
-+  echo "HADOOP_CONF_DIR='${HADOOP_CONF_DIR}'"
-+  echo "HADOOP_TOOLS_PATH='${TOOL_PATH}'"
-+  exit 0
- elif [ "$COMMAND" = "fsck" ] ; then
-   CLASS=org.apache.hadoop.hdfs.tools.DFSck
-   HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
-diff --git a/hadoop-mapreduce-project/bin/mapred b/hadoop-mapreduce-project/bin/mapred
-index 6f3bc04..fd1300d 100755
---- a/hadoop-mapreduce-project/bin/mapred
-+++ b/hadoop-mapreduce-project/bin/mapred
-@@ -38,6 +38,7 @@ function print_usage(){
-   echo "                       mapreduce subcommands"
-   echo "  historyserver        run job history servers as a standalone daemon"
-   echo "  distcp <srcurl> <desturl> copy file or directories recursively"
-+  echo "  envvars              display computed Hadoop environment variables"
-   echo "  archive -archiveName NAME -p <parent path> <src>* <dest> create a hadoop archive"
-   echo "  archive-logs         combine aggregated logs into hadoop archives"
-   echo "  hsadmin              job history server admin interface"
-@@ -75,6 +76,14 @@ elif [ "$COMMAND" = "sampler" ] ; then
-   HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
- elif [ "$COMMAND" = "classpath" ] ; then
-   echo -n 
-+elif [ "$COMMAND" = "envvars" ] ; then
-+  echo "JAVA_HOME='${JAVA_HOME}'"
-+  echo "HADOOP_MAPRED_HOME='${HADOOP_MAPRED_HOME}'"
-+  echo "MAPRED_DIR='${MAPRED_DIR}'"
-+  echo "MAPRED_LIB_JARS_DIR='${MAPRED_LIB_JARS_DIR}'"
-+  echo "HADOOP_CONF_DIR=${HADOOP_CONF_DIR}'"
-+  echo "HADOOP_TOOLS_PATH='${TOOL_PATH}'"
-+  exit 0
- elif [ "$COMMAND" = "historyserver" ] ; then
-   CLASS=org.apache.hadoop.mapreduce.v2.hs.JobHistoryServer
-   HADOOP_OPTS="$HADOOP_OPTS -Dmapred.jobsummary.logger=${HADOOP_JHS_LOGGER:-INFO,console} $HADOOP_JOB_HISTORYSERVER_OPTS"
-diff --git a/hadoop-yarn-project/hadoop-yarn/bin/yarn b/hadoop-yarn-project/hadoop-yarn/bin/yarn
-index 552cef4..5495387 100644
---- a/hadoop-yarn-project/hadoop-yarn/bin/yarn
-+++ b/hadoop-yarn-project/hadoop-yarn/bin/yarn
-@@ -89,6 +89,7 @@ function print_usage(){
-   echo "  cluster                               prints cluster information"
-   echo "  daemonlog                             get/set the log level for each"
-   echo "                                        daemon"
-+  echo "  envvars                               display computed Hadoop environment variables"
-   echo "  top                                   run cluster usage tool"
-   echo ""
-   echo "Most commands print help when invoked w/o parameters."
-@@ -217,6 +218,14 @@ if [ "$COMMAND" = "classpath" ] ; then
-     echo $CLASSPATH
-     exit 0
-   fi
-+elif [ "$COMMAND" = "envvars" ] ; then
-+  echo "JAVA_HOME='${JAVA_HOME}'"
-+  echo "HADOOP_YARN_HOME='${HADOOP_YARN_HOME}'"
-+  echo "YARN_DIR='${YARN_DIR}'"
-+  echo "YARN_LIB_JARS_DIR='${YARN_LIB_JARS_DIR}'"
-+  echo "HADOOP_CONF_DIR='${HADOOP_CONF_DIR}'"
-+  echo "HADOOP_TOOLS_PATH='${TOOL_PATH}'"
-+  exit 0
- elif [ "$COMMAND" = "rmadmin" ] ; then
-   CLASS='org.apache.hadoop.yarn.client.cli.RMAdminCLI'
-   YARN_OPTS="$YARN_OPTS $YARN_CLIENT_OPTS"
--- 
-2.7.4
-
diff --git a/bigtop-packages/src/common/hadoop/patch0-HADOOP-15122-trunk.002.diff b/bigtop-packages/src/common/hadoop/patch0-HADOOP-15122-trunk.002.diff
new file mode 100644
index 0000000..73b153f
--- /dev/null
+++ b/bigtop-packages/src/common/hadoop/patch0-HADOOP-15122-trunk.002.diff
@@ -0,0 +1,13 @@
+diff --git a/pom.xml b/pom.xml
+index f0b3c8e3ccc..d7766783978 100644
+--- a/pom.xml
++++ b/pom.xml
+@@ -187,7 +187,7 @@ xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xs
+             <dependency>
+               <groupId>org.apache.maven.doxia</groupId>
+               <artifactId>doxia-module-markdown</artifactId>
+-              <version>1.8-SNAPSHOT</version>
++              <version>1.8</version>
+             </dependency>
+           </dependencies>
+         </plugin>
diff --git a/bigtop-packages/src/common/hadoop/patch0-maven-enforcer.diff b/bigtop-packages/src/common/hadoop/patch0-maven-enforcer.diff
new file mode 100644
index 0000000..da419e5
--- /dev/null
+++ b/bigtop-packages/src/common/hadoop/patch0-maven-enforcer.diff
@@ -0,0 +1,40 @@
+--- hadoop-3.1.2-src/hadoop-client-modules/hadoop-client-check-invariants/pom.xml	2019-01-23 16:07:50.000000000 +0100
++++ hadoop-3.1.2-src/hadoop-client-modules/hadoop-client-check-invariants/pom.xml_	2019-06-22 23:36:50.531792133 +0200
+@@ -92,7 +92,7 @@
+                     <exclude>com.google.code.findbugs:jsr305</exclude>
+                   </excludes>
+                 </banTransitiveDependencies>
+-                <banDuplicateClasses>
++                <!-- <banDuplicateClasses>
+                   <findAllDuplicates>true</findAllDuplicates>
+                   <dependencies>
+                     <dependency>
+@@ -103,7 +103,7 @@
+                       </ignoreClasses>
+                     </dependency>
+                   </dependencies>
+-                </banDuplicateClasses>
++                </banDuplicateClasses> -->
+               </rules>
+ <!-- TODO we need a rule for "the constants in this set of classes haven't been shaded / don't have this prefix"
+      Manually checking the set of Keys that look like packages we relocate:
+--- hadoop-3.1.2-src/hadoop-client-modules/hadoop-client-check-test-invariants/pom.xml	2019-01-23 16:07:50.000000000 +0100
++++ hadoop-3.1.2-src/hadoop-client-modules/hadoop-client-check-test-invariants/pom.xml_	2019-06-22 23:37:00.251792363 +0200
+@@ -100,7 +100,7 @@
+                     <exclude>com.google.code.findbugs:jsr305</exclude>
+                   </excludes>
+                 </banTransitiveDependencies>
+-                <banDuplicateClasses>
++               <!-- <banDuplicateClasses>
+                   <findAllDuplicates>true</findAllDuplicates>
+                   <dependencies>
+                     <dependency>
+@@ -111,7 +111,7 @@
+                       </ignoreClasses>
+                     </dependency>
+                   </dependencies>
+-                </banDuplicateClasses>
++                </banDuplicateClasses> -->
+               </rules>
+ <!-- TODO we need a rule for "the constants in this set of classes haven't been shaded / don't have this prefix"
+      Manually checking the set of Keys that look like packages we relocate:
diff --git a/bigtop-packages/src/common/hadoop/patch2-HADOOP-14597.diff b/bigtop-packages/src/common/hadoop/patch2-HADOOP-14597.diff
deleted file mode 100644
index cbd428d..0000000
--- a/bigtop-packages/src/common/hadoop/patch2-HADOOP-14597.diff
+++ /dev/null
@@ -1,143 +0,0 @@
-commit 94ca52ae9ec0ae04854d726bf2ac1bc457b96a9c
-Author: Ravi Prakash <raviprak@apache.org>
-Date:   Mon Jul 24 16:01:45 2017 -0700
-
-    HADOOP-14597. Native compilation broken with OpenSSL-1.1.0. Contributed by Ravi Prakash.
-
-diff --git a/hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/crypto/OpensslCipher.c b/hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/crypto/OpensslCipher.c
-index 5cb5bba..c7984a3 100644
---- a/hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/crypto/OpensslCipher.c
-+++ b/hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/crypto/OpensslCipher.c
-@@ -30,6 +30,11 @@ static void (*dlsym_EVP_CIPHER_CTX_free)(EVP_CIPHER_CTX *);
- static int (*dlsym_EVP_CIPHER_CTX_cleanup)(EVP_CIPHER_CTX *);
- static void (*dlsym_EVP_CIPHER_CTX_init)(EVP_CIPHER_CTX *);
- static int (*dlsym_EVP_CIPHER_CTX_set_padding)(EVP_CIPHER_CTX *, int);
-+static int (*dlsym_EVP_CIPHER_CTX_test_flags)(const EVP_CIPHER_CTX *, int);
-+static int (*dlsym_EVP_CIPHER_CTX_block_size)(const EVP_CIPHER_CTX *);
-+#if OPENSSL_VERSION_NUMBER >= 0x10100000L
-+static int (*dlsym_EVP_CIPHER_CTX_encrypting)(const EVP_CIPHER_CTX *);
-+#endif
- static int (*dlsym_EVP_CipherInit_ex)(EVP_CIPHER_CTX *, const EVP_CIPHER *,  \
-            ENGINE *, const unsigned char *, const unsigned char *, int);
- static int (*dlsym_EVP_CipherUpdate)(EVP_CIPHER_CTX *, unsigned char *,  \
-@@ -46,6 +51,11 @@ typedef void (__cdecl *__dlsym_EVP_CIPHER_CTX_free)(EVP_CIPHER_CTX *);
- typedef int (__cdecl *__dlsym_EVP_CIPHER_CTX_cleanup)(EVP_CIPHER_CTX *);
- typedef void (__cdecl *__dlsym_EVP_CIPHER_CTX_init)(EVP_CIPHER_CTX *);
- typedef int (__cdecl *__dlsym_EVP_CIPHER_CTX_set_padding)(EVP_CIPHER_CTX *, int);
-+typedef int (__cdecl *__dlsym_EVP_CIPHER_CTX_test_flags)(const EVP_CIPHER_CTX *, int);
-+typedef int (__cdecl *__dlsym_EVP_CIPHER_CTX_block_size)(const EVP_CIPHER_CTX *);
-+#if OPENSSL_VERSION_NUMBER >= 0x10100000L
-+typedef int (__cdecl *__dlsym_EVP_CIPHER_CTX_encrypting)(const EVP_CIPHER_CTX *);
-+#endif
- typedef int (__cdecl *__dlsym_EVP_CipherInit_ex)(EVP_CIPHER_CTX *,  \
-              const EVP_CIPHER *, ENGINE *, const unsigned char *,  \
-              const unsigned char *, int);
-@@ -60,6 +70,11 @@ static __dlsym_EVP_CIPHER_CTX_free dlsym_EVP_CIPHER_CTX_free;
- static __dlsym_EVP_CIPHER_CTX_cleanup dlsym_EVP_CIPHER_CTX_cleanup;
- static __dlsym_EVP_CIPHER_CTX_init dlsym_EVP_CIPHER_CTX_init;
- static __dlsym_EVP_CIPHER_CTX_set_padding dlsym_EVP_CIPHER_CTX_set_padding;
-+static __dlsym_EVP_CIPHER_CTX_test_flags dlsym_EVP_CIPHER_CTX_test_flags;
-+static __dlsym_EVP_CIPHER_CTX_block_size dlsym_EVP_CIPHER_CTX_block_size;
-+#if OPENSSL_VERSION_NUMBER >= 0x10100000L
-+static __dlsym_EVP_CIPHER_CTX_encrypting dlsym_EVP_CIPHER_CTX_encrypting;
-+#endif
- static __dlsym_EVP_CipherInit_ex dlsym_EVP_CipherInit_ex;
- static __dlsym_EVP_CipherUpdate dlsym_EVP_CipherUpdate;
- static __dlsym_EVP_CipherFinal_ex dlsym_EVP_CipherFinal_ex;
-@@ -114,6 +129,14 @@ JNIEXPORT void JNICALL Java_org_apache_hadoop_crypto_OpensslCipher_initIDs
-                       "EVP_CIPHER_CTX_init");
-   LOAD_DYNAMIC_SYMBOL(dlsym_EVP_CIPHER_CTX_set_padding, env, openssl,  \
-                       "EVP_CIPHER_CTX_set_padding");
-+  LOAD_DYNAMIC_SYMBOL(dlsym_EVP_CIPHER_CTX_test_flags, env, openssl,  \
-+                      "EVP_CIPHER_CTX_test_flags");
-+  LOAD_DYNAMIC_SYMBOL(dlsym_EVP_CIPHER_CTX_block_size, env, openssl,  \
-+                      "EVP_CIPHER_CTX_block_size");
-+#if OPENSSL_VERSION_NUMBER >= 0x10100000L
-+  LOAD_DYNAMIC_SYMBOL(dlsym_EVP_CIPHER_CTX_encrypting, env, openssl,  \
-+                      "EVP_CIPHER_CTX_encrypting");
-+#endif
-   LOAD_DYNAMIC_SYMBOL(dlsym_EVP_CipherInit_ex, env, openssl,  \
-                       "EVP_CipherInit_ex");
-   LOAD_DYNAMIC_SYMBOL(dlsym_EVP_CipherUpdate, env, openssl,  \
-@@ -135,6 +158,17 @@ JNIEXPORT void JNICALL Java_org_apache_hadoop_crypto_OpensslCipher_initIDs
-   LOAD_DYNAMIC_SYMBOL(__dlsym_EVP_CIPHER_CTX_set_padding,  \
-                       dlsym_EVP_CIPHER_CTX_set_padding, env,  \
-                       openssl, "EVP_CIPHER_CTX_set_padding");
-+  LOAD_DYNAMIC_SYMBOL(__dlsym_EVP_CIPHER_CTX_test_flags,  \
-+                      dlsym_EVP_CIPHER_CTX_test_flags, env,  \
-+                      openssl, "EVP_CIPHER_CTX_test_flags");
-+  LOAD_DYNAMIC_SYMBOL(__dlsym_EVP_CIPHER_CTX_block_size,  \
-+                      dlsym_EVP_CIPHER_CTX_block_size, env,  \
-+                      openssl, "EVP_CIPHER_CTX_block_size");
-+#if OPENSSL_VERSION_NUMBER >= 0x10100000L
-+  LOAD_DYNAMIC_SYMBOL(__dlsym_EVP_CIPHER_CTX_encrypting,  \
-+                      dlsym_EVP_CIPHER_CTX_encrypting, env,  \
-+                      openssl, "EVP_CIPHER_CTX_encrypting");
-+#endif
-   LOAD_DYNAMIC_SYMBOL(__dlsym_EVP_CipherInit_ex, dlsym_EVP_CipherInit_ex,  \
-                       env, openssl, "EVP_CipherInit_ex");
-   LOAD_DYNAMIC_SYMBOL(__dlsym_EVP_CipherUpdate, dlsym_EVP_CipherUpdate,  \
-@@ -253,14 +287,18 @@ JNIEXPORT jlong JNICALL Java_org_apache_hadoop_crypto_OpensslCipher_init
- static int check_update_max_output_len(EVP_CIPHER_CTX *context, int input_len, 
-     int max_output_len)
- {
--  if (context->flags & EVP_CIPH_NO_PADDING) {
-+  if (  dlsym_EVP_CIPHER_CTX_test_flags(context, EVP_CIPH_NO_PADDING) ) {
-     if (max_output_len >= input_len) {
-       return 1;
-     }
-     return 0;
-   } else {
--    int b = context->cipher->block_size;
-+    int b = dlsym_EVP_CIPHER_CTX_block_size(context);
-+#if OPENSSL_VERSION_NUMBER < 0x10100000L
-     if (context->encrypt) {
-+#else
-+    if (dlsym_EVP_CIPHER_CTX_encrypting(context)) {
-+#endif
-       if (max_output_len >= input_len + b - 1) {
-         return 1;
-       }
-@@ -307,10 +345,10 @@ JNIEXPORT jint JNICALL Java_org_apache_hadoop_crypto_OpensslCipher_update
- static int check_doFinal_max_output_len(EVP_CIPHER_CTX *context, 
-     int max_output_len)
- {
--  if (context->flags & EVP_CIPH_NO_PADDING) {
-+  if (  dlsym_EVP_CIPHER_CTX_test_flags(context, EVP_CIPH_NO_PADDING) ) {
-     return 1;
-   } else {
--    int b = context->cipher->block_size;
-+    int b = dlsym_EVP_CIPHER_CTX_block_size(context);
-     if (max_output_len >= b) {
-       return 1;
-     }
-diff --git a/hadoop-tools/hadoop-pipes/src/main/native/pipes/impl/HadoopPipes.cc b/hadoop-tools/hadoop-pipes/src/main/native/pipes/impl/HadoopPipes.cc
-index 91fb5a4..45cb8c2 100644
---- a/hadoop-tools/hadoop-pipes/src/main/native/pipes/impl/HadoopPipes.cc
-+++ b/hadoop-tools/hadoop-pipes/src/main/native/pipes/impl/HadoopPipes.cc
-@@ -420,6 +420,7 @@ namespace HadoopPipes {
-     }
- 
-     string createDigest(string &password, string& msg) {
-+#if OPENSSL_VERSION_NUMBER < 0x10100000L
-       HMAC_CTX ctx;
-       unsigned char digest[EVP_MAX_MD_SIZE];
-       HMAC_Init(&ctx, (const unsigned char *)password.c_str(), 
-@@ -428,7 +429,16 @@ namespace HadoopPipes {
-       unsigned int digestLen;
-       HMAC_Final(&ctx, digest, &digestLen);
-       HMAC_cleanup(&ctx);
--
-+#else
-+      HMAC_CTX *ctx = HMAC_CTX_new();
-+      unsigned char digest[EVP_MAX_MD_SIZE];
-+      HMAC_Init_ex(ctx, (const unsigned char *)password.c_str(),
-+          password.length(), EVP_sha1(), NULL);
-+      HMAC_Update(ctx, (const unsigned char *)msg.c_str(), msg.length());
-+      unsigned int digestLen;
-+      HMAC_Final(ctx, digest, &digestLen);
-+      HMAC_CTX_free(ctx);
-+#endif
-       //now apply base64 encoding
-       BIO *bmem, *b64;
-       BUF_MEM *bptr;
diff --git a/bigtop-packages/src/common/zookeeper/do-component-build b/bigtop-packages/src/common/zookeeper/do-component-build
index 57fd09e..f35816e 100644
--- a/bigtop-packages/src/common/zookeeper/do-component-build
+++ b/bigtop-packages/src/common/zookeeper/do-component-build
@@ -22,10 +22,10 @@ set -ex
 sed -i -e 's@AM_PATH_CPPUNIT(1.10.2)@PKG_CHECK_MODULES([CPPUNIT], [cppunit])@' src/c/configure.ac
 
 ANT_OPTS="-Dversion=$ZOOKEEPER_VERSION -f build.xml $@"
-sed -i.orig -e 's#test-jar,api-report#test-jar#g' build.xml
+#sed -i.orig -e 's#test-jar,api-report#test-jar#g' build.xml
 ant compile ${ANT_OPTS}
 (cd src/contrib/rest && ant jar ${ANT_OPTS})
 ant package package-native tar ${ANT_OPTS}
 
-mvn install:install-file -DcreateChecksum=true -DgroupId=org.apache.zookeeper -DartifactId=zookeeper -Dversion=$ZOOKEEPER_VERSION -Dpackaging=jar -Dfile=build/zookeeper-$ZOOKEEPER_VERSION.jar
-mvn install:install-file -DcreateChecksum=true -DgroupId=org.apache.zookeeper -DartifactId=zookeeper -Dversion=$ZOOKEEPER_VERSION -Dclassifier=tests -Dpackaging=jar -Dfile=build/zookeeper-$ZOOKEEPER_VERSION-test.jar
+#mvn install:install-file -DcreateChecksum=true -DgroupId=org.apache.zookeeper -DartifactId=zookeeper -Dversion=$ZOOKEEPER_VERSION -Dpackaging=jar -Dfile=build/zookeeper-$ZOOKEEPER_VERSION.jar
+#mvn install:install-file -DcreateChecksum=true -DgroupId=org.apache.zookeeper -DartifactId=zookeeper -Dversion=$ZOOKEEPER_VERSION -Dclassifier=tests -Dpackaging=jar -Dfile=build/zookeeper-$ZOOKEEPER_VERSION-test.jar
diff --git a/bigtop-packages/src/common/zookeeper/patch0-ZOOKEEPER-1897v3.diff b/bigtop-packages/src/common/zookeeper/patch0-ZOOKEEPER-1897v3.diff
deleted file mode 100644
index 802647c..0000000
--- a/bigtop-packages/src/common/zookeeper/patch0-ZOOKEEPER-1897v3.diff
+++ /dev/null
@@ -1,14 +0,0 @@
-diff --git a/src/java/main/org/apache/zookeeper/ZooKeeperMain.java b/src/java/main/org/apache/zookeeper/ZooKeeperMain.java
-index b1b4c27..39fc465 100644
---- a/src/java/main/org/apache/zookeeper/ZooKeeperMain.java
-+++ b/src/java/main/org/apache/zookeeper/ZooKeeperMain.java
-@@ -349,6 +349,9 @@ public class ZooKeeperMain {
-                     executeLine(line);
-                 }
-             }
-+        } else {
-+            // Command line args non-null.  Run what was passed.
-+            processCmd(cl);
-         }
-     }
- 
diff --git a/bigtop-packages/src/common/zookeeper/patch1-ZOOKEEPER-1911.diff b/bigtop-packages/src/common/zookeeper/patch1-ZOOKEEPER-1911.diff
deleted file mode 100644
index 89579e7..0000000
--- a/bigtop-packages/src/common/zookeeper/patch1-ZOOKEEPER-1911.diff
+++ /dev/null
@@ -1,25 +0,0 @@
---- zookeeper-3.4.6/src/contrib/rest/build.xml.orig	2015-04-14 22:04:16.060000000 +0200
-+++ zookeeper-3.4.6/src/contrib/rest/build.xml	2015-04-14 22:04:20.268000000 +0200
-@@ -48,6 +48,22 @@
-       </sequential>
-     </macrodef>
- 
-+  <target name="package" depends="jar" unless="skip.contrib">
-+    <echo message="contrib: ${name}"/>
-+    <mkdir dir="${dist.dir}${package.share}/contrib/${name}"/>
-+    <copy todir="${dist.dir}${package.share}/contrib/${name}">
-+      <fileset dir="${build.dir}">
-+        <include name="zookeeper-${version}-${name}.jar" />
-+      </fileset>
-+    </copy>
-+    <copy todir="${dist.dir}${package.share}/contrib/${name}/lib">
-+      <fileset dir="${build.dir}/lib" />
-+    </copy>
-+    <copy todir="${dist.dir}${package.share}/contrib/${name}/conf">
-+      <fileset dir="conf" />
-+    </copy>
-+  </target>
-+
-   <target name="setjarname">
-     <property name="jarname"
-               value="${build.dir}/zookeeper-${version}-${name}.jar"/>
diff --git a/bigtop-packages/src/common/zookeeper/patch2-ZOOKEEPER-1926.diff b/bigtop-packages/src/common/zookeeper/patch2-ZOOKEEPER-1926.diff
deleted file mode 100644
index a0e5e6f..0000000
--- a/bigtop-packages/src/common/zookeeper/patch2-ZOOKEEPER-1926.diff
+++ /dev/null
@@ -1,177 +0,0 @@
-diff -ur zookeeper-3.4.6/src_/java/systest/org/apache/zookeeper/test/system/BaseSysTest.java zookeeper-3.4.6/src/java/systest/org/apache/zookeeper/test/system/BaseSysTest.java
---- zookeeper-3.4.6/src_/java/systest/org/apache/zookeeper/test/system/BaseSysTest.java	2015-04-14 22:07:14.464000000 +0200
-+++ zookeeper-3.4.6/src/java/systest/org/apache/zookeeper/test/system/BaseSysTest.java	2015-04-14 22:07:18.728000000 +0200
-@@ -37,6 +37,8 @@
- 
- @Ignore("No tests in this class.")
- public class BaseSysTest extends TestCase {
-+    private static final File testData = new File(
-+            System.getProperty("test.data.dir", "build/test/data"));
-     private static int fakeBasePort = 33222;
-     private static String zkHostPort;
-     protected String prefix = "/sysTest";
-@@ -149,7 +151,7 @@
-         }
-         StringBuilder sb = new StringBuilder();
-         for(int i = 0; i < count; i++) {
--            qpsDirs[i] = File.createTempFile("sysTest", ".tmp");
-+            qpsDirs[i] = File.createTempFile("sysTest", ".tmp", testData);
-             qpsDirs[i].delete();
-             qpsDirs[i].mkdir();
-             int port = fakeBasePort+10+i;
-diff -ur zookeeper-3.4.6/src_/java/systest/org/apache/zookeeper/test/system/QuorumPeerInstance.java zookeeper-3.4.6/src/java/systest/org/apache/zookeeper/test/system/QuorumPeerInstance.java
---- zookeeper-3.4.6/src_/java/systest/org/apache/zookeeper/test/system/QuorumPeerInstance.java	2015-04-14 22:07:14.464000000 +0200
-+++ zookeeper-3.4.6/src/java/systest/org/apache/zookeeper/test/system/QuorumPeerInstance.java	2015-04-14 22:07:18.728000000 +0200
-@@ -36,6 +36,9 @@
- 
- class QuorumPeerInstance implements Instance {
-     final private static Logger LOG = LoggerFactory.getLogger(QuorumPeerInstance.class);
-+    private static final File testData = new File(
-+        System.getProperty("test.data.dir", "build/test/data"));
-+
-     private static final int syncLimit = 3;
-     private static final int initLimit = 3;
-     private static final int tickTime = 2000;
-@@ -55,7 +58,7 @@
- 
-     public QuorumPeerInstance() {
-         try {
--            File tmpFile = File.createTempFile("test", ".dir");
-+            File tmpFile = File.createTempFile("test", ".dir", testData);
-             File tmpDir = tmpFile.getParentFile();
-             tmpFile.delete();
-             File zkDirs = new File(tmpDir, "zktmp.cfg");
-diff -ur zookeeper-3.4.6/src_/java/test/org/apache/zookeeper/server/quorum/LearnerTest.java zookeeper-3.4.6/src/java/test/org/apache/zookeeper/server/quorum/LearnerTest.java
---- zookeeper-3.4.6/src_/java/test/org/apache/zookeeper/server/quorum/LearnerTest.java	2015-04-14 22:07:14.464000000 +0200
-+++ zookeeper-3.4.6/src/java/test/org/apache/zookeeper/server/quorum/LearnerTest.java	2015-04-14 22:07:18.728000000 +0200
-@@ -47,6 +47,9 @@
- import org.junit.Test;
- 
- public class LearnerTest extends ZKTestCase {
-+  private static final File testData = new File(
-+    System.getProperty("test.data.dir", "build/test/data"));
-+
- 	class SimpleLearnerZooKeeperServer extends LearnerZooKeeperServer {
- 		boolean startupCalled;
- 		
-@@ -84,7 +87,7 @@
- 	}
- 	@Test
- 	public void syncTest() throws Exception {
--		File tmpFile = File.createTempFile("test", ".dir");
-+		File tmpFile = File.createTempFile("test", ".dir", testData);
- 		tmpFile.delete();
- 		try {
- 			FileTxnSnapLog ftsl = new FileTxnSnapLog(tmpFile, tmpFile);
-diff -ur zookeeper-3.4.6/src_/java/test/org/apache/zookeeper/server/quorum/Zab1_0Test.java zookeeper-3.4.6/src/java/test/org/apache/zookeeper/server/quorum/Zab1_0Test.java
---- zookeeper-3.4.6/src_/java/test/org/apache/zookeeper/server/quorum/Zab1_0Test.java	2015-04-14 22:07:14.464000000 +0200
-+++ zookeeper-3.4.6/src/java/test/org/apache/zookeeper/server/quorum/Zab1_0Test.java	2015-04-14 22:07:18.728000000 +0200
-@@ -80,6 +80,9 @@
- 
-     private static final Logger LOG = LoggerFactory.getLogger(Zab1_0Test.class);
- 
-+    private static final File testData = new File(
-+            System.getProperty("test.data.dir", "build/test/data"));
-+
-     private static final class LeadThread extends Thread {
-         private final Leader leader;
- 
-@@ -149,7 +152,7 @@
-     }
-     @Test
-     public void testLeaderInConnectingFollowers() throws Exception {    
--        File tmpDir = File.createTempFile("test", "dir");
-+        File tmpDir = File.createTempFile("test", "dir", testData);
-         tmpDir.delete();
-         tmpDir.mkdir();
-         Leader leader = null;
-@@ -203,7 +206,7 @@
-         
-     @Test
-     public void testLastAcceptedEpoch() throws Exception {    
--        File tmpDir = File.createTempFile("test", "dir");
-+        File tmpDir = File.createTempFile("test", "dir", testData);
-         tmpDir.delete();
-         tmpDir.mkdir();
-         Leader leader = null;
-@@ -241,7 +244,7 @@
-         
-     @Test
-     public void testLeaderInElectingFollowers() throws Exception {    
--        File tmpDir = File.createTempFile("test", "dir");
-+        File tmpDir = File.createTempFile("test", "dir", testData);
-         tmpDir.delete();
-         tmpDir.mkdir();
-         Leader leader = null;
-@@ -348,7 +351,7 @@
-         Socket pair[] = getSocketPair();
-         Socket leaderSocket = pair[0];
-         Socket followerSocket = pair[1];
--        File tmpDir = File.createTempFile("test", "dir");
-+        File tmpDir = File.createTempFile("test", "dir", testData);
-         tmpDir.delete();
-         tmpDir.mkdir();
-         LeadThread leadThread = null;
-@@ -390,7 +393,7 @@
-         Socket pair[] = getSocketPair();
-         Socket leaderSocket = pair[0];
-         Socket followerSocket = pair[1];
--        File tmpDir = File.createTempFile("test", "dir");
-+        File tmpDir = File.createTempFile("test", "dir", testData);
-         tmpDir.delete();
-         tmpDir.mkdir();
-         LeadThread leadThread = null;
-@@ -457,7 +460,7 @@
-     
-     
-     public void testFollowerConversation(FollowerConversation conversation) throws Exception {
--        File tmpDir = File.createTempFile("test", "dir");
-+        File tmpDir = File.createTempFile("test", "dir", testData);
-         tmpDir.delete();
-         tmpDir.mkdir();
-         Thread followerThread = null;
-@@ -509,7 +512,7 @@
-     }
- 
-     public void testObserverConversation(ObserverConversation conversation) throws Exception {
--        File tmpDir = File.createTempFile("test", "dir");
-+        File tmpDir = File.createTempFile("test", "dir", testData);
-         tmpDir.delete();
-         tmpDir.mkdir();
-         Thread observerThread = null;
-@@ -629,7 +632,7 @@
-             @Override
-             public void converseWithFollower(InputArchive ia, OutputArchive oa,
-                     Follower f) throws Exception {
--                File tmpDir = File.createTempFile("test", "dir");
-+                File tmpDir = File.createTempFile("test", "dir", testData);
-                 tmpDir.delete();
-                 tmpDir.mkdir();
-                 File logDir = f.fzk.getTxnLogFactory().getDataDir().getParentFile();
-@@ -757,7 +760,7 @@
-             @Override
-             public void converseWithFollower(InputArchive ia, OutputArchive oa,
-                     Follower f) throws Exception {
--                File tmpDir = File.createTempFile("test", "dir");
-+                File tmpDir = File.createTempFile("test", "dir", testData);
-                 tmpDir.delete();
-                 tmpDir.mkdir();
-                 File logDir = f.fzk.getTxnLogFactory().getDataDir().getParentFile();
-@@ -995,7 +998,7 @@
-             @Override
-             public void converseWithObserver(InputArchive ia, OutputArchive oa,
-                     Observer o) throws Exception {
--                File tmpDir = File.createTempFile("test", "dir");
-+                File tmpDir = File.createTempFile("test", "dir", testData);
-                 tmpDir.delete();
-                 tmpDir.mkdir();
-                 File logDir = o.zk.getTxnLogFactory().getDataDir().getParentFile();
-@@ -1413,7 +1416,7 @@
- 
-     @Test
-     public void testInitialAcceptedCurrent() throws Exception {
--        File tmpDir = File.createTempFile("test", ".dir");
-+        File tmpDir = File.createTempFile("test", ".dir", testData);
-         tmpDir.delete();
-         tmpDir.mkdir();
-         try {
diff --git a/bigtop-packages/src/common/zookeeper/patch3-ZOOKEEPER-2064.diff b/bigtop-packages/src/common/zookeeper/patch3-ZOOKEEPER-2064.diff
deleted file mode 100644
index 56003c8..0000000
--- a/bigtop-packages/src/common/zookeeper/patch3-ZOOKEEPER-2064.diff
+++ /dev/null
@@ -1,121 +0,0 @@
-diff -ur zookeeper-3.4.6/src_/contrib/loggraph/src/java/org/apache/zookeeper/graph/Log4JSource.java zookeeper-3.4.6/src/contrib/loggraph/src/java/org/apache/zookeeper/graph/Log4JSource.java
---- zookeeper-3.4.6/src_/contrib/loggraph/src/java/org/apache/zookeeper/graph/Log4JSource.java	2015-04-14 22:26:36.528000000 +0200
-+++ zookeeper-3.4.6/src/contrib/loggraph/src/java/org/apache/zookeeper/graph/Log4JSource.java	2015-04-14 22:26:41.896000000 +0200
-@@ -351,11 +351,21 @@
- 		
- 		LogIterator iter = s.iterator(starttime, endtime);
- 		System.out.println(iter);
-+		try {
-+		  iter.close();
-+		} catch (IOException ioe) {
-+		  System.out.println(ioe.getMessage());
-+		}
- 	    }; };
- 	    Thread t2 = new Thread() { public void run () { 
- 		
- 		LogIterator iter = s.iterator(starttime, endtime);
- 		System.out.println(iter);
-+		try {
-+		  iter.close();
-+		} catch (IOException ioe) {
-+		  System.out.println(ioe.getMessage());
-+		}
- 	    }; };
- 	    Thread t3 = new Thread() { public void run () { 
- 		
-diff -ur zookeeper-3.4.6/src_/contrib/loggraph/src/java/org/apache/zookeeper/graph/RandomAccessFileReader.java zookeeper-3.4.6/src/contrib/loggraph/src/java/org/apache/zookeeper/graph/RandomAccessFileReader.java
---- zookeeper-3.4.6/src_/contrib/loggraph/src/java/org/apache/zookeeper/graph/RandomAccessFileReader.java	2015-04-14 22:26:36.528000000 +0200
-+++ zookeeper-3.4.6/src/contrib/loggraph/src/java/org/apache/zookeeper/graph/RandomAccessFileReader.java	2015-04-14 22:26:41.896000000 +0200
-@@ -324,5 +324,6 @@
- 	System.out.println(f.readLine());
- 	f.seek(pos2);
- 	System.out.println(f.readLine());
-+	f.close();
-     }
- };
-diff -ur zookeeper-3.4.6/src_/contrib/loggraph/src/java/org/apache/zookeeper/graph/servlets/NumEvents.java zookeeper-3.4.6/src/contrib/loggraph/src/java/org/apache/zookeeper/graph/servlets/NumEvents.java
---- zookeeper-3.4.6/src_/contrib/loggraph/src/java/org/apache/zookeeper/graph/servlets/NumEvents.java	2015-04-14 22:26:36.528000000 +0200
-+++ zookeeper-3.4.6/src/contrib/loggraph/src/java/org/apache/zookeeper/graph/servlets/NumEvents.java	2015-04-14 22:26:41.896000000 +0200
-@@ -81,6 +81,7 @@
- 	if (LOG.isDebugEnabled()) {
- 	    LOG.debug("handle(start= " + starttime + ", end=" + endtime + ", numEntries=" + size +")");
- 	}
-+	iter.close();
- 	return JSONValue.toJSONString(data);
-     }
- }
-diff -ur zookeeper-3.4.6/src_/contrib/loggraph/src/java/org/apache/zookeeper/graph/servlets/StaticContent.java zookeeper-3.4.6/src/contrib/loggraph/src/java/org/apache/zookeeper/graph/servlets/StaticContent.java
---- zookeeper-3.4.6/src_/contrib/loggraph/src/java/org/apache/zookeeper/graph/servlets/StaticContent.java	2015-04-14 22:26:36.528000000 +0200
-+++ zookeeper-3.4.6/src/contrib/loggraph/src/java/org/apache/zookeeper/graph/servlets/StaticContent.java	2015-04-14 22:26:41.896000000 +0200
-@@ -39,9 +39,12 @@
- 	    response.setStatus(HttpServletResponse.SC_NOT_FOUND);
- 	    return;
- 	}
--	
--	while (resource.available() > 0) {
-+	try {
-+	  while (resource.available() > 0) {
- 	    response.getWriter().write(resource.read());
-+	  }
-+	} finally {
-+	  resource.close();
- 	}
- 	//        response.setContentType("text/plain;charset=utf-8");
-         response.setStatus(HttpServletResponse.SC_OK);
-diff -ur zookeeper-3.4.6/src_/contrib/rest/src/java/org/apache/zookeeper/server/jersey/cfg/RestCfg.java zookeeper-3.4.6/src/contrib/rest/src/java/org/apache/zookeeper/server/jersey/cfg/RestCfg.java
---- zookeeper-3.4.6/src_/contrib/rest/src/java/org/apache/zookeeper/server/jersey/cfg/RestCfg.java	2015-04-14 22:26:36.532000000 +0200
-+++ zookeeper-3.4.6/src/contrib/rest/src/java/org/apache/zookeeper/server/jersey/cfg/RestCfg.java	2015-04-14 22:26:41.900000000 +0200
-@@ -36,9 +36,13 @@
-    }
- 
-    public RestCfg(InputStream io) throws IOException {
-+     try {
-        cfg.load(io);
-        extractEndpoints();
-        extractCredentials();
-+     } finally {
-+       io.close();
-+     }
-    }
- 
-    private void extractCredentials() {
-diff -ur zookeeper-3.4.6/src_/java/systest/org/apache/zookeeper/test/system/GenerateLoad.java zookeeper-3.4.6/src/java/systest/org/apache/zookeeper/test/system/GenerateLoad.java
---- zookeeper-3.4.6/src_/java/systest/org/apache/zookeeper/test/system/GenerateLoad.java	2015-04-14 22:26:36.524000000 +0200
-+++ zookeeper-3.4.6/src/java/systest/org/apache/zookeeper/test/system/GenerateLoad.java	2015-04-14 22:26:41.900000000 +0200
-@@ -699,12 +699,16 @@
-         s.getOutputStream().write("stat".getBytes());
-         BufferedReader br = new BufferedReader(new InputStreamReader(s.getInputStream()));
-         String line;
--        while((line = br.readLine()) != null) {
-+        try {
-+          while((line = br.readLine()) != null) {
-             if (line.startsWith("Mode: ")) {
--                return line.substring(6);
-+              return line.substring(6);
-             }
-+          }
-+          return "unknown";
-+        } finally {
-+          s.close();
-         }
--        return "unknown";
-     }
- 
-     private static void doUsage() {
-diff -ur zookeeper-3.4.6/src_/java/systest/org/apache/zookeeper/test/system/QuorumPeerInstance.java zookeeper-3.4.6/src/java/systest/org/apache/zookeeper/test/system/QuorumPeerInstance.java
---- zookeeper-3.4.6/src_/java/systest/org/apache/zookeeper/test/system/QuorumPeerInstance.java	2015-04-14 22:26:36.524000000 +0200
-+++ zookeeper-3.4.6/src/java/systest/org/apache/zookeeper/test/system/QuorumPeerInstance.java	2015-04-14 22:26:41.900000000 +0200
-@@ -67,7 +67,12 @@
-             Properties p;
-             if (zkDirs.exists()) {
-                 p = new Properties();
--                p.load(new FileInputStream(zkDirs));
-+                FileInputStream input = new FileInputStream(zkDirs);
-+                try {
-+                  p.load(input);
-+                } finally {
-+                  input.close();
-+                }
-             } else {
-                 p = System.getProperties();
-             }
diff --git a/bigtop-packages/src/common/zookeeper/patch4-removeasm.diff b/bigtop-packages/src/common/zookeeper/patch4-removeasm.diff
deleted file mode 100644
index 6599617..0000000
--- a/bigtop-packages/src/common/zookeeper/patch4-removeasm.diff
+++ /dev/null
@@ -1,17 +0,0 @@
---- zookeeper-3.4.6/src/c/src/mt_adaptor.c.ORG	2015-10-05 20:50:36.662228733 +0000
-+++ zookeeper-3.4.6/src/c/src/mt_adaptor.c	2015-10-05 20:51:35.722773398 +0000
-@@ -483,13 +483,7 @@
- int32_t fetch_and_add(volatile int32_t* operand, int incr)
- {
- #ifndef WIN32
--    int32_t result;
--    asm __volatile__(
--         "lock xaddl %0,%1\n"
--         : "=r"(result), "=m"(*(int *)operand)
--         : "0"(incr)
--         : "memory");
--   return result;
-+    return __sync_fetch_and_add(operand, incr);
- #else
-     volatile int32_t result;
-     _asm
diff --git a/bigtop-packages/src/common/zookeeper/patch5-ZOOKEEPER-2594.diff b/bigtop-packages/src/common/zookeeper/patch5-ZOOKEEPER-2594.diff
deleted file mode 100644
index 1d5af01..0000000
--- a/bigtop-packages/src/common/zookeeper/patch5-ZOOKEEPER-2594.diff
+++ /dev/null
@@ -1,66 +0,0 @@
-diff -ur zookeeper-3.4.6/build.xml zookeeper-3.4.6-new/build.xml
---- zookeeper-3.4.6/build.xml	2014-02-20 10:14:08.000000000 +0000
-+++ zookeeper-3.4.6-new/build.xml	2016-10-02 08:15:52.140641227 +0000
-@@ -107,7 +107,7 @@
- 
-     <property name="ivy.version" value="2.2.0"/>
-     <property name="ivy.url"
--              value="http://repo2.maven.org/maven2/org/apache/ivy/ivy" />
-+              value="https://repo1.maven.org/maven2/org/apache/ivy/ivy" />
-     <property name="ivy.home" value="${user.home}/.ant" />
-     <property name="ivy.lib" value="${build.dir}/lib"/>
-     <property name="ivy.package.lib" value="${build.dir}/package/lib"/>
-@@ -115,7 +115,7 @@
-     <property name="ivy.jdiff.lib" value="${build.dir}/jdiff/lib"/>
-     <property name="ivysettings.xml" value="${basedir}/ivysettings.xml"/>
-     
--    <property name="mvnrepo" value="http://repo2.maven.org/maven2"/>
-+    <property name="mvnrepo" value="https://repo1.maven.org/maven2"/>
-     <property name="tsk.org" value="/org/apache/maven/maven-ant-tasks/"/>
-     <property name="ant-task.version" value="2.1.3"/>
-     <property name="ant_task_repo_url"
-Only in zookeeper-3.4.6-new/: build.xml.orig
-diff -ur zookeeper-3.4.6/ivysettings.xml zookeeper-3.4.6-new/ivysettings.xml
---- zookeeper-3.4.6/ivysettings.xml	2014-02-20 10:14:08.000000000 +0000
-+++ zookeeper-3.4.6-new/ivysettings.xml	2016-10-02 08:15:52.140641227 +0000
-@@ -18,11 +18,9 @@
- -->
- 
-   <property name="repo.maven.org"
--    value="http://repo1.maven.org/maven2/" override="false"/>
-+    value="https://repo1.maven.org/maven2/" override="false"/>
-   <property name="repo.jboss.org"
--    value="http://repository.jboss.org/nexus/content/groups/public/" override="false"/>
--  <property name="repo.sun.org"
--    value="http://download.java.net/maven/2/" override="false"/>
-+    value="https://repository.jboss.org/nexus/content/groups/public/" override="false"/>
-   <property name="maven2.pattern"
-     value="[organisation]/[module]/[revision]/[module]-[revision]"/>
-   <property name="maven2.pattern.ext" value="${maven2.pattern}.[ext]"/>
-@@ -33,13 +31,10 @@
-       pattern="${maven2.pattern.ext}" m2compatible="true"/>
-     <ibiblio name="jboss-maven2" root="${repo.jboss.org}"
-       pattern="${maven2.pattern.ext}" m2compatible="true"/>
--    <ibiblio name="sun-maven2" root="${repo.sun.org}"
--      pattern="${maven2.pattern.ext}" m2compatible="true"/>
- 
-     <chain name="default" dual="true">
-       <resolver ref="maven2"/>
-       <resolver ref="jboss-maven2"/>
--      <resolver ref="sun-maven2"/>
-     </chain>
- 
-   </resolvers>
-diff -ur zookeeper-3.4.6/src/contrib/build-contrib.xml zookeeper-3.4.6-new/src/contrib/build-contrib.xml
---- zookeeper-3.4.6/src/contrib/build-contrib.xml	2014-02-20 10:14:09.000000000 +0000
-+++ zookeeper-3.4.6-new/src/contrib/build-contrib.xml	2016-10-02 08:15:52.140641227 +0000
-@@ -43,7 +43,7 @@
- 
-   <property name="ivy.version" value="2.2.0"/>
-   <property name="ivy.url"
--            value="http://repo2.maven.org/maven2/org/apache/ivy/ivy" />
-+            value="https://repo1.maven.org/maven2/org/apache/ivy/ivy" />
-   <property name="ivy.home" value="${user.home}/.ant" />
-   <property name="ivy.lib" value="${build.dir}/lib"/>
-   <property name="ivy.test.lib" value="${build.test}/lib"/>
-Only in zookeeper-3.4.6-new/src/contrib: build-contrib.xml.orig
diff --git a/bigtop-packages/src/deb/hadoop/control b/bigtop-packages/src/deb/hadoop/control
index 9ca5870..fc3f3c9 100644
--- a/bigtop-packages/src/deb/hadoop/control
+++ b/bigtop-packages/src/deb/hadoop/control
@@ -23,7 +23,7 @@ Homepage: http://hadoop.apache.org/core/
 
 Package: hadoop
 Architecture: any
-Depends: ${shlibs:Depends}, ${misc:Depends}, adduser, bigtop-utils (>= 0.7), zookeeper (>= 3.4.0), psmisc, netcat-openbsd
+Depends: ${shlibs:Depends}, ${misc:Depends}, adduser, bigtop-utils (>= 0.7), psmisc, netcat-openbsd
 Description: Hadoop is a software platform for processing vast amounts of data
  Hadoop is a software platform that lets one easily write and
  run applications that process vast amounts of data.
@@ -195,7 +195,7 @@ Description: Development support for libhdfs0
 
 Package: hadoop-httpfs
 Architecture: any
-Depends: hadoop-hdfs (= ${binary:Version}), bigtop-tomcat
+Depends: hadoop-hdfs (= ${binary:Version})
 Description: HTTPFS for Hadoop
   The server providing HTTP REST API support for the complete FileSystem/FileContext
   interface in HDFS.
diff --git a/bigtop-packages/src/deb/hadoop/hadoop-httpfs.install b/bigtop-packages/src/deb/hadoop/hadoop-httpfs.install
index a594bc0..19cd6f5 100644
--- a/bigtop-packages/src/deb/hadoop/hadoop-httpfs.install
+++ b/bigtop-packages/src/deb/hadoop/hadoop-httpfs.install
@@ -1,5 +1,4 @@
 /etc/default/hadoop-httpfs
 /etc/hadoop-httpfs
-/usr/lib/hadoop/libexec/httpfs-config.sh
 /usr/lib/hadoop-httpfs
 /var/lib/hadoop-httpfs
diff --git a/bigtop-packages/src/deb/hadoop/hadoop-mapreduce.install b/bigtop-packages/src/deb/hadoop/hadoop-mapreduce.install
index 34a3796..ee971a9 100644
--- a/bigtop-packages/src/deb/hadoop/hadoop-mapreduce.install
+++ b/bigtop-packages/src/deb/hadoop/hadoop-mapreduce.install
@@ -2,7 +2,6 @@
 /etc/hadoop/conf.empty/mapred-site.xml
 /etc/hadoop/conf.empty/mapred-env.sh
 /etc/hadoop/conf.empty/mapred-queues.xml.template
-/etc/hadoop/conf.empty/mapred-site.xml.template
 /usr/lib/hadoop-mapreduce
 /usr/lib/hadoop/libexec/mapred-config.sh
 /usr/bin/mapred
diff --git a/bigtop-packages/src/deb/hadoop/hadoop.install b/bigtop-packages/src/deb/hadoop/hadoop.install
index 2700cb6..7fd0f34 100644
--- a/bigtop-packages/src/deb/hadoop/hadoop.install
+++ b/bigtop-packages/src/deb/hadoop/hadoop.install
@@ -1,7 +1,5 @@
-/etc/hadoop/conf.empty/hadoop-metrics.properties
 /etc/hadoop/conf.empty/hadoop-metrics2.properties
 /etc/hadoop/conf.empty/log4j.properties
-/etc/hadoop/conf.empty/slaves
 /etc/hadoop/conf.empty/ssl-client.xml.example
 /etc/hadoop/conf.empty/ssl-server.xml.example
 /etc/hadoop/conf.empty/core-site.xml
diff --git a/bigtop-packages/src/deb/hadoop/rules b/bigtop-packages/src/deb/hadoop/rules
index 58d4edc..b0f2167 100755
--- a/bigtop-packages/src/deb/hadoop/rules
+++ b/bigtop-packages/src/deb/hadoop/rules
@@ -67,10 +67,6 @@ override_dh_auto_install:
 	  --example-dir=debian/tmp/usr/share/doc/hadoop/examples \
 	  --native-build-string=${native_dir} \
 	  --installed-lib-dir=/usr/lib/hadoop
-	# Forcing Zookeeper dependency to be on the packaged jar
-	ln -sf /usr/lib/zookeeper/zookeeper.jar debian/tmp/usr/lib/hadoop/lib/zookeeper*.jar
-	# Workaround for BIGTOP-583
-	rm -f debian/tmp/usr/lib/hadoop-*/lib/slf4j-log4j12-*.jar
 	# FIXME: BIGTOP-463
 	mkdir -p debian/tmp/etc/default
 	echo 'export JSVC_HOME=/usr/lib/bigtop-utils' >> debian/hadoop.default
diff --git a/bigtop-packages/src/rpm/hadoop/SPECS/hadoop.spec b/bigtop-packages/src/rpm/hadoop/SPECS/hadoop.spec
index 1223dea..262b2a6 100644
--- a/bigtop-packages/src/rpm/hadoop/SPECS/hadoop.spec
+++ b/bigtop-packages/src/rpm/hadoop/SPECS/hadoop.spec
@@ -172,7 +172,7 @@ Source29: hadoop-yarn-timelineserver.svc
 #BIGTOP_PATCH_FILES
 Buildroot: %{_tmppath}/%{name}-%{version}-%{release}-root-%(%{__id} -u -n)
 BuildRequires: fuse-devel, fuse, cmake
-Requires: coreutils, /usr/sbin/useradd, /usr/sbin/usermod, /sbin/chkconfig, /sbin/service, bigtop-utils >= 0.7, zookeeper >= 3.4.0
+Requires: coreutils, /usr/sbin/useradd, /usr/sbin/usermod, /sbin/chkconfig, /sbin/service, bigtop-utils >= 0.7
 Requires: psmisc, %{netcat_package}
 # Sadly, Sun/Oracle JDK in RPM form doesn't provide libjvm.so, which means we have
 # to set AutoReq to no in order to minimize confusion. Not ideal, but seems to work.
@@ -326,7 +326,7 @@ blocks of data over the network to Hadoop Distributed Filesystem
 %package httpfs
 Summary: HTTPFS for Hadoop
 Group: System/Daemons
-Requires: %{name}-hdfs = %{version}-%{release}, bigtop-tomcat
+Requires: %{name}-hdfs = %{version}-%{release}
 Requires(pre): %{name} = %{version}-%{release}
 Requires(pre): %{name}-hdfs = %{version}-%{release}
 
@@ -502,11 +502,6 @@ env HADOOP_VERSION=%{hadoop_base_version} bash %{SOURCE2} \
   --installed-lib-dir=%{lib_hadoop} \
   --man-dir=$RPM_BUILD_ROOT%{man_hadoop} \
 
-# Forcing Zookeeper dependency to be on the packaged jar
-%__ln_s -f /usr/lib/zookeeper/zookeeper.jar $RPM_BUILD_ROOT/%{lib_hadoop}/lib/zookeeper*.jar
-# Workaround for BIGTOP-583
-%__rm -f $RPM_BUILD_ROOT/%{lib_hadoop}-*/lib/slf4j-log4j12-*.jar
-
 # Init.d scripts
 %__install -d -m 0755 $RPM_BUILD_ROOT/%{initd_dir}/
 
@@ -632,7 +627,7 @@ fi
 
 %files mapreduce
 %defattr(-,root,root)
-%config(noreplace) %{etc_hadoop}/conf.empty/mapred-site.xml
+#%config(noreplace) %{etc_hadoop}/conf.empty/mapred-site.xml
 %config(noreplace) %{etc_hadoop}/conf.empty/mapred-env.sh
 %config(noreplace) %{etc_hadoop}/conf.empty/mapred-queues.xml.template
 %config(noreplace) %{etc_hadoop}/conf.empty/mapred-site.xml.template
@@ -649,10 +644,10 @@ fi
 %files
 %defattr(-,root,root)
 %config(noreplace) %{etc_hadoop}/conf.empty/core-site.xml
-%config(noreplace) %{etc_hadoop}/conf.empty/hadoop-metrics.properties
+#%config(noreplace) %{etc_hadoop}/conf.empty/hadoop-metrics.properties
 %config(noreplace) %{etc_hadoop}/conf.empty/hadoop-metrics2.properties
 %config(noreplace) %{etc_hadoop}/conf.empty/log4j.properties
-%config(noreplace) %{etc_hadoop}/conf.empty/slaves
+#%config(noreplace) %{etc_hadoop}/conf.empty/slaves
 %config(noreplace) %{etc_hadoop}/conf.empty/ssl-client.xml.example
 %config(noreplace) %{etc_hadoop}/conf.empty/ssl-server.xml.example
 %config(noreplace) %{etc_hadoop}/conf.empty/configuration.xsl
@@ -689,7 +684,7 @@ fi
 %defattr(-,root,root)
 %config(noreplace) %{etc_httpfs}
 %config(noreplace) /etc/default/%{name}-httpfs
-%{lib_hadoop}/libexec/httpfs-config.sh
+#%{lib_hadoop}/libexec/httpfs-config.sh
 %{initd_dir}/%{name}-httpfs
 %{lib_httpfs}
 %attr(0775,httpfs,httpfs) %{run_httpfs}
diff --git a/bigtop.bom b/bigtop.bom
index 20858fa..6c20da4 100644
--- a/bigtop.bom
+++ b/bigtop.bom
@@ -109,16 +109,6 @@ bigtop {
       To build a component with its dependencies, run build with -Dbuildwithdeps=true
   */
   dependencies = [
-    zookeeper:['hadoop', 'hbase'],
-    hadoop:['ignite-hadoop', 'hbase', 'hive', 'tez', 'sqoop', 'sqoop2',
-      'oozie', 'mahout', 'flume', 'giraph', 'solr', 'spark',
-      'phoenix', 'alluxio', 'kafka', 'ycsb', 'hama', 'zeppelin',
-      'tajo', 'apex'
-    ],
-    hbase:['phoenix','giraph','ycsb','hive'],
-    hive:['oozie', 'zeppelin'],
-    'ignite-hadoop':['zeppelin'],
-    spark:['zeppelin']
   ]
 
   components {
@@ -126,7 +116,7 @@ bigtop {
       name    = 'zookeeper'
       pkg     = name
       version {
-        base  = '3.4.6'
+        base  = '3.4.11'
         pkg   = base
         release = 1
       }
@@ -143,7 +133,7 @@ bigtop {
     'hadoop' {
       name    = 'hadoop'
       relNotes = 'Apache Hadoop'
-      version { base = '2.8.5'; pkg = base; release = 1 }
+      version { base = '3.1.2'; pkg = base; release = 1 }
       tarball { destination = "${name}-${version.base}.tar.gz"
                 source      = "${name}-${version.base}-src.tar.gz" }
       url     { download_path = "/$name/common/$name-${version.base}"


Mime
View raw message