Return-Path: X-Original-To: apmail-hadoop-common-commits-archive@www.apache.org Delivered-To: apmail-hadoop-common-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 4C174173C4 for ; Tue, 27 Jan 2015 22:32:40 +0000 (UTC) Received: (qmail 9010 invoked by uid 500); 27 Jan 2015 22:32:40 -0000 Delivered-To: apmail-hadoop-common-commits-archive@hadoop.apache.org Received: (qmail 8940 invoked by uid 500); 27 Jan 2015 22:32:40 -0000 Mailing-List: contact common-commits-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: common-dev@hadoop.apache.org Delivered-To: mailing list common-commits@hadoop.apache.org Received: (qmail 8931 invoked by uid 99); 27 Jan 2015 22:32:40 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 27 Jan 2015 22:32:40 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id 6B27FE0404; Tue, 27 Jan 2015 22:32:40 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: aw@apache.org To: common-commits@hadoop.apache.org Message-Id: <19e37201e4ad46ed84a5c358f62aec80@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: hadoop git commit: HDFS-7566. Remove obsolete entries from hdfs-default.xml (Ray Chiang via aw) Date: Tue, 27 Jan 2015 22:32:40 +0000 (UTC) Repository: hadoop Updated Branches: refs/heads/branch-2 c95e8951f -> 874e38d81 HDFS-7566. Remove obsolete entries from hdfs-default.xml (Ray Chiang via aw) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/874e38d8 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/874e38d8 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/874e38d8 Branch: refs/heads/branch-2 Commit: 874e38d81f4a70af6f887ba86c57c7650f319bf4 Parents: c95e895 Author: Allen Wittenauer Authored: Tue Jan 27 14:32:21 2015 -0800 Committer: Allen Wittenauer Committed: Tue Jan 27 14:32:21 2015 -0800 ---------------------------------------------------------------------- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 3 +++ .../src/main/resources/hdfs-default.xml | 26 -------------------- .../resources/job_1329348432655_0001_conf.xml | 2 -- .../src/main/data/2jobs2min-rumen-jh.json | 6 ----- 4 files changed, 3 insertions(+), 34 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hadoop/blob/874e38d8/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index ad4e369..7ebc31e 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -511,6 +511,9 @@ Release 2.7.0 - UNRELEASED HDFS-49. MiniDFSCluster.stopDataNode will always shut down a node in the cluster if a matching name is not found. (stevel) + HDFS-7566. Remove obsolete entries from hdfs-default.xml (Ray Chiang + via aw) + Release 2.6.1 - UNRELEASED INCOMPATIBLE CHANGES http://git-wip-us.apache.org/repos/asf/hadoop/blob/874e38d8/hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml b/hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml index 51d9ed6..8296414 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml @@ -31,16 +31,6 @@ - dfs.namenode.logging.level - info - - The logging level for dfs namenode. Other values are "dir" (trace - namespace mutations), "block" (trace block under/over replications - and block creations/deletions), or "all". - - - - dfs.namenode.rpc-address @@ -155,14 +145,6 @@ - dfs.https.enable - false - - Deprecated. Use "dfs.http.policy" instead. - - - - dfs.http.policy HTTP_ONLY Decide if HTTPS(SSL) is supported on HDFS @@ -1225,14 +1207,6 @@ - dfs.support.append - true - - Does HDFS allow appends to files? - - - - dfs.client.use.datanode.hostname false Whether clients should use datanode hostnames when http://git-wip-us.apache.org/repos/asf/hadoop/blob/874e38d8/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/test/resources/job_1329348432655_0001_conf.xml ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/test/resources/job_1329348432655_0001_conf.xml b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/test/resources/job_1329348432655_0001_conf.xml index f988cf6..191aae8 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/test/resources/job_1329348432655_0001_conf.xml +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/test/resources/job_1329348432655_0001_conf.xml @@ -113,7 +113,6 @@ hadoop.proxyuser.user.groupsusers dfs.namenode.name.dir.restorefalse io.seqfile.lazydecompresstrue -dfs.https.enablefalse mapreduce.reduce.merge.inmem.threshold1000 mapreduce.input.fileinputformat.split.minsize0 dfs.replication3 @@ -210,7 +209,6 @@ mapreduce.job.dir/tmp/hadoop-yarn/staging/user/.staging/job_1329348432655_0001 io.map.index.skip0 net.topology.node.switch.mapping.implorg.apache.hadoop.net.ScriptBasedMapping -dfs.namenode.logging.levelinfo fs.s3.maxRetries4 s3native.client-write-packet-size65536 yarn.resourcemanager.amliveliness-monitor.interval-ms1000 http://git-wip-us.apache.org/repos/asf/hadoop/blob/874e38d8/hadoop-tools/hadoop-sls/src/main/data/2jobs2min-rumen-jh.json ---------------------------------------------------------------------- diff --git a/hadoop-tools/hadoop-sls/src/main/data/2jobs2min-rumen-jh.json b/hadoop-tools/hadoop-sls/src/main/data/2jobs2min-rumen-jh.json index 59ae8d7..70ff8af 100644 --- a/hadoop-tools/hadoop-sls/src/main/data/2jobs2min-rumen-jh.json +++ b/hadoop-tools/hadoop-sls/src/main/data/2jobs2min-rumen-jh.json @@ -4657,7 +4657,6 @@ "mapreduce.tasktracker.taskcontroller" : "org.apache.hadoop.mapred.DefaultTaskController", "yarn.scheduler.fair.preemption" : "true", "mapreduce.reduce.shuffle.parallelcopies" : "5", - "dfs.support.append" : "true", "yarn.nodemanager.env-whitelist" : "JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,YARN_HOME", "mapreduce.jobtracker.heartbeats.in.second" : "100", "mapreduce.job.maxtaskfailures.per.tracker" : "3", @@ -4674,7 +4673,6 @@ "dfs.datanode.hdfs-blocks-metadata.enabled" : "true", "ha.zookeeper.parent-znode" : "/hadoop-ha", "io.seqfile.lazydecompress" : "true", - "dfs.https.enable" : "false", "mapreduce.reduce.merge.inmem.threshold" : "1000", "mapreduce.input.fileinputformat.split.minsize" : "0", "dfs.replication" : "3", @@ -4783,7 +4781,6 @@ "io.map.index.skip" : "0", "net.topology.node.switch.mapping.impl" : "org.apache.hadoop.net.ScriptBasedMapping", "fs.s3.maxRetries" : "4", - "dfs.namenode.logging.level" : "info", "ha.failover-controller.new-active.rpc-timeout.ms" : "60000", "s3native.client-write-packet-size" : "65536", "yarn.resourcemanager.amliveliness-monitor.interval-ms" : "1000", @@ -9770,7 +9767,6 @@ "mapreduce.tasktracker.taskcontroller" : "org.apache.hadoop.mapred.DefaultTaskController", "yarn.scheduler.fair.preemption" : "true", "mapreduce.reduce.shuffle.parallelcopies" : "5", - "dfs.support.append" : "true", "yarn.nodemanager.env-whitelist" : "JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,YARN_HOME", "mapreduce.jobtracker.heartbeats.in.second" : "100", "mapreduce.job.maxtaskfailures.per.tracker" : "3", @@ -9787,7 +9783,6 @@ "dfs.datanode.hdfs-blocks-metadata.enabled" : "true", "ha.zookeeper.parent-znode" : "/hadoop-ha", "io.seqfile.lazydecompress" : "true", - "dfs.https.enable" : "false", "mapreduce.reduce.merge.inmem.threshold" : "1000", "mapreduce.input.fileinputformat.split.minsize" : "0", "dfs.replication" : "3", @@ -9896,7 +9891,6 @@ "io.map.index.skip" : "0", "net.topology.node.switch.mapping.impl" : "org.apache.hadoop.net.ScriptBasedMapping", "fs.s3.maxRetries" : "4", - "dfs.namenode.logging.level" : "info", "ha.failover-controller.new-active.rpc-timeout.ms" : "60000", "s3native.client-write-packet-size" : "65536", "yarn.resourcemanager.amliveliness-monitor.interval-ms" : "1000",