Return-Path: X-Original-To: apmail-hadoop-common-commits-archive@www.apache.org Delivered-To: apmail-hadoop-common-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 96AED1837B for ; Mon, 4 May 2015 20:04:30 +0000 (UTC) Received: (qmail 14861 invoked by uid 500); 4 May 2015 20:04:21 -0000 Delivered-To: apmail-hadoop-common-commits-archive@hadoop.apache.org Received: (qmail 14621 invoked by uid 500); 4 May 2015 20:04:21 -0000 Mailing-List: contact common-commits-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: common-dev@hadoop.apache.org Delivered-To: mailing list common-commits@hadoop.apache.org Received: (qmail 13353 invoked by uid 99); 4 May 2015 20:04:20 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 04 May 2015 20:04:20 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id DDE41E0A05; Mon, 4 May 2015 20:04:20 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: zjshen@apache.org To: common-commits@hadoop.apache.org Date: Mon, 04 May 2015 20:04:45 -0000 Message-Id: <268c3b2639d54024b76c08941bfacdec@git.apache.org> In-Reply-To: <6c861728d79543b6973a8facd17148c6@git.apache.org> References: <6c861728d79543b6973a8facd17148c6@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [26/50] [abbrv] hadoop git commit: HDFS-8213. DFSClient should use hdfs.client.htrace HTrace configuration prefix rather than hadoop.htrace (cmccabe) HDFS-8213. DFSClient should use hdfs.client.htrace HTrace configuration prefix rather than hadoop.htrace (cmccabe) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/fe95065a Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/fe95065a Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/fe95065a Branch: refs/heads/YARN-2928 Commit: fe95065abfb481c9b38fc32b0a0c15dbb137aed8 Parents: 45f0120 Author: Colin Patrick Mccabe Authored: Fri May 1 11:19:40 2015 -0700 Committer: Zhijie Shen Committed: Mon May 4 12:58:57 2015 -0700 ---------------------------------------------------------------------- .../apache/hadoop/tracing/SpanReceiverHost.java | 61 ++++++++++---------- .../org/apache/hadoop/tracing/TraceUtils.java | 14 ++--- .../apache/hadoop/tracing/TestTraceUtils.java | 10 ++-- hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt | 3 + .../java/org/apache/hadoop/hdfs/DFSClient.java | 5 +- .../org/apache/hadoop/hdfs/DFSConfigKeys.java | 7 +++ .../hadoop/hdfs/server/datanode/DataNode.java | 3 +- .../hadoop/hdfs/server/namenode/NameNode.java | 3 +- .../apache/hadoop/tracing/TestTraceAdmin.java | 4 +- .../org/apache/hadoop/tracing/TestTracing.java | 10 +--- .../TestTracingShortCircuitLocalRead.java | 4 +- 11 files changed, 69 insertions(+), 55 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hadoop/blob/fe95065a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/tracing/SpanReceiverHost.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/tracing/SpanReceiverHost.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/tracing/SpanReceiverHost.java index f2de0a0..bf9479b 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/tracing/SpanReceiverHost.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/tracing/SpanReceiverHost.java @@ -25,6 +25,7 @@ import java.io.FileInputStream; import java.io.IOException; import java.io.InputStreamReader; import java.util.Collections; +import java.util.HashMap; import java.util.List; import java.util.Map; import java.util.TreeMap; @@ -52,41 +53,36 @@ import org.apache.htrace.Trace; */ @InterfaceAudience.Private public class SpanReceiverHost implements TraceAdminProtocol { - public static final String SPAN_RECEIVERS_CONF_KEY = - "hadoop.htrace.spanreceiver.classes"; + public static final String SPAN_RECEIVERS_CONF_SUFFIX = + "spanreceiver.classes"; private static final Log LOG = LogFactory.getLog(SpanReceiverHost.class); + private static final HashMap hosts = + new HashMap(1); private final TreeMap receivers = new TreeMap(); + private final String confPrefix; private Configuration config; private boolean closed = false; private long highestId = 1; - private final static String LOCAL_FILE_SPAN_RECEIVER_PATH = - "hadoop.htrace.local-file-span-receiver.path"; + private final static String LOCAL_FILE_SPAN_RECEIVER_PATH_SUFFIX = + "local-file-span-receiver.path"; - private static enum SingletonHolder { - INSTANCE; - Object lock = new Object(); - SpanReceiverHost host = null; - } - - public static SpanReceiverHost getInstance(Configuration conf) { - if (SingletonHolder.INSTANCE.host != null) { - return SingletonHolder.INSTANCE.host; - } - synchronized (SingletonHolder.INSTANCE.lock) { - if (SingletonHolder.INSTANCE.host != null) { - return SingletonHolder.INSTANCE.host; + public static SpanReceiverHost get(Configuration conf, String confPrefix) { + synchronized (SpanReceiverHost.class) { + SpanReceiverHost host = hosts.get(confPrefix); + if (host != null) { + return host; } - SpanReceiverHost host = new SpanReceiverHost(); - host.loadSpanReceivers(conf); - SingletonHolder.INSTANCE.host = host; + final SpanReceiverHost newHost = new SpanReceiverHost(confPrefix); + newHost.loadSpanReceivers(conf); ShutdownHookManager.get().addShutdownHook(new Runnable() { public void run() { - SingletonHolder.INSTANCE.host.closeReceivers(); + newHost.closeReceivers(); } }, 0); - return SingletonHolder.INSTANCE.host; + hosts.put(confPrefix, newHost); + return newHost; } } @@ -119,6 +115,10 @@ public class SpanReceiverHost implements TraceAdminProtocol { return new File(tmp, nonce).getAbsolutePath(); } + private SpanReceiverHost(String confPrefix) { + this.confPrefix = confPrefix; + } + /** * Reads the names of classes specified in the * "hadoop.htrace.spanreceiver.classes" property and instantiates and registers @@ -131,22 +131,22 @@ public class SpanReceiverHost implements TraceAdminProtocol { */ public synchronized void loadSpanReceivers(Configuration conf) { config = new Configuration(conf); - String[] receiverNames = - config.getTrimmedStrings(SPAN_RECEIVERS_CONF_KEY); + String receiverKey = confPrefix + SPAN_RECEIVERS_CONF_SUFFIX; + String[] receiverNames = config.getTrimmedStrings(receiverKey); if (receiverNames == null || receiverNames.length == 0) { if (LOG.isTraceEnabled()) { - LOG.trace("No span receiver names found in " + - SPAN_RECEIVERS_CONF_KEY + "."); + LOG.trace("No span receiver names found in " + receiverKey + "."); } return; } // It's convenient to have each daemon log to a random trace file when // testing. - if (config.get(LOCAL_FILE_SPAN_RECEIVER_PATH) == null) { + String pathKey = confPrefix + LOCAL_FILE_SPAN_RECEIVER_PATH_SUFFIX; + if (config.get(pathKey) == null) { String uniqueFile = getUniqueLocalTraceFileName(); - config.set(LOCAL_FILE_SPAN_RECEIVER_PATH, uniqueFile); + config.set(pathKey, uniqueFile); if (LOG.isTraceEnabled()) { - LOG.trace("Set " + LOCAL_FILE_SPAN_RECEIVER_PATH + " to " + uniqueFile); + LOG.trace("Set " + pathKey + " to " + uniqueFile); } } for (String className : receiverNames) { @@ -164,7 +164,8 @@ public class SpanReceiverHost implements TraceAdminProtocol { private synchronized SpanReceiver loadInstance(String className, List extraConfig) throws IOException { SpanReceiverBuilder builder = - new SpanReceiverBuilder(TraceUtils.wrapHadoopConf(config, extraConfig)); + new SpanReceiverBuilder(TraceUtils. + wrapHadoopConf(confPrefix, config, extraConfig)); SpanReceiver rcvr = builder.spanReceiverClass(className.trim()).build(); if (rcvr == null) { throw new IOException("Failed to load SpanReceiver " + className); http://git-wip-us.apache.org/repos/asf/hadoop/blob/fe95065a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/tracing/TraceUtils.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/tracing/TraceUtils.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/tracing/TraceUtils.java index 11797e6..fa52ac6 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/tracing/TraceUtils.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/tracing/TraceUtils.java @@ -31,15 +31,15 @@ import org.apache.htrace.HTraceConfiguration; */ @InterfaceAudience.Private public class TraceUtils { - public static final String HTRACE_CONF_PREFIX = "hadoop.htrace."; private static List EMPTY = Collections.emptyList(); - public static HTraceConfiguration wrapHadoopConf(final Configuration conf) { - return wrapHadoopConf(conf, EMPTY); + public static HTraceConfiguration wrapHadoopConf(final String prefix, + final Configuration conf) { + return wrapHadoopConf(prefix, conf, EMPTY); } - public static HTraceConfiguration wrapHadoopConf(final Configuration conf, - List extraConfig) { + public static HTraceConfiguration wrapHadoopConf(final String prefix, + final Configuration conf, List extraConfig) { final HashMap extraMap = new HashMap(); for (ConfigurationPair pair : extraConfig) { extraMap.put(pair.getKey(), pair.getValue()); @@ -50,7 +50,7 @@ public class TraceUtils { if (extraMap.containsKey(key)) { return extraMap.get(key); } - return conf.get(HTRACE_CONF_PREFIX + key, ""); + return conf.get(prefix + key, ""); } @Override @@ -58,7 +58,7 @@ public class TraceUtils { if (extraMap.containsKey(key)) { return extraMap.get(key); } - return conf.get(HTRACE_CONF_PREFIX + key, defaultValue); + return conf.get(prefix + key, defaultValue); } }; } http://git-wip-us.apache.org/repos/asf/hadoop/blob/fe95065a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/tracing/TestTraceUtils.java ---------------------------------------------------------------------- diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/tracing/TestTraceUtils.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/tracing/TestTraceUtils.java index 9ef3483..80d64b1 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/tracing/TestTraceUtils.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/tracing/TestTraceUtils.java @@ -25,13 +25,15 @@ import org.apache.htrace.HTraceConfiguration; import org.junit.Test; public class TestTraceUtils { + private static String TEST_PREFIX = "test.prefix.htrace."; + @Test public void testWrappedHadoopConf() { String key = "sampler"; String value = "ProbabilitySampler"; Configuration conf = new Configuration(); - conf.set(TraceUtils.HTRACE_CONF_PREFIX + key, value); - HTraceConfiguration wrapped = TraceUtils.wrapHadoopConf(conf); + conf.set(TEST_PREFIX + key, value); + HTraceConfiguration wrapped = TraceUtils.wrapHadoopConf(TEST_PREFIX, conf); assertEquals(value, wrapped.get(key)); } @@ -41,11 +43,11 @@ public class TestTraceUtils { String oldValue = "old value"; String newValue = "new value"; Configuration conf = new Configuration(); - conf.set(TraceUtils.HTRACE_CONF_PREFIX + key, oldValue); + conf.set(TEST_PREFIX + key, oldValue); LinkedList extraConfig = new LinkedList(); extraConfig.add(new ConfigurationPair(key, newValue)); - HTraceConfiguration wrapped = TraceUtils.wrapHadoopConf(conf, extraConfig); + HTraceConfiguration wrapped = TraceUtils.wrapHadoopConf(TEST_PREFIX, conf, extraConfig); assertEquals(newValue, wrapped.get(key)); } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/fe95065a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt index 9accdc0..16094a2 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt +++ b/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt @@ -617,6 +617,9 @@ Release 2.7.1 - UNRELEASED HDFS-7770. Need document for storage type label of data node storage locations under dfs.data.dir. (Xiaoyu Yao via aajisaka) + HDFS-8213. DFSClient should use hdfs.client.htrace HTrace configuration + prefix rather than hadoop.htrace (cmccabe) + OPTIMIZATIONS BUG FIXES http://git-wip-us.apache.org/repos/asf/hadoop/blob/fe95065a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSClient.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSClient.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSClient.java index 8fc9e77..d47992b 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSClient.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSClient.java @@ -299,8 +299,9 @@ public class DFSClient implements java.io.Closeable, RemotePeerFactory, public DFSClient(URI nameNodeUri, ClientProtocol rpcNamenode, Configuration conf, FileSystem.Statistics stats) throws IOException { - SpanReceiverHost.getInstance(conf); - traceSampler = new SamplerBuilder(TraceUtils.wrapHadoopConf(conf)).build(); + SpanReceiverHost.get(conf, DFSConfigKeys.DFS_CLIENT_HTRACE_PREFIX); + traceSampler = new SamplerBuilder(TraceUtils. + wrapHadoopConf(DFSConfigKeys.DFS_CLIENT_HTRACE_PREFIX, conf)).build(); // Copy only the required DFSClient configuration this.dfsClientConf = new DfsClientConf(conf); this.conf = conf; http://git-wip-us.apache.org/repos/asf/hadoop/blob/fe95065a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java index 1e3a5b6..4356b9b 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java @@ -55,6 +55,13 @@ public class DFSConfigKeys extends CommonConfigurationKeys { public static final String DFS_WEBHDFS_ACL_PERMISSION_PATTERN_DEFAULT = HdfsClientConfigKeys.DFS_WEBHDFS_ACL_PERMISSION_PATTERN_DEFAULT; + // HDFS HTrace configuration is controlled by dfs.htrace.spanreceiver.classes, + // etc. + public static final String DFS_SERVER_HTRACE_PREFIX = "dfs.htrace."; + + // HDFS client HTrace configuration. + public static final String DFS_CLIENT_HTRACE_PREFIX = "dfs.client.htrace."; + // HA related configuration public static final String DFS_DATANODE_RESTART_REPLICA_EXPIRY_KEY = "dfs.datanode.restart.replica.expiration"; public static final long DFS_DATANODE_RESTART_REPLICA_EXPIRY_DEFAULT = 50; http://git-wip-us.apache.org/repos/asf/hadoop/blob/fe95065a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java index 2401d9c..f042dff 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java @@ -1099,7 +1099,8 @@ public class DataNode extends ReconfigurableBase this.dnConf = new DNConf(conf); checkSecureConfig(dnConf, conf, resources); - this.spanReceiverHost = SpanReceiverHost.getInstance(conf); + this.spanReceiverHost = + SpanReceiverHost.get(conf, DFSConfigKeys.DFS_SERVER_HTRACE_PREFIX); if (dnConf.maxLockedMemory > 0) { if (!NativeIO.POSIX.getCacheManipulator().verifyCanMlock()) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/fe95065a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java index 1e94923..132b93e 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java @@ -639,7 +639,8 @@ public class NameNode implements NameNodeStatusMXBean { startHttpServer(conf); } - this.spanReceiverHost = SpanReceiverHost.getInstance(conf); + this.spanReceiverHost = + SpanReceiverHost.get(conf, DFSConfigKeys.DFS_SERVER_HTRACE_PREFIX); loadNamesystem(conf); http://git-wip-us.apache.org/repos/asf/hadoop/blob/fe95065a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTraceAdmin.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTraceAdmin.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTraceAdmin.java index 7b3568d..4a102a3 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTraceAdmin.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTraceAdmin.java @@ -18,6 +18,7 @@ package org.apache.hadoop.tracing; import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.hdfs.DFSConfigKeys; import org.apache.hadoop.hdfs.MiniDFSCluster; import org.apache.hadoop.net.unix.TemporarySocketDirectory; import org.junit.Assert; @@ -57,7 +58,8 @@ public class TestTraceAdmin { public void testCreateAndDestroySpanReceiver() throws Exception { Configuration conf = new Configuration(); conf = new Configuration(); - conf.set(SpanReceiverHost.SPAN_RECEIVERS_CONF_KEY, ""); + conf.set(DFSConfigKeys.DFS_SERVER_HTRACE_PREFIX + + SpanReceiverHost.SPAN_RECEIVERS_CONF_SUFFIX, ""); MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).numDataNodes(3).build(); cluster.waitActive(); http://git-wip-us.apache.org/repos/asf/hadoop/blob/fe95065a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTracing.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTracing.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTracing.java index f6fef5a..59d1238 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTracing.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTracing.java @@ -22,6 +22,7 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.Path; +import org.apache.hadoop.hdfs.DFSConfigKeys; import org.apache.hadoop.hdfs.DistributedFileSystem; import org.apache.hadoop.hdfs.MiniDFSCluster; import org.apache.hadoop.test.GenericTestUtils; @@ -53,14 +54,9 @@ public class TestTracing { private static Configuration conf; private static MiniDFSCluster cluster; private static DistributedFileSystem dfs; - private static SpanReceiverHost spanReceiverHost; @Test public void testTracing() throws Exception { - // getting instance already loaded. - Assert.assertEquals(spanReceiverHost, - SpanReceiverHost.getInstance(new Configuration())); - // write and read without tracing started String fileName = "testTracingDisabled.dat"; writeTestFile(fileName); @@ -196,9 +192,9 @@ public class TestTracing { public static void setup() throws IOException { conf = new Configuration(); conf.setLong("dfs.blocksize", 100 * 1024); - conf.set(SpanReceiverHost.SPAN_RECEIVERS_CONF_KEY, + conf.set(DFSConfigKeys.DFS_CLIENT_HTRACE_PREFIX + + SpanReceiverHost.SPAN_RECEIVERS_CONF_SUFFIX, SetSpanReceiver.class.getName()); - spanReceiverHost = SpanReceiverHost.getInstance(conf); } @Before http://git-wip-us.apache.org/repos/asf/hadoop/blob/fe95065a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTracingShortCircuitLocalRead.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTracingShortCircuitLocalRead.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTracingShortCircuitLocalRead.java index 5d6db16..09ab350 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTracingShortCircuitLocalRead.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/tracing/TestTracingShortCircuitLocalRead.java @@ -64,7 +64,8 @@ public class TestTracingShortCircuitLocalRead { public void testShortCircuitTraceHooks() throws IOException { assumeTrue(NativeCodeLoader.isNativeCodeLoaded() && !Path.WINDOWS); conf = new Configuration(); - conf.set(SpanReceiverHost.SPAN_RECEIVERS_CONF_KEY, + conf.set(DFSConfigKeys.DFS_CLIENT_HTRACE_PREFIX + + SpanReceiverHost.SPAN_RECEIVERS_CONF_SUFFIX, TestTracing.SetSpanReceiver.class.getName()); conf.setLong("dfs.blocksize", 100 * 1024); conf.setBoolean(HdfsClientConfigKeys.Read.ShortCircuit.KEY, true); @@ -78,7 +79,6 @@ public class TestTracingShortCircuitLocalRead { dfs = cluster.getFileSystem(); try { - spanReceiverHost = SpanReceiverHost.getInstance(conf); DFSTestUtil.createFile(dfs, TEST_PATH, TEST_LENGTH, (short)1, 5678L); TraceScope ts = Trace.startSpan("testShortCircuitTraceHooks", Sampler.ALWAYS);