Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id B00C4200B43 for ; Wed, 8 Jun 2016 01:06:52 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id AE96E160A4F; Tue, 7 Jun 2016 23:06:52 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id B9F17160A5D for ; Wed, 8 Jun 2016 01:06:50 +0200 (CEST) Received: (qmail 73792 invoked by uid 500); 7 Jun 2016 23:06:49 -0000 Mailing-List: contact commits-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list commits@hbase.apache.org Received: (qmail 73349 invoked by uid 99); 7 Jun 2016 23:06:49 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 07 Jun 2016 23:06:49 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id 4922DE083B; Tue, 7 Jun 2016 23:06:49 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: apurtell@apache.org To: commits@hbase.apache.org Date: Tue, 07 Jun 2016 23:06:55 -0000 Message-Id: <50e9dc1eb0de4548a05f244527e7a47c@git.apache.org> In-Reply-To: <2b6e0d86cd454b4d8f4750522b2633d2@git.apache.org> References: <2b6e0d86cd454b4d8f4750522b2633d2@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [7/8] hbase git commit: HBASE-15889. String case conversions are locale-sensitive, used without locale archived-at: Tue, 07 Jun 2016 23:06:52 -0000 HBASE-15889. String case conversions are locale-sensitive, used without locale Signed-off-by: Sean Busbey Conflicts: hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcChannel.java hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RpcClientImpl.java hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaTableLocator.java hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcExecutor.java hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormat.java hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java hbase-server/src/test/java/org/apache/hadoop/hbase/TimedOutTestsListener.java hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestMultiTableInputFormat.java Amending-Author: Andrew Purtell Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/e26800c0 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/e26800c0 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/e26800c0 Branch: refs/heads/0.98 Commit: e26800c027fc010fd72c4bb2b78997ea30466606 Parents: 60958f5 Author: Sean Mackrory Authored: Wed May 25 07:49:01 2016 -0600 Committer: Andrew Purtell Committed: Tue Jun 7 15:35:12 2016 -0700 ---------------------------------------------------------------------- .../classification/tools/StabilityOptions.java | 5 +++-- .../apache/hadoop/hbase/HColumnDescriptor.java | 19 ++++++++++--------- .../java/org/apache/hadoop/hbase/ServerName.java | 3 ++- .../hadoop/hbase/catalog/CatalogTracker.java | 3 ++- .../hadoop/hbase/filter/SubstringComparator.java | 7 ++++--- .../org/apache/hadoop/hbase/ipc/RpcClient.java | 7 ++++--- .../apache/hadoop/hbase/security/SaslUtil.java | 7 ++++--- .../org/apache/hadoop/hbase/util/PoolMap.java | 3 ++- .../hbase/io/crypto/KeyStoreKeyProvider.java | 3 ++- .../ipc/MetricsHBaseServerSourceFactoryImpl.java | 3 ++- .../hbase/test/MetricsAssertHelperImpl.java | 3 ++- .../ipc/MetricsHBaseServerSourceFactoryImpl.java | 3 ++- .../hbase/test/MetricsAssertHelperImpl.java | 3 ++- .../apache/hadoop/hbase/HBaseClusterManager.java | 7 ++++--- .../hadoop/hbase/RESTApiClusterManager.java | 9 +++++---- .../StripeCompactionsPerformanceEvaluation.java | 3 ++- .../hadoop/hbase/rest/filter/GzipFilter.java | 5 +++-- .../apache/hadoop/hbase/mapreduce/Import.java | 3 ++- .../hbase/regionserver/BaseRowProcessor.java | 3 ++- .../hadoop/hbase/regionserver/HRegion.java | 3 ++- .../hbase/security/HBaseSaslRpcServer.java | 3 ++- .../hadoop/hbase/snapshot/CreateSnapshot.java | 3 ++- .../hadoop/hbase/util/CompressionTest.java | 5 +++-- .../hadoop/hbase/util/DirectMemoryUtils.java | 3 ++- .../org/apache/hadoop/hbase/util/FSUtils.java | 3 ++- .../org/apache/hadoop/hbase/util/HBaseFsck.java | 3 ++- .../hadoop/hbase/util/ServerCommandLine.java | 5 +++-- .../hadoop/hbase/PerformanceEvaluation.java | 5 +++-- .../hbase/TestIPv6NIOServerSocketChannel.java | 3 ++- .../mapreduce/MultiTableInputFormatTestBase.java | 5 +++-- .../mapreduce/TestLoadIncrementalHFiles.java | 3 ++- .../mapreduce/TestTableInputFormatScanBase.java | 9 +++++---- .../regionserver/DataBlockEncodingTool.java | 9 +++++---- .../apache/hadoop/hbase/util/LoadTestTool.java | 3 ++- .../thrift/HThreadedSelectorServerArgs.java | 3 ++- .../hadoop/hbase/thrift/ThriftUtilities.java | 3 ++- 36 files changed, 103 insertions(+), 67 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java ---------------------------------------------------------------------- diff --git a/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java b/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java index f3169ba..217ad12 100644 --- a/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java +++ b/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java @@ -21,6 +21,7 @@ import com.sun.javadoc.DocErrorReporter; import java.util.ArrayList; import java.util.List; +import java.util.Locale; class StabilityOptions { public static final String STABLE_OPTION = "-stable"; @@ -28,7 +29,7 @@ class StabilityOptions { public static final String UNSTABLE_OPTION = "-unstable"; public static Integer optionLength(String option) { - String opt = option.toLowerCase(); + String opt = option.toLowerCase(Locale.ROOT); if (opt.equals(UNSTABLE_OPTION)) return 1; if (opt.equals(EVOLVING_OPTION)) return 1; if (opt.equals(STABLE_OPTION)) return 1; @@ -38,7 +39,7 @@ class StabilityOptions { public static void validOptions(String[][] options, DocErrorReporter reporter) { for (int i = 0; i < options.length; i++) { - String opt = options[i][0].toLowerCase(); + String opt = options[i][0].toLowerCase(Locale.ROOT); if (opt.equals(UNSTABLE_OPTION)) { RootDocProcessor.stability = UNSTABLE_OPTION; } else if (opt.equals(EVOLVING_OPTION)) { http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java index 15aee9d..d83487c 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java @@ -24,6 +24,7 @@ import java.io.IOException; import java.util.Collections; import java.util.HashMap; import java.util.HashSet; +import java.util.Locale; import java.util.Map; import java.util.Set; @@ -445,11 +446,11 @@ public class HColumnDescriptor implements WritableComparable setBlockCacheEnabled(blockCacheEnabled); setTimeToLive(timeToLive); setCompressionType(Compression.Algorithm. - valueOf(compression.toUpperCase())); + valueOf(compression.toUpperCase(Locale.ROOT))); setDataBlockEncoding(DataBlockEncoding. - valueOf(dataBlockEncoding.toUpperCase())); + valueOf(dataBlockEncoding.toUpperCase(Locale.ROOT))); setBloomFilterType(BloomType. - valueOf(bloomFilter.toUpperCase())); + valueOf(bloomFilter.toUpperCase(Locale.ROOT))); setBlocksize(blocksize); setScope(scope); } @@ -568,7 +569,7 @@ public class HColumnDescriptor implements WritableComparable if (n == null) { return Compression.Algorithm.NONE; } - return Compression.Algorithm.valueOf(n.toUpperCase()); + return Compression.Algorithm.valueOf(n.toUpperCase(Locale.ROOT)); } /** @return compression type being used for the column family for major @@ -578,7 +579,7 @@ public class HColumnDescriptor implements WritableComparable if (n == null) { return getCompression(); } - return Compression.Algorithm.valueOf(n.toUpperCase()); + return Compression.Algorithm.valueOf(n.toUpperCase(Locale.ROOT)); } /** @return maximum number of versions */ @@ -673,7 +674,7 @@ public class HColumnDescriptor implements WritableComparable * @return this (for chained invocation) */ public HColumnDescriptor setCompressionType(Compression.Algorithm type) { - return setValue(COMPRESSION, type.getName().toUpperCase()); + return setValue(COMPRESSION, type.getName().toUpperCase(Locale.ROOT)); } /** @return data block encoding algorithm used on disk */ @@ -760,7 +761,7 @@ public class HColumnDescriptor implements WritableComparable */ public HColumnDescriptor setCompactionCompressionType( Compression.Algorithm type) { - return setValue(COMPRESSION_COMPACT, type.getName().toUpperCase()); + return setValue(COMPRESSION_COMPACT, type.getName().toUpperCase(Locale.ROOT)); } /** @@ -786,7 +787,7 @@ public class HColumnDescriptor implements WritableComparable String value = getValue(KEEP_DELETED_CELLS); if (value != null) { // toUpperCase for backwards compatibility - return KeepDeletedCells.valueOf(value.toUpperCase()); + return KeepDeletedCells.valueOf(value.toUpperCase(Locale.ROOT)); } return DEFAULT_KEEP_DELETED; } @@ -871,7 +872,7 @@ public class HColumnDescriptor implements WritableComparable if (n == null) { n = DEFAULT_BLOOMFILTER; } - return BloomType.valueOf(n.toUpperCase()); + return BloomType.valueOf(n.toUpperCase(Locale.ROOT)); } /** http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerName.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerName.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerName.java index 36a67fe..ec8fcb2 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerName.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerName.java @@ -33,6 +33,7 @@ import org.apache.hadoop.hbase.util.Bytes; import java.io.Serializable; import java.util.ArrayList; import java.util.List; +import java.util.Locale; import java.util.regex.Pattern; /** @@ -229,7 +230,7 @@ public class ServerName implements Comparable, Serializable { */ static String getServerName(String hostName, int port, long startcode) { final StringBuilder name = new StringBuilder(hostName.length() + 1 + 5 + 1 + 13); - name.append(hostName.toLowerCase()); + name.append(hostName.toLowerCase(Locale.ROOT)); name.append(SERVERNAME_SEPARATOR); name.append(port); name.append(SERVERNAME_SEPARATOR); http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-client/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java index 08b7ae7..4c9e20f 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java @@ -50,6 +50,7 @@ import java.net.NoRouteToHostException; import java.net.SocketException; import java.net.SocketTimeoutException; import java.net.UnknownHostException; +import java.util.Locale; /** * Tracks the availability of the catalog tables @@ -383,7 +384,7 @@ public class CatalogTracker { } else if (cause != null && cause instanceof EOFException) { // Catch. Other end disconnected us. } else if (cause != null && cause.getMessage() != null && - cause.getMessage().toLowerCase().contains("connection reset")) { + cause.getMessage().toLowerCase(Locale.ROOT).contains("connection reset")) { // Catch. Connection reset. } else { throw ioe; http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java index 5eb3703..1f0043c 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java @@ -19,6 +19,7 @@ package org.apache.hadoop.hbase.filter; import com.google.protobuf.InvalidProtocolBufferException; +import java.util.Locale; import org.apache.hadoop.hbase.classification.InterfaceAudience; import org.apache.hadoop.hbase.classification.InterfaceStability; import org.apache.hadoop.hbase.exceptions.DeserializationException; @@ -52,8 +53,8 @@ public class SubstringComparator extends ByteArrayComparable { * @param substr the substring */ public SubstringComparator(String substr) { - super(Bytes.toBytes(substr.toLowerCase())); - this.substr = substr.toLowerCase(); + super(Bytes.toBytes(substr.toLowerCase(Locale.ROOT))); + this.substr = substr.toLowerCase(Locale.ROOT); } @Override @@ -63,7 +64,7 @@ public class SubstringComparator extends ByteArrayComparable { @Override public int compareTo(byte[] value, int offset, int length) { - return Bytes.toString(value, offset, length).toLowerCase().contains(substr) ? 0 + return Bytes.toString(value, offset, length).toLowerCase(Locale.ROOT).contains(substr) ? 0 : 1; } http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RpcClient.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RpcClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RpcClient.java index a06faf9..edeb328 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RpcClient.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/RpcClient.java @@ -39,6 +39,7 @@ import java.security.PrivilegedExceptionAction; import java.util.HashMap; import java.util.Iterator; import java.util.LinkedList; +import java.util.Locale; import java.util.Map; import java.util.Map.Entry; import java.util.Random; @@ -411,8 +412,8 @@ public class RpcClient { throw new IOException( "Can't obtain server Kerberos config key from SecurityInfo"); } - serverPrincipal = SecurityUtil.getServerPrincipal( - conf.get(serverKey), server.getAddress().getCanonicalHostName().toLowerCase()); + serverPrincipal = SecurityUtil.getServerPrincipal(conf.get(serverKey), + server.getAddress().getCanonicalHostName().toLowerCase(Locale.ROOT)); if (LOG.isDebugEnabled()) { LOG.debug("RPC Server Kerberos principal name for service=" + remoteId.getServiceName() + " is " + serverPrincipal); @@ -770,7 +771,7 @@ public class RpcClient { final OutputStream out2) throws IOException { saslRpcClient = new HBaseSaslRpcClient(authMethod, token, serverPrincipal, fallbackAllowed, conf.get("hbase.rpc.protection", - QualityOfProtection.AUTHENTICATION.name().toLowerCase())); + QualityOfProtection.AUTHENTICATION.name().toLowerCase(Locale.ROOT))); return saslRpcClient.saslConnect(in2, out2); } http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SaslUtil.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SaslUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SaslUtil.java index 1c7a77d..9a18cf9 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SaslUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/SaslUtil.java @@ -23,6 +23,7 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hbase.classification.InterfaceAudience; +import java.util.Locale; import java.util.Map; import java.util.TreeMap; @@ -76,13 +77,13 @@ public class SaslUtil { */ public static QualityOfProtection getQop(String stringQop) { QualityOfProtection qop = null; - if (QualityOfProtection.AUTHENTICATION.name().toLowerCase().equals(stringQop) + if (QualityOfProtection.AUTHENTICATION.name().toLowerCase(Locale.ROOT).equals(stringQop) || QualityOfProtection.AUTHENTICATION.saslQop.equals(stringQop)) { qop = QualityOfProtection.AUTHENTICATION; - } else if (QualityOfProtection.INTEGRITY.name().toLowerCase().equals(stringQop) + } else if (QualityOfProtection.INTEGRITY.name().toLowerCase(Locale.ROOT).equals(stringQop) || QualityOfProtection.INTEGRITY.saslQop.equals(stringQop)) { qop = QualityOfProtection.INTEGRITY; - } else if (QualityOfProtection.PRIVACY.name().toLowerCase().equals(stringQop) + } else if (QualityOfProtection.PRIVACY.name().toLowerCase(Locale.ROOT).equals(stringQop) || QualityOfProtection.PRIVACY.saslQop.equals(stringQop)) { qop = QualityOfProtection.PRIVACY; } http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-client/src/main/java/org/apache/hadoop/hbase/util/PoolMap.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/util/PoolMap.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/util/PoolMap.java index 7d27b7a..086cd1c 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/util/PoolMap.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/util/PoolMap.java @@ -24,6 +24,7 @@ import java.util.HashMap; import java.util.HashSet; import java.util.LinkedList; import java.util.List; +import java.util.Locale; import java.util.Map; import java.util.Set; import java.util.concurrent.ConcurrentHashMap; @@ -255,7 +256,7 @@ public class PoolMap implements Map { } public static String fuzzyNormalize(String name) { - return name != null ? name.replaceAll("-", "").trim().toLowerCase() : ""; + return name != null ? name.replaceAll("-", "").trim().toLowerCase(Locale.ROOT) : ""; } public static PoolType fuzzyMatch(String name) { http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-common/src/main/java/org/apache/hadoop/hbase/io/crypto/KeyStoreKeyProvider.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/crypto/KeyStoreKeyProvider.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/crypto/KeyStoreKeyProvider.java index 24e03c0..1e35351 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/crypto/KeyStoreKeyProvider.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/crypto/KeyStoreKeyProvider.java @@ -30,6 +30,7 @@ import java.security.KeyStoreException; import java.security.NoSuchAlgorithmException; import java.security.UnrecoverableKeyException; import java.security.cert.CertificateException; +import java.util.Locale; import java.util.Properties; import org.apache.hadoop.hbase.classification.InterfaceAudience; @@ -142,7 +143,7 @@ public class KeyStoreKeyProvider implements KeyProvider { throw new RuntimeException("KeyProvider scheme should specify KeyStore type"); } // KeyStore expects instance type specifications in uppercase - store = KeyStore.getInstance(storeType.toUpperCase()); + store = KeyStore.getInstance(storeType.toUpperCase(Locale.ROOT)); processParameters(uri); load(uri); } catch (URISyntaxException e) { http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-hadoop1-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java ---------------------------------------------------------------------- diff --git a/hbase-hadoop1-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java b/hbase-hadoop1-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java index 96de011..90550dc 100644 --- a/hbase-hadoop1-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java +++ b/hbase-hadoop1-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.ipc; import java.util.HashMap; +import java.util.Locale; import org.apache.hadoop.hbase.classification.InterfaceAudience; @@ -48,7 +49,7 @@ public class MetricsHBaseServerSourceFactoryImpl extends MetricsHBaseServerSourc source = new MetricsHBaseServerSourceImpl( METRICS_NAME, METRICS_DESCRIPTION, - context.toLowerCase(), + context.toLowerCase(Locale.ROOT), context + METRICS_JMX_CONTEXT_SUFFIX, wrapper); //Store back in storage http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-hadoop1-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java ---------------------------------------------------------------------- diff --git a/hbase-hadoop1-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java b/hbase-hadoop1-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java index dd85ae0..4ed194f 100644 --- a/hbase-hadoop1-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java +++ b/hbase-hadoop1-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java @@ -28,6 +28,7 @@ import org.apache.hadoop.metrics2.MetricsTag; import org.apache.hadoop.metrics2.lib.DefaultMetricsSystem; import java.util.HashMap; +import java.util.Locale; import java.util.Map; import static org.junit.Assert.*; @@ -226,6 +227,6 @@ public class MetricsAssertHelperImpl implements MetricsAssertHelper { } private String canonicalizeMetricName(String in) { - return in.toLowerCase().replaceAll("[^A-Za-z0-9 ]", ""); + return in.toLowerCase(Locale.ROOT).replaceAll("[^A-Za-z0-9 ]", ""); } } http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java ---------------------------------------------------------------------- diff --git a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java index 96de011..90550dc 100644 --- a/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java +++ b/hbase-hadoop2-compat/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerSourceFactoryImpl.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.ipc; import java.util.HashMap; +import java.util.Locale; import org.apache.hadoop.hbase.classification.InterfaceAudience; @@ -48,7 +49,7 @@ public class MetricsHBaseServerSourceFactoryImpl extends MetricsHBaseServerSourc source = new MetricsHBaseServerSourceImpl( METRICS_NAME, METRICS_DESCRIPTION, - context.toLowerCase(), + context.toLowerCase(Locale.ROOT), context + METRICS_JMX_CONTEXT_SUFFIX, wrapper); //Store back in storage http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java ---------------------------------------------------------------------- diff --git a/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java b/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java index 36e2696..85b2cee 100644 --- a/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java +++ b/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java @@ -29,6 +29,7 @@ import org.apache.hadoop.metrics2.MetricsTag; import org.apache.hadoop.metrics2.lib.DefaultMetricsSystem; import java.util.HashMap; +import java.util.Locale; import java.util.Map; import static org.junit.Assert.*; @@ -247,6 +248,6 @@ public class MetricsAssertHelperImpl implements MetricsAssertHelper { } private String canonicalizeMetricName(String in) { - return in.toLowerCase().replaceAll("[^A-Za-z0-9 ]", ""); + return in.toLowerCase(Locale.ROOT).replaceAll("[^A-Za-z0-9 ]", ""); } } http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-it/src/test/java/org/apache/hadoop/hbase/HBaseClusterManager.java ---------------------------------------------------------------------- diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/HBaseClusterManager.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/HBaseClusterManager.java index ba6a4a9..b0281fb 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/HBaseClusterManager.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/HBaseClusterManager.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase; import java.io.File; import java.io.IOException; +import java.util.Locale; import java.util.Map; import org.apache.commons.lang.StringUtils; @@ -205,7 +206,7 @@ public class HBaseClusterManager extends Configured implements ClusterManager { @Override public String getCommand(ServiceType service, Operation op) { return String.format("%s/bin/hbase-daemon.sh %s %s %s", hbaseHome, confDir, - op.toString().toLowerCase(), service); + op.toString().toLowerCase(Locale.ROOT), service); } } @@ -235,7 +236,7 @@ public class HBaseClusterManager extends Configured implements ClusterManager { @Override public String getCommand(ServiceType service, Operation op) { return String.format("%s/sbin/hadoop-daemon.sh %s %s %s", hadoopHome, confDir, - op.toString().toLowerCase(), service); + op.toString().toLowerCase(Locale.ROOT), service); } } @@ -264,7 +265,7 @@ public class HBaseClusterManager extends Configured implements ClusterManager { @Override public String getCommand(ServiceType service, Operation op) { - return String.format("%s/bin/zkServer.sh %s", zookeeperHome, op.toString().toLowerCase()); + return String.format("%s/bin/zkServer.sh %s", zookeeperHome, op.toString().toLowerCase(Locale.ROOT)); } @Override http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-it/src/test/java/org/apache/hadoop/hbase/RESTApiClusterManager.java ---------------------------------------------------------------------- diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/RESTApiClusterManager.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/RESTApiClusterManager.java index 6c5c45e..ce7d645 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/RESTApiClusterManager.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/RESTApiClusterManager.java @@ -37,6 +37,7 @@ import javax.xml.ws.http.HTTPException; import java.io.IOException; import java.net.URI; import java.util.HashMap; +import java.util.Locale; import java.util.Map; /** @@ -273,8 +274,8 @@ public class RESTApiClusterManager extends Configured implements ClusterManager if (role.get("hostRef").get("hostId").getTextValue().equals(hostId) && role.get("type") .getTextValue() - .toLowerCase() - .equals(roleType.toLowerCase())) { + .toLowerCase(Locale.ROOT) + .equals(roleType.toLowerCase(Locale.ROOT))) { roleValue = role.get(property).getTextValue(); break; } @@ -327,7 +328,7 @@ public class RESTApiClusterManager extends Configured implements ClusterManager // APIs tend to take commands in lowercase, so convert them to save the trouble later. @Override public String toString() { - return name().toLowerCase(); + return name().toLowerCase(Locale.ROOT); } } @@ -347,4 +348,4 @@ public class RESTApiClusterManager extends Configured implements ClusterManager private enum Service { HBASE, HDFS, MAPREDUCE } -} \ No newline at end of file +} http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-it/src/test/java/org/apache/hadoop/hbase/StripeCompactionsPerformanceEvaluation.java ---------------------------------------------------------------------- diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/StripeCompactionsPerformanceEvaluation.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/StripeCompactionsPerformanceEvaluation.java index c48734c..7071c73 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/StripeCompactionsPerformanceEvaluation.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/StripeCompactionsPerformanceEvaluation.java @@ -19,6 +19,7 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import java.util.Locale; import java.util.Set; import org.apache.commons.cli.CommandLine; @@ -111,7 +112,7 @@ public class StripeCompactionsPerformanceEvaluation extends AbstractHBaseTool { } else { minValueSize = maxValueSize = Integer.parseInt(valueSize); } - String datagen = cmd.getOptionValue(DATAGEN_KEY, "default").toLowerCase(); + String datagen = cmd.getOptionValue(DATAGEN_KEY, "default").toLowerCase(Locale.ROOT); if ("default".equals(datagen)) { dataGen = new MultiThreadedAction.DefaultDataGenerator( minValueSize, maxValueSize, 1, 1, new byte[][] { COLUMN_FAMILY }); http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/filter/GzipFilter.java ---------------------------------------------------------------------- diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/filter/GzipFilter.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/filter/GzipFilter.java index 4995b86..094ae0b 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/filter/GzipFilter.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/filter/GzipFilter.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.rest.filter; import java.io.IOException; import java.io.OutputStream; import java.util.HashSet; +import java.util.Locale; import java.util.Set; import java.util.StringTokenizer; @@ -65,11 +66,11 @@ public class GzipFilter implements Filter { String acceptEncoding = request.getHeader("accept-encoding"); String contentType = request.getHeader("content-type"); if ((contentEncoding != null) && - (contentEncoding.toLowerCase().indexOf("gzip") > -1)) { + (contentEncoding.toLowerCase(Locale.ROOT).indexOf("gzip") > -1)) { request = new GZIPRequestWrapper(request); } if (((acceptEncoding != null) && - (acceptEncoding.toLowerCase().indexOf("gzip") > -1)) || + (acceptEncoding.toLowerCase(Locale.ROOT).indexOf("gzip") > -1)) || ((contentType != null) && mimeTypes.contains(contentType))) { response = new GZIPResponseWrapper(response); } http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java index 22a59a5..d14f339 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java @@ -28,6 +28,7 @@ import java.lang.reflect.Method; import java.util.ArrayList; import java.util.Collections; import java.util.List; +import java.util.Locale; import java.util.Map; import java.util.TreeMap; import java.util.UUID; @@ -400,7 +401,7 @@ public class Import { filter = instantiateFilter(conf); String durabilityStr = conf.get(WAL_DURABILITY); if(durabilityStr != null){ - durability = Durability.valueOf(durabilityStr.toUpperCase()); + durability = Durability.valueOf(durabilityStr.toUpperCase(Locale.ROOT)); } // TODO: This is kind of ugly doing setup of ZKW just to read the clusterid. ZooKeeperWatcher zkw = null; http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/BaseRowProcessor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/BaseRowProcessor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/BaseRowProcessor.java index 65375b8..be2bd91 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/BaseRowProcessor.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/BaseRowProcessor.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; import java.util.ArrayList; import java.util.List; +import java.util.Locale; import java.util.UUID; import org.apache.hadoop.hbase.classification.InterfaceAudience; @@ -61,7 +62,7 @@ implements RowProcessor { @Override public String getName() { - return this.getClass().getSimpleName().toLowerCase(); + return this.getClass().getSimpleName().toLowerCase(Locale.ROOT); } @Override http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java index c337c9c..90e9296 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java @@ -33,6 +33,7 @@ import java.util.Collections; import java.util.HashMap; import java.util.Iterator; import java.util.List; +import java.util.Locale; import java.util.Map; import java.util.NavigableMap; import java.util.NavigableSet; @@ -6480,7 +6481,7 @@ public class HRegion implements HeapSize { // , Writable{ } boolean majorCompact = false; if (args.length > 1) { - if (!args[1].toLowerCase().startsWith("major")) { + if (!args[1].toLowerCase(Locale.ROOT).startsWith("major")) { printUsageAndExit("ERROR: Unrecognized option <" + args[1] + ">"); } majorCompact = true; http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/main/java/org/apache/hadoop/hbase/security/HBaseSaslRpcServer.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/HBaseSaslRpcServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/HBaseSaslRpcServer.java index 3ee7a9a..1b37efc 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/HBaseSaslRpcServer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/HBaseSaslRpcServer.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.security; import java.io.ByteArrayInputStream; import java.io.DataInputStream; import java.io.IOException; +import java.util.Locale; import javax.security.auth.callback.Callback; import javax.security.auth.callback.CallbackHandler; @@ -50,7 +51,7 @@ public class HBaseSaslRpcServer { public static void init(Configuration conf) { SaslUtil.initSaslProperties(conf.get("hbase.rpc.protection", - QualityOfProtection.AUTHENTICATION.name().toLowerCase())); + QualityOfProtection.AUTHENTICATION.name().toLowerCase(Locale.ROOT))); } public static T getIdentifier(String id, http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/CreateSnapshot.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/CreateSnapshot.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/CreateSnapshot.java index e135143..75e8dea 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/CreateSnapshot.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/snapshot/CreateSnapshot.java @@ -25,6 +25,7 @@ import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; import org.apache.hadoop.hbase.util.AbstractHBaseTool; import java.util.Arrays; +import java.util.Locale; /** @@ -63,7 +64,7 @@ public class CreateSnapshot extends AbstractHBaseTool { admin = new HBaseAdmin(conf); HBaseProtos.SnapshotDescription.Type type = HBaseProtos.SnapshotDescription.Type.FLUSH; if (snapshotType != null) { - type = HBaseProtos.SnapshotDescription.Type.valueOf(snapshotName.toUpperCase()); + type = HBaseProtos.SnapshotDescription.Type.valueOf(snapshotName.toUpperCase(Locale.ROOT)); } admin.snapshot(snapshotName, TableName.valueOf(tableName), type); http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java index 355719d..2c6fadd 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java @@ -19,6 +19,7 @@ package org.apache.hadoop.hbase.util; import java.io.IOException; +import java.util.Locale; import org.apache.commons.lang.StringUtils; import org.apache.commons.logging.Log; @@ -47,7 +48,7 @@ public class CompressionTest { static final Log LOG = LogFactory.getLog(CompressionTest.class); public static boolean testCompression(String codec) { - codec = codec.toLowerCase(); + codec = codec.toLowerCase(Locale.ROOT); Compression.Algorithm a; @@ -103,7 +104,7 @@ public class CompressionTest { System.err.println( "Usage: CompressionTest " + - StringUtils.join( Compression.Algorithm.values(), "|").toLowerCase() + + StringUtils.join( Compression.Algorithm.values(), "|").toLowerCase(Locale.ROOT) + "\n" + "For example:\n" + " hbase " + CompressionTest.class + " file:///tmp/testfile gz\n"); http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/main/java/org/apache/hadoop/hbase/util/DirectMemoryUtils.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/DirectMemoryUtils.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/DirectMemoryUtils.java index bc901ab..bd8837d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/DirectMemoryUtils.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/DirectMemoryUtils.java @@ -25,6 +25,7 @@ import java.lang.reflect.InvocationTargetException; import java.lang.reflect.Method; import java.nio.ByteBuffer; import java.util.List; +import java.util.Locale; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; @@ -87,7 +88,7 @@ public class DirectMemoryUtils { long multiplier = 1; //for the byte case. for (String s : arguments) { if (s.contains("-XX:MaxDirectMemorySize=")) { - String memSize = s.toLowerCase() + String memSize = s.toLowerCase(Locale.ROOT) .replace("-xx:maxdirectmemorysize=", "").trim(); if (memSize.contains("k")) { http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java index d0443b8..74dad43 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java @@ -35,6 +35,7 @@ import java.util.Collections; import java.util.HashMap; import java.util.LinkedList; import java.util.List; +import java.util.Locale; import java.util.Map; import java.util.concurrent.ArrayBlockingQueue; import java.util.concurrent.ConcurrentHashMap; @@ -1885,7 +1886,7 @@ public abstract class FSUtils { return false; } - if (!regionName.toLowerCase().matches("[0-9a-f]+")) { + if (!regionName.toLowerCase(Locale.ROOT).matches("[0-9a-f]+")) { return false; } return true; http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java index 4311ca5..feb75b7 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java @@ -32,6 +32,7 @@ import java.util.HashMap; import java.util.HashSet; import java.util.Iterator; import java.util.List; +import java.util.Locale; import java.util.Map; import java.util.Map.Entry; import java.util.Set; @@ -3759,7 +3760,7 @@ public class HBaseFsck extends Configured { errors.progress(); String encodedName = regionDir.getPath().getName(); // ignore directories that aren't hexadecimal - if (!encodedName.toLowerCase().matches("[0-9a-f]+")) { + if (!encodedName.toLowerCase(Locale.ROOT).matches("[0-9a-f]+")) { continue; } http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ServerCommandLine.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ServerCommandLine.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ServerCommandLine.java index 983d49c..e6b746c 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ServerCommandLine.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/ServerCommandLine.java @@ -22,6 +22,7 @@ import java.lang.management.ManagementFactory; import java.lang.management.RuntimeMXBean; import java.util.Arrays; import java.util.HashSet; +import java.util.Locale; import java.util.Map.Entry; import java.util.Set; @@ -103,8 +104,8 @@ public abstract class ServerCommandLine extends Configured implements Tool { nextEnv: for (Entry entry : System.getenv().entrySet()) { - String key = entry.getKey().toLowerCase(); - String value = entry.getValue().toLowerCase(); + String key = entry.getKey().toLowerCase(Locale.ROOT); + String value = entry.getValue().toLowerCase(Locale.ROOT); // exclude variables which may contain skip words for(String skipWord : skipWords) { if (key.contains(skipWord) || value.contains(skipWord)) http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java index 6a144a0..2b462c9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java @@ -31,6 +31,7 @@ import java.util.ArrayList; import java.util.Arrays; import java.util.Date; import java.util.LinkedList; +import java.util.Locale; import java.util.Map; import java.util.Queue; import java.util.Random; @@ -308,8 +309,8 @@ public class PerformanceEvaluation extends Configured implements Tool { static boolean checkTable(HBaseAdmin admin, TestOptions opts) throws IOException { TableName tableName = TableName.valueOf(opts.tableName); boolean needsDelete = false, exists = admin.tableExists(tableName); - boolean isReadCmd = opts.cmdName.toLowerCase().contains("read") - || opts.cmdName.toLowerCase().contains("scan"); + boolean isReadCmd = opts.cmdName.toLowerCase(Locale.ROOT).contains("read") + || opts.cmdName.toLowerCase(Locale.ROOT).contains("scan"); if (!exists && isReadCmd) { throw new IllegalStateException( "Must specify an existing table for read commands. Run a write command first."); http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/test/java/org/apache/hadoop/hbase/TestIPv6NIOServerSocketChannel.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestIPv6NIOServerSocketChannel.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestIPv6NIOServerSocketChannel.java index 0baf5de3..09fbe16 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestIPv6NIOServerSocketChannel.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestIPv6NIOServerSocketChannel.java @@ -24,6 +24,7 @@ import java.net.InetAddress; import java.net.InetSocketAddress; import java.net.ServerSocket; import java.nio.channels.ServerSocketChannel; +import java.util.Locale; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; @@ -118,7 +119,7 @@ public class TestIPv6NIOServerSocketChannel { //java.net.SocketException: Address family not supported by protocol family //or java.net.SocketException: Protocol family not supported Assert.assertFalse(ex.getClass().isInstance(BindException.class)); - Assert.assertTrue(ex.getMessage().toLowerCase().contains("protocol family")); + Assert.assertTrue(ex.getMessage().toLowerCase(Locale.ROOT).contains("protocol family")); LOG.info("Received expected exception:"); LOG.info(ex); http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatTestBase.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatTestBase.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatTestBase.java index 9a6314a..c96d5fb 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatTestBase.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatTestBase.java @@ -44,6 +44,7 @@ import java.io.File; import java.io.IOException; import java.util.ArrayList; import java.util.List; +import java.util.Locale; import java.util.Map; import java.util.NavigableMap; @@ -226,8 +227,8 @@ public abstract class MultiTableInputFormatTestBase { private void testScan(String start, String stop, String last) throws IOException, InterruptedException, ClassNotFoundException { String jobName = - "Scan" + (start != null ? start.toUpperCase() : "Empty") + "To" + - (stop != null ? stop.toUpperCase() : "Empty"); + "Scan" + (start != null ? start.toUpperCase(Locale.ROOT) : "Empty") + "To" + + (stop != null ? stop.toUpperCase(Locale.ROOT) : "Empty"); LOG.info("Before map/reduce startup - job " + jobName); Configuration c = new Configuration(TEST_UTIL.getConfiguration()); http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFiles.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFiles.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFiles.java index 8ed7b29..04b8bd9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFiles.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestLoadIncrementalHFiles.java @@ -24,6 +24,7 @@ import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; import java.io.IOException; +import java.util.Locale; import java.util.TreeMap; import org.apache.hadoop.conf.Configuration; @@ -353,7 +354,7 @@ public class TestLoadIncrementalHFiles { // set real family name to upper case in purpose to simulate the case that // family name in HFiles is invalid HColumnDescriptor family = - new HColumnDescriptor(Bytes.toBytes(new String(FAMILY).toUpperCase())); + new HColumnDescriptor(Bytes.toBytes(new String(FAMILY).toUpperCase(Locale.ROOT))); htd.addFamily(family); try { http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableInputFormatScanBase.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableInputFormatScanBase.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableInputFormatScanBase.java index a29b53c..e167be9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableInputFormatScanBase.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableInputFormatScanBase.java @@ -24,6 +24,7 @@ import static org.junit.Assert.assertTrue; import java.io.IOException; import java.util.Arrays; import java.util.List; +import java.util.Locale; import java.util.Map; import java.util.NavigableMap; @@ -181,8 +182,8 @@ public abstract class TestTableInputFormatScanBase { */ protected void testScanFromConfiguration(String start, String stop, String last) throws IOException, InterruptedException, ClassNotFoundException { - String jobName = "ScanFromConfig" + (start != null ? start.toUpperCase() : "Empty") + - "To" + (stop != null ? stop.toUpperCase() : "Empty"); + String jobName = "ScanFromConfig" + (start != null ? start.toUpperCase(Locale.ROOT) : "Empty") + + "To" + (stop != null ? stop.toUpperCase(Locale.ROOT) : "Empty"); Configuration c = new Configuration(TEST_UTIL.getConfiguration()); c.set(TableInputFormat.INPUT_TABLE, Bytes.toString(TABLE_NAME)); c.set(TableInputFormat.SCAN_COLUMN_FAMILY, Bytes.toString(INPUT_FAMILY)); @@ -218,8 +219,8 @@ public abstract class TestTableInputFormatScanBase { */ protected void testScan(String start, String stop, String last) throws IOException, InterruptedException, ClassNotFoundException { - String jobName = "Scan" + (start != null ? start.toUpperCase() : "Empty") + - "To" + (stop != null ? stop.toUpperCase() : "Empty"); + String jobName = "Scan" + (start != null ? start.toUpperCase(Locale.ROOT) : "Empty") + + "To" + (stop != null ? stop.toUpperCase(Locale.ROOT) : "Empty"); LOG.info("Before map/reduce startup - job " + jobName); Configuration c = new Configuration(TEST_UTIL.getConfiguration()); Scan scan = new Scan(); http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/DataBlockEncodingTool.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/DataBlockEncodingTool.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/DataBlockEncodingTool.java index ead913a..826ffbc 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/DataBlockEncodingTool.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/DataBlockEncodingTool.java @@ -24,6 +24,7 @@ import java.text.DecimalFormat; import java.util.ArrayList; import java.util.Iterator; import java.util.List; +import java.util.Locale; import org.apache.commons.cli.CommandLine; import org.apache.commons.cli.CommandLineParser; @@ -142,7 +143,7 @@ public class DataBlockEncodingTool { String s = super.toString(); StringBuilder sb = new StringBuilder(); sb.append(s.charAt(0)); - sb.append(s.substring(1).toLowerCase()); + sb.append(s.substring(1).toLowerCase(Locale.ROOT)); return sb.toString(); } } @@ -372,7 +373,7 @@ public class DataBlockEncodingTool { private void benchmarkDefaultCompression(int totalSize, byte[] rawBuffer) throws IOException { benchmarkAlgorithm(compressionAlgorithm, - compressionAlgorithmName.toUpperCase(), rawBuffer, 0, totalSize); + compressionAlgorithmName.toUpperCase(Locale.ROOT), rawBuffer, 0, totalSize); } /** @@ -526,7 +527,7 @@ public class DataBlockEncodingTool { * @throws IOException */ public void displayStatistics() throws IOException { - final String comprAlgo = compressionAlgorithmName.toUpperCase(); + final String comprAlgo = compressionAlgorithmName.toUpperCase(Locale.ROOT); long rawBytes = totalKeyLength + totalPrefixLength + totalValueLength; System.out.println("Raw data size:"); @@ -694,7 +695,7 @@ public class DataBlockEncodingTool { String compressionName = DEFAULT_COMPRESSION.getName(); if (cmd.hasOption(OPT_ENCODING_ALGORITHM)) { compressionName = - cmd.getOptionValue(OPT_ENCODING_ALGORITHM).toLowerCase(); + cmd.getOptionValue(OPT_ENCODING_ALGORITHM).toLowerCase(Locale.ROOT); } boolean doBenchmark = cmd.hasOption(OPT_MEASURE_THROUGHPUT); boolean doVerify = !cmd.hasOption(OPT_OMIT_CORRECTNESS_TEST); http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-server/src/test/java/org/apache/hadoop/hbase/util/LoadTestTool.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/LoadTestTool.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/LoadTestTool.java index 1ca75be..afeb0ad 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/LoadTestTool.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/LoadTestTool.java @@ -24,6 +24,7 @@ import java.security.SecureRandom; import java.util.ArrayList; import java.util.Arrays; import java.util.List; +import java.util.Locale; import java.util.Properties; import java.util.Random; import java.util.concurrent.atomic.AtomicReference; @@ -116,7 +117,7 @@ public class LoadTestTool extends AbstractHBaseTool { public static final String OPT_DEFERRED_LOG_FLUSH_USAGE = "Enable deferred log flush."; public static final String OPT_DATA_BLOCK_ENCODING = - HColumnDescriptor.DATA_BLOCK_ENCODING.toLowerCase(); + HColumnDescriptor.DATA_BLOCK_ENCODING.toLowerCase(Locale.ROOT); public static final String OPT_INMEMORY = "in_memory"; public static final String OPT_USAGE_IN_MEMORY = "Tries to keep the HFiles of the CF " + http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/HThreadedSelectorServerArgs.java ---------------------------------------------------------------------- diff --git a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/HThreadedSelectorServerArgs.java b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/HThreadedSelectorServerArgs.java index 5fbde7a..da33cc0 100644 --- a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/HThreadedSelectorServerArgs.java +++ b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/HThreadedSelectorServerArgs.java @@ -19,6 +19,7 @@ package org.apache.hadoop.hbase.thrift; +import java.util.Locale; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hbase.classification.InterfaceAudience; @@ -79,7 +80,7 @@ public class HThreadedSelectorServerArgs extends TThreadedSelectorServer.Args { int acceptQueueSizePerThread = conf.getInt( ACCEPT_QUEUE_SIZE_PER_THREAD_CONF_KEY, getAcceptQueueSizePerThread()); AcceptPolicy acceptPolicy = AcceptPolicy.valueOf(conf.get( - ACCEPT_POLICY_CONF_KEY, getAcceptPolicy().toString()).toUpperCase()); + ACCEPT_POLICY_CONF_KEY, getAcceptPolicy().toString()).toUpperCase(Locale.ROOT)); super.selectorThreads(selectorThreads) .workerThreads(workerThreads) http://git-wip-us.apache.org/repos/asf/hbase/blob/e26800c0/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/ThriftUtilities.java ---------------------------------------------------------------------- diff --git a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/ThriftUtilities.java b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/ThriftUtilities.java index 0340ccf..6c2b7b5 100644 --- a/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/ThriftUtilities.java +++ b/hbase-thrift/src/main/java/org/apache/hadoop/hbase/thrift/ThriftUtilities.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.thrift; import java.nio.ByteBuffer; import java.util.ArrayList; import java.util.List; +import java.util.Locale; import java.util.TreeMap; import org.apache.hadoop.hbase.classification.InterfaceAudience; @@ -55,7 +56,7 @@ public class ThriftUtilities { static public HColumnDescriptor colDescFromThrift(ColumnDescriptor in) throws IllegalArgument { Compression.Algorithm comp = - Compression.getCompressionAlgorithmByName(in.compression.toLowerCase()); + Compression.getCompressionAlgorithmByName(in.compression.toLowerCase(Locale.ROOT)); BloomType bt = BloomType.valueOf(in.bloomFilterType);