Return-Path: X-Original-To: apmail-hbase-commits-archive@www.apache.org Delivered-To: apmail-hbase-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 16A1410409 for ; Sat, 29 Nov 2014 17:24:21 +0000 (UTC) Received: (qmail 32280 invoked by uid 500); 29 Nov 2014 17:24:20 -0000 Delivered-To: apmail-hbase-commits-archive@hbase.apache.org Received: (qmail 32039 invoked by uid 500); 29 Nov 2014 17:24:20 -0000 Mailing-List: contact commits-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list commits@hbase.apache.org Received: (qmail 32023 invoked by uid 99); 29 Nov 2014 17:24:20 -0000 Received: from tyr.zones.apache.org (HELO tyr.zones.apache.org) (140.211.11.114) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 29 Nov 2014 17:24:20 +0000 Received: by tyr.zones.apache.org (Postfix, from userid 65534) id 7B2BF94DAA2; Sat, 29 Nov 2014 17:24:20 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: stack@apache.org To: commits@hbase.apache.org Date: Sat, 29 Nov 2014 17:24:21 -0000 Message-Id: <9997360e26b94065b1202d57f7ab7183@git.apache.org> In-Reply-To: <32118a0f13e9474ba3b11e054983d626@git.apache.org> References: <32118a0f13e9474ba3b11e054983d626@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [2/2] hbase git commit: HBASE-12519 Remove tabs used as whitespace (Varun Saxena) HBASE-12519 Remove tabs used as whitespace (Varun Saxena) Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/b12d5778 Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/b12d5778 Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/b12d5778 Branch: refs/heads/branch-1 Commit: b12d57783fe70102fcaf896a635c945c4c2cd97b Parents: 7eefa36 Author: stack Authored: Sat Nov 29 09:24:08 2014 -0800 Committer: stack Committed: Sat Nov 29 09:24:08 2014 -0800 ---------------------------------------------------------------------- ...ExcludePrivateAnnotationsStandardDoclet.java | 2 +- .../classification/tools/RootDocProcessor.java | 184 ++++---- .../classification/tools/StabilityOptions.java | 15 +- .../apache/hadoop/hbase/HTableDescriptor.java | 2 +- .../org/apache/hadoop/hbase/client/HTable.java | 4 +- .../apache/hadoop/hbase/client/HTableUtil.java | 2 +- .../client/UnmodifyableHTableDescriptor.java | 2 +- .../client/coprocessor/AggregationClient.java | 2 +- .../hbase/filter/DependentColumnFilter.java | 16 +- .../hadoop/hbase/protobuf/RequestConverter.java | 2 +- .../apache/hadoop/hbase/zookeeper/ZKAssign.java | 4 +- .../apache/hadoop/hbase/HBaseConfiguration.java | 2 +- .../org/apache/hadoop/hbase/io/TimeRange.java | 12 +- .../hadoop/hbase/util/CollectionUtils.java | 6 +- .../apache/hadoop/hbase/util/JenkinsHash.java | 2 +- .../hadoop/hbase/TestHBaseConfiguration.java | 4 +- ...egrationTestBigLinkedListWithVisibility.java | 4 +- .../codec/prefixtree/PrefixTreeBlockMeta.java | 6 +- .../decode/PrefixTreeArraySearcher.java | 2 +- .../encode/column/ColumnSectionWriter.java | 4 +- .../prefixtree/encode/other/LongEncoder.java | 4 +- .../prefixtree/encode/tokenize/Tokenizer.java | 2 +- .../encode/tokenize/TokenizerNode.java | 6 +- .../tokenize/TokenizerRowSearchPosition.java | 8 +- .../hadoop/hbase/util/vint/UFIntTool.java | 14 +- .../prefixtree/row/TestPrefixTreeSearcher.java | 2 +- .../codec/prefixtree/row/TestRowEncoder.java | 4 +- .../row/data/TestRowDataComplexQualifiers.java | 2 +- .../prefixtree/row/data/TestRowDataDeeper.java | 48 +-- .../data/TestRowDataDifferentTimestamps.java | 72 ++-- .../row/data/TestRowDataExerciseFInts.java | 84 ++-- .../prefixtree/row/data/TestRowDataNub.java | 52 +-- .../data/TestRowDataQualifierByteOrdering.java | 18 +- .../row/data/TestRowDataSearcherRowMiss.java | 24 +- .../row/data/TestRowDataSingleQualifier.java | 12 +- .../prefixtree/row/data/TestRowDataTrivial.java | 12 +- .../prefixtree/row/data/TestRowDataUrls.java | 2 +- .../row/data/TestRowDataUrlsExample.java | 60 +-- .../rest/model/StorageClusterVersionModel.java | 44 +- .../hadoop/hbase/rest/model/TableListModel.java | 108 ++--- .../hadoop/hbase/rest/model/TableModel.java | 72 ++-- .../hadoop/hbase/rest/model/VersionModel.java | 84 ++-- .../rest/provider/JAXBContextResolver.java | 36 +- .../producer/PlainTextMessageBodyProducer.java | 24 +- .../producer/ProtobufMessageBodyProducer.java | 44 +- .../hadoop/hbase/ipc/HBaseRPCErrorHandler.java | 10 +- .../hbase/master/RegionPlacementMaintainer.java | 2 +- .../SnapshotOfRegionAssignmentFromMeta.java | 2 +- .../hbase/regionserver/HRegionServer.java | 2 +- .../hbase/regionserver/SplitLogWorker.java | 2 +- .../handler/HLogSplitterHandler.java | 2 +- .../hadoop/hbase/regionserver/wal/FSHLog.java | 2 +- .../org/apache/hadoop/hbase/util/HBaseFsck.java | 4 +- .../hbase/zookeeper/RegionServerTracker.java | 2 +- .../hadoop/hbase/HBaseTestingUtility.java | 4 +- .../hadoop/hbase/client/TestHTablePool.java | 426 +++++++++---------- .../TestBigDecimalColumnInterpreter.java | 68 +-- .../hbase/filter/TestDependentColumnFilter.java | 24 +- .../hbase/io/hfile/TestHFilePerformance.java | 122 +++--- .../hbase/master/TestClockSkewDetection.java | 2 +- .../hbase/regionserver/TestMajorCompaction.java | 2 +- .../TestSplitTransactionOnCluster.java | 2 +- .../hadoop/hbase/regionserver/TestTags.java | 10 +- .../replication/TestReplicationSmallTests.java | 2 +- .../apache/hadoop/hbase/util/LoadTestTool.java | 2 +- .../hadoop/hbase/util/MultiThreadedReader.java | 6 +- .../hadoop/hbase/util/MultiThreadedUpdater.java | 2 +- 67 files changed, 897 insertions(+), 918 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/ExcludePrivateAnnotationsStandardDoclet.java ---------------------------------------------------------------------- diff --git a/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/ExcludePrivateAnnotationsStandardDoclet.java b/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/ExcludePrivateAnnotationsStandardDoclet.java index eb03a21..f93e13f 100644 --- a/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/ExcludePrivateAnnotationsStandardDoclet.java +++ b/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/ExcludePrivateAnnotationsStandardDoclet.java @@ -39,7 +39,7 @@ public class ExcludePrivateAnnotationsStandardDoclet { public static boolean start(RootDoc root) { System.out.println( - ExcludePrivateAnnotationsStandardDoclet.class.getSimpleName()); + ExcludePrivateAnnotationsStandardDoclet.class.getSimpleName()); return Standard.start(RootDocProcessor.process(root)); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/RootDocProcessor.java ---------------------------------------------------------------------- diff --git a/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/RootDocProcessor.java b/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/RootDocProcessor.java index 78e8163..c6fb74a 100644 --- a/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/RootDocProcessor.java +++ b/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/RootDocProcessor.java @@ -65,10 +65,10 @@ class RootDocProcessor { return getProxy(obj); } else if (obj instanceof Object[]) { Class componentType = type.isArray() ? type.getComponentType() - : cls.getComponentType(); + : cls.getComponentType(); Object[] array = (Object[]) obj; Object[] newArray = (Object[]) Array.newInstance(componentType, - array.length); + array.length); for (int i = 0; i < array.length; ++i) { newArray[i] = process(array[i], componentType); } @@ -98,116 +98,99 @@ class RootDocProcessor { } @Override - public Object invoke(Object proxy, Method method, Object[] args) - throws Throwable { + public Object invoke(Object proxy, Method method, Object[] args) throws Throwable { String methodName = method.getName(); if (target instanceof Doc) { - if (methodName.equals("isIncluded")) { - Doc doc = (Doc) target; - return !exclude(doc) && doc.isIncluded(); - } - if (target instanceof RootDoc) { - if (methodName.equals("classes")) { - return filter(((RootDoc) target).classes(), ClassDoc.class); - } else if (methodName.equals("specifiedClasses")) { - return filter(((RootDoc) target).specifiedClasses(), ClassDoc.class); - } else if (methodName.equals("specifiedPackages")) { - return filter(((RootDoc) target).specifiedPackages(), PackageDoc.class); - } - } else if (target instanceof ClassDoc) { - if (isFiltered(args)) { - if (methodName.equals("methods")) { - return filter(((ClassDoc) target).methods(true), MethodDoc.class); - } else if (methodName.equals("fields")) { - return filter(((ClassDoc) target).fields(true), FieldDoc.class); - } else if (methodName.equals("innerClasses")) { - return filter(((ClassDoc) target).innerClasses(true), - ClassDoc.class); - } else if (methodName.equals("constructors")) { - return filter(((ClassDoc) target).constructors(true), - ConstructorDoc.class); - } - } - } else if (target instanceof PackageDoc) { - if (methodName.equals("allClasses")) { - if (isFiltered(args)) { - return filter(((PackageDoc) target).allClasses(true), - ClassDoc.class); - } else { - return filter(((PackageDoc) target).allClasses(), ClassDoc.class); - } - } else if (methodName.equals("annotationTypes")) { - return filter(((PackageDoc) target).annotationTypes(), - AnnotationTypeDoc.class); - } else if (methodName.equals("enums")) { - return filter(((PackageDoc) target).enums(), - ClassDoc.class); - } else if (methodName.equals("errors")) { - return filter(((PackageDoc) target).errors(), - ClassDoc.class); - } else if (methodName.equals("exceptions")) { - return filter(((PackageDoc) target).exceptions(), - ClassDoc.class); - } else if (methodName.equals("interfaces")) { - return filter(((PackageDoc) target).interfaces(), - ClassDoc.class); - } else if (methodName.equals("ordinaryClasses")) { - return filter(((PackageDoc) target).ordinaryClasses(), - ClassDoc.class); - } - } + if (methodName.equals("isIncluded")) { + Doc doc = (Doc) target; + return !exclude(doc) && doc.isIncluded(); + } + if (target instanceof RootDoc) { + if (methodName.equals("classes")) { + return filter(((RootDoc) target).classes(), ClassDoc.class); + } else if (methodName.equals("specifiedClasses")) { + return filter(((RootDoc) target).specifiedClasses(), ClassDoc.class); + } else if (methodName.equals("specifiedPackages")) { + return filter(((RootDoc) target).specifiedPackages(), PackageDoc.class); + } + } else if (target instanceof ClassDoc) { + if (isFiltered(args)) { + if (methodName.equals("methods")) { + return filter(((ClassDoc) target).methods(true), MethodDoc.class); + } else if (methodName.equals("fields")) { + return filter(((ClassDoc) target).fields(true), FieldDoc.class); + } else if (methodName.equals("innerClasses")) { + return filter(((ClassDoc) target).innerClasses(true), ClassDoc.class); + } else if (methodName.equals("constructors")) { + return filter(((ClassDoc) target).constructors(true), ConstructorDoc.class); + } + } + } else if (target instanceof PackageDoc) { + if (methodName.equals("allClasses")) { + if (isFiltered(args)) { + return filter(((PackageDoc) target).allClasses(true), ClassDoc.class); + } else { + return filter(((PackageDoc) target).allClasses(), ClassDoc.class); + } + } else if (methodName.equals("annotationTypes")) { + return filter(((PackageDoc) target).annotationTypes(), AnnotationTypeDoc.class); + } else if (methodName.equals("enums")) { + return filter(((PackageDoc) target).enums(), ClassDoc.class); + } else if (methodName.equals("errors")) { + return filter(((PackageDoc) target).errors(), ClassDoc.class); + } else if (methodName.equals("exceptions")) { + return filter(((PackageDoc) target).exceptions(), ClassDoc.class); + } else if (methodName.equals("interfaces")) { + return filter(((PackageDoc) target).interfaces(), ClassDoc.class); + } else if (methodName.equals("ordinaryClasses")) { + return filter(((PackageDoc) target).ordinaryClasses(), ClassDoc.class); + } + } } if (args != null) { - if (methodName.equals("compareTo") || methodName.equals("equals") - || methodName.equals("overrides") - || methodName.equals("subclassOf")) { - args[0] = unwrap(args[0]); - } + if (methodName.equals("compareTo") || methodName.equals("equals") + || methodName.equals("overrides") || methodName.equals("subclassOf")) { + args[0] = unwrap(args[0]); + } } try { - return process(method.invoke(target, args), method.getReturnType()); + return process(method.invoke(target, args), method.getReturnType()); } catch (InvocationTargetException e) { - throw e.getTargetException(); + throw e.getTargetException(); } } private static boolean exclude(Doc doc) { AnnotationDesc[] annotations = null; if (doc instanceof ProgramElementDoc) { - annotations = ((ProgramElementDoc) doc).annotations(); + annotations = ((ProgramElementDoc) doc).annotations(); } else if (doc instanceof PackageDoc) { - annotations = ((PackageDoc) doc).annotations(); + annotations = ((PackageDoc) doc).annotations(); } if (annotations != null) { - for (AnnotationDesc annotation : annotations) { - String qualifiedTypeName = annotation.annotationType().qualifiedTypeName(); - if (qualifiedTypeName.equals( - InterfaceAudience.Private.class.getCanonicalName()) - || qualifiedTypeName.equals( - InterfaceAudience.LimitedPrivate.class.getCanonicalName())) { - return true; - } - if (stability.equals(StabilityOptions.EVOLVING_OPTION)) { - if (qualifiedTypeName.equals( - InterfaceStability.Unstable.class.getCanonicalName())) { - return true; - } - } - if (stability.equals(StabilityOptions.STABLE_OPTION)) { - if (qualifiedTypeName.equals( - InterfaceStability.Unstable.class.getCanonicalName()) - || qualifiedTypeName.equals( - InterfaceStability.Evolving.class.getCanonicalName())) { - return true; - } - } - } for (AnnotationDesc annotation : annotations) { - String qualifiedTypeName = - annotation.annotationType().qualifiedTypeName(); - if (qualifiedTypeName.equals( - InterfaceAudience.Public.class.getCanonicalName())) { + String qualifiedTypeName = annotation.annotationType().qualifiedTypeName(); + if (qualifiedTypeName.equals(InterfaceAudience.Private.class.getCanonicalName()) + || qualifiedTypeName + .equals(InterfaceAudience.LimitedPrivate.class.getCanonicalName())) { + return true; + } + if (stability.equals(StabilityOptions.EVOLVING_OPTION)) { + if (qualifiedTypeName.equals(InterfaceStability.Unstable.class.getCanonicalName())) { + return true; + } + } + if (stability.equals(StabilityOptions.STABLE_OPTION)) { + if (qualifiedTypeName.equals(InterfaceStability.Unstable.class.getCanonicalName()) + || qualifiedTypeName.equals(InterfaceStability.Evolving.class.getCanonicalName())) { + return true; + } + } + } + for (AnnotationDesc annotation : annotations) { + String qualifiedTypeName = annotation.annotationType().qualifiedTypeName(); + if (qualifiedTypeName.equals(InterfaceAudience.Public.class.getCanonicalName())) { return false; } } @@ -220,21 +203,20 @@ class RootDocProcessor { private static Object[] filter(Doc[] array, Class componentType) { if (array == null || array.length == 0) { - return array; + return array; } List list = new ArrayList(array.length); for (Doc entry : array) { - if (!exclude(entry)) { - list.add(process(entry, componentType)); - } + if (!exclude(entry)) { + list.add(process(entry, componentType)); + } } - return list.toArray((Object[]) Array.newInstance(componentType, list - .size())); + return list.toArray((Object[]) Array.newInstance(componentType, list.size())); } private Object unwrap(Object proxy) { if (proxy instanceof Proxy) - return ((ExcludeHandler) Proxy.getInvocationHandler(proxy)).target; + return ((ExcludeHandler) Proxy.getInvocationHandler(proxy)).target; return proxy; } http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java ---------------------------------------------------------------------- diff --git a/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java b/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java index f3169ba..b79f645 100644 --- a/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java +++ b/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java @@ -35,16 +35,15 @@ class StabilityOptions { return null; } - public static void validOptions(String[][] options, - DocErrorReporter reporter) { + public static void validOptions(String[][] options, DocErrorReporter reporter) { for (int i = 0; i < options.length; i++) { String opt = options[i][0].toLowerCase(); if (opt.equals(UNSTABLE_OPTION)) { - RootDocProcessor.stability = UNSTABLE_OPTION; + RootDocProcessor.stability = UNSTABLE_OPTION; } else if (opt.equals(EVOLVING_OPTION)) { - RootDocProcessor.stability = EVOLVING_OPTION; + RootDocProcessor.stability = EVOLVING_OPTION; } else if (opt.equals(STABLE_OPTION)) { - RootDocProcessor.stability = STABLE_OPTION; + RootDocProcessor.stability = STABLE_OPTION; } } } @@ -53,9 +52,9 @@ class StabilityOptions { List optionsList = new ArrayList(); for (int i = 0; i < options.length; i++) { if (!options[i][0].equalsIgnoreCase(UNSTABLE_OPTION) - && !options[i][0].equalsIgnoreCase(EVOLVING_OPTION) - && !options[i][0].equalsIgnoreCase(STABLE_OPTION)) { - optionsList.add(options[i]); + && !options[i][0].equalsIgnoreCase(EVOLVING_OPTION) + && !options[i][0].equalsIgnoreCase(STABLE_OPTION)) { + optionsList.add(options[i]); } } String[][] filteredOptions = new String[optionsList.size()][]; http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java index fd0de03..95f1771 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java @@ -1034,7 +1034,7 @@ public class HTableDescriptor implements WritableComparable { @Deprecated @Override public void write(DataOutput out) throws IOException { - out.writeInt(TABLE_DESCRIPTOR_VERSION); + out.writeInt(TABLE_DESCRIPTOR_VERSION); Bytes.writeByteArray(out, name.toBytes()); out.writeBoolean(isRootRegion()); out.writeBoolean(isMetaRegion()); http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java index 3d4c113..7dcf3e9 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java @@ -429,7 +429,7 @@ public class HTable implements HTableInterface, RegionLocator { * @param tableName Name of table to check. * @return {@code true} if table is online. * @throws IOException if a remote or network exception occurs - * @deprecated use {@link HBaseAdmin#isTableEnabled(byte[])} + * @deprecated use {@link HBaseAdmin#isTableEnabled(byte[])} */ @Deprecated public static boolean isTableEnabled(Configuration conf, String tableName) @@ -443,7 +443,7 @@ public class HTable implements HTableInterface, RegionLocator { * @param tableName Name of table to check. * @return {@code true} if table is online. * @throws IOException if a remote or network exception occurs - * @deprecated use {@link HBaseAdmin#isTableEnabled(byte[])} + * @deprecated use {@link HBaseAdmin#isTableEnabled(byte[])} */ @Deprecated public static boolean isTableEnabled(Configuration conf, byte[] tableName) http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java index f3b6563..ab77ceb 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java @@ -111,7 +111,7 @@ public class HTableUtil { List recs = putMap.get( hostname); if (recs == null) { recs = new ArrayList(INITIAL_LIST_SIZE); - putMap.put( hostname, recs); + putMap.put( hostname, recs); } recs.add(put); } http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHTableDescriptor.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHTableDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHTableDescriptor.java index 361e63e..c5a93e1 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHTableDescriptor.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHTableDescriptor.java @@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.HTableDescriptor; public class UnmodifyableHTableDescriptor extends HTableDescriptor { /** Default constructor */ public UnmodifyableHTableDescriptor() { - super(); + super(); } /* http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java index df79f8b..1e378e7 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java @@ -184,7 +184,7 @@ public class AggregationClient implements Closeable { || (Bytes.equals(scan.getStartRow(), scan.getStopRow()) && !Bytes .equals(scan.getStartRow(), HConstants.EMPTY_START_ROW)) || ((Bytes.compareTo(scan.getStartRow(), scan.getStopRow()) > 0) && - !Bytes.equals(scan.getStopRow(), HConstants.EMPTY_END_ROW))) { + !Bytes.equals(scan.getStopRow(), HConstants.EMPTY_END_ROW))) { throw new IOException( "Agg client Exception: Startrow should be smaller than Stoprow"); } else if (!canFamilyBeAbsent) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java index 5cfedaa..2843751 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java @@ -68,8 +68,8 @@ public class DependentColumnFilter extends CompareFilter { * @param valueComparator comparator */ public DependentColumnFilter(final byte [] family, final byte[] qualifier, - final boolean dropDependentColumn, final CompareOp valueCompareOp, - final ByteArrayComparable valueComparator) { + final boolean dropDependentColumn, final CompareOp valueCompareOp, + final ByteArrayComparable valueComparator) { // set up the comparator super(valueCompareOp, valueComparator); this.columnFamily = family; @@ -136,19 +136,19 @@ public class DependentColumnFilter extends CompareFilter { @Override public ReturnCode filterKeyValue(Cell c) { // Check if the column and qualifier match - if (!CellUtil.matchingColumn(c, this.columnFamily, this.columnQualifier)) { - // include non-matches for the time being, they'll be discarded afterwards - return ReturnCode.INCLUDE; - } + if (!CellUtil.matchingColumn(c, this.columnFamily, this.columnQualifier)) { + // include non-matches for the time being, they'll be discarded afterwards + return ReturnCode.INCLUDE; + } // If it doesn't pass the op, skip it if (comparator != null && doCompare(compareOp, comparator, c.getValueArray(), c.getValueOffset(), c.getValueLength())) return ReturnCode.SKIP; - + stampSet.add(c.getTimestamp()); if(dropDependentColumn) { - return ReturnCode.SKIP; + return ReturnCode.SKIP; } return ReturnCode.INCLUDE; } http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java index f5c128b..c7c9038 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java @@ -1069,7 +1069,7 @@ public final class RequestConverter { public static MoveRegionRequest buildMoveRegionRequest( final byte [] encodedRegionName, final byte [] destServerName) throws DeserializationException { - MoveRegionRequest.Builder builder = MoveRegionRequest.newBuilder(); + MoveRegionRequest.Builder builder = MoveRegionRequest.newBuilder(); builder.setRegion( buildRegionSpecifier(RegionSpecifierType.ENCODED_REGION_NAME,encodedRegionName)); if (destServerName != null) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java ---------------------------------------------------------------------- diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java index 7747ffa..ab12467 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java @@ -441,8 +441,8 @@ public class ZKAssign { EventType expectedState, ServerName serverName, int expectedVersion) throws KeeperException, KeeperException.NoNodeException { if (LOG.isTraceEnabled()) { - LOG.trace(zkw.prefix("Deleting existing unassigned " + - "node " + encodedRegionName + " in expected state " + expectedState)); + LOG.trace(zkw.prefix("Deleting existing unassigned " + + "node " + encodedRegionName + " in expected state " + expectedState)); } String node = getNodeName(zkw, encodedRegionName); zkw.sync(node); http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java index 808c4c1..7779399 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java @@ -181,7 +181,7 @@ public class HBaseConfiguration extends Configuration { char[] p = (char[]) m.invoke(conf, alias); if (p != null) { LOG.debug(String.format("Config option \"%s\" was found through" + - " the Configuration getPassword method.", alias)); + " the Configuration getPassword method.", alias)); passwd = new String(p); } else { http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java index b23b2f1..8c16389 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java @@ -61,7 +61,7 @@ public class TimeRange { * @param minStamp the minimum timestamp value, inclusive */ public TimeRange(byte [] minStamp) { - this.minStamp = Bytes.toLong(minStamp); + this.minStamp = Bytes.toLong(minStamp); } /** @@ -126,8 +126,8 @@ public class TimeRange { * @return true if within TimeRange, false if not */ public boolean withinTimeRange(byte [] bytes, int offset) { - if(allTime) return true; - return withinTimeRange(Bytes.toLong(bytes, offset)); + if(allTime) return true; + return withinTimeRange(Bytes.toLong(bytes, offset)); } /** @@ -139,9 +139,9 @@ public class TimeRange { * @return true if within TimeRange, false if not */ public boolean withinTimeRange(long timestamp) { - if(allTime) return true; - // check if >= minStamp - return (minStamp <= timestamp && timestamp < maxStamp); + if(allTime) return true; + // check if >= minStamp + return (minStamp <= timestamp && timestamp < maxStamp); } /** http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java index b15b2e0..b7b9beb 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java @@ -43,7 +43,7 @@ public class CollectionUtils { return in; } - /************************ size ************************************/ + /************************ size ************************************/ public static int nullSafeSize(Collection collection) { if (collection == null) { @@ -56,7 +56,7 @@ public class CollectionUtils { return nullSafeSize(a) == nullSafeSize(b); } - /*************************** empty ****************************************/ + /*************************** empty ****************************************/ public static boolean isEmpty(Collection collection) { return collection == null || collection.isEmpty(); @@ -66,7 +66,7 @@ public class CollectionUtils { return !isEmpty(collection); } - /************************ first/last **************************/ + /************************ first/last **************************/ public static T getFirst(Collection collection) { if (CollectionUtils.isEmpty(collection)) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-common/src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java index 42d8b11..359e7a9 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java @@ -164,7 +164,7 @@ public class JenkinsHash extends Hash { //-------------------------------- last block: affect all 32 bits of (c) switch (length) { // all the case statements fall through case 12: - c += ((key[offset + 11] & BYTE_MASK) << 24); + c += ((key[offset + 11] & BYTE_MASK) << 24); case 11: c += ((key[offset + 10] & BYTE_MASK) << 16); case 10: http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-common/src/test/java/org/apache/hadoop/hbase/TestHBaseConfiguration.java ---------------------------------------------------------------------- diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/TestHBaseConfiguration.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/TestHBaseConfiguration.java index 60fa3b3..445db84 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/TestHBaseConfiguration.java +++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/TestHBaseConfiguration.java @@ -199,7 +199,7 @@ public class TestHBaseConfiguration { hadoopClassesAvailable = true; LOG.info("Credential provider classes have been" + - " loaded and initialized successfully through reflection."); + " loaded and initialized successfully through reflection."); return true; } @@ -278,7 +278,7 @@ public class TestHBaseConfiguration { List providers = getCredentialProviders(conf); if (null == providers) { throw new IOException("Could not fetch any CredentialProviders, " + - "is the implementation available?"); + "is the implementation available?"); } Object provider = providers.get(0); http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java ---------------------------------------------------------------------- diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java index f80036c..9a8f2ab 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java @@ -583,8 +583,8 @@ public class IntegrationTestBigLinkedListWithVisibility extends IntegrationTestB if (args.length < 5) { System.err .println("Usage: Loop " + - " " + - " [ ]"); + " " + + " [ ]"); return 1; } LOG.info("Running Loop with args:" + Arrays.deepToString(args)); http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/PrefixTreeBlockMeta.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/PrefixTreeBlockMeta.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/PrefixTreeBlockMeta.java index 620a6ef..8410cf3 100644 --- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/PrefixTreeBlockMeta.java +++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/PrefixTreeBlockMeta.java @@ -45,7 +45,7 @@ public class PrefixTreeBlockMeta { public static final int MAX_FAMILY_LENGTH = Byte.MAX_VALUE;// hard-coded in KeyValue public static final int - NUM_LONGS = 2, + NUM_LONGS = 2, NUM_INTS = 28, NUM_SHORTS = 0,//keyValueTypeWidth not persisted NUM_SINGLE_BYTES = 2, @@ -135,7 +135,7 @@ public class PrefixTreeBlockMeta { } - /**************** operate on each field **********************/ + /**************** operate on each field **********************/ public int calculateNumMetaBytes(){ int numBytes = 0; @@ -339,7 +339,7 @@ public class PrefixTreeBlockMeta { position += UVIntTool.numBytes(numUniqueTags); } - //TODO method that can read directly from ByteBuffer instead of InputStream + //TODO method that can read directly from ByteBuffer instead of InputStream /*************** methods *************************/ http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeArraySearcher.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeArraySearcher.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeArraySearcher.java index 8ea6e85..ec54c2a 100644 --- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeArraySearcher.java +++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeArraySearcher.java @@ -306,7 +306,7 @@ public class PrefixTreeArraySearcher extends PrefixTreeArrayReversibleScanner im } - /****************** complete seek when token mismatch ******************/ + /****************** complete seek when token mismatch ******************/ /** * @param searcherIsAfterInputKey <0: input key is before the searcher's position
http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/column/ColumnSectionWriter.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/column/ColumnSectionWriter.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/column/ColumnSectionWriter.java index 4793335..3ceae63 100644 --- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/column/ColumnSectionWriter.java +++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/column/ColumnSectionWriter.java @@ -71,7 +71,7 @@ public class ColumnSectionWriter { private List outputArrayOffsets; - /*********************** construct *********************/ + /*********************** construct *********************/ public ColumnSectionWriter() { this.nonLeaves = Lists.newArrayList(); @@ -100,7 +100,7 @@ public class ColumnSectionWriter { } - /****************** methods *******************************/ + /****************** methods *******************************/ public ColumnSectionWriter compile() { if (this.nodeType == ColumnNodeType.FAMILY) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java index c6ae347..3291d72 100644 --- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java +++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java @@ -75,7 +75,7 @@ public class LongEncoder { } - /************* methods ***************************/ + /************* methods ***************************/ public void add(long timestamp) { uniqueValues.add(timestamp); @@ -158,7 +158,7 @@ public class LongEncoder { } - /******************** get/set **************************/ + /******************** get/set **************************/ public long getMin() { return min; http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java index cf13add..75a11ad 100644 --- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java +++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java @@ -179,7 +179,7 @@ public class Tokenizer{ } - /********************** write ***************************/ + /********************** write ***************************/ public Tokenizer setNodeFirstInsertionIndexes() { root.setInsertionIndexes(0); http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerNode.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerNode.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerNode.java index eaf8ab7..e51d5be 100644 --- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerNode.java +++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerNode.java @@ -289,7 +289,7 @@ public class TokenizerNode{ } - /************************ byte[] utils *************************/ + /************************ byte[] utils *************************/ protected boolean partiallyMatchesToken(ByteRange bytes) { return numIdenticalBytes(bytes) > 0; @@ -304,7 +304,7 @@ public class TokenizerNode{ } - /***************** moving nodes around ************************/ + /***************** moving nodes around ************************/ public void appendNodesToExternalList(List appendTo, boolean includeNonLeaves, boolean includeLeaves) { @@ -462,7 +462,7 @@ public class TokenizerNode{ } - /********************** count different node types ********************/ + /********************** count different node types ********************/ public int getNumBranchNodesIncludingThisNode() { if (isLeaf()) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerRowSearchPosition.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerRowSearchPosition.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerRowSearchPosition.java index 6054a46..1166baa 100644 --- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerRowSearchPosition.java +++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerRowSearchPosition.java @@ -30,9 +30,9 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience; @InterfaceAudience.Private public enum TokenizerRowSearchPosition { - AFTER,//the key is after this tree node, so keep searching - BEFORE,//in a binary search, this tells us to back up - MATCH,//the current node is a full match - NO_MATCH,//might as well return a value more informative than null + AFTER,//the key is after this tree node, so keep searching + BEFORE,//in a binary search, this tells us to back up + MATCH,//the current node is a full match + NO_MATCH,//might as well return a value more informative than null } http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/vint/UFIntTool.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/vint/UFIntTool.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/vint/UFIntTool.java index 8b63fd1..fc7c107 100644 --- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/vint/UFIntTool.java +++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/vint/UFIntTool.java @@ -78,13 +78,13 @@ public class UFIntTool { private static final long[] MASKS = new long[] { (long) 255, - (long) 255 << 8, - (long) 255 << 16, - (long) 255 << 24, - (long) 255 << 32, - (long) 255 << 40, - (long) 255 << 48, - (long) 255 << 56 + (long) 255 << 8, + (long) 255 << 16, + (long) 255 << 24, + (long) 255 << 32, + (long) 255 << 40, + (long) 255 << 48, + (long) 255 << 56 }; public static void writeBytes(int outputWidth, final long value, OutputStream os) throws IOException { http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestPrefixTreeSearcher.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestPrefixTreeSearcher.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestPrefixTreeSearcher.java index 6a9e52c..945db32 100644 --- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestPrefixTreeSearcher.java +++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestPrefixTreeSearcher.java @@ -47,7 +47,7 @@ import org.junit.runners.Parameterized.Parameters; @RunWith(Parameterized.class) public class TestPrefixTreeSearcher { - protected static int BLOCK_START = 7; + protected static int BLOCK_START = 7; @Parameters public static Collection parameters() { http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestRowEncoder.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestRowEncoder.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestRowEncoder.java index 74d5cfc..a1c5a25 100644 --- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestRowEncoder.java +++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestRowEncoder.java @@ -76,7 +76,7 @@ public class TestRowEncoder { this.rows = testRows; } - @Before + @Before public void compile() throws IOException { // Always run with tags. But should also ensure that KVs without tags work fine os = new ByteArrayOutputStream(1 << 20); @@ -174,7 +174,7 @@ public class TestRowEncoder { } - /**************** helper **************************/ + /**************** helper **************************/ protected void assertKeyAndValueEqual(Cell expected, Cell actual) { // assert keys are equal (doesn't compare values) http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataComplexQualifiers.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataComplexQualifiers.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataComplexQualifiers.java index bd6f02b..66fe3f3 100644 --- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataComplexQualifiers.java +++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataComplexQualifiers.java @@ -29,7 +29,7 @@ import com.google.common.collect.Lists; public class TestRowDataComplexQualifiers extends BaseTestRowData{ - static byte[] + static byte[] Arow = Bytes.toBytes("Arow"), cf = PrefixTreeTestConstants.TEST_CF, v0 = Bytes.toBytes("v0"); http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDeeper.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDeeper.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDeeper.java index cb3913c..4d05742 100644 --- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDeeper.java +++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDeeper.java @@ -36,36 +36,36 @@ import com.google.common.collect.Lists; */ public class TestRowDataDeeper extends BaseTestRowData{ - static byte[] - cdc = Bytes.toBytes("cdc"), - cf6 = Bytes.toBytes("cf6"), - cfc = Bytes.toBytes("cfc"), - f = Bytes.toBytes("f"), - q = Bytes.toBytes("q"), - v = Bytes.toBytes("v"); + static byte[] + cdc = Bytes.toBytes("cdc"), + cf6 = Bytes.toBytes("cf6"), + cfc = Bytes.toBytes("cfc"), + f = Bytes.toBytes("f"), + q = Bytes.toBytes("q"), + v = Bytes.toBytes("v"); - static long - ts = 55L; + static long + ts = 55L; - static List d = Lists.newArrayList(); - static{ - d.add(new KeyValue(cdc, f, q, ts, v)); + static List d = Lists.newArrayList(); + static{ + d.add(new KeyValue(cdc, f, q, ts, v)); d.add(new KeyValue(cf6, f, q, ts, v)); d.add(new KeyValue(cfc, f, q, ts, v)); - } + } - @Override - public List getInputs() { - return d; - } + @Override + public List getInputs() { + return d; + } - @Override - public void individualBlockMetaAssertions(PrefixTreeBlockMeta blockMeta) { - //0: token:c; fan:d,f - //1: token:f; fan:6,c - //2: leaves - Assert.assertEquals(3, blockMeta.getRowTreeDepth()); - } + @Override + public void individualBlockMetaAssertions(PrefixTreeBlockMeta blockMeta) { + //0: token:c; fan:d,f + //1: token:f; fan:6,c + //2: leaves + Assert.assertEquals(3, blockMeta.getRowTreeDepth()); + } @Override public void individualSearcherAssertions(CellSearcher searcher) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDifferentTimestamps.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDifferentTimestamps.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDifferentTimestamps.java index 2668f2a..8639e8f 100644 --- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDifferentTimestamps.java +++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDifferentTimestamps.java @@ -33,62 +33,62 @@ import com.google.common.collect.Lists; */ public class TestRowDataDifferentTimestamps extends BaseTestRowData{ - static byte[] - Arow = Bytes.toBytes("Arow"), - Brow = Bytes.toBytes("Brow"), - cf = Bytes.toBytes("fammy"), - cq0 = Bytes.toBytes("cq0"), - cq1 = Bytes.toBytes("cq1"), - v0 = Bytes.toBytes("v0"); - - static List d = Lists.newArrayList(); - static{ - KeyValue kv0 = new KeyValue(Arow, cf, cq0, 0L, v0); - kv0.setSequenceId(123456789L); - d.add(kv0); - - KeyValue kv1 = new KeyValue(Arow, cf, cq1, 1L, v0); + static byte[] + Arow = Bytes.toBytes("Arow"), + Brow = Bytes.toBytes("Brow"), + cf = Bytes.toBytes("fammy"), + cq0 = Bytes.toBytes("cq0"), + cq1 = Bytes.toBytes("cq1"), + v0 = Bytes.toBytes("v0"); + + static List d = Lists.newArrayList(); + static{ + KeyValue kv0 = new KeyValue(Arow, cf, cq0, 0L, v0); + kv0.setSequenceId(123456789L); + d.add(kv0); + + KeyValue kv1 = new KeyValue(Arow, cf, cq1, 1L, v0); kv1.setSequenceId(3L); d.add(kv1); - KeyValue kv2 = new KeyValue(Brow, cf, cq0, 12345678L, v0); + KeyValue kv2 = new KeyValue(Brow, cf, cq0, 12345678L, v0); kv2.setSequenceId(65537L); d.add(kv2); - //watch out... Long.MAX_VALUE comes back as 1332221664203, even with other encoders -// d.add(new KeyValue(Brow, cf, cq1, Long.MAX_VALUE, v0)); - KeyValue kv3 = new KeyValue(Brow, cf, cq1, Long.MAX_VALUE-1, v0); + //watch out... Long.MAX_VALUE comes back as 1332221664203, even with other encoders + //d.add(new KeyValue(Brow, cf, cq1, Long.MAX_VALUE, v0)); + KeyValue kv3 = new KeyValue(Brow, cf, cq1, Long.MAX_VALUE-1, v0); kv3.setSequenceId(1L); d.add(kv3); - KeyValue kv4 = new KeyValue(Brow, cf, cq1, 999999999, v0); + KeyValue kv4 = new KeyValue(Brow, cf, cq1, 999999999, v0); //don't set memstoreTS - d.add(kv4); + d.add(kv4); - KeyValue kv5 = new KeyValue(Brow, cf, cq1, 12345, v0); + KeyValue kv5 = new KeyValue(Brow, cf, cq1, 12345, v0); kv5.setSequenceId(0L); d.add(kv5); - } + } - @Override - public List getInputs() { - return d; - } + @Override + public List getInputs() { + return d; + } - @Override - public void individualBlockMetaAssertions(PrefixTreeBlockMeta blockMeta) { - Assert.assertTrue(blockMeta.getNumMvccVersionBytes() > 0); - Assert.assertEquals(12, blockMeta.getNumValueBytes()); + @Override + public void individualBlockMetaAssertions(PrefixTreeBlockMeta blockMeta) { + Assert.assertTrue(blockMeta.getNumMvccVersionBytes() > 0); + Assert.assertEquals(12, blockMeta.getNumValueBytes()); - Assert.assertFalse(blockMeta.isAllSameTimestamp()); - Assert.assertNotNull(blockMeta.getMinTimestamp()); - Assert.assertTrue(blockMeta.getTimestampIndexWidth() > 0); - Assert.assertTrue(blockMeta.getTimestampDeltaWidth() > 0); + Assert.assertFalse(blockMeta.isAllSameTimestamp()); + Assert.assertNotNull(blockMeta.getMinTimestamp()); + Assert.assertTrue(blockMeta.getTimestampIndexWidth() > 0); + Assert.assertTrue(blockMeta.getTimestampDeltaWidth() > 0); Assert.assertFalse(blockMeta.isAllSameMvccVersion()); Assert.assertNotNull(blockMeta.getMinMvccVersion()); Assert.assertTrue(blockMeta.getMvccVersionIndexWidth() > 0); Assert.assertTrue(blockMeta.getMvccVersionDeltaWidth() > 0); - } + } } http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java index 184f537..c49db13 100644 --- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java +++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java @@ -42,49 +42,49 @@ import com.google.common.collect.Lists; public class TestRowDataExerciseFInts extends BaseTestRowData{ static List rows; - static{ - List rowStrings = new ArrayList(); - rowStrings.add("com.edsBlog/directoryAa/pageAaa"); - rowStrings.add("com.edsBlog/directoryAa/pageBbb"); - rowStrings.add("com.edsBlog/directoryAa/pageCcc"); - rowStrings.add("com.edsBlog/directoryAa/pageDdd"); - rowStrings.add("com.edsBlog/directoryBb/pageEee"); - rowStrings.add("com.edsBlog/directoryBb/pageFff"); - rowStrings.add("com.edsBlog/directoryBb/pageGgg"); - rowStrings.add("com.edsBlog/directoryBb/pageHhh"); - rowStrings.add("com.isabellasBlog/directoryAa/pageAaa"); - rowStrings.add("com.isabellasBlog/directoryAa/pageBbb"); - rowStrings.add("com.isabellasBlog/directoryAa/pageCcc"); - rowStrings.add("com.isabellasBlog/directoryAa/pageDdd"); - rowStrings.add("com.isabellasBlog/directoryBb/pageEee"); - rowStrings.add("com.isabellasBlog/directoryBb/pageFff"); - rowStrings.add("com.isabellasBlog/directoryBb/pageGgg"); - rowStrings.add("com.isabellasBlog/directoryBb/pageHhh"); - ByteRangeTreeSet ba = new ByteRangeTreeSet(); - for(String row : rowStrings){ - ba.add(new SimpleMutableByteRange(Bytes.toBytes(row))); - } - rows = ba.compile().getSortedRanges(); - } + static{ + List rowStrings = new ArrayList(); + rowStrings.add("com.edsBlog/directoryAa/pageAaa"); + rowStrings.add("com.edsBlog/directoryAa/pageBbb"); + rowStrings.add("com.edsBlog/directoryAa/pageCcc"); + rowStrings.add("com.edsBlog/directoryAa/pageDdd"); + rowStrings.add("com.edsBlog/directoryBb/pageEee"); + rowStrings.add("com.edsBlog/directoryBb/pageFff"); + rowStrings.add("com.edsBlog/directoryBb/pageGgg"); + rowStrings.add("com.edsBlog/directoryBb/pageHhh"); + rowStrings.add("com.isabellasBlog/directoryAa/pageAaa"); + rowStrings.add("com.isabellasBlog/directoryAa/pageBbb"); + rowStrings.add("com.isabellasBlog/directoryAa/pageCcc"); + rowStrings.add("com.isabellasBlog/directoryAa/pageDdd"); + rowStrings.add("com.isabellasBlog/directoryBb/pageEee"); + rowStrings.add("com.isabellasBlog/directoryBb/pageFff"); + rowStrings.add("com.isabellasBlog/directoryBb/pageGgg"); + rowStrings.add("com.isabellasBlog/directoryBb/pageHhh"); + ByteRangeTreeSet ba = new ByteRangeTreeSet(); + for(String row : rowStrings){ + ba.add(new SimpleMutableByteRange(Bytes.toBytes(row))); + } + rows = ba.compile().getSortedRanges(); + } - static List cols = Lists.newArrayList(); - static{ - cols.add("Chrome"); - cols.add("Chromeb"); - cols.add("Firefox"); - cols.add("InternetExplorer"); - cols.add("Opera"); - cols.add("Safari"); - cols.add("Z1stBrowserWithHuuuuuuuuuuuugeQualifier"); - cols.add("Z2ndBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); - cols.add("Z3rdBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); - cols.add("Z4thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); - cols.add("Z5thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); - cols.add("Z6thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); - cols.add("Z7thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); - cols.add("Z8thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); - cols.add("Z9thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); - } + static List cols = Lists.newArrayList(); + static{ + cols.add("Chrome"); + cols.add("Chromeb"); + cols.add("Firefox"); + cols.add("InternetExplorer"); + cols.add("Opera"); + cols.add("Safari"); + cols.add("Z1stBrowserWithHuuuuuuuuuuuugeQualifier"); + cols.add("Z2ndBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); + cols.add("Z3rdBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); + cols.add("Z4thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); + cols.add("Z5thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); + cols.add("Z6thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); + cols.add("Z7thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); + cols.add("Z8thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); + cols.add("Z9thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore"); + } static long ts = 1234567890; http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataNub.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataNub.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataNub.java index ad19cd4..a818f5e 100644 --- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataNub.java +++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataNub.java @@ -29,31 +29,31 @@ import com.google.common.collect.Lists; public class TestRowDataNub extends BaseTestRowData{ - static byte[] - rowA = Bytes.toBytes("rowA"), - rowB = Bytes.toBytes("rowB"),//nub - rowBB = Bytes.toBytes("rowBB"), - cf = PrefixTreeTestConstants.TEST_CF, - cq0 = Bytes.toBytes("cq0"), - cq1 = Bytes.toBytes("cq1"), - v0 = Bytes.toBytes("v0"); - - static long - ts = 55L; - - static List d = Lists.newArrayList(); - static{ - d.add(new KeyValue(rowA, cf, cq0, ts, v0)); - d.add(new KeyValue(rowA, cf, cq1, ts, v0)); - d.add(new KeyValue(rowB, cf, cq0, ts, v0)); - d.add(new KeyValue(rowB, cf, cq1, ts, v0)); - d.add(new KeyValue(rowBB, cf, cq0, ts, v0)); - d.add(new KeyValue(rowBB, cf, cq1, ts, v0)); - } - - @Override - public List getInputs() { - return d; - } + static byte[] + rowA = Bytes.toBytes("rowA"), + rowB = Bytes.toBytes("rowB"),//nub + rowBB = Bytes.toBytes("rowBB"), + cf = PrefixTreeTestConstants.TEST_CF, + cq0 = Bytes.toBytes("cq0"), + cq1 = Bytes.toBytes("cq1"), + v0 = Bytes.toBytes("v0"); + + static long + ts = 55L; + + static List d = Lists.newArrayList(); + static{ + d.add(new KeyValue(rowA, cf, cq0, ts, v0)); + d.add(new KeyValue(rowA, cf, cq1, ts, v0)); + d.add(new KeyValue(rowB, cf, cq0, ts, v0)); + d.add(new KeyValue(rowB, cf, cq1, ts, v0)); + d.add(new KeyValue(rowBB, cf, cq0, ts, v0)); + d.add(new KeyValue(rowBB, cf, cq1, ts, v0)); + } + + @Override + public List getInputs() { + return d; + } } http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataQualifierByteOrdering.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataQualifierByteOrdering.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataQualifierByteOrdering.java index a8c4646..6d3918f 100644 --- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataQualifierByteOrdering.java +++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataQualifierByteOrdering.java @@ -28,15 +28,15 @@ import com.google.common.collect.Lists; public class TestRowDataQualifierByteOrdering extends BaseTestRowData{ - static byte[] - Arow = Bytes.toBytes("Arow"), - Brow = Bytes.toBytes("Brow"), - Brow2 = Bytes.toBytes("Brow2"), - fam = Bytes.toBytes("HappyFam"), - cq0 = Bytes.toBytes("cq0"), - cq1 = Bytes.toBytes("cq1tail"),//make sure tail does not come back as liat - cq2 = Bytes.toBytes("cq2"), - v0 = Bytes.toBytes("v0"); + static byte[] + Arow = Bytes.toBytes("Arow"), + Brow = Bytes.toBytes("Brow"), + Brow2 = Bytes.toBytes("Brow2"), + fam = Bytes.toBytes("HappyFam"), + cq0 = Bytes.toBytes("cq0"), + cq1 = Bytes.toBytes("cq1tail"),//make sure tail does not come back as liat + cq2 = Bytes.toBytes("cq2"), + v0 = Bytes.toBytes("v0"); static long ts = 55L; http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSearcherRowMiss.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSearcherRowMiss.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSearcherRowMiss.java index 5f319fc..edea305 100644 --- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSearcherRowMiss.java +++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSearcherRowMiss.java @@ -55,13 +55,13 @@ public class TestRowDataSearcherRowMiss extends BaseTestRowData{ d.add(new KeyValue(B, cf, cq, ts, v)); } - @Override - public List getInputs() { - return d; - } + @Override + public List getInputs() { + return d; + } - @Override - public void individualSearcherAssertions(CellSearcher searcher) { + @Override + public void individualSearcherAssertions(CellSearcher searcher) { assertRowOffsetsCorrect(); searcher.resetToBeforeFirstEntry(); @@ -83,13 +83,13 @@ public class TestRowDataSearcherRowMiss extends BaseTestRowData{ testBetween2and3(searcher); } - /************ private methods, call from above *******************/ + /************ private methods, call from above *******************/ - private void assertRowOffsetsCorrect(){ - Assert.assertEquals(4, getRowStartIndexes().size()); - } + private void assertRowOffsetsCorrect(){ + Assert.assertEquals(4, getRowStartIndexes().size()); + } - private void testBetween1and2(CellSearcher searcher){ + private void testBetween1and2(CellSearcher searcher){ CellScannerPosition p;//reuse Cell betweenAAndAAA = new KeyValue(AA, cf, cq, ts-2, v); @@ -105,7 +105,7 @@ public class TestRowDataSearcherRowMiss extends BaseTestRowData{ p = searcher.positionAtOrAfter(betweenAAndAAA); Assert.assertEquals(CellScannerPosition.AFTER, p); Assert.assertTrue(CellComparator.equals(searcher.current(), d.get(2))); - } + } private void testBetween2and3(CellSearcher searcher){ CellScannerPosition p;//reuse http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSingleQualifier.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSingleQualifier.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSingleQualifier.java index 9944057..35ba2c1 100644 --- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSingleQualifier.java +++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSingleQualifier.java @@ -29,12 +29,12 @@ import com.google.common.collect.Lists; public class TestRowDataSingleQualifier extends BaseTestRowData{ - static byte[] - rowA = Bytes.toBytes("rowA"), - rowB = Bytes.toBytes("rowB"), - cf = PrefixTreeTestConstants.TEST_CF, - cq0 = Bytes.toBytes("cq0"), - v0 = Bytes.toBytes("v0"); + static byte[] + rowA = Bytes.toBytes("rowA"), + rowB = Bytes.toBytes("rowB"), + cf = PrefixTreeTestConstants.TEST_CF, + cq0 = Bytes.toBytes("cq0"), + v0 = Bytes.toBytes("v0"); static long ts = 55L; http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivial.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivial.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivial.java index df785fa..2a53976 100644 --- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivial.java +++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivial.java @@ -33,12 +33,12 @@ import com.google.common.collect.Lists; public class TestRowDataTrivial extends BaseTestRowData{ - static byte[] - rA = Bytes.toBytes("rA"), - rB = Bytes.toBytes("rB"),//turn "r" into a branch for the Searcher tests - cf = Bytes.toBytes("fam"), - cq0 = Bytes.toBytes("q0"), - v0 = Bytes.toBytes("v0"); + static byte[] + rA = Bytes.toBytes("rA"), + rB = Bytes.toBytes("rB"),//turn "r" into a branch for the Searcher tests + cf = Bytes.toBytes("fam"), + cq0 = Bytes.toBytes("q0"), + v0 = Bytes.toBytes("v0"); static long ts = 55L; http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java index d0801c1..74ac85a 100644 --- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java +++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java @@ -40,7 +40,7 @@ import com.google.common.collect.Lists; public class TestRowDataUrls extends BaseTestRowData{ static List rows; - static{ + static{ List rowStrings = new ArrayList(); rowStrings.add("com.edsBlog/directoryAa/pageAaa"); rowStrings.add("com.edsBlog/directoryAa/pageBbb"); http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrlsExample.java ---------------------------------------------------------------------- diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrlsExample.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrlsExample.java index 543afb6..56d4e85 100644 --- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrlsExample.java +++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrlsExample.java @@ -51,33 +51,33 @@ public class TestRowDataUrlsExample extends BaseTestRowData{ static String FAMILY = "hits"; static List BROWSERS = Lists.newArrayList( "Chrome", "IE8", "IE9beta");//, "Opera", "Safari"); - static long TIMESTAMP = 1234567890; - - static int MAX_VALUE = 50; - - static List kvs = Lists.newArrayList(); - static{ - for(String rowKey : URLS){ - for(String qualifier : BROWSERS){ - KeyValue kv = new KeyValue( - Bytes.toBytes(rowKey), - Bytes.toBytes(FAMILY), - Bytes.toBytes(qualifier), - TIMESTAMP, - KeyValue.Type.Put, - Bytes.toBytes("VvvV")); - kvs.add(kv); - } - } - } - - /** - * Used for generating docs. - */ - public static void main(String... args) throws IOException{ + static long TIMESTAMP = 1234567890; + + static int MAX_VALUE = 50; + + static List kvs = Lists.newArrayList(); + static{ + for(String rowKey : URLS){ + for(String qualifier : BROWSERS){ + KeyValue kv = new KeyValue( + Bytes.toBytes(rowKey), + Bytes.toBytes(FAMILY), + Bytes.toBytes(qualifier), + TIMESTAMP, + KeyValue.Type.Put, + Bytes.toBytes("VvvV")); + kvs.add(kv); + } + } + } + + /** + * Used for generating docs. + */ + public static void main(String... args) throws IOException{ System.out.println("-- inputs --"); System.out.println(KeyValueTestUtil.toStringWithPadding(kvs, true)); - ByteArrayOutputStream os = new ByteArrayOutputStream(1<<20); + ByteArrayOutputStream os = new ByteArrayOutputStream(1<<20); PrefixTreeEncoder encoder = new PrefixTreeEncoder(os, false); for(KeyValue kv : kvs){ @@ -116,11 +116,11 @@ public class TestRowDataUrlsExample extends BaseTestRowData{ System.out.println("-- concatenated values --"); System.out.println(Bytes.toStringBinary(encoder.getValueByteRange().deepCopyToNewArray())); - } + } - @Override - public List getInputs() { - return kvs; - } + @Override + public List getInputs() { + return kvs; + } } http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java ---------------------------------------------------------------------- diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java index 4321a8e..54fc8de 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java @@ -40,33 +40,33 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience; @XmlRootElement(name="ClusterVersion") @InterfaceAudience.Private public class StorageClusterVersionModel implements Serializable { - private static final long serialVersionUID = 1L; + private static final long serialVersionUID = 1L; - private String version; + private String version; - /** - * @return the storage cluster version - */ - @XmlValue - public String getVersion() { - return version; - } + /** + * @return the storage cluster version + */ + @XmlValue + public String getVersion() { + return version; + } - /** - * @param version the storage cluster version - */ - public void setVersion(String version) { - this.version = version; - } + /** + * @param version the storage cluster version + */ + public void setVersion(String version) { + this.version = version; + } - /* (non-Javadoc) - * @see java.lang.Object#toString() - */ + /* (non-Javadoc) + * @see java.lang.Object#toString() + */ @JsonValue - @Override - public String toString() { - return version; - } + @Override + public String toString() { + return version; + } //needed for jackson deserialization private static StorageClusterVersionModel valueOf(String value) { http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java ---------------------------------------------------------------------- diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java index 596adac..2ed4e80 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java @@ -38,67 +38,67 @@ import org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableLis @InterfaceAudience.Private public class TableListModel implements Serializable, ProtobufMessageHandler { - private static final long serialVersionUID = 1L; + private static final long serialVersionUID = 1L; - private List tables = new ArrayList(); + private List tables = new ArrayList(); - /** - * Default constructor - */ - public TableListModel() {} + /** + * Default constructor + */ + public TableListModel() {} - /** - * Add the table name model to the list - * @param table the table model - */ - public void add(TableModel table) { - tables.add(table); - } - - /** - * @param index the index - * @return the table model - */ - public TableModel get(int index) { - return tables.get(index); - } + /** + * Add the table name model to the list + * @param table the table model + */ + public void add(TableModel table) { + tables.add(table); + } + + /** + * @param index the index + * @return the table model + */ + public TableModel get(int index) { + return tables.get(index); + } - /** - * @return the tables - */ - @XmlElementRef(name="table") - public List getTables() { - return tables; - } + /** + * @return the tables + */ + @XmlElementRef(name="table") + public List getTables() { + return tables; + } - /** - * @param tables the tables to set - */ - public void setTables(List tables) { - this.tables = tables; - } + /** + * @param tables the tables to set + */ + public void setTables(List tables) { + this.tables = tables; + } - /* (non-Javadoc) - * @see java.lang.Object#toString() - */ - @Override - public String toString() { - StringBuilder sb = new StringBuilder(); - for(TableModel aTable : tables) { - sb.append(aTable.toString()); - sb.append('\n'); - } - return sb.toString(); - } + /* (non-Javadoc) + * @see java.lang.Object#toString() + */ + @Override + public String toString() { + StringBuilder sb = new StringBuilder(); + for(TableModel aTable : tables) { + sb.append(aTable.toString()); + sb.append('\n'); + } + return sb.toString(); + } - @Override - public byte[] createProtobufOutput() { - TableList.Builder builder = TableList.newBuilder(); - for (TableModel aTable : tables) { - builder.addName(aTable.getName()); - } - return builder.build().toByteArray(); - } + @Override + public byte[] createProtobufOutput() { + TableList.Builder builder = TableList.newBuilder(); + for (TableModel aTable : tables) { + builder.addName(aTable.getName()); + } + return builder.build().toByteArray(); + } @Override public ProtobufMessageHandler getObjectFromMessage(byte[] message) http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java ---------------------------------------------------------------------- diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java index 0fb0d6e..3be0da9 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java @@ -41,44 +41,44 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience; @InterfaceAudience.Private public class TableModel implements Serializable { - private static final long serialVersionUID = 1L; - - private String name; - - /** - * Default constructor - */ - public TableModel() {} + private static final long serialVersionUID = 1L; + + private String name; + + /** + * Default constructor + */ + public TableModel() {} - /** - * Constructor - * @param name - */ - public TableModel(String name) { - super(); - this.name = name; - } + /** + * Constructor + * @param name + */ + public TableModel(String name) { + super(); + this.name = name; + } - /** - * @return the name - */ - @XmlAttribute - public String getName() { - return name; - } + /** + * @return the name + */ + @XmlAttribute + public String getName() { + return name; + } - /** - * @param name the name to set - */ - public void setName(String name) { - this.name = name; - } + /** + * @param name the name to set + */ + public void setName(String name) { + this.name = name; + } - /* (non-Javadoc) - * @see java.lang.Object#toString() - */ - @Override - public String toString() { - return this.name; - } + /* (non-Javadoc) + * @see java.lang.Object#toString() + */ + @Override + public String toString() { + return this.name; + } } http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java ---------------------------------------------------------------------- diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java index 0938803..8b08279 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java @@ -48,9 +48,9 @@ import com.sun.jersey.spi.container.servlet.ServletContainer; @InterfaceAudience.Private public class VersionModel implements Serializable, ProtobufMessageHandler { - private static final long serialVersionUID = 1L; + private static final long serialVersionUID = 1L; - private String restVersion; + private String restVersion; private String jvmVersion; private String osVersion; private String serverVersion; @@ -65,30 +65,30 @@ public class VersionModel implements Serializable, ProtobufMessageHandler { * Constructor * @param context the servlet context */ - public VersionModel(ServletContext context) { - restVersion = RESTServlet.VERSION_STRING; - jvmVersion = System.getProperty("java.vm.vendor") + ' ' + + public VersionModel(ServletContext context) { + restVersion = RESTServlet.VERSION_STRING; + jvmVersion = System.getProperty("java.vm.vendor") + ' ' + System.getProperty("java.version") + '-' + System.getProperty("java.vm.version"); - osVersion = System.getProperty("os.name") + ' ' + + osVersion = System.getProperty("os.name") + ' ' + System.getProperty("os.version") + ' ' + System.getProperty("os.arch"); - serverVersion = context.getServerInfo(); - jerseyVersion = ServletContainer.class.getPackage() + serverVersion = context.getServerInfo(); + jerseyVersion = ServletContainer.class.getPackage() .getImplementationVersion(); - } + } - /** - * @return the REST gateway version - */ - @XmlAttribute(name="REST") - public String getRESTVersion() { + /** + * @return the REST gateway version + */ + @XmlAttribute(name="REST") + public String getRESTVersion() { return restVersion; } - /** - * @return the JVM vendor and version - */ + /** + * @return the JVM vendor and version + */ @XmlAttribute(name="JVM") public String getJVMVersion() { return jvmVersion; @@ -154,34 +154,34 @@ public class VersionModel implements Serializable, ProtobufMessageHandler { } /* (non-Javadoc) - * @see java.lang.Object#toString() - */ - @Override - public String toString() { - StringBuilder sb = new StringBuilder(); - sb.append("rest "); - sb.append(restVersion); - sb.append(" [JVM: "); - sb.append(jvmVersion); - sb.append("] [OS: "); - sb.append(osVersion); - sb.append("] [Server: "); - sb.append(serverVersion); - sb.append("] [Jersey: "); + * @see java.lang.Object#toString() + */ + @Override + public String toString() { + StringBuilder sb = new StringBuilder(); + sb.append("rest "); + sb.append(restVersion); + sb.append(" [JVM: "); + sb.append(jvmVersion); + sb.append("] [OS: "); + sb.append(osVersion); + sb.append("] [Server: "); + sb.append(serverVersion); + sb.append("] [Jersey: "); sb.append(jerseyVersion); - sb.append("]\n"); - return sb.toString(); - } + sb.append("]\n"); + return sb.toString(); + } - @Override + @Override public byte[] createProtobufOutput() { - Version.Builder builder = Version.newBuilder(); - builder.setRestVersion(restVersion); - builder.setJvmVersion(jvmVersion); - builder.setOsVersion(osVersion); - builder.setServerVersion(serverVersion); - builder.setJerseyVersion(jerseyVersion); - return builder.build().toByteArray(); + Version.Builder builder = Version.newBuilder(); + builder.setRestVersion(restVersion); + builder.setJvmVersion(jvmVersion); + builder.setOsVersion(osVersion); + builder.setServerVersion(serverVersion); + builder.setJerseyVersion(jerseyVersion); + return builder.build().toByteArray(); } @Override http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java ---------------------------------------------------------------------- diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java index 3b8b8ca..216f830 100644 --- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java +++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java @@ -55,12 +55,12 @@ import com.sun.jersey.api.json.JSONJAXBContext; @InterfaceAudience.Private public class JAXBContextResolver implements ContextResolver { - private final JAXBContext context; + private final JAXBContext context; - private final Set> types; + private final Set> types; - private final Class[] cTypes = { - CellModel.class, + private final Class[] cTypes = { + CellModel.class, CellSetModel.class, ColumnSchemaModel.class, RowModel.class, @@ -68,22 +68,22 @@ public class JAXBContextResolver implements ContextResolver { StorageClusterStatusModel.class, StorageClusterVersionModel.class, TableInfoModel.class, - TableListModel.class, - TableModel.class, - TableRegionModel.class, - TableSchemaModel.class, - VersionModel.class - }; + TableListModel.class, + TableModel.class, + TableRegionModel.class, + TableSchemaModel.class, + VersionModel.class + }; - @SuppressWarnings("unchecked") + @SuppressWarnings("unchecked") public JAXBContextResolver() throws Exception { - this.types = new HashSet(Arrays.asList(cTypes)); - this.context = new JSONJAXBContext(JSONConfiguration.natural().build(), - cTypes); - } + this.types = new HashSet(Arrays.asList(cTypes)); + this.context = new JSONJAXBContext(JSONConfiguration.natural().build(), + cTypes); + } - @Override - public JAXBContext getContext(Class objectType) { - return (types.contains(objectType)) ? context : null; + @Override + public JAXBContext getContext(Class objectType) { + return (types.contains(objectType)) ? context : null; } }