hbase-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From st...@apache.org
Subject [2/2] hbase git commit: HBASE-12519 Remove tabs used as whitespace (Varun Saxena)
Date Sat, 29 Nov 2014 17:24:21 GMT
HBASE-12519 Remove tabs used as whitespace (Varun Saxena)


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/b12d5778
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/b12d5778
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/b12d5778

Branch: refs/heads/branch-1
Commit: b12d57783fe70102fcaf896a635c945c4c2cd97b
Parents: 7eefa36
Author: stack <stack@apache.org>
Authored: Sat Nov 29 09:24:08 2014 -0800
Committer: stack <stack@apache.org>
Committed: Sat Nov 29 09:24:08 2014 -0800

----------------------------------------------------------------------
 ...ExcludePrivateAnnotationsStandardDoclet.java |   2 +-
 .../classification/tools/RootDocProcessor.java  | 184 ++++----
 .../classification/tools/StabilityOptions.java  |  15 +-
 .../apache/hadoop/hbase/HTableDescriptor.java   |   2 +-
 .../org/apache/hadoop/hbase/client/HTable.java  |   4 +-
 .../apache/hadoop/hbase/client/HTableUtil.java  |   2 +-
 .../client/UnmodifyableHTableDescriptor.java    |   2 +-
 .../client/coprocessor/AggregationClient.java   |   2 +-
 .../hbase/filter/DependentColumnFilter.java     |  16 +-
 .../hadoop/hbase/protobuf/RequestConverter.java |   2 +-
 .../apache/hadoop/hbase/zookeeper/ZKAssign.java |   4 +-
 .../apache/hadoop/hbase/HBaseConfiguration.java |   2 +-
 .../org/apache/hadoop/hbase/io/TimeRange.java   |  12 +-
 .../hadoop/hbase/util/CollectionUtils.java      |   6 +-
 .../apache/hadoop/hbase/util/JenkinsHash.java   |   2 +-
 .../hadoop/hbase/TestHBaseConfiguration.java    |   4 +-
 ...egrationTestBigLinkedListWithVisibility.java |   4 +-
 .../codec/prefixtree/PrefixTreeBlockMeta.java   |   6 +-
 .../decode/PrefixTreeArraySearcher.java         |   2 +-
 .../encode/column/ColumnSectionWriter.java      |   4 +-
 .../prefixtree/encode/other/LongEncoder.java    |   4 +-
 .../prefixtree/encode/tokenize/Tokenizer.java   |   2 +-
 .../encode/tokenize/TokenizerNode.java          |   6 +-
 .../tokenize/TokenizerRowSearchPosition.java    |   8 +-
 .../hadoop/hbase/util/vint/UFIntTool.java       |  14 +-
 .../prefixtree/row/TestPrefixTreeSearcher.java  |   2 +-
 .../codec/prefixtree/row/TestRowEncoder.java    |   4 +-
 .../row/data/TestRowDataComplexQualifiers.java  |   2 +-
 .../prefixtree/row/data/TestRowDataDeeper.java  |  48 +--
 .../data/TestRowDataDifferentTimestamps.java    |  72 ++--
 .../row/data/TestRowDataExerciseFInts.java      |  84 ++--
 .../prefixtree/row/data/TestRowDataNub.java     |  52 +--
 .../data/TestRowDataQualifierByteOrdering.java  |  18 +-
 .../row/data/TestRowDataSearcherRowMiss.java    |  24 +-
 .../row/data/TestRowDataSingleQualifier.java    |  12 +-
 .../prefixtree/row/data/TestRowDataTrivial.java |  12 +-
 .../prefixtree/row/data/TestRowDataUrls.java    |   2 +-
 .../row/data/TestRowDataUrlsExample.java        |  60 +--
 .../rest/model/StorageClusterVersionModel.java  |  44 +-
 .../hadoop/hbase/rest/model/TableListModel.java | 108 ++---
 .../hadoop/hbase/rest/model/TableModel.java     |  72 ++--
 .../hadoop/hbase/rest/model/VersionModel.java   |  84 ++--
 .../rest/provider/JAXBContextResolver.java      |  36 +-
 .../producer/PlainTextMessageBodyProducer.java  |  24 +-
 .../producer/ProtobufMessageBodyProducer.java   |  44 +-
 .../hadoop/hbase/ipc/HBaseRPCErrorHandler.java  |  10 +-
 .../hbase/master/RegionPlacementMaintainer.java |   2 +-
 .../SnapshotOfRegionAssignmentFromMeta.java     |   2 +-
 .../hbase/regionserver/HRegionServer.java       |   2 +-
 .../hbase/regionserver/SplitLogWorker.java      |   2 +-
 .../handler/HLogSplitterHandler.java            |   2 +-
 .../hadoop/hbase/regionserver/wal/FSHLog.java   |   2 +-
 .../org/apache/hadoop/hbase/util/HBaseFsck.java |   4 +-
 .../hbase/zookeeper/RegionServerTracker.java    |   2 +-
 .../hadoop/hbase/HBaseTestingUtility.java       |   4 +-
 .../hadoop/hbase/client/TestHTablePool.java     | 426 +++++++++----------
 .../TestBigDecimalColumnInterpreter.java        |  68 +--
 .../hbase/filter/TestDependentColumnFilter.java |  24 +-
 .../hbase/io/hfile/TestHFilePerformance.java    | 122 +++---
 .../hbase/master/TestClockSkewDetection.java    |   2 +-
 .../hbase/regionserver/TestMajorCompaction.java |   2 +-
 .../TestSplitTransactionOnCluster.java          |   2 +-
 .../hadoop/hbase/regionserver/TestTags.java     |  10 +-
 .../replication/TestReplicationSmallTests.java  |   2 +-
 .../apache/hadoop/hbase/util/LoadTestTool.java  |   2 +-
 .../hadoop/hbase/util/MultiThreadedReader.java  |   6 +-
 .../hadoop/hbase/util/MultiThreadedUpdater.java |   2 +-
 67 files changed, 897 insertions(+), 918 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/ExcludePrivateAnnotationsStandardDoclet.java
----------------------------------------------------------------------
diff --git a/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/ExcludePrivateAnnotationsStandardDoclet.java b/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/ExcludePrivateAnnotationsStandardDoclet.java
index eb03a21..f93e13f 100644
--- a/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/ExcludePrivateAnnotationsStandardDoclet.java
+++ b/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/ExcludePrivateAnnotationsStandardDoclet.java
@@ -39,7 +39,7 @@ public class ExcludePrivateAnnotationsStandardDoclet {
 
   public static boolean start(RootDoc root) {
     System.out.println(
-	ExcludePrivateAnnotationsStandardDoclet.class.getSimpleName());
+      ExcludePrivateAnnotationsStandardDoclet.class.getSimpleName());
     return Standard.start(RootDocProcessor.process(root));
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/RootDocProcessor.java
----------------------------------------------------------------------
diff --git a/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/RootDocProcessor.java b/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/RootDocProcessor.java
index 78e8163..c6fb74a 100644
--- a/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/RootDocProcessor.java
+++ b/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/RootDocProcessor.java
@@ -65,10 +65,10 @@ class RootDocProcessor {
       return getProxy(obj);
     } else if (obj instanceof Object[]) {
       Class<?> componentType = type.isArray() ? type.getComponentType()
-	  : cls.getComponentType();
+        : cls.getComponentType();
       Object[] array = (Object[]) obj;
       Object[] newArray = (Object[]) Array.newInstance(componentType,
-	  array.length);
+        array.length);
       for (int i = 0; i < array.length; ++i) {
         newArray[i] = process(array[i], componentType);
       }
@@ -98,116 +98,99 @@ class RootDocProcessor {
     }
 
     @Override
-    public Object invoke(Object proxy, Method method, Object[] args)
-	throws Throwable {
+    public Object invoke(Object proxy, Method method, Object[] args) throws Throwable {
       String methodName = method.getName();
       if (target instanceof Doc) {
-	if (methodName.equals("isIncluded")) {
-	  Doc doc = (Doc) target;
-	  return !exclude(doc) && doc.isIncluded();
-	}
-	if (target instanceof RootDoc) {
-	  if (methodName.equals("classes")) {
-	    return filter(((RootDoc) target).classes(), ClassDoc.class);
-	  } else if (methodName.equals("specifiedClasses")) {
-	    return filter(((RootDoc) target).specifiedClasses(), ClassDoc.class);
-	  } else if (methodName.equals("specifiedPackages")) {
-	    return filter(((RootDoc) target).specifiedPackages(), PackageDoc.class);
-	  }
-	} else if (target instanceof ClassDoc) {
-	  if (isFiltered(args)) {
-	    if (methodName.equals("methods")) {
-	      return filter(((ClassDoc) target).methods(true), MethodDoc.class);
-	    } else if (methodName.equals("fields")) {
-	      return filter(((ClassDoc) target).fields(true), FieldDoc.class);
-	    } else if (methodName.equals("innerClasses")) {
-	      return filter(((ClassDoc) target).innerClasses(true),
-		  ClassDoc.class);
-	    } else if (methodName.equals("constructors")) {
-	      return filter(((ClassDoc) target).constructors(true),
-		  ConstructorDoc.class);
-	    }
-	  }
-	} else if (target instanceof PackageDoc) {
-	  if (methodName.equals("allClasses")) {
-	    if (isFiltered(args)) {
-	      return filter(((PackageDoc) target).allClasses(true),
-		ClassDoc.class);
-	    } else {
-	      return filter(((PackageDoc) target).allClasses(), ClassDoc.class);
-	    }
-	  } else if (methodName.equals("annotationTypes")) {
-	    return filter(((PackageDoc) target).annotationTypes(),
-		AnnotationTypeDoc.class);
-	  } else if (methodName.equals("enums")) {
-	    return filter(((PackageDoc) target).enums(),
-		ClassDoc.class);
-	  } else if (methodName.equals("errors")) {
-	    return filter(((PackageDoc) target).errors(),
-		ClassDoc.class);
-	  } else if (methodName.equals("exceptions")) {
-	    return filter(((PackageDoc) target).exceptions(),
-		ClassDoc.class);
-	  } else if (methodName.equals("interfaces")) {
-	    return filter(((PackageDoc) target).interfaces(),
-		ClassDoc.class);
-	  } else if (methodName.equals("ordinaryClasses")) {
-	    return filter(((PackageDoc) target).ordinaryClasses(),
-		ClassDoc.class);
-	  }
-	}
+        if (methodName.equals("isIncluded")) {
+          Doc doc = (Doc) target;
+          return !exclude(doc) && doc.isIncluded();
+        }
+        if (target instanceof RootDoc) {
+          if (methodName.equals("classes")) {
+            return filter(((RootDoc) target).classes(), ClassDoc.class);
+          } else if (methodName.equals("specifiedClasses")) {
+            return filter(((RootDoc) target).specifiedClasses(), ClassDoc.class);
+          } else if (methodName.equals("specifiedPackages")) {
+            return filter(((RootDoc) target).specifiedPackages(), PackageDoc.class);
+          }
+        } else if (target instanceof ClassDoc) {
+          if (isFiltered(args)) {
+            if (methodName.equals("methods")) {
+              return filter(((ClassDoc) target).methods(true), MethodDoc.class);
+            } else if (methodName.equals("fields")) {
+              return filter(((ClassDoc) target).fields(true), FieldDoc.class);
+            } else if (methodName.equals("innerClasses")) {
+              return filter(((ClassDoc) target).innerClasses(true), ClassDoc.class);
+            } else if (methodName.equals("constructors")) {
+              return filter(((ClassDoc) target).constructors(true), ConstructorDoc.class);
+            }
+          }
+        } else if (target instanceof PackageDoc) {
+          if (methodName.equals("allClasses")) {
+            if (isFiltered(args)) {
+              return filter(((PackageDoc) target).allClasses(true), ClassDoc.class);
+            } else {
+              return filter(((PackageDoc) target).allClasses(), ClassDoc.class);
+            }
+          } else if (methodName.equals("annotationTypes")) {
+            return filter(((PackageDoc) target).annotationTypes(), AnnotationTypeDoc.class);
+          } else if (methodName.equals("enums")) {
+            return filter(((PackageDoc) target).enums(), ClassDoc.class);
+          } else if (methodName.equals("errors")) {
+            return filter(((PackageDoc) target).errors(), ClassDoc.class);
+          } else if (methodName.equals("exceptions")) {
+            return filter(((PackageDoc) target).exceptions(), ClassDoc.class);
+          } else if (methodName.equals("interfaces")) {
+            return filter(((PackageDoc) target).interfaces(), ClassDoc.class);
+          } else if (methodName.equals("ordinaryClasses")) {
+            return filter(((PackageDoc) target).ordinaryClasses(), ClassDoc.class);
+          }
+        }
       }
 
       if (args != null) {
-	if (methodName.equals("compareTo") || methodName.equals("equals")
-	    || methodName.equals("overrides")
-	    || methodName.equals("subclassOf")) {
-	  args[0] = unwrap(args[0]);
-	}
+        if (methodName.equals("compareTo") || methodName.equals("equals")
+            || methodName.equals("overrides") || methodName.equals("subclassOf")) {
+          args[0] = unwrap(args[0]);
+        }
       }
       try {
-	return process(method.invoke(target, args), method.getReturnType());
+        return process(method.invoke(target, args), method.getReturnType());
       } catch (InvocationTargetException e) {
-	throw e.getTargetException();
+        throw e.getTargetException();
       }
     }
 
     private static boolean exclude(Doc doc) {
       AnnotationDesc[] annotations = null;
       if (doc instanceof ProgramElementDoc) {
-	annotations = ((ProgramElementDoc) doc).annotations();
+        annotations = ((ProgramElementDoc) doc).annotations();
       } else if (doc instanceof PackageDoc) {
-	annotations = ((PackageDoc) doc).annotations();
+        annotations = ((PackageDoc) doc).annotations();
       }
       if (annotations != null) {
-	for (AnnotationDesc annotation : annotations) {
-	  String qualifiedTypeName = annotation.annotationType().qualifiedTypeName();
-	  if (qualifiedTypeName.equals(
-	        InterfaceAudience.Private.class.getCanonicalName())
-	    || qualifiedTypeName.equals(
-                InterfaceAudience.LimitedPrivate.class.getCanonicalName())) {
-	    return true;
-	  }
-	  if (stability.equals(StabilityOptions.EVOLVING_OPTION)) {
-	    if (qualifiedTypeName.equals(
-		InterfaceStability.Unstable.class.getCanonicalName())) {
-	      return true;
-	    }
-	  }
-	  if (stability.equals(StabilityOptions.STABLE_OPTION)) {
-	    if (qualifiedTypeName.equals(
-		InterfaceStability.Unstable.class.getCanonicalName())
-              || qualifiedTypeName.equals(
-  		InterfaceStability.Evolving.class.getCanonicalName())) {
-	      return true;
-	    }
-	  }
-	}
         for (AnnotationDesc annotation : annotations) {
-          String qualifiedTypeName =
-            annotation.annotationType().qualifiedTypeName();
-          if (qualifiedTypeName.equals(
-              InterfaceAudience.Public.class.getCanonicalName())) {
+          String qualifiedTypeName = annotation.annotationType().qualifiedTypeName();
+          if (qualifiedTypeName.equals(InterfaceAudience.Private.class.getCanonicalName())
+              || qualifiedTypeName
+                  .equals(InterfaceAudience.LimitedPrivate.class.getCanonicalName())) {
+            return true;
+          }
+          if (stability.equals(StabilityOptions.EVOLVING_OPTION)) {
+            if (qualifiedTypeName.equals(InterfaceStability.Unstable.class.getCanonicalName())) {
+              return true;
+            }
+          }
+          if (stability.equals(StabilityOptions.STABLE_OPTION)) {
+            if (qualifiedTypeName.equals(InterfaceStability.Unstable.class.getCanonicalName())
+                || qualifiedTypeName.equals(InterfaceStability.Evolving.class.getCanonicalName())) {
+              return true;
+            }
+          }
+        }
+        for (AnnotationDesc annotation : annotations) {
+          String qualifiedTypeName = annotation.annotationType().qualifiedTypeName();
+          if (qualifiedTypeName.equals(InterfaceAudience.Public.class.getCanonicalName())) {
             return false;
           }
         }
@@ -220,21 +203,20 @@ class RootDocProcessor {
 
     private static Object[] filter(Doc[] array, Class<?> componentType) {
       if (array == null || array.length == 0) {
-	return array;
+        return array;
       }
       List<Object> list = new ArrayList<Object>(array.length);
       for (Doc entry : array) {
-	if (!exclude(entry)) {
-	  list.add(process(entry, componentType));
-	}
+        if (!exclude(entry)) {
+          list.add(process(entry, componentType));
+        }
       }
-      return list.toArray((Object[]) Array.newInstance(componentType, list
-	  .size()));
+      return list.toArray((Object[]) Array.newInstance(componentType, list.size()));
     }
 
     private Object unwrap(Object proxy) {
       if (proxy instanceof Proxy)
-	return ((ExcludeHandler) Proxy.getInvocationHandler(proxy)).target;
+        return ((ExcludeHandler) Proxy.getInvocationHandler(proxy)).target;
       return proxy;
     }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java
----------------------------------------------------------------------
diff --git a/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java b/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java
index f3169ba..b79f645 100644
--- a/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java
+++ b/hbase-annotations/src/main/java/org/apache/hadoop/hbase/classification/tools/StabilityOptions.java
@@ -35,16 +35,15 @@ class StabilityOptions {
     return null;
   }
 
-  public static void validOptions(String[][] options,
-      DocErrorReporter reporter) {
+  public static void validOptions(String[][] options, DocErrorReporter reporter) {
     for (int i = 0; i < options.length; i++) {
       String opt = options[i][0].toLowerCase();
       if (opt.equals(UNSTABLE_OPTION)) {
-	RootDocProcessor.stability = UNSTABLE_OPTION;
+        RootDocProcessor.stability = UNSTABLE_OPTION;
       } else if (opt.equals(EVOLVING_OPTION)) {
-	RootDocProcessor.stability = EVOLVING_OPTION;
+        RootDocProcessor.stability = EVOLVING_OPTION;
       } else if (opt.equals(STABLE_OPTION)) {
-	RootDocProcessor.stability = STABLE_OPTION;
+        RootDocProcessor.stability = STABLE_OPTION;
       }
     }
   }
@@ -53,9 +52,9 @@ class StabilityOptions {
     List<String[]> optionsList = new ArrayList<String[]>();
     for (int i = 0; i < options.length; i++) {
       if (!options[i][0].equalsIgnoreCase(UNSTABLE_OPTION)
-	  && !options[i][0].equalsIgnoreCase(EVOLVING_OPTION)
-	  && !options[i][0].equalsIgnoreCase(STABLE_OPTION)) {
-	optionsList.add(options[i]);
+          && !options[i][0].equalsIgnoreCase(EVOLVING_OPTION)
+          && !options[i][0].equalsIgnoreCase(STABLE_OPTION)) {
+        optionsList.add(options[i]);
       }
     }
     String[][] filteredOptions = new String[optionsList.size()][];

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
index fd0de03..95f1771 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
@@ -1034,7 +1034,7 @@ public class HTableDescriptor implements WritableComparable<HTableDescriptor> {
   @Deprecated
   @Override
   public void write(DataOutput out) throws IOException {
-	  out.writeInt(TABLE_DESCRIPTOR_VERSION);
+    out.writeInt(TABLE_DESCRIPTOR_VERSION);
     Bytes.writeByteArray(out, name.toBytes());
     out.writeBoolean(isRootRegion());
     out.writeBoolean(isMetaRegion());

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java
index 3d4c113..7dcf3e9 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTable.java
@@ -429,7 +429,7 @@ public class HTable implements HTableInterface, RegionLocator {
    * @param tableName Name of table to check.
    * @return {@code true} if table is online.
    * @throws IOException if a remote or network exception occurs
-	 * @deprecated use {@link HBaseAdmin#isTableEnabled(byte[])}
+   * @deprecated use {@link HBaseAdmin#isTableEnabled(byte[])}
    */
   @Deprecated
   public static boolean isTableEnabled(Configuration conf, String tableName)
@@ -443,7 +443,7 @@ public class HTable implements HTableInterface, RegionLocator {
    * @param tableName Name of table to check.
    * @return {@code true} if table is online.
    * @throws IOException if a remote or network exception occurs
-	 * @deprecated use {@link HBaseAdmin#isTableEnabled(byte[])}
+   * @deprecated use {@link HBaseAdmin#isTableEnabled(byte[])}
    */
   @Deprecated
   public static boolean isTableEnabled(Configuration conf, byte[] tableName)

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java
index f3b6563..ab77ceb 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java
@@ -111,7 +111,7 @@ public class HTableUtil {
       List<Put> recs = putMap.get( hostname);
       if (recs == null) {
         recs = new ArrayList<Put>(INITIAL_LIST_SIZE);
-    		putMap.put( hostname, recs);
+        putMap.put( hostname, recs);
       }
       recs.add(put);
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHTableDescriptor.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHTableDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHTableDescriptor.java
index 361e63e..c5a93e1 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHTableDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHTableDescriptor.java
@@ -32,7 +32,7 @@ import org.apache.hadoop.hbase.HTableDescriptor;
 public class UnmodifyableHTableDescriptor extends HTableDescriptor {
   /** Default constructor */
   public UnmodifyableHTableDescriptor() {
-	  super();
+    super();
   }
 
   /*

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
index df79f8b..1e378e7 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
@@ -184,7 +184,7 @@ public class AggregationClient implements Closeable {
         || (Bytes.equals(scan.getStartRow(), scan.getStopRow()) && !Bytes
             .equals(scan.getStartRow(), HConstants.EMPTY_START_ROW))
         || ((Bytes.compareTo(scan.getStartRow(), scan.getStopRow()) > 0) &&
-        	!Bytes.equals(scan.getStopRow(), HConstants.EMPTY_END_ROW))) {
+          !Bytes.equals(scan.getStopRow(), HConstants.EMPTY_END_ROW))) {
       throw new IOException(
           "Agg client Exception: Startrow should be smaller than Stoprow");
     } else if (!canFamilyBeAbsent) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java
index 5cfedaa..2843751 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java
@@ -68,8 +68,8 @@ public class DependentColumnFilter extends CompareFilter {
    * @param valueComparator comparator
    */
   public DependentColumnFilter(final byte [] family, final byte[] qualifier,
-		  final boolean dropDependentColumn, final CompareOp valueCompareOp,
-	      final ByteArrayComparable valueComparator) {
+      final boolean dropDependentColumn, final CompareOp valueCompareOp,
+      final ByteArrayComparable valueComparator) {
     // set up the comparator   
     super(valueCompareOp, valueComparator);
     this.columnFamily = family;
@@ -136,19 +136,19 @@ public class DependentColumnFilter extends CompareFilter {
   @Override
   public ReturnCode filterKeyValue(Cell c) {
     // Check if the column and qualifier match
-  	if (!CellUtil.matchingColumn(c, this.columnFamily, this.columnQualifier)) {
-        // include non-matches for the time being, they'll be discarded afterwards
-        return ReturnCode.INCLUDE;
-  	}
+    if (!CellUtil.matchingColumn(c, this.columnFamily, this.columnQualifier)) {
+      // include non-matches for the time being, they'll be discarded afterwards
+      return ReturnCode.INCLUDE;
+    }
     // If it doesn't pass the op, skip it
     if (comparator != null
         && doCompare(compareOp, comparator, c.getValueArray(), c.getValueOffset(),
             c.getValueLength()))
       return ReturnCode.SKIP;
-	
+
     stampSet.add(c.getTimestamp());
     if(dropDependentColumn) {
-    	return ReturnCode.SKIP;
+      return ReturnCode.SKIP;
     }
     return ReturnCode.INCLUDE;
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java
index f5c128b..c7c9038 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/RequestConverter.java
@@ -1069,7 +1069,7 @@ public final class RequestConverter {
   public static MoveRegionRequest buildMoveRegionRequest(
       final byte [] encodedRegionName, final byte [] destServerName) throws
       DeserializationException {
-	MoveRegionRequest.Builder builder = MoveRegionRequest.newBuilder();
+    MoveRegionRequest.Builder builder = MoveRegionRequest.newBuilder();
     builder.setRegion(
       buildRegionSpecifier(RegionSpecifierType.ENCODED_REGION_NAME,encodedRegionName));
     if (destServerName != null) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java
index 7747ffa..ab12467 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java
@@ -441,8 +441,8 @@ public class ZKAssign {
       EventType expectedState, ServerName serverName, int expectedVersion)
   throws KeeperException, KeeperException.NoNodeException {
     if (LOG.isTraceEnabled()) {
-    	LOG.trace(zkw.prefix("Deleting existing unassigned " +
-      "node " + encodedRegionName + " in expected state " + expectedState));
+      LOG.trace(zkw.prefix("Deleting existing unassigned " +
+        "node " + encodedRegionName + " in expected state " + expectedState));
     }
     String node = getNodeName(zkw, encodedRegionName);
     zkw.sync(node);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java
index 808c4c1..7779399 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java
@@ -181,7 +181,7 @@ public class HBaseConfiguration extends Configuration {
       char[] p = (char[]) m.invoke(conf, alias);
       if (p != null) {
         LOG.debug(String.format("Config option \"%s\" was found through" +
-        		" the Configuration getPassword method.", alias));
+            " the Configuration getPassword method.", alias));
         passwd = new String(p);
       }
       else {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java
index b23b2f1..8c16389 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java
@@ -61,7 +61,7 @@ public class TimeRange {
    * @param minStamp the minimum timestamp value, inclusive
    */
   public TimeRange(byte [] minStamp) {
-  	this.minStamp = Bytes.toLong(minStamp);
+    this.minStamp = Bytes.toLong(minStamp);
   }
 
   /**
@@ -126,8 +126,8 @@ public class TimeRange {
    * @return true if within TimeRange, false if not
    */
   public boolean withinTimeRange(byte [] bytes, int offset) {
-  	if(allTime) return true;
-  	return withinTimeRange(Bytes.toLong(bytes, offset));
+    if(allTime) return true;
+    return withinTimeRange(Bytes.toLong(bytes, offset));
   }
 
   /**
@@ -139,9 +139,9 @@ public class TimeRange {
    * @return true if within TimeRange, false if not
    */
   public boolean withinTimeRange(long timestamp) {
-  	if(allTime) return true;
-  	// check if >= minStamp
-  	return (minStamp <= timestamp && timestamp < maxStamp);
+    if(allTime) return true;
+    // check if >= minStamp
+    return (minStamp <= timestamp && timestamp < maxStamp);
   }
 
   /**

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java
index b15b2e0..b7b9beb 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/CollectionUtils.java
@@ -43,7 +43,7 @@ public class CollectionUtils {
     return in;
   }
 
-	/************************ size ************************************/
+  /************************ size ************************************/
 
   public static <T> int nullSafeSize(Collection<T> collection) {
     if (collection == null) {
@@ -56,7 +56,7 @@ public class CollectionUtils {
     return nullSafeSize(a) == nullSafeSize(b);
   }
 
-	/*************************** empty ****************************************/
+  /*************************** empty ****************************************/
 
   public static <T> boolean isEmpty(Collection<T> collection) {
     return collection == null || collection.isEmpty();
@@ -66,7 +66,7 @@ public class CollectionUtils {
     return !isEmpty(collection);
   }
 
-	/************************ first/last **************************/
+  /************************ first/last **************************/
 
   public static <T> T getFirst(Collection<T> collection) {
     if (CollectionUtils.isEmpty(collection)) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-common/src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java
index 42d8b11..359e7a9 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java
@@ -164,7 +164,7 @@ public class JenkinsHash extends Hash {
     //-------------------------------- last block: affect all 32 bits of (c)
     switch (length) {                   // all the case statements fall through
     case 12:
-    	c += ((key[offset + 11] & BYTE_MASK) << 24);
+      c += ((key[offset + 11] & BYTE_MASK) << 24);
     case 11:
       c += ((key[offset + 10] & BYTE_MASK) << 16);
     case 10:

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-common/src/test/java/org/apache/hadoop/hbase/TestHBaseConfiguration.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/TestHBaseConfiguration.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/TestHBaseConfiguration.java
index 60fa3b3..445db84 100644
--- a/hbase-common/src/test/java/org/apache/hadoop/hbase/TestHBaseConfiguration.java
+++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/TestHBaseConfiguration.java
@@ -199,7 +199,7 @@ public class TestHBaseConfiguration {
 
       hadoopClassesAvailable = true;
       LOG.info("Credential provider classes have been" +
-      		" loaded and initialized successfully through reflection.");
+          " loaded and initialized successfully through reflection.");
       return true;
 
     }
@@ -278,7 +278,7 @@ public class TestHBaseConfiguration {
       List<Object> providers = getCredentialProviders(conf);
       if (null == providers) {
         throw new IOException("Could not fetch any CredentialProviders, " +
-        		"is the implementation available?");
+            "is the implementation available?");
       }
 
       Object provider = providers.get(0);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java
----------------------------------------------------------------------
diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java
index f80036c..9a8f2ab 100644
--- a/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java
+++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/test/IntegrationTestBigLinkedListWithVisibility.java
@@ -583,8 +583,8 @@ public class IntegrationTestBigLinkedListWithVisibility extends IntegrationTestB
       if (args.length < 5) {
         System.err
             .println("Usage: Loop <num iterations> " +
-            		"<num mappers> <num nodes per mapper> <output dir> " +
-            		"<num reducers> [<width> <wrap multiplier>]");
+                "<num mappers> <num nodes per mapper> <output dir> " +
+                "<num reducers> [<width> <wrap multiplier>]");
         return 1;
       }
       LOG.info("Running Loop with args:" + Arrays.deepToString(args));

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/PrefixTreeBlockMeta.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/PrefixTreeBlockMeta.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/PrefixTreeBlockMeta.java
index 620a6ef..8410cf3 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/PrefixTreeBlockMeta.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/PrefixTreeBlockMeta.java
@@ -45,7 +45,7 @@ public class PrefixTreeBlockMeta {
   public static final int MAX_FAMILY_LENGTH = Byte.MAX_VALUE;// hard-coded in KeyValue
 
   public static final int
-	  NUM_LONGS = 2,
+    NUM_LONGS = 2,
     NUM_INTS = 28,
     NUM_SHORTS = 0,//keyValueTypeWidth not persisted
     NUM_SINGLE_BYTES = 2,
@@ -135,7 +135,7 @@ public class PrefixTreeBlockMeta {
   }
 
 
-	/**************** operate on each field **********************/
+  /**************** operate on each field **********************/
 
   public int calculateNumMetaBytes(){
     int numBytes = 0;
@@ -339,7 +339,7 @@ public class PrefixTreeBlockMeta {
     position += UVIntTool.numBytes(numUniqueTags);
   }
 
-	//TODO method that can read directly from ByteBuffer instead of InputStream
+  //TODO method that can read directly from ByteBuffer instead of InputStream
 
 
   /*************** methods *************************/

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeArraySearcher.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeArraySearcher.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeArraySearcher.java
index 8ea6e85..ec54c2a 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeArraySearcher.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/decode/PrefixTreeArraySearcher.java
@@ -306,7 +306,7 @@ public class PrefixTreeArraySearcher extends PrefixTreeArrayReversibleScanner im
   }
 
 
-	/****************** complete seek when token mismatch ******************/
+  /****************** complete seek when token mismatch ******************/
 
   /**
    * @param searcherIsAfterInputKey <0: input key is before the searcher's position<br/>

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/column/ColumnSectionWriter.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/column/ColumnSectionWriter.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/column/ColumnSectionWriter.java
index 4793335..3ceae63 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/column/ColumnSectionWriter.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/column/ColumnSectionWriter.java
@@ -71,7 +71,7 @@ public class ColumnSectionWriter {
   private List<Integer> outputArrayOffsets;
 
 
-	/*********************** construct *********************/
+  /*********************** construct *********************/
 
   public ColumnSectionWriter() {
     this.nonLeaves = Lists.newArrayList();
@@ -100,7 +100,7 @@ public class ColumnSectionWriter {
   }
 
 
-	/****************** methods *******************************/
+  /****************** methods *******************************/
 
   public ColumnSectionWriter compile() {
     if (this.nodeType == ColumnNodeType.FAMILY) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java
index c6ae347..3291d72 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/other/LongEncoder.java
@@ -75,7 +75,7 @@ public class LongEncoder {
   }
 
 
-	/************* methods ***************************/
+  /************* methods ***************************/
 
   public void add(long timestamp) {
     uniqueValues.add(timestamp);
@@ -158,7 +158,7 @@ public class LongEncoder {
   }
 
 
-	/******************** get/set **************************/
+  /******************** get/set **************************/
 
   public long getMin() {
     return min;

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java
index cf13add..75a11ad 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/Tokenizer.java
@@ -179,7 +179,7 @@ public class Tokenizer{
   }
 
 
-	/********************** write ***************************/
+  /********************** write ***************************/
 
   public Tokenizer setNodeFirstInsertionIndexes() {
     root.setInsertionIndexes(0);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerNode.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerNode.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerNode.java
index eaf8ab7..e51d5be 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerNode.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerNode.java
@@ -289,7 +289,7 @@ public class TokenizerNode{
   }
 
 
-	/************************ byte[] utils *************************/
+  /************************ byte[] utils *************************/
 
   protected boolean partiallyMatchesToken(ByteRange bytes) {
     return numIdenticalBytes(bytes) > 0;
@@ -304,7 +304,7 @@ public class TokenizerNode{
   }
 
 
-	/***************** moving nodes around ************************/
+  /***************** moving nodes around ************************/
 
   public void appendNodesToExternalList(List<TokenizerNode> appendTo, boolean includeNonLeaves,
       boolean includeLeaves) {
@@ -462,7 +462,7 @@ public class TokenizerNode{
   }
 
 
-	/********************** count different node types ********************/
+  /********************** count different node types ********************/
 
   public int getNumBranchNodesIncludingThisNode() {
     if (isLeaf()) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerRowSearchPosition.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerRowSearchPosition.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerRowSearchPosition.java
index 6054a46..1166baa 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerRowSearchPosition.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/codec/prefixtree/encode/tokenize/TokenizerRowSearchPosition.java
@@ -30,9 +30,9 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
 @InterfaceAudience.Private
 public enum TokenizerRowSearchPosition {
 
-	AFTER,//the key is after this tree node, so keep searching
-	BEFORE,//in a binary search, this tells us to back up
-	MATCH,//the current node is a full match
-	NO_MATCH,//might as well return a value more informative than null
+  AFTER,//the key is after this tree node, so keep searching
+  BEFORE,//in a binary search, this tells us to back up
+  MATCH,//the current node is a full match
+  NO_MATCH,//might as well return a value more informative than null
 
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/vint/UFIntTool.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/vint/UFIntTool.java b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/vint/UFIntTool.java
index 8b63fd1..fc7c107 100644
--- a/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/vint/UFIntTool.java
+++ b/hbase-prefix-tree/src/main/java/org/apache/hadoop/hbase/util/vint/UFIntTool.java
@@ -78,13 +78,13 @@ public class UFIntTool {
 
   private static final long[] MASKS = new long[] {
     (long) 255,
-	  (long) 255 << 8,
-	  (long) 255 << 16,
-	  (long) 255 << 24,
-	  (long) 255 << 32,
-	  (long) 255 << 40,
-	  (long) 255 << 48,
-	  (long) 255 << 56
+    (long) 255 << 8,
+    (long) 255 << 16,
+    (long) 255 << 24,
+    (long) 255 << 32,
+    (long) 255 << 40,
+    (long) 255 << 48,
+    (long) 255 << 56
   };
 
   public static void writeBytes(int outputWidth, final long value, OutputStream os) throws IOException {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestPrefixTreeSearcher.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestPrefixTreeSearcher.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestPrefixTreeSearcher.java
index 6a9e52c..945db32 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestPrefixTreeSearcher.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestPrefixTreeSearcher.java
@@ -47,7 +47,7 @@ import org.junit.runners.Parameterized.Parameters;
 @RunWith(Parameterized.class)
 public class TestPrefixTreeSearcher {
 
-	protected static int BLOCK_START = 7;
+  protected static int BLOCK_START = 7;
 
   @Parameters
   public static Collection<Object[]> parameters() {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestRowEncoder.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestRowEncoder.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestRowEncoder.java
index 74d5cfc..a1c5a25 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestRowEncoder.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/TestRowEncoder.java
@@ -76,7 +76,7 @@ public class TestRowEncoder {
     this.rows = testRows;
   }
 
-	@Before
+  @Before
   public void compile() throws IOException {
     // Always run with tags. But should also ensure that KVs without tags work fine
     os = new ByteArrayOutputStream(1 << 20);
@@ -174,7 +174,7 @@ public class TestRowEncoder {
   }
 
 
-	/**************** helper **************************/
+  /**************** helper **************************/
 
   protected void assertKeyAndValueEqual(Cell expected, Cell actual) {
     // assert keys are equal (doesn't compare values)

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataComplexQualifiers.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataComplexQualifiers.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataComplexQualifiers.java
index bd6f02b..66fe3f3 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataComplexQualifiers.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataComplexQualifiers.java
@@ -29,7 +29,7 @@ import com.google.common.collect.Lists;
 
 public class TestRowDataComplexQualifiers extends BaseTestRowData{
 
-	static byte[]
+  static byte[]
     Arow = Bytes.toBytes("Arow"),
     cf = PrefixTreeTestConstants.TEST_CF,
     v0 = Bytes.toBytes("v0");

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDeeper.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDeeper.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDeeper.java
index cb3913c..4d05742 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDeeper.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDeeper.java
@@ -36,36 +36,36 @@ import com.google.common.collect.Lists;
  */
 public class TestRowDataDeeper extends BaseTestRowData{
 
-	static byte[]
-        cdc = Bytes.toBytes("cdc"),
-        cf6 = Bytes.toBytes("cf6"),
-        cfc = Bytes.toBytes("cfc"),
-        f = Bytes.toBytes("f"),
-        q = Bytes.toBytes("q"),
-        v = Bytes.toBytes("v");
+  static byte[]
+    cdc = Bytes.toBytes("cdc"),
+    cf6 = Bytes.toBytes("cf6"),
+    cfc = Bytes.toBytes("cfc"),
+    f = Bytes.toBytes("f"),
+    q = Bytes.toBytes("q"),
+    v = Bytes.toBytes("v");
 
-	static long
-		ts = 55L;
+  static long
+    ts = 55L;
 
-	static List<KeyValue> d = Lists.newArrayList();
-	static{
-		d.add(new KeyValue(cdc, f, q, ts, v));
+  static List<KeyValue> d = Lists.newArrayList();
+  static{
+    d.add(new KeyValue(cdc, f, q, ts, v));
     d.add(new KeyValue(cf6, f, q, ts, v));
     d.add(new KeyValue(cfc, f, q, ts, v));
-	}
+  }
 
-	@Override
-	public List<KeyValue> getInputs() {
-		return d;
-	}
+  @Override
+  public List<KeyValue> getInputs() {
+    return d;
+  }
 
-	@Override
-	public void individualBlockMetaAssertions(PrefixTreeBlockMeta blockMeta) {
-	  //0: token:c; fan:d,f
-	  //1: token:f; fan:6,c
-	  //2: leaves
-		Assert.assertEquals(3, blockMeta.getRowTreeDepth());
-	}
+  @Override
+  public void individualBlockMetaAssertions(PrefixTreeBlockMeta blockMeta) {
+    //0: token:c; fan:d,f
+    //1: token:f; fan:6,c
+    //2: leaves
+    Assert.assertEquals(3, blockMeta.getRowTreeDepth());
+  }
 
   @Override
   public void individualSearcherAssertions(CellSearcher searcher) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDifferentTimestamps.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDifferentTimestamps.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDifferentTimestamps.java
index 2668f2a..8639e8f 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDifferentTimestamps.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataDifferentTimestamps.java
@@ -33,62 +33,62 @@ import com.google.common.collect.Lists;
  */
 public class TestRowDataDifferentTimestamps extends BaseTestRowData{
 
-	static byte[]
-        Arow = Bytes.toBytes("Arow"),
-        Brow = Bytes.toBytes("Brow"),
-        cf = Bytes.toBytes("fammy"),
-        cq0 = Bytes.toBytes("cq0"),
-        cq1 = Bytes.toBytes("cq1"),
-        v0 = Bytes.toBytes("v0");
-
-	static List<KeyValue> d = Lists.newArrayList();
-	static{
-	  KeyValue kv0 = new KeyValue(Arow, cf, cq0, 0L, v0);
-	  kv0.setSequenceId(123456789L);
-	  d.add(kv0);
-
-	  KeyValue kv1 = new KeyValue(Arow, cf, cq1, 1L, v0);
+  static byte[]
+    Arow = Bytes.toBytes("Arow"),
+    Brow = Bytes.toBytes("Brow"),
+    cf = Bytes.toBytes("fammy"),
+    cq0 = Bytes.toBytes("cq0"),
+    cq1 = Bytes.toBytes("cq1"),
+    v0 = Bytes.toBytes("v0");
+
+  static List<KeyValue> d = Lists.newArrayList();
+  static{
+    KeyValue kv0 = new KeyValue(Arow, cf, cq0, 0L, v0);
+    kv0.setSequenceId(123456789L);
+    d.add(kv0);
+
+    KeyValue kv1 = new KeyValue(Arow, cf, cq1, 1L, v0);
     kv1.setSequenceId(3L);
     d.add(kv1);
 
-	  KeyValue kv2 = new KeyValue(Brow, cf, cq0, 12345678L, v0);
+    KeyValue kv2 = new KeyValue(Brow, cf, cq0, 12345678L, v0);
     kv2.setSequenceId(65537L);
     d.add(kv2);
 
-		//watch out... Long.MAX_VALUE comes back as 1332221664203, even with other encoders
-//		d.add(new KeyValue(Brow, cf, cq1, Long.MAX_VALUE, v0));
-	  KeyValue kv3 = new KeyValue(Brow, cf, cq1, Long.MAX_VALUE-1, v0);
+    //watch out... Long.MAX_VALUE comes back as 1332221664203, even with other encoders
+    //d.add(new KeyValue(Brow, cf, cq1, Long.MAX_VALUE, v0));
+    KeyValue kv3 = new KeyValue(Brow, cf, cq1, Long.MAX_VALUE-1, v0);
     kv3.setSequenceId(1L);
     d.add(kv3);
 
-	  KeyValue kv4 = new KeyValue(Brow, cf, cq1, 999999999, v0);
+    KeyValue kv4 = new KeyValue(Brow, cf, cq1, 999999999, v0);
     //don't set memstoreTS
-	  d.add(kv4);
+    d.add(kv4);
 
-	  KeyValue kv5 = new KeyValue(Brow, cf, cq1, 12345, v0);
+    KeyValue kv5 = new KeyValue(Brow, cf, cq1, 12345, v0);
     kv5.setSequenceId(0L);
     d.add(kv5);
-	}
+  }
 
-	@Override
-	public List<KeyValue> getInputs() {
-		return d;
-	}
+  @Override
+  public List<KeyValue> getInputs() {
+    return d;
+  }
 
-	@Override
-	public void individualBlockMetaAssertions(PrefixTreeBlockMeta blockMeta) {
-	  Assert.assertTrue(blockMeta.getNumMvccVersionBytes() > 0);
-	  Assert.assertEquals(12, blockMeta.getNumValueBytes());
+  @Override
+  public void individualBlockMetaAssertions(PrefixTreeBlockMeta blockMeta) {
+    Assert.assertTrue(blockMeta.getNumMvccVersionBytes() > 0);
+    Assert.assertEquals(12, blockMeta.getNumValueBytes());
 
-		Assert.assertFalse(blockMeta.isAllSameTimestamp());
-		Assert.assertNotNull(blockMeta.getMinTimestamp());
-		Assert.assertTrue(blockMeta.getTimestampIndexWidth() > 0);
-		Assert.assertTrue(blockMeta.getTimestampDeltaWidth() > 0);
+    Assert.assertFalse(blockMeta.isAllSameTimestamp());
+    Assert.assertNotNull(blockMeta.getMinTimestamp());
+    Assert.assertTrue(blockMeta.getTimestampIndexWidth() > 0);
+    Assert.assertTrue(blockMeta.getTimestampDeltaWidth() > 0);
 
     Assert.assertFalse(blockMeta.isAllSameMvccVersion());
     Assert.assertNotNull(blockMeta.getMinMvccVersion());
     Assert.assertTrue(blockMeta.getMvccVersionIndexWidth() > 0);
     Assert.assertTrue(blockMeta.getMvccVersionDeltaWidth() > 0);
-	}
+  }
 
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java
index 184f537..c49db13 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataExerciseFInts.java
@@ -42,49 +42,49 @@ import com.google.common.collect.Lists;
 public class TestRowDataExerciseFInts extends BaseTestRowData{
 
   static List<ByteRange> rows;
-	static{
-		List<String> rowStrings = new ArrayList<String>();
-        rowStrings.add("com.edsBlog/directoryAa/pageAaa");
-        rowStrings.add("com.edsBlog/directoryAa/pageBbb");
-        rowStrings.add("com.edsBlog/directoryAa/pageCcc");
-        rowStrings.add("com.edsBlog/directoryAa/pageDdd");
-        rowStrings.add("com.edsBlog/directoryBb/pageEee");
-        rowStrings.add("com.edsBlog/directoryBb/pageFff");
-        rowStrings.add("com.edsBlog/directoryBb/pageGgg");
-        rowStrings.add("com.edsBlog/directoryBb/pageHhh");
-        rowStrings.add("com.isabellasBlog/directoryAa/pageAaa");
-        rowStrings.add("com.isabellasBlog/directoryAa/pageBbb");
-        rowStrings.add("com.isabellasBlog/directoryAa/pageCcc");
-        rowStrings.add("com.isabellasBlog/directoryAa/pageDdd");
-        rowStrings.add("com.isabellasBlog/directoryBb/pageEee");
-        rowStrings.add("com.isabellasBlog/directoryBb/pageFff");
-        rowStrings.add("com.isabellasBlog/directoryBb/pageGgg");
-        rowStrings.add("com.isabellasBlog/directoryBb/pageHhh");
-        ByteRangeTreeSet ba = new ByteRangeTreeSet();
-        for(String row : rowStrings){
-        	ba.add(new SimpleMutableByteRange(Bytes.toBytes(row)));
-        }
-        rows = ba.compile().getSortedRanges();
-	}
+  static{
+    List<String> rowStrings = new ArrayList<String>();
+    rowStrings.add("com.edsBlog/directoryAa/pageAaa");
+    rowStrings.add("com.edsBlog/directoryAa/pageBbb");
+    rowStrings.add("com.edsBlog/directoryAa/pageCcc");
+    rowStrings.add("com.edsBlog/directoryAa/pageDdd");
+    rowStrings.add("com.edsBlog/directoryBb/pageEee");
+    rowStrings.add("com.edsBlog/directoryBb/pageFff");
+    rowStrings.add("com.edsBlog/directoryBb/pageGgg");
+    rowStrings.add("com.edsBlog/directoryBb/pageHhh");
+    rowStrings.add("com.isabellasBlog/directoryAa/pageAaa");
+    rowStrings.add("com.isabellasBlog/directoryAa/pageBbb");
+    rowStrings.add("com.isabellasBlog/directoryAa/pageCcc");
+    rowStrings.add("com.isabellasBlog/directoryAa/pageDdd");
+    rowStrings.add("com.isabellasBlog/directoryBb/pageEee");
+    rowStrings.add("com.isabellasBlog/directoryBb/pageFff");
+    rowStrings.add("com.isabellasBlog/directoryBb/pageGgg");
+    rowStrings.add("com.isabellasBlog/directoryBb/pageHhh");
+    ByteRangeTreeSet ba = new ByteRangeTreeSet();
+    for(String row : rowStrings){
+      ba.add(new SimpleMutableByteRange(Bytes.toBytes(row)));
+    }
+    rows = ba.compile().getSortedRanges();
+  }
 
-	static List<String> cols = Lists.newArrayList();
-	static{
-		cols.add("Chrome");
-		cols.add("Chromeb");
-		cols.add("Firefox");
-		cols.add("InternetExplorer");
-		cols.add("Opera");
-		cols.add("Safari");
-		cols.add("Z1stBrowserWithHuuuuuuuuuuuugeQualifier");
-		cols.add("Z2ndBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
-		cols.add("Z3rdBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
-		cols.add("Z4thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
-		cols.add("Z5thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
-		cols.add("Z6thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
-		cols.add("Z7thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
-		cols.add("Z8thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
-		cols.add("Z9thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
-	}
+  static List<String> cols = Lists.newArrayList();
+  static{
+    cols.add("Chrome");
+    cols.add("Chromeb");
+    cols.add("Firefox");
+    cols.add("InternetExplorer");
+    cols.add("Opera");
+    cols.add("Safari");
+    cols.add("Z1stBrowserWithHuuuuuuuuuuuugeQualifier");
+    cols.add("Z2ndBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
+    cols.add("Z3rdBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
+    cols.add("Z4thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
+    cols.add("Z5thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
+    cols.add("Z6thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
+    cols.add("Z7thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
+    cols.add("Z8thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
+    cols.add("Z9thBrowserWithEvenBiggerQualifierMoreMoreMoreMoreMore");
+  }
 
   static long ts = 1234567890;
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataNub.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataNub.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataNub.java
index ad19cd4..a818f5e 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataNub.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataNub.java
@@ -29,31 +29,31 @@ import com.google.common.collect.Lists;
 
 public class TestRowDataNub extends BaseTestRowData{
 
-	static byte[]
-        rowA = Bytes.toBytes("rowA"),
-        rowB = Bytes.toBytes("rowB"),//nub
-        rowBB = Bytes.toBytes("rowBB"),
-        cf = PrefixTreeTestConstants.TEST_CF,
-        cq0 = Bytes.toBytes("cq0"),
-        cq1 = Bytes.toBytes("cq1"),
-        v0 = Bytes.toBytes("v0");
-
-	static long
-		ts = 55L;
-
-	static List<KeyValue> d = Lists.newArrayList();
-	static{
-		d.add(new KeyValue(rowA, cf, cq0, ts, v0));
-		d.add(new KeyValue(rowA, cf, cq1, ts, v0));
-		d.add(new KeyValue(rowB, cf, cq0, ts, v0));
-		d.add(new KeyValue(rowB, cf, cq1, ts, v0));
-		d.add(new KeyValue(rowBB, cf, cq0, ts, v0));
-		d.add(new KeyValue(rowBB, cf, cq1, ts, v0));
-	}
-
-	@Override
-	public List<KeyValue> getInputs() {
-		return d;
-	}
+  static byte[]
+    rowA = Bytes.toBytes("rowA"),
+    rowB = Bytes.toBytes("rowB"),//nub
+    rowBB = Bytes.toBytes("rowBB"),
+    cf = PrefixTreeTestConstants.TEST_CF,
+    cq0 = Bytes.toBytes("cq0"),
+    cq1 = Bytes.toBytes("cq1"),
+    v0 = Bytes.toBytes("v0");
+
+  static long
+    ts = 55L;
+
+  static List<KeyValue> d = Lists.newArrayList();
+  static{
+    d.add(new KeyValue(rowA, cf, cq0, ts, v0));
+    d.add(new KeyValue(rowA, cf, cq1, ts, v0));
+    d.add(new KeyValue(rowB, cf, cq0, ts, v0));
+    d.add(new KeyValue(rowB, cf, cq1, ts, v0));
+    d.add(new KeyValue(rowBB, cf, cq0, ts, v0));
+    d.add(new KeyValue(rowBB, cf, cq1, ts, v0));
+  }
+
+  @Override
+  public List<KeyValue> getInputs() {
+    return d;
+  }
 
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataQualifierByteOrdering.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataQualifierByteOrdering.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataQualifierByteOrdering.java
index a8c4646..6d3918f 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataQualifierByteOrdering.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataQualifierByteOrdering.java
@@ -28,15 +28,15 @@ import com.google.common.collect.Lists;
 
 public class TestRowDataQualifierByteOrdering extends BaseTestRowData{
 
-	static byte[]
-        Arow = Bytes.toBytes("Arow"),
-        Brow = Bytes.toBytes("Brow"),
-        Brow2 = Bytes.toBytes("Brow2"),
-        fam = Bytes.toBytes("HappyFam"),
-        cq0 = Bytes.toBytes("cq0"),
-        cq1 = Bytes.toBytes("cq1tail"),//make sure tail does not come back as liat
-        cq2 = Bytes.toBytes("cq2"),
-        v0 = Bytes.toBytes("v0");
+  static byte[]
+    Arow = Bytes.toBytes("Arow"),
+    Brow = Bytes.toBytes("Brow"),
+    Brow2 = Bytes.toBytes("Brow2"),
+    fam = Bytes.toBytes("HappyFam"),
+    cq0 = Bytes.toBytes("cq0"),
+    cq1 = Bytes.toBytes("cq1tail"),//make sure tail does not come back as liat
+    cq2 = Bytes.toBytes("cq2"),
+    v0 = Bytes.toBytes("v0");
 
   static long ts = 55L;
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSearcherRowMiss.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSearcherRowMiss.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSearcherRowMiss.java
index 5f319fc..edea305 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSearcherRowMiss.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSearcherRowMiss.java
@@ -55,13 +55,13 @@ public class TestRowDataSearcherRowMiss extends BaseTestRowData{
     d.add(new KeyValue(B, cf, cq, ts, v));
   }
 
-	@Override
-	public List<KeyValue> getInputs() {
-		return d;
-	}
+  @Override
+  public List<KeyValue> getInputs() {
+    return d;
+  }
 
-	@Override
-	public void individualSearcherAssertions(CellSearcher searcher) {
+  @Override
+  public void individualSearcherAssertions(CellSearcher searcher) {
     assertRowOffsetsCorrect();
 
     searcher.resetToBeforeFirstEntry();
@@ -83,13 +83,13 @@ public class TestRowDataSearcherRowMiss extends BaseTestRowData{
     testBetween2and3(searcher);
   }
 
-	/************ private methods, call from above *******************/
+  /************ private methods, call from above *******************/
 
-	private void assertRowOffsetsCorrect(){
-	  Assert.assertEquals(4, getRowStartIndexes().size());
-	}
+  private void assertRowOffsetsCorrect(){
+    Assert.assertEquals(4, getRowStartIndexes().size());
+  }
 
-	private void testBetween1and2(CellSearcher searcher){
+  private void testBetween1and2(CellSearcher searcher){
     CellScannerPosition p;//reuse
     Cell betweenAAndAAA = new KeyValue(AA, cf, cq, ts-2, v);
 
@@ -105,7 +105,7 @@ public class TestRowDataSearcherRowMiss extends BaseTestRowData{
     p = searcher.positionAtOrAfter(betweenAAndAAA);
     Assert.assertEquals(CellScannerPosition.AFTER, p);
     Assert.assertTrue(CellComparator.equals(searcher.current(), d.get(2)));
-	}
+  }
 
   private void testBetween2and3(CellSearcher searcher){
     CellScannerPosition p;//reuse

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSingleQualifier.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSingleQualifier.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSingleQualifier.java
index 9944057..35ba2c1 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSingleQualifier.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataSingleQualifier.java
@@ -29,12 +29,12 @@ import com.google.common.collect.Lists;
 
 public class TestRowDataSingleQualifier extends BaseTestRowData{
 
-	static byte[]
-        rowA = Bytes.toBytes("rowA"),
-        rowB = Bytes.toBytes("rowB"),
-        cf = PrefixTreeTestConstants.TEST_CF,
-        cq0 = Bytes.toBytes("cq0"),
-        v0 = Bytes.toBytes("v0");
+  static byte[]
+    rowA = Bytes.toBytes("rowA"),
+    rowB = Bytes.toBytes("rowB"),
+    cf = PrefixTreeTestConstants.TEST_CF,
+    cq0 = Bytes.toBytes("cq0"),
+    v0 = Bytes.toBytes("v0");
 
   static long ts = 55L;
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivial.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivial.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivial.java
index df785fa..2a53976 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivial.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataTrivial.java
@@ -33,12 +33,12 @@ import com.google.common.collect.Lists;
 
 public class TestRowDataTrivial extends BaseTestRowData{
 
-	static byte[]
-        rA = Bytes.toBytes("rA"),
-        rB = Bytes.toBytes("rB"),//turn "r" into a branch for the Searcher tests
-        cf = Bytes.toBytes("fam"),
-        cq0 = Bytes.toBytes("q0"),
-        v0 = Bytes.toBytes("v0");
+  static byte[]
+    rA = Bytes.toBytes("rA"),
+    rB = Bytes.toBytes("rB"),//turn "r" into a branch for the Searcher tests
+    cf = Bytes.toBytes("fam"),
+    cq0 = Bytes.toBytes("q0"),
+    v0 = Bytes.toBytes("v0");
 
   static long ts = 55L;
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java
index d0801c1..74ac85a 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrls.java
@@ -40,7 +40,7 @@ import com.google.common.collect.Lists;
 public class TestRowDataUrls extends BaseTestRowData{
 
   static List<ByteRange> rows;
-	static{
+  static{
     List<String> rowStrings = new ArrayList<String>();
     rowStrings.add("com.edsBlog/directoryAa/pageAaa");
     rowStrings.add("com.edsBlog/directoryAa/pageBbb");

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrlsExample.java
----------------------------------------------------------------------
diff --git a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrlsExample.java b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrlsExample.java
index 543afb6..56d4e85 100644
--- a/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrlsExample.java
+++ b/hbase-prefix-tree/src/test/java/org/apache/hadoop/hbase/codec/prefixtree/row/data/TestRowDataUrlsExample.java
@@ -51,33 +51,33 @@ public class TestRowDataUrlsExample extends BaseTestRowData{
   static String FAMILY = "hits";
   static List<String> BROWSERS = Lists.newArrayList(
       "Chrome", "IE8", "IE9beta");//, "Opera", "Safari");
-	static long TIMESTAMP = 1234567890;
-
-	static int MAX_VALUE = 50;
-
-	static List<KeyValue> kvs = Lists.newArrayList();
-	static{
-		for(String rowKey : URLS){
-			for(String qualifier : BROWSERS){
-			  KeyValue kv = new KeyValue(
-						Bytes.toBytes(rowKey), 
-						Bytes.toBytes(FAMILY), 
-						Bytes.toBytes(qualifier), 
-						TIMESTAMP, 
-						KeyValue.Type.Put, 
-						Bytes.toBytes("VvvV"));
-				kvs.add(kv);
-			}
-		}
-	}
-
-	/**
-	 * Used for generating docs.
-	 */
-	public static void main(String... args) throws IOException{
+  static long TIMESTAMP = 1234567890;
+
+  static int MAX_VALUE = 50;
+
+  static List<KeyValue> kvs = Lists.newArrayList();
+  static{
+    for(String rowKey : URLS){
+      for(String qualifier : BROWSERS){
+        KeyValue kv = new KeyValue(
+            Bytes.toBytes(rowKey), 
+            Bytes.toBytes(FAMILY), 
+            Bytes.toBytes(qualifier), 
+            TIMESTAMP, 
+            KeyValue.Type.Put, 
+            Bytes.toBytes("VvvV"));
+        kvs.add(kv);
+      }
+    }
+  }
+
+  /**
+   * Used for generating docs.
+   */
+  public static void main(String... args) throws IOException{
     System.out.println("-- inputs --");
     System.out.println(KeyValueTestUtil.toStringWithPadding(kvs, true));
-		ByteArrayOutputStream os = new ByteArrayOutputStream(1<<20);
+    ByteArrayOutputStream os = new ByteArrayOutputStream(1<<20);
     PrefixTreeEncoder encoder = new PrefixTreeEncoder(os, false);
 
     for(KeyValue kv : kvs){
@@ -116,11 +116,11 @@ public class TestRowDataUrlsExample extends BaseTestRowData{
 
     System.out.println("-- concatenated values --");
     System.out.println(Bytes.toStringBinary(encoder.getValueByteRange().deepCopyToNewArray()));
-	}
+  }
 
-	@Override
-	public List<KeyValue> getInputs() {
-		return kvs;
-	}
+  @Override
+  public List<KeyValue> getInputs() {
+    return kvs;
+  }
 
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java
----------------------------------------------------------------------
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java
index 4321a8e..54fc8de 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java
@@ -40,33 +40,33 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
 @XmlRootElement(name="ClusterVersion")
 @InterfaceAudience.Private
 public class StorageClusterVersionModel implements Serializable {
-	private static final long serialVersionUID = 1L;
+  private static final long serialVersionUID = 1L;
 
-	private String version;
+  private String version;
 
-	/**
-	 * @return the storage cluster version
-	 */
-	@XmlValue
-	public String getVersion() {
-	  return version;
-	}
+  /**
+   * @return the storage cluster version
+   */
+  @XmlValue
+  public String getVersion() {
+    return version;
+  }
 
-	/**
-	 * @param version the storage cluster version
-	 */
-	public void setVersion(String version) {
-	  this.version = version;
-	}
+  /**
+   * @param version the storage cluster version
+   */
+  public void setVersion(String version) {
+    this.version = version;
+  }
 
-	/* (non-Javadoc)
-	 * @see java.lang.Object#toString()
-	 */
+  /* (non-Javadoc)
+   * @see java.lang.Object#toString()
+   */
   @JsonValue
-	@Override
-	public String toString() {
-	  return version;
-	}
+  @Override
+  public String toString() {
+    return version;
+  }
 
     //needed for jackson deserialization
     private static StorageClusterVersionModel valueOf(String value) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java
----------------------------------------------------------------------
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java
index 596adac..2ed4e80 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java
@@ -38,67 +38,67 @@ import org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableLis
 @InterfaceAudience.Private
 public class TableListModel implements Serializable, ProtobufMessageHandler {
 
-	private static final long serialVersionUID = 1L;
+  private static final long serialVersionUID = 1L;
 
-	private List<TableModel> tables = new ArrayList<TableModel>();
+  private List<TableModel> tables = new ArrayList<TableModel>();
 
-	/**
-	 * Default constructor
-	 */
-	public TableListModel() {}
+  /**
+   * Default constructor
+   */
+  public TableListModel() {}
 
-	/**
-	 * Add the table name model to the list
-	 * @param table the table model
-	 */
-	public void add(TableModel table) {
-		tables.add(table);
-	}
-	
-	/**
-	 * @param index the index
-	 * @return the table model
-	 */
-	public TableModel get(int index) {
-		return tables.get(index);
-	}
+  /**
+   * Add the table name model to the list
+   * @param table the table model
+   */
+  public void add(TableModel table) {
+    tables.add(table);
+  }
+  
+  /**
+   * @param index the index
+   * @return the table model
+   */
+  public TableModel get(int index) {
+    return tables.get(index);
+  }
 
-	/**
-	 * @return the tables
-	 */
-	@XmlElementRef(name="table")
-	public List<TableModel> getTables() {
-		return tables;
-	}
+  /**
+   * @return the tables
+   */
+  @XmlElementRef(name="table")
+  public List<TableModel> getTables() {
+    return tables;
+  }
 
-	/**
-	 * @param tables the tables to set
-	 */
-	public void setTables(List<TableModel> tables) {
-		this.tables = tables;
-	}
+  /**
+   * @param tables the tables to set
+   */
+  public void setTables(List<TableModel> tables) {
+    this.tables = tables;
+  }
 
-	/* (non-Javadoc)
-	 * @see java.lang.Object#toString()
-	 */
-	@Override
-	public String toString() {
-		StringBuilder sb = new StringBuilder();
-		for(TableModel aTable : tables) {
-			sb.append(aTable.toString());
-			sb.append('\n');
-		}
-		return sb.toString();
-	}
+  /* (non-Javadoc)
+   * @see java.lang.Object#toString()
+   */
+  @Override
+  public String toString() {
+    StringBuilder sb = new StringBuilder();
+    for(TableModel aTable : tables) {
+      sb.append(aTable.toString());
+      sb.append('\n');
+    }
+    return sb.toString();
+  }
 
-	@Override
-	public byte[] createProtobufOutput() {
-		TableList.Builder builder = TableList.newBuilder();
-		for (TableModel aTable : tables) {
-			builder.addName(aTable.getName());
-		}
-		return builder.build().toByteArray();
-	}
+  @Override
+  public byte[] createProtobufOutput() {
+    TableList.Builder builder = TableList.newBuilder();
+    for (TableModel aTable : tables) {
+      builder.addName(aTable.getName());
+    }
+    return builder.build().toByteArray();
+  }
 
   @Override
   public ProtobufMessageHandler getObjectFromMessage(byte[] message)

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java
----------------------------------------------------------------------
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java
index 0fb0d6e..3be0da9 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java
@@ -41,44 +41,44 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
 @InterfaceAudience.Private
 public class TableModel implements Serializable {
 
-	private static final long serialVersionUID = 1L;
-	
-	private String name;
-	
-	/**
-	 * Default constructor
-	 */
-	public TableModel() {}
+  private static final long serialVersionUID = 1L;
+  
+  private String name;
+  
+  /**
+   * Default constructor
+   */
+  public TableModel() {}
 
-	/**
-	 * Constructor
-	 * @param name
-	 */
-	public TableModel(String name) {
-		super();
-		this.name = name;
-	}
+  /**
+   * Constructor
+   * @param name
+   */
+  public TableModel(String name) {
+    super();
+    this.name = name;
+  }
 
-	/**
-	 * @return the name
-	 */
-	@XmlAttribute
-	public String getName() {
-		return name;
-	}
+  /**
+   * @return the name
+   */
+  @XmlAttribute
+  public String getName() {
+    return name;
+  }
 
-	/**
-	 * @param name the name to set
-	 */
-	public void setName(String name) {
-		this.name = name;
-	}
+  /**
+   * @param name the name to set
+   */
+  public void setName(String name) {
+    this.name = name;
+  }
 
-	/* (non-Javadoc)
-	 * @see java.lang.Object#toString()
-	 */
-	@Override
-	public String toString() {
-		return this.name;
-	}
+  /* (non-Javadoc)
+   * @see java.lang.Object#toString()
+   */
+  @Override
+  public String toString() {
+    return this.name;
+  }
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java
----------------------------------------------------------------------
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java
index 0938803..8b08279 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java
@@ -48,9 +48,9 @@ import com.sun.jersey.spi.container.servlet.ServletContainer;
 @InterfaceAudience.Private
 public class VersionModel implements Serializable, ProtobufMessageHandler {
 
-	private static final long serialVersionUID = 1L;
+  private static final long serialVersionUID = 1L;
 
-	private String restVersion;
+  private String restVersion;
   private String jvmVersion;
   private String osVersion;
   private String serverVersion;
@@ -65,30 +65,30 @@ public class VersionModel implements Serializable, ProtobufMessageHandler {
    * Constructor
    * @param context the servlet context
    */
-	public VersionModel(ServletContext context) {
-	  restVersion = RESTServlet.VERSION_STRING;
-	  jvmVersion = System.getProperty("java.vm.vendor") + ' ' +
+  public VersionModel(ServletContext context) {
+    restVersion = RESTServlet.VERSION_STRING;
+    jvmVersion = System.getProperty("java.vm.vendor") + ' ' +
       System.getProperty("java.version") + '-' +
       System.getProperty("java.vm.version");
-	  osVersion = System.getProperty("os.name") + ' ' +
+    osVersion = System.getProperty("os.name") + ' ' +
       System.getProperty("os.version") + ' ' +
       System.getProperty("os.arch");
-	  serverVersion = context.getServerInfo();
-	  jerseyVersion = ServletContainer.class.getPackage()
+    serverVersion = context.getServerInfo();
+    jerseyVersion = ServletContainer.class.getPackage()
       .getImplementationVersion();
-	}
+  }
 
-	/**
-	 * @return the REST gateway version
-	 */
-	@XmlAttribute(name="REST")
-	public String getRESTVersion() {
+  /**
+   * @return the REST gateway version
+   */
+  @XmlAttribute(name="REST")
+  public String getRESTVersion() {
     return restVersion;
   }
 
-	/**
-	 * @return the JVM vendor and version
-	 */
+  /**
+   * @return the JVM vendor and version
+   */
   @XmlAttribute(name="JVM")
   public String getJVMVersion() {
     return jvmVersion;
@@ -154,34 +154,34 @@ public class VersionModel implements Serializable, ProtobufMessageHandler {
   }
 
   /* (non-Javadoc)
-	 * @see java.lang.Object#toString()
-	 */
-	@Override
-	public String toString() {
-	  StringBuilder sb = new StringBuilder();
-	  sb.append("rest ");
-	  sb.append(restVersion);
-	  sb.append(" [JVM: ");
-	  sb.append(jvmVersion);
-	  sb.append("] [OS: ");
-	  sb.append(osVersion);
-	  sb.append("] [Server: ");
-	  sb.append(serverVersion);
-	  sb.append("] [Jersey: ");
+   * @see java.lang.Object#toString()
+   */
+  @Override
+  public String toString() {
+    StringBuilder sb = new StringBuilder();
+    sb.append("rest ");
+    sb.append(restVersion);
+    sb.append(" [JVM: ");
+    sb.append(jvmVersion);
+    sb.append("] [OS: ");
+    sb.append(osVersion);
+    sb.append("] [Server: ");
+    sb.append(serverVersion);
+    sb.append("] [Jersey: ");
     sb.append(jerseyVersion);
-	  sb.append("]\n");
-	  return sb.toString();
-	}
+    sb.append("]\n");
+    return sb.toString();
+  }
 
-	@Override
+  @Override
   public byte[] createProtobufOutput() {
-	  Version.Builder builder = Version.newBuilder();
-	  builder.setRestVersion(restVersion);
-	  builder.setJvmVersion(jvmVersion);
-	  builder.setOsVersion(osVersion);
-	  builder.setServerVersion(serverVersion);
-	  builder.setJerseyVersion(jerseyVersion);
-	  return builder.build().toByteArray();
+    Version.Builder builder = Version.newBuilder();
+    builder.setRestVersion(restVersion);
+    builder.setJvmVersion(jvmVersion);
+    builder.setOsVersion(osVersion);
+    builder.setServerVersion(serverVersion);
+    builder.setJerseyVersion(jerseyVersion);
+    return builder.build().toByteArray();
   }
 
   @Override

http://git-wip-us.apache.org/repos/asf/hbase/blob/b12d5778/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java
----------------------------------------------------------------------
diff --git a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java
index 3b8b8ca..216f830 100644
--- a/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java
+++ b/hbase-rest/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java
@@ -55,12 +55,12 @@ import com.sun.jersey.api.json.JSONJAXBContext;
 @InterfaceAudience.Private
 public class JAXBContextResolver implements ContextResolver<JAXBContext> {
 
-	private final JAXBContext context;
+  private final JAXBContext context;
 
-	private final Set<Class<?>> types;
+  private final Set<Class<?>> types;
 
-	private final Class<?>[] cTypes = {
-	  CellModel.class,
+  private final Class<?>[] cTypes = {
+    CellModel.class,
     CellSetModel.class,
     ColumnSchemaModel.class,
     RowModel.class,
@@ -68,22 +68,22 @@ public class JAXBContextResolver implements ContextResolver<JAXBContext> {
     StorageClusterStatusModel.class,
     StorageClusterVersionModel.class,
     TableInfoModel.class,
-	  TableListModel.class,
-	  TableModel.class,
-	  TableRegionModel.class,
-	  TableSchemaModel.class,
-	  VersionModel.class
-	};
+    TableListModel.class,
+    TableModel.class,
+    TableRegionModel.class,
+    TableSchemaModel.class,
+    VersionModel.class
+  };
 
-	@SuppressWarnings("unchecked")
+  @SuppressWarnings("unchecked")
   public JAXBContextResolver() throws Exception {
-		this.types = new HashSet(Arrays.asList(cTypes));
-		this.context = new JSONJAXBContext(JSONConfiguration.natural().build(),
-		  cTypes);
-	}
+    this.types = new HashSet(Arrays.asList(cTypes));
+    this.context = new JSONJAXBContext(JSONConfiguration.natural().build(),
+      cTypes);
+  }
 
-	@Override
-	public JAXBContext getContext(Class<?> objectType) {
-		return (types.contains(objectType)) ? context : null;
+  @Override
+  public JAXBContext getContext(Class<?> objectType) {
+    return (types.contains(objectType)) ? context : null;
   }
 }


Mime
View raw message