hbase-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From st...@apache.org
Subject [2/2] git commit: HBASE-12038 Replace internal uses of signatures with byte[] and String tableNames to use the TableName equivalents (Solomon Duskis)
Date Tue, 30 Sep 2014 23:36:05 GMT
HBASE-12038 Replace internal uses of signatures with byte[] and String tableNames to use the TableName equivalents (Solomon Duskis)


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/b5783795
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/b5783795
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/b5783795

Branch: refs/heads/master
Commit: b5783795cded0f5d52dd9c524d964ec6efc0046a
Parents: c4107d5
Author: stack <stack@apache.org>
Authored: Tue Sep 30 16:35:55 2014 -0700
Committer: stack <stack@apache.org>
Committed: Tue Sep 30 16:35:55 2014 -0700

----------------------------------------------------------------------
 .../hadoop/hbase/client/HTableFactory.java      |  3 +-
 .../security/access/AccessControlClient.java    |  6 ++--
 .../security/visibility/VisibilityClient.java   |  6 ++--
 .../example/TestBulkDeleteProtocol.java         | 26 ++++++++---------
 .../example/TestRowCountEndpoint.java           |  4 +--
 .../hadoop/hbase/mapred/HRegionPartitioner.java |  3 +-
 .../hadoop/hbase/mapred/TableInputFormat.java   |  4 ++-
 .../hadoop/hbase/mapred/TableOutputFormat.java  |  3 +-
 .../mapreduce/MultiTableInputFormatBase.java    |  7 +++--
 .../hbase/mapreduce/MultiTableOutputFormat.java |  3 +-
 .../hbase/mapreduce/TableMapReduceUtil.java     | 30 ++++++++++++++++++++
 .../org/apache/hadoop/hbase/tool/Canary.java    |  4 +--
 .../hadoop/hbase/util/RegionSplitter.java       |  8 +++---
 .../hadoop/hbase/HBaseTestingUtility.java       | 16 +++++------
 .../apache/hadoop/hbase/TestAcidGuarantees.java |  2 +-
 .../apache/hadoop/hbase/TestMultiVersions.java  |  4 +--
 .../apache/hadoop/hbase/client/TestAdmin.java   | 20 ++++++-------
 .../hadoop/hbase/client/TestFromClientSide.java |  7 ++---
 .../org/apache/hadoop/hbase/client/TestHCM.java | 18 ++++++------
 .../hadoop/hbase/client/TestMultiParallel.java  |  2 +-
 .../hbase/coprocessor/TestClassLoading.java     |  2 +-
 .../hbase/filter/TestFilterWithScanLimits.java  |  4 +--
 .../hadoop/hbase/filter/TestFilterWrapper.java  |  4 +--
 .../hbase/mapreduce/TestHFileOutputFormat.java  |  6 ++--
 .../hbase/mapreduce/TestHFileOutputFormat2.java |  4 +--
 .../TestSecureLoadIncrementalHFiles.java        |  2 +-
 ...ecureLoadIncrementalHFilesSplitRecovery.java |  2 +-
 .../hbase/mapreduce/TestTableMapReduceBase.java |  2 +-
 .../hbase/mapreduce/TestTimeRangeMapRed.java    |  6 ++--
 .../TestMasterOperationsForRegionReplicas.java  |  8 +++---
 .../hbase/master/TestRegionPlacement.java       |  6 ++--
 .../hadoop/hbase/master/TestRestartCluster.java |  6 ++--
 .../hbase/master/TestTableLockManager.java      |  2 +-
 .../hadoop/hbase/quotas/TestQuotaAdmin.java     |  2 +-
 .../hadoop/hbase/quotas/TestQuotaTableUtil.java |  2 +-
 .../hadoop/hbase/quotas/TestQuotaThrottle.java  |  2 +-
 .../regionserver/TestEncryptionKeyRotation.java |  2 +-
 .../hbase/regionserver/TestFSErrorsExposed.java |  4 +--
 .../hbase/regionserver/TestJoinedScanners.java  |  4 +--
 .../replication/TestMasterReplication.java      |  6 ++--
 .../replication/TestMultiSlaveReplication.java  |  4 +--
 .../replication/TestPerTableCFReplication.java  | 10 +++----
 .../hbase/replication/TestReplicationBase.java  |  4 +--
 .../replication/TestReplicationSmallTests.java  |  2 +-
 .../replication/TestReplicationSyncUpTool.java  | 12 ++++----
 .../replication/TestReplicationWithTags.java    |  4 +--
 .../hbase/security/access/SecureTestUtil.java   |  6 ++--
 .../security/access/TestAccessController.java   |  4 +--
 .../security/access/TestAccessController2.java  |  4 +--
 .../access/TestCellACLWithMultipleVersions.java |  4 +--
 .../hbase/security/access/TestCellACLs.java     |  4 +--
 .../access/TestScanEarlyTermination.java        |  4 +--
 .../security/access/TestTablePermissions.java   |  2 +-
 .../visibility/TestVisibilityLabelsWithACL.java |  8 +++---
 ...ibilityLabelsWithDefaultVisLabelService.java |  2 +-
 .../snapshot/TestSecureExportSnapshot.java      |  2 +-
 .../hbase/util/TestProcessBasedCluster.java     |  2 +-
 .../hadoop/hbase/util/TestRegionSplitter.java   |  6 ++--
 .../hadoop/hbase/thrift2/TestHTablePool.java    | 24 ++++++++--------
 59 files changed, 198 insertions(+), 162 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableFactory.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableFactory.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableFactory.java
index c38c386..d053e66 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableFactory.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableFactory.java
@@ -18,6 +18,7 @@
  */
 package org.apache.hadoop.hbase.client;
 
+import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.classification.InterfaceAudience;
 import org.apache.hadoop.hbase.classification.InterfaceStability;
 import org.apache.hadoop.conf.Configuration;
@@ -37,7 +38,7 @@ public class HTableFactory implements HTableInterfaceFactory {
   public HTableInterface createHTableInterface(Configuration config,
       byte[] tableName) {
     try {
-      return new HTable(config, tableName);
+      return new HTable(config, TableName.valueOf(tableName));
     } catch (IOException ioe) {
       throw new RuntimeException(ioe);
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java
index a2afe55..6a743da 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlClient.java
@@ -76,7 +76,7 @@ public class AccessControlClient {
     try {
       TableName aclTableName =
           TableName.valueOf(NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR, "acl");
-      ht = new HTable(conf, aclTableName.getName());
+      ht = new HTable(conf, aclTableName);
       Batch.Call<AccessControlService, GrantResponse> callable =
           new Batch.Call<AccessControlService, GrantResponse>() {
         ServerRpcController controller = new ServerRpcController();
@@ -156,7 +156,7 @@ public class AccessControlClient {
     try {
       TableName aclTableName = TableName.valueOf(NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR,
           "acl");
-      ht = new HTable(conf, aclTableName.getName());
+      ht = new HTable(conf, aclTableName);
       Batch.Call<AccessControlService, AccessControlProtos.RevokeResponse> callable =
           new Batch.Call<AccessControlService, AccessControlProtos.RevokeResponse>() {
         ServerRpcController controller = new ServerRpcController();
@@ -219,7 +219,7 @@ public class AccessControlClient {
       TableName aclTableName =
           TableName.valueOf(NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR, "acl");
       ha = new HBaseAdmin(conf);
-      ht = new HTable(conf, aclTableName.getName());
+      ht = new HTable(conf, aclTableName);
       CoprocessorRpcChannel service = ht.coprocessorService(HConstants.EMPTY_START_ROW);
       BlockingInterface protocol =
           AccessControlProtos.AccessControlService.newBlockingStub(service);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java
index a56ddc3..6ac1d3f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/VisibilityClient.java
@@ -75,7 +75,7 @@ public class VisibilityClient {
       throws Throwable {
     Table ht = null;
     try {
-      ht = new HTable(conf, LABELS_TABLE_NAME.getName());
+      ht = new HTable(conf, LABELS_TABLE_NAME);
       Batch.Call<VisibilityLabelsService, VisibilityLabelsResponse> callable = 
           new Batch.Call<VisibilityLabelsService, VisibilityLabelsResponse>() {
         ServerRpcController controller = new ServerRpcController();
@@ -129,7 +129,7 @@ public class VisibilityClient {
   public static GetAuthsResponse getAuths(Configuration conf, final String user) throws Throwable {
     Table ht = null;
     try {
-      ht = new HTable(conf, LABELS_TABLE_NAME.getName());
+      ht = new HTable(conf, LABELS_TABLE_NAME);
       Batch.Call<VisibilityLabelsService, GetAuthsResponse> callable = 
           new Batch.Call<VisibilityLabelsService, GetAuthsResponse>() {
         ServerRpcController controller = new ServerRpcController();
@@ -171,7 +171,7 @@ public class VisibilityClient {
       final String user, final boolean setOrClear) throws IOException, ServiceException, Throwable {
     Table ht = null;
     try {
-      ht = new HTable(conf, LABELS_TABLE_NAME.getName());
+      ht = new HTable(conf, LABELS_TABLE_NAME);
       Batch.Call<VisibilityLabelsService, VisibilityLabelsResponse> callable = 
           new Batch.Call<VisibilityLabelsService, VisibilityLabelsResponse>() {
         ServerRpcController controller = new ServerRpcController();

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestBulkDeleteProtocol.java
----------------------------------------------------------------------
diff --git a/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestBulkDeleteProtocol.java b/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestBulkDeleteProtocol.java
index 391360f..87e655e 100644
--- a/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestBulkDeleteProtocol.java
+++ b/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestBulkDeleteProtocol.java
@@ -79,7 +79,7 @@ public class TestBulkDeleteProtocol {
 
   // @Ignore @Test
   public void testBulkDeleteEndpoint() throws Throwable {
-    byte[] tableName = Bytes.toBytes("testBulkDeleteEndpoint");
+    TableName tableName = TableName.valueOf("testBulkDeleteEndpoint");
     Table ht = createTable(tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
@@ -102,8 +102,8 @@ public class TestBulkDeleteProtocol {
   // @Ignore @Test
   public void testBulkDeleteEndpointWhenRowBatchSizeLessThanRowsToDeleteFromARegion()
       throws Throwable {
-    byte[] tableName = Bytes
-        .toBytes("testBulkDeleteEndpointWhenRowBatchSizeLessThanRowsToDeleteFromARegion");
+    TableName tableName = TableName
+        .valueOf("testBulkDeleteEndpointWhenRowBatchSizeLessThanRowsToDeleteFromARegion");
     Table ht = createTable(tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
@@ -123,7 +123,7 @@ public class TestBulkDeleteProtocol {
     ht.close();
   }
 
-  private long invokeBulkDeleteProtocol(byte[] tableName, final Scan scan, final int rowBatchSize,
+  private long invokeBulkDeleteProtocol(TableName tableName, final Scan scan, final int rowBatchSize,
       final DeleteType deleteType, final Long timeStamp) throws Throwable {
     Table ht = new HTable(TEST_UTIL.getConfiguration(), tableName);
     long noOfDeletedRows = 0L;
@@ -156,7 +156,7 @@ public class TestBulkDeleteProtocol {
 
   // @Ignore @Test
   public void testBulkDeleteWithConditionBasedDelete() throws Throwable {
-    byte[] tableName = Bytes.toBytes("testBulkDeleteWithConditionBasedDelete");
+    TableName tableName = TableName.valueOf("testBulkDeleteWithConditionBasedDelete");
     Table ht = createTable(tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
@@ -186,7 +186,7 @@ public class TestBulkDeleteProtocol {
 
   // @Ignore @Test
   public void testBulkDeleteColumn() throws Throwable {
-    byte[] tableName = Bytes.toBytes("testBulkDeleteColumn");
+    TableName tableName = TableName.valueOf("testBulkDeleteColumn");
     Table ht = createTable(tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
@@ -215,8 +215,8 @@ public class TestBulkDeleteProtocol {
 
   // @Ignore @Test
   public void testBulkDeleteFamily() throws Throwable {
-    byte[] tableName = Bytes.toBytes("testBulkDeleteFamily");
-    HTableDescriptor htd = new HTableDescriptor(TableName.valueOf(tableName));
+    TableName tableName = TableName.valueOf("testBulkDeleteFamily");
+    HTableDescriptor htd = new HTableDescriptor(tableName);
     htd.addFamily(new HColumnDescriptor(FAMILY1));
     htd.addFamily(new HColumnDescriptor(FAMILY2));
     TEST_UTIL.getHBaseAdmin().createTable(htd, Bytes.toBytes(0), Bytes.toBytes(120), 5);
@@ -246,7 +246,7 @@ public class TestBulkDeleteProtocol {
 
   // @Ignore @Test
   public void testBulkDeleteColumnVersion() throws Throwable {
-    byte[] tableName = Bytes.toBytes("testBulkDeleteColumnVersion");
+    TableName tableName = TableName.valueOf("testBulkDeleteColumnVersion");
     Table ht = createTable(tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
@@ -294,7 +294,7 @@ public class TestBulkDeleteProtocol {
 
   // @Ignore @Test
   public void testBulkDeleteColumnVersionBasedOnTS() throws Throwable {
-    byte[] tableName = Bytes.toBytes("testBulkDeleteColumnVersionBasedOnTS");
+    TableName tableName = TableName.valueOf("testBulkDeleteColumnVersionBasedOnTS");
     Table ht = createTable(tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
@@ -341,7 +341,7 @@ public class TestBulkDeleteProtocol {
 
   // @Ignore @Test
   public void testBulkDeleteWithNumberOfVersions() throws Throwable {
-    byte[] tableName = Bytes.toBytes("testBulkDeleteWithNumberOfVersions");
+    TableName tableName = TableName.valueOf("testBulkDeleteWithNumberOfVersions");
     Table ht = createTable(tableName);
     List<Put> puts = new ArrayList<Put>(100);
     for (int j = 0; j < 100; j++) {
@@ -424,8 +424,8 @@ public class TestBulkDeleteProtocol {
     ht.close();
   }
 
-  private Table createTable(byte[] tableName) throws IOException {
-    HTableDescriptor htd = new HTableDescriptor(TableName.valueOf(tableName));
+  private Table createTable(TableName tableName) throws IOException {
+    HTableDescriptor htd = new HTableDescriptor(tableName);
     HColumnDescriptor hcd = new HColumnDescriptor(FAMILY1);
     hcd.setMaxVersions(10);// Just setting 10 as I am not testing with more than 10 versions here
     htd.addFamily(hcd);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestRowCountEndpoint.java
----------------------------------------------------------------------
diff --git a/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestRowCountEndpoint.java b/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestRowCountEndpoint.java
index b28c247..ddc5847 100644
--- a/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestRowCountEndpoint.java
+++ b/hbase-examples/src/test/java/org/apache/hadoop/hbase/coprocessor/example/TestRowCountEndpoint.java
@@ -46,7 +46,7 @@ import static junit.framework.Assert.*;
  */
 @Category({CoprocessorTests.class, MediumTests.class})
 public class TestRowCountEndpoint {
-  private static final byte[] TEST_TABLE = Bytes.toBytes("testrowcounter");
+  private static final TableName TEST_TABLE = TableName.valueOf("testrowcounter");
   private static final byte[] TEST_FAMILY = Bytes.toBytes("f");
   private static final byte[] TEST_COLUMN = Bytes.toBytes("col");
 
@@ -61,7 +61,7 @@ public class TestRowCountEndpoint {
         RowCountEndpoint.class.getName());
 
     TEST_UTIL.startMiniCluster();
-    TEST_UTIL.createTable(TableName.valueOf(TEST_TABLE), new byte[][]{TEST_FAMILY});
+    TEST_UTIL.createTable(TEST_TABLE, new byte[][]{TEST_FAMILY});
   }
 
   // @Ignore @AfterClass

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/HRegionPartitioner.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/HRegionPartitioner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/HRegionPartitioner.java
index 60a2c99..11acea0 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/HRegionPartitioner.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/HRegionPartitioner.java
@@ -25,6 +25,7 @@ import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.hbase.classification.InterfaceAudience;
 import org.apache.hadoop.hbase.classification.InterfaceStability;
 import org.apache.hadoop.hbase.HBaseConfiguration;
+import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.RegionLocator;
 import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
@@ -52,7 +53,7 @@ implements Partitioner<ImmutableBytesWritable, V2> {
   public void configure(JobConf job) {
     try {
       this.table = new HTable(HBaseConfiguration.create(job),
-        job.get(TableOutputFormat.OUTPUT_TABLE));
+        TableName.valueOf(job.get(TableOutputFormat.OUTPUT_TABLE)));
     } catch (IOException e) {
       LOG.error(e);
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableInputFormat.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableInputFormat.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableInputFormat.java
index 6c96f86..e1220fb 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableInputFormat.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableInputFormat.java
@@ -26,6 +26,7 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
 import org.apache.hadoop.hbase.classification.InterfaceStability;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hbase.HBaseConfiguration;
+import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.hadoop.mapred.FileInputFormat;
@@ -57,7 +58,8 @@ public class TableInputFormat extends TableInputFormatBase implements
     }
     setInputColumns(m_cols);
     try {
-      setHTable(new HTable(HBaseConfiguration.create(job), tableNames[0].getName()));
+      setHTable(
+          new HTable(HBaseConfiguration.create(job), TableName.valueOf(tableNames[0].getName())));
     } catch (Exception e) {
       LOG.error(StringUtils.stringifyException(e));
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableOutputFormat.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableOutputFormat.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableOutputFormat.java
index 0f7cded..5a5f544 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableOutputFormat.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapred/TableOutputFormat.java
@@ -26,6 +26,7 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
 import org.apache.hadoop.hbase.classification.InterfaceStability;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.hbase.HBaseConfiguration;
+import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Put;
 import org.apache.hadoop.hbase.client.Table;
@@ -85,7 +86,7 @@ FileOutputFormat<ImmutableBytesWritable, Put> {
 
     // expecting exactly one path
 
-    String tableName = job.get(OUTPUT_TABLE);
+    TableName tableName = TableName.valueOf(job.get(OUTPUT_TABLE));
     HTable table = null;
     try {
       table = new HTable(HBaseConfiguration.create(job), tableName);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java
index 761c46f..20169e0 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java
@@ -28,6 +28,7 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
 import org.apache.hadoop.hbase.classification.InterfaceStability;
 import org.apache.hadoop.hbase.HRegionInfo;
 import org.apache.hadoop.hbase.HRegionLocation;
+import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Result;
 import org.apache.hadoop.hbase.client.Scan;
@@ -80,13 +81,13 @@ public abstract class MultiTableInputFormatBase extends
     TableSplit tSplit = (TableSplit) split;
     LOG.info(MessageFormat.format("Input split length: {0} bytes.", tSplit.getLength()));
 
-    if (tSplit.getTableName() == null) {
+    if (tSplit.getTable() == null) {
       throw new IOException("Cannot create a record reader because of a"
           + " previous error. Please look at the previous logs lines from"
           + " the task's full log for more details.");
     }
     Table table =
-        new HTable(context.getConfiguration(), tSplit.getTableName());
+        new HTable(context.getConfiguration(), tSplit.getTable());
 
     TableRecordReader trr = this.tableRecordReader;
 
@@ -133,7 +134,7 @@ public abstract class MultiTableInputFormatBase extends
 
       HTable table = null;
       try {
-        table = new HTable(context.getConfiguration(), tableName);
+        table = new HTable(context.getConfiguration(), TableName.valueOf(tableName));
         Pair<byte[][], byte[][]> keys = table.getStartEndKeys();
         if (keys == null || keys.getFirst() == null ||
             keys.getFirst().length == 0) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java
index 5902a8e..c1d8373 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java
@@ -28,6 +28,7 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
 import org.apache.hadoop.hbase.classification.InterfaceStability;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hbase.HBaseConfiguration;
+import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.client.Delete;
 import org.apache.hadoop.hbase.client.HTable;
 import org.apache.hadoop.hbase.client.Mutation;
@@ -102,7 +103,7 @@ public class MultiTableOutputFormat extends OutputFormat<ImmutableBytesWritable,
     HTable getTable(ImmutableBytesWritable tableName) throws IOException {
       if (!tables.containsKey(tableName)) {
         LOG.debug("Opening HTable \"" + Bytes.toString(tableName.get())+ "\" for writing");
-        HTable table = new HTable(conf, tableName.get());
+        HTable table = new HTable(conf, TableName.valueOf(tableName.get()));
         table.setAutoFlush(false, true);
         tables.put(tableName, table);
       }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java
index 568384c..268f2b1 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java
@@ -44,6 +44,7 @@ import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.hbase.HBaseConfiguration;
 import org.apache.hadoop.hbase.HConstants;
 import org.apache.hadoop.hbase.MetaTableAccessor;
+import org.apache.hadoop.hbase.TableName;
 import org.apache.hadoop.hbase.client.Put;
 import org.apache.hadoop.hbase.client.Scan;
 import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
@@ -102,6 +103,35 @@ public class TableMapReduceUtil {
         job, true);
   }
 
+
+  /**
+   * Use this before submitting a TableMap job. It will appropriately set up
+   * the job.
+   *
+   * @param table  The table name to read from.
+   * @param scan  The scan instance with the columns, time range etc.
+   * @param mapper  The mapper class to use.
+   * @param outputKeyClass  The class of the output key.
+   * @param outputValueClass  The class of the output value.
+   * @param job  The current job to adjust.  Make sure the passed job is
+   * carrying all necessary HBase configuration.
+   * @throws IOException When setting up the details fails.
+   */
+  public static void initTableMapperJob(TableName table,
+      Scan scan,
+      Class<? extends TableMapper> mapper,
+      Class<?> outputKeyClass,
+      Class<?> outputValueClass,
+      Job job) throws IOException {
+    initTableMapperJob(table.getNameAsString(),
+        scan,
+        mapper,
+        outputKeyClass,
+        outputValueClass,
+        job,
+        true);
+  }
+
   /**
    * Use this before submitting a TableMap job. It will appropriately set up
    * the job.

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java
index d5cdb39..e6e975f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java
@@ -480,7 +480,7 @@ public final class Canary implements Tool {
     Table table = null;
 
     try {
-      table = new HTable(admin.getConfiguration(), tableDesc.getName());
+      table = new HTable(admin.getConfiguration(), tableDesc.getTableName());
     } catch (TableNotFoundException e) {
       return;
     }
@@ -679,7 +679,7 @@ public final class Canary implements Tool {
         HTableDescriptor[] tableDescs = this.admin.listTables();
         List<HRegionInfo> regions = null;
         for (HTableDescriptor tableDesc : tableDescs) {
-          table = new HTable(this.admin.getConfiguration(), tableDesc.getName());
+          table = new HTable(this.admin.getConfiguration(), tableDesc.getTableName());
 
           for (Map.Entry<HRegionInfo, ServerName> entry : table
               .getRegionLocations().entrySet()) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java
index 9055e51..483da8f 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java
@@ -339,7 +339,7 @@ public class RegionSplitter {
 		  "UniformSplit treats keys as arbitrary bytes.", opt);
       return;
     }
-    String tableName = cmd.getArgs()[0];
+    TableName tableName = TableName.valueOf(cmd.getArgs()[0]);
     String splitClass = cmd.getArgs()[1];
     SplitAlgorithm splitAlgo = newSplitAlgoInstance(conf, splitClass);
 
@@ -363,7 +363,7 @@ public class RegionSplitter {
     }
   }
 
-  static void createPresplitTable(String tableName, SplitAlgorithm splitAlgo,
+  static void createPresplitTable(TableName tableName, SplitAlgorithm splitAlgo,
           String[] columnFamilies, Configuration conf) throws IOException,
           InterruptedException {
     final int splitCount = conf.getInt("split.count", 0);
@@ -374,7 +374,7 @@ public class RegionSplitter {
     LOG.debug("Creating table " + tableName + " with " + columnFamilies.length
         + " column families.  Presplitting to " + splitCount + " regions");
 
-    HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(tableName));
+    HTableDescriptor desc = new HTableDescriptor(tableName);
     for (String cf : columnFamilies) {
       desc.addFamily(new HColumnDescriptor(Bytes.toBytes(cf)));
     }
@@ -399,7 +399,7 @@ public class RegionSplitter {
     LOG.debug("Finished creating table with " + splitCount + " regions");
   }
 
-  static void rollingSplit(String tableName, SplitAlgorithm splitAlgo,
+  static void rollingSplit(TableName tableName, SplitAlgorithm splitAlgo,
           Configuration conf) throws IOException, InterruptedException {
     final int minOS = conf.getInt("split.outstanding", 2);
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java
index de530aa..296ef4e 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java
@@ -1229,7 +1229,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility {
       desc.addFamily(hcd);
     }
     getHBaseAdmin().createTable(desc);
-    return new HTable(c, tableName);
+    return new HTable(c, desc.getTableName());
   }
 
   /**
@@ -1275,7 +1275,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility {
       desc.addFamily(hcd);
     }
     getHBaseAdmin().createTable(desc);
-    return new HTable(c, tableName);
+    return new HTable(c, desc.getTableName());
   }
 
   /**
@@ -1465,8 +1465,8 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility {
     }
     getHBaseAdmin().createTable(desc, splitRows);
     // HBaseAdmin only waits for regions to appear in hbase:meta we should wait until they are assigned
-    waitUntilAllRegionsAssigned(TableName.valueOf(tableName));
-    return new HTable(getConfiguration(), tableName);
+    waitUntilAllRegionsAssigned(desc.getTableName());
+    return new HTable(getConfiguration(), desc.getTableName());
   }
 
   /**
@@ -2713,9 +2713,9 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility {
    * @throws InterruptedException
    * @throws IOException
    */
-  public void waitTableAvailable(byte[] table)
+  public void waitTableAvailable(TableName table)
       throws InterruptedException, IOException {
-    waitTableAvailable(getHBaseAdmin(), table, 30000);
+    waitTableAvailable(getHBaseAdmin(), table.getName(), 30000);
   }
 
   public void waitTableAvailable(Admin admin, byte[] table)
@@ -2755,9 +2755,9 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility {
    * @throws InterruptedException
    * @throws IOException
    */
-  public void waitTableEnabled(byte[] table)
+  public void waitTableEnabled(TableName table)
       throws InterruptedException, IOException {
-    waitTableEnabled(getHBaseAdmin(), table, 30000);
+    waitTableEnabled(getHBaseAdmin(), table.getName(), 30000);
   }
 
   public void waitTableEnabled(Admin admin, byte[] table)

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/TestAcidGuarantees.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestAcidGuarantees.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestAcidGuarantees.java
index 0e62d77..3bcb0bb 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestAcidGuarantees.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestAcidGuarantees.java
@@ -58,7 +58,7 @@ import com.google.common.collect.Lists;
 @Category({FlakeyTests.class, MediumTests.class})
 public class TestAcidGuarantees implements Tool {
   protected static final Log LOG = LogFactory.getLog(TestAcidGuarantees.class);
-  public static final byte [] TABLE_NAME = Bytes.toBytes("TestAcidGuarantees");
+  public static final TableName TABLE_NAME = TableName.valueOf("TestAcidGuarantees");
   public static final byte [] FAMILY_A = Bytes.toBytes("A");
   public static final byte [] FAMILY_B = Bytes.toBytes("B");
   public static final byte [] FAMILY_C = Bytes.toBytes("C");

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMultiVersions.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMultiVersions.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMultiVersions.java
index 6c70384..2a9b953 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMultiVersions.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMultiVersions.java
@@ -197,8 +197,8 @@ public class TestMultiVersions {
    */
   @Test
   public void testScanMultipleVersions() throws Exception {
-    final byte [] tableName = Bytes.toBytes("testScanMultipleVersions");
-    final HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(tableName));
+    final TableName tableName = TableName.valueOf("testScanMultipleVersions");
+    final HTableDescriptor desc = new HTableDescriptor(tableName);
     desc.addFamily(new HColumnDescriptor(HConstants.CATALOG_FAMILY));
     final byte [][] rows = new byte[][] {
       Bytes.toBytes("row_0200"),

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin.java
index 7f93780..08dab86 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin.java
@@ -615,8 +615,8 @@ public class TestAdmin {
 
   @Test (timeout=300000)
   public void testCreateTableNumberOfRegions() throws IOException, InterruptedException {
-    byte[] tableName = Bytes.toBytes("testCreateTableNumberOfRegions");
-    HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(tableName));
+    TableName tableName = TableName.valueOf("testCreateTableNumberOfRegions");
+    HTableDescriptor desc = new HTableDescriptor(tableName);
     desc.addFamily(new HColumnDescriptor(HConstants.CATALOG_FAMILY));
     admin.createTable(desc);
     HTable ht = new HTable(TEST_UTIL.getConfiguration(), tableName);
@@ -624,8 +624,8 @@ public class TestAdmin {
     assertEquals("Table should have only 1 region", 1, regions.size());
     ht.close();
 
-    byte [] TABLE_2 = Bytes.add(tableName, Bytes.toBytes("_2"));
-    desc = new HTableDescriptor(TableName.valueOf(TABLE_2));
+    TableName TABLE_2 = TableName.valueOf(tableName.getNameAsString() + "_2");
+    desc = new HTableDescriptor(TABLE_2);
     desc.addFamily(new HColumnDescriptor(HConstants.CATALOG_FAMILY));
     admin.createTable(desc, new byte[][]{new byte[]{42}});
     HTable ht2 = new HTable(TEST_UTIL.getConfiguration(), TABLE_2);
@@ -633,8 +633,8 @@ public class TestAdmin {
     assertEquals("Table should have only 2 region", 2, regions.size());
     ht2.close();
 
-    byte [] TABLE_3 = Bytes.add(tableName, Bytes.toBytes("_3"));
-    desc = new HTableDescriptor(TableName.valueOf(TABLE_3));
+    TableName TABLE_3 = TableName.valueOf(tableName.getNameAsString() + "_3");
+    desc = new HTableDescriptor(TABLE_3);
     desc.addFamily(new HColumnDescriptor(HConstants.CATALOG_FAMILY));
     admin.createTable(desc, "a".getBytes(), "z".getBytes(), 3);
     HTable ht3 = new HTable(TEST_UTIL.getConfiguration(), TABLE_3);
@@ -642,8 +642,8 @@ public class TestAdmin {
     assertEquals("Table should have only 3 region", 3, regions.size());
     ht3.close();
 
-    byte [] TABLE_4 = Bytes.add(tableName, Bytes.toBytes("_4"));
-    desc = new HTableDescriptor(TableName.valueOf(TABLE_4));
+    TableName TABLE_4 = TableName.valueOf(tableName.getNameAsString() + "_4");
+    desc = new HTableDescriptor(TABLE_4);
     desc.addFamily(new HColumnDescriptor(HConstants.CATALOG_FAMILY));
     try {
       admin.createTable(desc, "a".getBytes(), "z".getBytes(), 2);
@@ -652,8 +652,8 @@ public class TestAdmin {
     // Expected
     }
 
-    byte [] TABLE_5 = Bytes.add(tableName, Bytes.toBytes("_5"));
-    desc = new HTableDescriptor(TableName.valueOf(TABLE_5));
+    TableName TABLE_5 = TableName.valueOf(tableName.getNameAsString() + "_5");
+    desc = new HTableDescriptor(TABLE_5);
     desc.addFamily(new HColumnDescriptor(HConstants.CATALOG_FAMILY));
     admin.createTable(desc, new byte[] {1}, new byte[] {127}, 16);
     HTable ht5 = new HTable(TEST_UTIL.getConfiguration(), TABLE_5);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
index aa44647..bfece26 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
@@ -172,7 +172,7 @@ public class TestFromClientSide {
    */
    @Test
    public void testKeepDeletedCells() throws Exception {
-     final byte[] TABLENAME = Bytes.toBytes("testKeepDeletesCells");
+     final TableName TABLENAME = TableName.valueOf("testKeepDeletesCells");
      final byte[] FAMILY = Bytes.toBytes("family");
      final byte[] C0 = Bytes.toBytes("c0");
 
@@ -400,7 +400,7 @@ public class TestFromClientSide {
    */
   @Test
   public void testWeirdCacheBehaviour() throws Exception {
-    byte [] TABLE = Bytes.toBytes("testWeirdCacheBehaviour");
+    TableName TABLE = TableName.valueOf("testWeirdCacheBehaviour");
     byte [][] FAMILIES = new byte[][] { Bytes.toBytes("trans-blob"),
         Bytes.toBytes("trans-type"), Bytes.toBytes("trans-date"),
         Bytes.toBytes("trans-tags"), Bytes.toBytes("trans-group") };
@@ -4209,8 +4209,7 @@ public class TestFromClientSide {
 
     // Test that attribute changes were applied
     desc = a.getTableDescriptor();
-    assertTrue("wrong table descriptor returned",
-      desc.getTableName().equals(tableAname));
+    assertEquals("wrong table descriptor returned", desc.getTableName(), tableAname);
     // check HTD attribute
     value = desc.getValue(attrName);
     assertFalse("missing HTD attribute value", value == null);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java
index ffe1e19..3e29725 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java
@@ -236,7 +236,7 @@ public class TestHCM {
     final ServerName sn = rs.getRegionServer().getServerName();
 
     HTable t = TEST_UTIL.createTable(tn, cf);
-    TEST_UTIL.waitTableAvailable(tn.getName());
+    TEST_UTIL.waitTableAvailable(tn);
 
     while(TEST_UTIL.getHBaseCluster().getMaster().getAssignmentManager().
         getRegionStates().isRegionsInTransition()){
@@ -333,8 +333,8 @@ public class TestHCM {
 
 
   private void testConnectionClose(boolean allowsInterrupt) throws Exception {
-    String tableName = "HCM-testConnectionClose" + allowsInterrupt;
-    TEST_UTIL.createTable(tableName.getBytes(), FAM_NAM).close();
+    TableName tableName = TableName.valueOf("HCM-testConnectionClose" + allowsInterrupt);
+    TEST_UTIL.createTable(tableName, FAM_NAM).close();
 
     boolean previousBalance = TEST_UTIL.getHBaseAdmin().setBalancerRunning(false, true);
 
@@ -346,7 +346,7 @@ public class TestHCM {
     c2.setInt(RpcClient.FAILED_SERVER_EXPIRY_KEY, 0); // Server do not really expire
     c2.setBoolean(RpcClient.SPECIFIC_WRITE_THREAD, allowsInterrupt);
 
-    final HTable table = new HTable(c2, tableName.getBytes());
+    final HTable table = new HTable(c2, tableName);
 
     Put put = new Put(ROW);
     put.add(FAM_NAM, ROW, ROW);
@@ -415,8 +415,8 @@ public class TestHCM {
    */
   @Test
   public void testConnectionIdle() throws Exception {
-    String tableName = "HCM-testConnectionIdle";
-    TEST_UTIL.createTable(tableName.getBytes(), FAM_NAM).close();
+    TableName tableName = TableName.valueOf("HCM-testConnectionIdle");
+    TEST_UTIL.createTable(tableName, FAM_NAM).close();
     int idleTime =  20000;
     boolean previousBalance = TEST_UTIL.getHBaseAdmin().setBalancerRunning(false, true);
 
@@ -426,7 +426,7 @@ public class TestHCM {
     c2.setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 1); // Don't retry: retry = test failed
     c2.setInt(RpcClient.IDLE_TIME, idleTime);
 
-    final Table table = new HTable(c2, tableName.getBytes());
+    final Table table = new HTable(c2, tableName);
 
     Put put = new Put(ROW);
     put.add(FAM_NAM, ROW, ROW);
@@ -478,9 +478,9 @@ public class TestHCM {
       return;
     }
 
-    String tableName = "HCM-testConnectionCut";
+    TableName tableName = TableName.valueOf("HCM-testConnectionCut");
 
-    TEST_UTIL.createTable(tableName.getBytes(), FAM_NAM).close();
+    TEST_UTIL.createTable(tableName, FAM_NAM).close();
     boolean previousBalance = TEST_UTIL.getHBaseAdmin().setBalancerRunning(false, true);
 
     Configuration c2 = new Configuration(TEST_UTIL.getConfiguration());

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiParallel.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiParallel.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiParallel.java
index e4bf256..b911f6e 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiParallel.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiParallel.java
@@ -82,7 +82,7 @@ public class TestMultiParallel {
     UTIL.startMiniCluster(slaves);
     HTable t = UTIL.createTable(TEST_TABLE, Bytes.toBytes(FAMILY));
     UTIL.createMultiRegions(t, Bytes.toBytes(FAMILY));
-    UTIL.waitTableEnabled(TEST_TABLE.getName());
+    UTIL.waitTableEnabled(TEST_TABLE);
     t.close();
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestClassLoading.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestClassLoading.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestClassLoading.java
index accd764..2fe121d 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestClassLoading.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestClassLoading.java
@@ -533,7 +533,7 @@ public class TestClassLoading {
 
   private void waitForTable(TableName name) throws InterruptedException, IOException {
     // First wait until all regions are online
-    TEST_UTIL.waitTableEnabled(name.getName());
+    TEST_UTIL.waitTableEnabled(name);
     // Now wait a bit longer for the coprocessor hosts to load the CPs
     Thread.sleep(1000);
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWithScanLimits.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWithScanLimits.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWithScanLimits.java
index 504877d..fac66c1 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWithScanLimits.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWithScanLimits.java
@@ -71,7 +71,7 @@ public class TestFilterWithScanLimits {
   private static final HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility();
   private static Configuration conf = null;
   private static HBaseAdmin admin = null;
-  private static byte[] name = Bytes.toBytes("test");
+  private static TableName name = TableName.valueOf("test");
 
   @Test
   public void testScanWithLimit() {
@@ -141,7 +141,7 @@ public class TestFilterWithScanLimits {
     assertNotNull("HBaseAdmin is not initialized successfully.", admin);
     if (admin != null) {
 
-      HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(name));
+      HTableDescriptor desc = new HTableDescriptor(name);
       HColumnDescriptor coldef = new HColumnDescriptor(Bytes.toBytes("f1"));
       desc.addFamily(coldef);
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java
index 615fabb..1cffe1d 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java
@@ -65,7 +65,7 @@ public class TestFilterWrapper {
   private static final HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility();
   private static Configuration conf = null;
   private static HBaseAdmin admin = null;
-  private static byte[] name = Bytes.toBytes("test");
+  private static TableName name = TableName.valueOf("test");
 
   @Test
   public void testFilterWrapper() {
@@ -144,7 +144,7 @@ public class TestFilterWrapper {
     assertNotNull("HBaseAdmin is not initialized successfully.", admin);
     if (admin != null) {
 
-      HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(name));
+      HTableDescriptor desc = new HTableDescriptor(name);
       HColumnDescriptor coldef = new HColumnDescriptor(Bytes.toBytes("f1"));
       desc.addFamily(coldef);
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat.java
index 8f6e167..9bdebe6 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat.java
@@ -449,7 +449,7 @@ public class TestHFileOutputFormat  {
         LOG.info("Waiting for table to disable");
       }
       admin.enableTable(TABLE_NAME);
-      util.waitTableAvailable(TABLE_NAME.getName());
+      util.waitTableAvailable(TABLE_NAME);
       assertEquals("Data should remain after reopening of regions",
           tableDigestBefore, util.checksumRows(table));
     } finally {
@@ -1049,7 +1049,7 @@ public class TestHFileOutputFormat  {
     Configuration conf = HBaseConfiguration.create();
     util = new HBaseTestingUtility(conf);
     if ("newtable".equals(args[0])) {
-      byte[] tname = args[1].getBytes();
+      TableName tname = TableName.valueOf(args[1]);
       HTable table = util.createTable(tname, FAMILIES);
       HBaseAdmin admin = new HBaseAdmin(conf);
       admin.disableTable(tname);
@@ -1057,7 +1057,7 @@ public class TestHFileOutputFormat  {
       util.createMultiRegions(conf, table, FAMILIES[0], startKeys);
       admin.enableTable(tname);
     } else if ("incremental".equals(args[0])) {
-      byte[] tname = args[1].getBytes();
+      TableName tname = TableName.valueOf(args[1]);
       HTable table = new HTable(conf, tname);
       Path outDir = new Path("incremental-out");
       runIncrementalPELoad(conf, table, outDir);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat2.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat2.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat2.java
index c14b9a6..6ab675f 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat2.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestHFileOutputFormat2.java
@@ -448,7 +448,7 @@ public class TestHFileOutputFormat2  {
         LOG.info("Waiting for table to disable");
       }
       admin.enableTable(TABLE_NAME);
-      util.waitTableAvailable(TABLE_NAME.getName());
+      util.waitTableAvailable(TABLE_NAME);
       assertEquals("Data should remain after reopening of regions",
           tableDigestBefore, util.checksumRows(table));
     } finally {
@@ -1057,7 +1057,7 @@ public class TestHFileOutputFormat2  {
       util.createMultiRegions(conf, table, FAMILIES[0], startKeys);
       admin.enableTable(tname);
     } else if ("incremental".equals(args[0])) {
-      byte[] tname = args[1].getBytes();
+      TableName tname = TableName.valueOf(args[1]);
       HTable table = new HTable(conf, tname);
       Path outDir = new Path("incremental-out");
       runIncrementalPELoad(conf, table, outDir);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestSecureLoadIncrementalHFiles.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestSecureLoadIncrementalHFiles.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestSecureLoadIncrementalHFiles.java
index fcd8216..e8aca29 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestSecureLoadIncrementalHFiles.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestSecureLoadIncrementalHFiles.java
@@ -56,7 +56,7 @@ public class TestSecureLoadIncrementalHFiles extends  TestLoadIncrementalHFiles{
     util.startMiniCluster();
 
     // Wait for the ACL table to become available
-    util.waitTableEnabled(AccessControlLists.ACL_TABLE_NAME.getName());
+    util.waitTableEnabled(AccessControlLists.ACL_TABLE_NAME);
 
     setupNamespace();
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestSecureLoadIncrementalHFilesSplitRecovery.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestSecureLoadIncrementalHFilesSplitRecovery.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestSecureLoadIncrementalHFilesSplitRecovery.java
index 76a6d57..fe9c132 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestSecureLoadIncrementalHFilesSplitRecovery.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestSecureLoadIncrementalHFilesSplitRecovery.java
@@ -58,7 +58,7 @@ public class TestSecureLoadIncrementalHFilesSplitRecovery extends TestLoadIncrem
     util.startMiniCluster();
 
     // Wait for the ACL table to become available
-    util.waitTableEnabled(AccessControlLists.ACL_TABLE_NAME.getName());
+    util.waitTableEnabled(AccessControlLists.ACL_TABLE_NAME);
   }
 
   //Disabling this test as it does not work in secure mode

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduceBase.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduceBase.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduceBase.java
index c218bc9..c82d82f 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduceBase.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableMapReduceBase.java
@@ -54,7 +54,7 @@ import org.junit.Test;
 public abstract class TestTableMapReduceBase {
 
   protected static final HBaseTestingUtility UTIL = new HBaseTestingUtility();
-  protected static final byte[] MULTI_REGION_TABLE_NAME = Bytes.toBytes("mrtest");
+  protected static final TableName MULTI_REGION_TABLE_NAME = TableName.valueOf("mrtest");
   protected static final byte[] INPUT_FAMILY = Bytes.toBytes("contents");
   protected static final byte[] OUTPUT_FAMILY = Bytes.toBytes("text");
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTimeRangeMapRed.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTimeRangeMapRed.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTimeRangeMapRed.java
index 58d1b6e..b701c35 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTimeRangeMapRed.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTimeRangeMapRed.java
@@ -78,7 +78,7 @@ public class TestTimeRangeMapRed {
   static final long MINSTAMP = 1245620005;
   static final long MAXSTAMP = 1245620100 + 1; // maxStamp itself is excluded. so increment it.
 
-  static final byte[] TABLE_NAME = Bytes.toBytes("table123");
+  static final TableName TABLE_NAME = TableName.valueOf("table123");
   static final byte[] FAMILY_NAME = Bytes.toBytes("text");
   static final byte[] COLUMN_NAME = Bytes.toBytes("input");
 
@@ -146,7 +146,7 @@ public class TestTimeRangeMapRed {
   @Test
   public void testTimeRangeMapRed()
   throws IOException, InterruptedException, ClassNotFoundException {
-    final HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(TABLE_NAME));
+    final HTableDescriptor desc = new HTableDescriptor(TABLE_NAME);
     final HColumnDescriptor col = new HColumnDescriptor(FAMILY_NAME);
     col.setMaxVersions(Integer.MAX_VALUE);
     desc.addFamily(col);
@@ -179,7 +179,7 @@ public class TestTimeRangeMapRed {
       scan.addColumn(FAMILY_NAME, COLUMN_NAME);
       scan.setTimeRange(MINSTAMP, MAXSTAMP);
       scan.setMaxVersions();
-      TableMapReduceUtil.initTableMapperJob(Bytes.toString(TABLE_NAME),
+      TableMapReduceUtil.initTableMapperJob(TABLE_NAME,
         scan, ProcessTimeRangeMapper.class, Text.class, Text.class, job);
       job.waitForCompletion(true);
     } catch (IOException e) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterOperationsForRegionReplicas.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterOperationsForRegionReplicas.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterOperationsForRegionReplicas.java
index 8d38bdd..4e6fe59 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterOperationsForRegionReplicas.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterOperationsForRegionReplicas.java
@@ -113,7 +113,7 @@ public class TestMasterOperationsForRegionReplicas {
       desc.setRegionReplication(numReplica);
       desc.addFamily(new HColumnDescriptor("family"));
       admin.createTable(desc, Bytes.toBytes("A"), Bytes.toBytes("Z"), numRegions);
-      TEST_UTIL.waitTableEnabled(table.getName());
+      TEST_UTIL.waitTableEnabled(table);
       validateNumberOfRowsInMeta(table, numRegions, admin.getConnection());
 
       List<HRegionInfo> hris = MetaTableAccessor.getTableRegions(admin.getConnection(), table);
@@ -169,7 +169,7 @@ public class TestMasterOperationsForRegionReplicas {
       TEST_UTIL.getConfiguration().setBoolean("hbase.master.startup.retainassign", true);
       TEST_UTIL.shutdownMiniHBaseCluster();
       TEST_UTIL.startMiniHBaseCluster(1, numSlaves);
-      TEST_UTIL.waitTableEnabled(table.getName());
+      TEST_UTIL.waitTableEnabled(table);
       admin.close();
       admin = new HBaseAdmin(conf); 
       validateFromSnapshotFromMeta(TEST_UTIL, table, numRegions, numReplica,
@@ -179,7 +179,7 @@ public class TestMasterOperationsForRegionReplicas {
       // one server running
       TEST_UTIL.shutdownMiniHBaseCluster();
       TEST_UTIL.startMiniHBaseCluster(1, 1);
-      TEST_UTIL.waitTableEnabled(table.getName());
+      TEST_UTIL.waitTableEnabled(table);
       admin.close();
       admin = new HBaseAdmin(conf);
       validateSingleRegionServerAssignment(admin.getConnection(), numRegions, numReplica);
@@ -242,7 +242,7 @@ public class TestMasterOperationsForRegionReplicas {
       desc.setRegionReplication(numReplica);
       desc.addFamily(new HColumnDescriptor("family"));
       admin.createTable(desc, Bytes.toBytes("A"), Bytes.toBytes("Z"), numRegions);
-      TEST_UTIL.waitTableEnabled(table.getName());
+      TEST_UTIL.waitTableEnabled(table);
       Set<byte[]> tableRows = new HashSet<byte[]>();
       List<HRegionInfo> hris = MetaTableAccessor.getTableRegions(admin.getConnection(), table);
       for (HRegionInfo hri : hris) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRegionPlacement.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRegionPlacement.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRegionPlacement.java
index c607459..c34fb48 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRegionPlacement.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRegionPlacement.java
@@ -212,7 +212,7 @@ public class TestRegionPlacement {
   @Test
   public void testRegionPlacement() throws Exception {
     String tableStr = "testRegionAssignment";
-    byte[] table = Bytes.toBytes(tableStr);
+    TableName table = TableName.valueOf(tableStr);
     // Create a table with REGION_NUM regions.
     createTable(table, REGION_NUM);
 
@@ -667,7 +667,7 @@ public class TestRegionPlacement {
    * @return
    * @throws IOException
    */
-  private static void createTable(byte[] tableName, int regionNum)
+  private static void createTable(TableName tableName, int regionNum)
       throws IOException {
     int expectedRegions = regionNum;
     byte[][] splitKeys = new byte[expectedRegions - 1][];
@@ -676,7 +676,7 @@ public class TestRegionPlacement {
       splitKeys[i - 1] = new byte[] { splitKey, splitKey, splitKey };
     }
 
-    HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(tableName));
+    HTableDescriptor desc = new HTableDescriptor(tableName);
     desc.addFamily(new HColumnDescriptor(HConstants.CATALOG_FAMILY));
     admin.createTable(desc, splitKeys);
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRestartCluster.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRestartCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRestartCluster.java
index 0b1ccb2..74afe7e 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRestartCluster.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestRestartCluster.java
@@ -72,7 +72,7 @@ public class TestRestartCluster {
       UTIL.createTable(TABLE, FAMILY);
     }
     for(TableName TABLE : TABLES) {
-      UTIL.waitTableEnabled(TABLE.getName());
+      UTIL.waitTableEnabled(TABLE);
     }
 
     List<HRegionInfo> allRegions =
@@ -101,7 +101,7 @@ public class TestRestartCluster {
       } catch(TableExistsException tee) {
         LOG.info("Table already exists as expected");
       }
-      UTIL.waitTableAvailable(TABLE.getName());
+      UTIL.waitTableAvailable(TABLE);
     }
   }
 
@@ -122,7 +122,7 @@ public class TestRestartCluster {
       UTIL.createTable(TABLE, FAMILY);
     }
     for(TableName TABLE : TABLES) {
-      UTIL.waitTableEnabled(TABLE.getName());
+      UTIL.waitTableEnabled(TABLE);
     }
 
     HMaster master = UTIL.getMiniHBaseCluster().getMaster();

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestTableLockManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestTableLockManager.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestTableLockManager.java
index 5d24368..54f0691 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestTableLockManager.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestTableLockManager.java
@@ -394,7 +394,7 @@ public class TestTableLockManager {
 
     alterThread.start();
     splitThread.start();
-    TEST_UTIL.waitTableEnabled(tableName.toBytes());
+    TEST_UTIL.waitTableEnabled(tableName);
     while (true) {
       List<HRegionInfo> regions = admin.getTableRegions(tableName);
       LOG.info(String.format("Table #regions: %d regions: %s:", regions.size(), regions));

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaAdmin.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaAdmin.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaAdmin.java
index 622cac2..18dd5ae 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaAdmin.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaAdmin.java
@@ -67,7 +67,7 @@ public class TestQuotaAdmin {
     TEST_UTIL.getConfiguration().setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 6);
     TEST_UTIL.getConfiguration().setBoolean("hbase.master.enabletable.roundrobin", true);
     TEST_UTIL.startMiniCluster(1);
-    TEST_UTIL.waitTableAvailable(QuotaTableUtil.QUOTA_TABLE_NAME.getName());
+    TEST_UTIL.waitTableAvailable(QuotaTableUtil.QUOTA_TABLE_NAME);
   }
 
   @AfterClass

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaTableUtil.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaTableUtil.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaTableUtil.java
index d4185a1..84d9155 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaTableUtil.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaTableUtil.java
@@ -58,7 +58,7 @@ public class TestQuotaTableUtil {
     TEST_UTIL.getConfiguration().setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 6);
     TEST_UTIL.getConfiguration().setBoolean("hbase.master.enabletable.roundrobin", true);
     TEST_UTIL.startMiniCluster(1);
-    TEST_UTIL.waitTableAvailable(QuotaTableUtil.QUOTA_TABLE_NAME.getName());
+    TEST_UTIL.waitTableAvailable(QuotaTableUtil.QUOTA_TABLE_NAME);
   }
 
   @AfterClass

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaThrottle.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaThrottle.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaThrottle.java
index 8dbb76e..0901d2f 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaThrottle.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/quotas/TestQuotaThrottle.java
@@ -75,7 +75,7 @@ public class TestQuotaThrottle {
     TEST_UTIL.getConfiguration().setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 6);
     TEST_UTIL.getConfiguration().setBoolean("hbase.master.enabletable.roundrobin", true);
     TEST_UTIL.startMiniCluster(1);
-    TEST_UTIL.waitTableAvailable(QuotaTableUtil.QUOTA_TABLE_NAME.getName());
+    TEST_UTIL.waitTableAvailable(QuotaTableUtil.QUOTA_TABLE_NAME);
     QuotaCache.TEST_FORCE_REFRESH = true;
 
     tables = new HTable[TABLE_NAMES.length];

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionKeyRotation.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionKeyRotation.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionKeyRotation.java
index 4d04dc6..941f6d2 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionKeyRotation.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestEncryptionKeyRotation.java
@@ -208,7 +208,7 @@ public class TestEncryptionKeyRotation {
     TEST_UTIL.getHBaseAdmin().createTable(htd);
     TEST_UTIL.waitTableAvailable(htd.getName(), 5000);
     // Create a store file
-    Table table = new HTable(conf, htd.getName());
+    Table table = new HTable(conf, htd.getTableName());
     try {
       table.put(new Put(Bytes.toBytes("testrow"))
         .add(hcd.getName(), Bytes.toBytes("q"), Bytes.toBytes("value")));

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestFSErrorsExposed.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestFSErrorsExposed.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestFSErrorsExposed.java
index f01b089..9a2cc82 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestFSErrorsExposed.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestFSErrorsExposed.java
@@ -187,11 +187,11 @@ public class TestFSErrorsExposed {
       util.getConfiguration().setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 3);
 
       util.startMiniCluster(1);
-      byte[] tableName = Bytes.toBytes("table");
+      TableName tableName = TableName.valueOf("table");
       byte[] fam = Bytes.toBytes("fam");
 
       Admin admin = new HBaseAdmin(util.getConfiguration());
-      HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(tableName));
+      HTableDescriptor desc = new HTableDescriptor(tableName);
       desc.addFamily(new HColumnDescriptor(fam)
           .setMaxVersions(1)
           .setBlockCacheEnabled(false)

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestJoinedScanners.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestJoinedScanners.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestJoinedScanners.java
index d56aaa0..b8e6382 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestJoinedScanners.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestJoinedScanners.java
@@ -93,8 +93,8 @@ public class TestJoinedScanners {
       cluster = htu.startMiniCluster(1, regionServersCount, dataNodeHosts);
       byte [][] families = {cf_essential, cf_joined};
 
-      byte[] tableName = Bytes.toBytes(this.getClass().getSimpleName());
-      HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(tableName));
+      TableName tableName = TableName.valueOf(this.getClass().getSimpleName());
+      HTableDescriptor desc = new HTableDescriptor(tableName);
       for(byte[] family : families) {
         HColumnDescriptor hcd = new HColumnDescriptor(family);
         hcd.setDataBlockEncoding(blockEncoding);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMasterReplication.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMasterReplication.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMasterReplication.java
index 1718fd1..a501af9 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMasterReplication.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMasterReplication.java
@@ -75,7 +75,7 @@ public class TestMasterReplication {
   private static final long SLEEP_TIME = 500;
   private static final int NB_RETRIES = 10;
 
-  private static final byte[] tableName = Bytes.toBytes("test");
+  private static final TableName tableName = TableName.valueOf("test");
   private static final byte[] famName = Bytes.toBytes("f");
   private static final byte[] row = Bytes.toBytes("row");
   private static final byte[] row1 = Bytes.toBytes("row1");
@@ -108,7 +108,7 @@ public class TestMasterReplication {
         CoprocessorHost.USER_REGION_COPROCESSOR_CONF_KEY,
         CoprocessorCounter.class.getName());
 
-    table = new HTableDescriptor(TableName.valueOf(tableName));
+    table = new HTableDescriptor(tableName);
     HColumnDescriptor fam = new HColumnDescriptor(famName);
     fam.setScope(HConstants.REPLICATION_SCOPE_GLOBAL);
     table.addFamily(fam);
@@ -346,7 +346,7 @@ public class TestMasterReplication {
   }
 
   @SuppressWarnings("resource")
-  private Table[] getHTablesOnClusters(byte[] tableName) throws Exception {
+  private Table[] getHTablesOnClusters(TableName tableName) throws Exception {
     int numClusters = utilities.length;
     Table[] htables = new Table[numClusters];
     for (int i = 0; i < numClusters; i++) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMultiSlaveReplication.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMultiSlaveReplication.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMultiSlaveReplication.java
index 8e4d700..6b241b9 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMultiSlaveReplication.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestMultiSlaveReplication.java
@@ -62,7 +62,7 @@ public class TestMultiSlaveReplication {
   private static final long SLEEP_TIME = 500;
   private static final int NB_RETRIES = 100;
 
-  private static final byte[] tableName = Bytes.toBytes("test");
+  private static final TableName tableName = TableName.valueOf("test");
   private static final byte[] famName = Bytes.toBytes("f");
   private static final byte[] row = Bytes.toBytes("row");
   private static final byte[] row1 = Bytes.toBytes("row1");
@@ -108,7 +108,7 @@ public class TestMultiSlaveReplication {
     utility3.setZkCluster(miniZK);
     new ZooKeeperWatcher(conf3, "cluster3", null, true);
 
-    table = new HTableDescriptor(TableName.valueOf(tableName));
+    table = new HTableDescriptor(tableName);
     HColumnDescriptor fam = new HColumnDescriptor(famName);
     fam.setScope(HConstants.REPLICATION_SCOPE_GLOBAL);
     table.addFamily(fam);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java
index 7778baa..ab5f136 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestPerTableCFReplication.java
@@ -67,10 +67,10 @@ public class TestPerTableCFReplication {
   private static final long SLEEP_TIME = 500;
   private static final int NB_RETRIES = 100;
 
-  private static final byte[] tableName = Bytes.toBytes("test");
-  private static final byte[] tabAName = Bytes.toBytes("TA");
-  private static final byte[] tabBName = Bytes.toBytes("TB");
-  private static final byte[] tabCName = Bytes.toBytes("TC");
+  private static final TableName tableName = TableName.valueOf("test");
+  private static final TableName tabAName = TableName.valueOf("TA");
+  private static final TableName tabBName = TableName.valueOf("TB");
+  private static final TableName tabCName = TableName.valueOf("TC");
   private static final byte[] famName = Bytes.toBytes("f");
   private static final byte[] f1Name = Bytes.toBytes("f1");
   private static final byte[] f2Name = Bytes.toBytes("f2");
@@ -121,7 +121,7 @@ public class TestPerTableCFReplication {
     utility3.setZkCluster(miniZK);
     new ZooKeeperWatcher(conf3, "cluster3", null, true);
 
-    table = new HTableDescriptor(TableName.valueOf(tableName));
+    table = new HTableDescriptor(tableName);
     HColumnDescriptor fam = new HColumnDescriptor(famName);
     fam.setScope(HConstants.REPLICATION_SCOPE_GLOBAL);
     table.addFamily(fam);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java
index 252d3a4..9fe2e0b 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationBase.java
@@ -75,7 +75,7 @@ public class TestReplicationBase {
   protected static final long SLEEP_TIME = 500;
   protected static final int NB_RETRIES = 10;
 
-  protected static final byte[] tableName = Bytes.toBytes("test");
+  protected static final TableName tableName = TableName.valueOf("test");
   protected static final byte[] famName = Bytes.toBytes("f");
   protected static final byte[] row = Bytes.toBytes("row");
   protected static final byte[] noRepfamName = Bytes.toBytes("norep");
@@ -129,7 +129,7 @@ public class TestReplicationBase {
     utility1.startMiniCluster(2);
     utility2.startMiniCluster(2);
 
-    HTableDescriptor table = new HTableDescriptor(TableName.valueOf(tableName));
+    HTableDescriptor table = new HTableDescriptor(tableName);
     HColumnDescriptor fam = new HColumnDescriptor(famName);
     fam.setMaxVersions(3);
     fam.setScope(HConstants.REPLICATION_SCOPE_GLOBAL);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java
index cd363e7..d698255 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSmallTests.java
@@ -446,7 +446,7 @@ public class TestReplicationSmallTests extends TestReplicationBase {
     // identical since it does the check
     testSmallBatch();
 
-    String[] args = new String[] {"2", Bytes.toString(tableName)};
+    String[] args = new String[] {"2", tableName.getNameAsString()};
     Job job = VerifyReplication.createSubmittableJob(CONF_WITH_LOCALFS, args);
     if (job == null) {
       fail("Job wasn't created, see the log");

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSyncUpTool.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSyncUpTool.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSyncUpTool.java
index 701c974..58eb19f 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSyncUpTool.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationSyncUpTool.java
@@ -45,8 +45,8 @@ public class TestReplicationSyncUpTool extends TestReplicationBase {
 
   private static final Log LOG = LogFactory.getLog(TestReplicationSyncUpTool.class);
 
-  private static final byte[] t1_su = Bytes.toBytes("t1_syncup");
-  private static final byte[] t2_su = Bytes.toBytes("t2_syncup");
+  private static final TableName t1_su = TableName.valueOf("t1_syncup");
+  private static final TableName t2_su = TableName.valueOf("t2_syncup");
 
   private static final byte[] famName = Bytes.toBytes("cf1");
   private static final byte[] qualName = Bytes.toBytes("q1");
@@ -63,27 +63,27 @@ public class TestReplicationSyncUpTool extends TestReplicationBase {
 
     HColumnDescriptor fam;
 
-    t1_syncupSource = new HTableDescriptor(TableName.valueOf(t1_su));
+    t1_syncupSource = new HTableDescriptor(t1_su);
     fam = new HColumnDescriptor(famName);
     fam.setScope(HConstants.REPLICATION_SCOPE_GLOBAL);
     t1_syncupSource.addFamily(fam);
     fam = new HColumnDescriptor(noRepfamName);
     t1_syncupSource.addFamily(fam);
 
-    t1_syncupTarget = new HTableDescriptor(TableName.valueOf(t1_su));
+    t1_syncupTarget = new HTableDescriptor(t1_su);
     fam = new HColumnDescriptor(famName);
     t1_syncupTarget.addFamily(fam);
     fam = new HColumnDescriptor(noRepfamName);
     t1_syncupTarget.addFamily(fam);
 
-    t2_syncupSource = new HTableDescriptor(TableName.valueOf(t2_su));
+    t2_syncupSource = new HTableDescriptor(t2_su);
     fam = new HColumnDescriptor(famName);
     fam.setScope(HConstants.REPLICATION_SCOPE_GLOBAL);
     t2_syncupSource.addFamily(fam);
     fam = new HColumnDescriptor(noRepfamName);
     t2_syncupSource.addFamily(fam);
 
-    t2_syncupTarget = new HTableDescriptor(TableName.valueOf(t2_su));
+    t2_syncupTarget = new HTableDescriptor(t2_su);
     fam = new HColumnDescriptor(famName);
     t2_syncupTarget.addFamily(fam);
     fam = new HColumnDescriptor(noRepfamName);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationWithTags.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationWithTags.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationWithTags.java
index 39067a1..43ee682 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationWithTags.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/TestReplicationWithTags.java
@@ -81,7 +81,7 @@ public class TestReplicationWithTags {
   private static final long SLEEP_TIME = 500;
   private static final int NB_RETRIES = 10;
 
-  private static final byte[] TABLE_NAME = Bytes.toBytes("TestReplicationWithTags");
+  private static final TableName TABLE_NAME = TableName.valueOf("TestReplicationWithTags");
   private static final byte[] FAMILY = Bytes.toBytes("f");
   private static final byte[] ROW = Bytes.toBytes("row");
 
@@ -132,7 +132,7 @@ public class TestReplicationWithTags {
     utility1.startMiniCluster(2);
     utility2.startMiniCluster(2);
 
-    HTableDescriptor table = new HTableDescriptor(TableName.valueOf(TABLE_NAME));
+    HTableDescriptor table = new HTableDescriptor(TABLE_NAME);
     HColumnDescriptor fam = new HColumnDescriptor(FAMILY);
     fam.setMaxVersions(3);
     fam.setScope(HConstants.REPLICATION_SCOPE_GLOBAL);

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/SecureTestUtil.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/SecureTestUtil.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/SecureTestUtil.java
index 3e834bd..1141c9f 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/SecureTestUtil.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/SecureTestUtil.java
@@ -103,17 +103,17 @@ public class SecureTestUtil {
     }
   }
 
-  public static void checkTablePerms(Configuration conf, byte[] table, byte[] family, byte[] column,
+  public static void checkTablePerms(Configuration conf, TableName table, byte[] family, byte[] column,
       Permission.Action... actions) throws IOException {
     Permission[] perms = new Permission[actions.length];
     for (int i = 0; i < actions.length; i++) {
-      perms[i] = new TablePermission(TableName.valueOf(table), family, column, actions[i]);
+      perms[i] = new TablePermission(table, family, column, actions[i]);
     }
 
     checkTablePerms(conf, table, perms);
   }
 
-  public static void checkTablePerms(Configuration conf, byte[] table, Permission... perms) throws IOException {
+  public static void checkTablePerms(Configuration conf, TableName table, Permission... perms) throws IOException {
     CheckPermissionsRequest.Builder request = CheckPermissionsRequest.newBuilder();
     for (Permission p : perms) {
       request.addPermission(ProtobufUtil.toPermission(p));

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java
index f94e6a3..3ed58cc 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController.java
@@ -192,7 +192,7 @@ public class TestAccessController extends SecureTestUtil {
       Coprocessor.PRIORITY_HIGHEST, 1, conf);
 
     // Wait for the ACL table to become available
-    TEST_UTIL.waitTableEnabled(AccessControlLists.ACL_TABLE_NAME.getName());
+    TEST_UTIL.waitTableEnabled(AccessControlLists.ACL_TABLE_NAME);
 
     // create a set of test users
     SUPERUSER = User.createUserForTesting(conf, "admin", new String[] { "supergroup" });
@@ -219,7 +219,7 @@ public class TestAccessController extends SecureTestUtil {
     htd.addFamily(hcd);
     htd.setOwner(USER_OWNER);
     admin.createTable(htd, new byte[][] { Bytes.toBytes("s") });
-    TEST_UTIL.waitTableEnabled(TEST_TABLE.getTableName().getName());
+    TEST_UTIL.waitTableEnabled(TEST_TABLE.getTableName());
 
     HRegion region = TEST_UTIL.getHBaseCluster().getRegions(TEST_TABLE.getTableName()).get(0);
     RegionCoprocessorHost rcpHost = region.getCoprocessorHost();

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java
index 62d35bf..a305e10 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestAccessController2.java
@@ -57,7 +57,7 @@ public class TestAccessController2 extends SecureTestUtil {
     verifyConfiguration(conf);
     TEST_UTIL.startMiniCluster();
     // Wait for the ACL table to become available
-    TEST_UTIL.waitTableEnabled(AccessControlLists.ACL_TABLE_NAME.getName());
+    TEST_UTIL.waitTableEnabled(AccessControlLists.ACL_TABLE_NAME);
   }
 
   @AfterClass
@@ -86,7 +86,7 @@ public class TestAccessController2 extends SecureTestUtil {
         return null;
       }
     }, testUser);
-    TEST_UTIL.waitTableEnabled(TEST_TABLE.getTableName().getName());
+    TEST_UTIL.waitTableEnabled(TEST_TABLE.getTableName());
     // Verify that owner permissions have been granted to the test user on the
     // table just created
     List<TablePermission> perms = AccessControlLists.getTablePermissions(conf, TEST_TABLE.getTableName())

http://git-wip-us.apache.org/repos/asf/hbase/blob/b5783795/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestCellACLWithMultipleVersions.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestCellACLWithMultipleVersions.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestCellACLWithMultipleVersions.java
index 0459b9f..9484698 100644
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestCellACLWithMultipleVersions.java
+++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestCellACLWithMultipleVersions.java
@@ -109,7 +109,7 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil {
     rsHost.createEnvironment(AccessController.class, ac, Coprocessor.PRIORITY_HIGHEST, 1, conf);
 
     // Wait for the ACL table to become available
-    TEST_UTIL.waitTableEnabled(AccessControlLists.ACL_TABLE_NAME.getName());
+    TEST_UTIL.waitTableEnabled(AccessControlLists.ACL_TABLE_NAME);
 
     // create a set of test users
     USER_OWNER = User.createUserForTesting(conf, "owner", new String[0]);
@@ -136,7 +136,7 @@ public class TestCellACLWithMultipleVersions extends SecureTestUtil {
     htd.setOwner(USER_OWNER);
     htd.addFamily(hcd);
     admin.createTable(htd, new byte[][] { Bytes.toBytes("s") });
-    TEST_UTIL.waitTableEnabled(TEST_TABLE.getTableName().getName());
+    TEST_UTIL.waitTableEnabled(TEST_TABLE.getTableName());
   }
 
   @Test


Mime
View raw message