Return-Path: X-Original-To: apmail-hbase-commits-archive@www.apache.org Delivered-To: apmail-hbase-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 4AA71117A5 for ; Tue, 9 Sep 2014 06:40:45 +0000 (UTC) Received: (qmail 36280 invoked by uid 500); 9 Sep 2014 06:40:44 -0000 Delivered-To: apmail-hbase-commits-archive@hbase.apache.org Received: (qmail 36181 invoked by uid 500); 9 Sep 2014 06:40:44 -0000 Mailing-List: contact commits-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list commits@hbase.apache.org Received: (qmail 35888 invoked by uid 99); 9 Sep 2014 06:40:44 -0000 Received: from tyr.zones.apache.org (HELO tyr.zones.apache.org) (140.211.11.114) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 09 Sep 2014 06:40:44 +0000 Received: by tyr.zones.apache.org (Postfix, from userid 65534) id 39BE6A0FB1A; Tue, 9 Sep 2014 06:40:44 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: enis@apache.org To: commits@hbase.apache.org Date: Tue, 09 Sep 2014 06:40:50 -0000 Message-Id: <8b6d69bfa50746c1b4e99c9ac69815c7@git.apache.org> In-Reply-To: References: X-Mailer: ASF-Git Admin Mailer Subject: [7/8] HBASE-11679 Replace HTable with HTableInterface where backwards-compatible (Carter) http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java index 0748481..de5f57a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java @@ -39,6 +39,7 @@ import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.TableName; +import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; @@ -469,7 +470,7 @@ public class ImportTsv extends Configured implements Tool { return job; } - private static void createTable(HBaseAdmin admin, String tableName, String[] columns) + private static void createTable(Admin admin, String tableName, String[] columns) throws IOException { HTableDescriptor htd = new HTableDescriptor(TableName.valueOf(tableName)); Set cfSet = new HashSet(); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java index 3e61d95..6fa39af 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java @@ -41,7 +41,6 @@ import java.util.concurrent.Future; import java.util.concurrent.LinkedBlockingQueue; import java.util.concurrent.ThreadPoolExecutor; import java.util.concurrent.TimeUnit; -import java.util.concurrent.atomic.AtomicLong; import org.apache.commons.lang.mutable.MutableInt; import org.apache.commons.logging.Log; @@ -63,11 +62,13 @@ import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.KeyValueUtil; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.TableNotFoundException; +import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HConnection; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.RegionServerCallable; import org.apache.hadoop.hbase.client.RpcRetryingCallerFactory; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.client.coprocessor.SecureBulkLoadClient; import org.apache.hadoop.hbase.io.HFileLink; import org.apache.hadoop.hbase.io.HalfStoreFileReader; @@ -106,7 +107,7 @@ import java.util.UUID; @InterfaceStability.Stable public class LoadIncrementalHFiles extends Configured implements Tool { private static final Log LOG = LogFactory.getLog(LoadIncrementalHFiles.class); - private HBaseAdmin hbAdmin; + private Admin hbAdmin; public static final String NAME = "completebulkload"; public static final String MAX_FILES_PER_REGION_PER_FAMILY @@ -359,7 +360,7 @@ public class LoadIncrementalHFiles extends Configured implements Tool { * them. Any failures are re-queued for another pass with the * groupOrSplitPhase. */ - protected void bulkLoadPhase(final HTable table, final HConnection conn, + protected void bulkLoadPhase(final Table table, final HConnection conn, ExecutorService pool, Deque queue, final Multimap regionGroups) throws IOException { // atomically bulk load the groups. @@ -482,7 +483,7 @@ public class LoadIncrementalHFiles extends Configured implements Tool { } protected List splitStoreFile(final LoadQueueItem item, - final HTable table, byte[] startKey, + final Table table, byte[] startKey, byte[] splitKey) throws IOException { final Path hfilePath = item.hfilePath; @@ -646,7 +647,7 @@ public class LoadIncrementalHFiles extends Configured implements Tool { if(!userProvider.isHBaseSecurityEnabled()) { success = ProtobufUtil.bulkLoadHFile(getStub(), famPaths, regionName, assignSeqIds); } else { - HTable table = new HTable(conn.getConfiguration(), getTableName()); + Table table = new HTable(conn.getConfiguration(), getTableName()); secureClient = new SecureBulkLoadClient(table); success = secureClient.bulkLoadHFiles(famPaths, fsDelegationToken.getUserToken(), bulkToken, getLocation().getRegionInfo().getStartKey()); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java index c50e3d5..15036ee 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableInputFormatBase.java @@ -31,6 +31,7 @@ import org.apache.hadoop.hbase.HRegionLocation; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Pair; @@ -84,7 +85,7 @@ public abstract class MultiTableInputFormatBase extends + " previous error. Please look at the previous logs lines from" + " the task's full log for more details."); } - HTable table = + Table table = new HTable(context.getConfiguration(), tSplit.getTableName()); TableRecordReader trr = this.tableRecordReader; http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java index e23ea37..7416093 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java @@ -19,10 +19,8 @@ package org.apache.hadoop.hbase.mapreduce; import java.io.IOException; -import java.io.InterruptedIOException; import java.net.InetAddress; import java.net.InetSocketAddress; -import java.text.MessageFormat; import java.util.ArrayList; import java.util.HashMap; import java.util.List; http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableOutputFormat.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableOutputFormat.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableOutputFormat.java index 9033871..4917045 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableOutputFormat.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableOutputFormat.java @@ -32,6 +32,7 @@ import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Mutation; import org.apache.hadoop.hbase.client.Put; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.zookeeper.ZKUtil; import org.apache.hadoop.mapreduce.JobContext; import org.apache.hadoop.mapreduce.OutputCommitter; @@ -88,14 +89,14 @@ implements Configurable { extends RecordWriter { /** The table to write to. */ - private HTable table; + private Table table; /** * Instantiate a TableRecordWriter with the HBase HClient for writing. * * @param table The table to write to. */ - public TableRecordWriter(HTable table) { + public TableRecordWriter(Table table) { this.table = table; } http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReader.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReader.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReader.java index 7db2c56..825729d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReader.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReader.java @@ -25,6 +25,7 @@ import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.mapreduce.InputSplit; import org.apache.hadoop.mapreduce.RecordReader; @@ -57,7 +58,7 @@ extends RecordReader { * * @param htable The {@link HTable} to scan. */ - public void setHTable(HTable htable) { + public void setHTable(Table htable) { this.recordReaderImpl.setHTable(htable); } http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.java index e8e6e8b..91b1c1c 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.java @@ -31,6 +31,7 @@ import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.client.ScannerCallable; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.client.metrics.ScanMetrics; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; @@ -58,7 +59,7 @@ public class TableRecordReaderImpl { private ResultScanner scanner = null; private Scan scan = null; private Scan currentScan = null; - private HTable htable = null; + private Table htable = null; private byte[] lastSuccessfulRow = null; private ImmutableBytesWritable key = null; private Result value = null; @@ -121,7 +122,7 @@ public class TableRecordReaderImpl { * * @param htable The {@link HTable} to scan. */ - public void setHTable(HTable htable) { + public void setHTable(Table htable) { Configuration conf = htable.getConfiguration(); logScannerActivity = conf.getBoolean( ScannerCallable.LOG_SCANNER_ACTIVITY, false); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/replication/VerifyReplication.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/replication/VerifyReplication.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/replication/VerifyReplication.java index 7748675..61caf58 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/replication/VerifyReplication.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/replication/VerifyReplication.java @@ -33,6 +33,7 @@ import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.hbase.mapreduce.TableInputFormat; import org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil; @@ -120,7 +121,7 @@ public class VerifyReplication extends Configured implements Tool { Configuration peerConf = HBaseConfiguration.create(conf); ZKUtil.applyClusterKeyToConf(peerConf, zkClusterKey); - HTable replicatedTable = new HTable(peerConf, conf.get(NAME + ".tableName")); + Table replicatedTable = new HTable(peerConf, conf.get(NAME + ".tableName")); scan.setStartRow(value.getRow()); replicatedScanner = replicatedTable.getScanner(scan); return null; http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java index 01a1bf1..5a03c43 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java @@ -30,7 +30,6 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.hbase.CoordinatedStateManager; import org.apache.hadoop.hbase.CoordinatedStateManagerFactory; import org.apache.hadoop.hbase.MasterNotRunningException; @@ -38,6 +37,7 @@ import org.apache.hadoop.hbase.ZNodeClearer; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.LocalHBaseCluster; import org.apache.hadoop.hbase.ZooKeeperConnectionException; +import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.regionserver.HRegionServer; import org.apache.hadoop.hbase.util.JVMClusterUtil; @@ -209,7 +209,7 @@ public class HMasterCommandLine extends ServerCommandLine { } private int stopMaster() { - HBaseAdmin adm = null; + Admin adm = null; try { Configuration conf = getConf(); // Don't try more than once http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java index 282641e..ef205be 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/RegionPlacementMaintainer.java @@ -47,6 +47,7 @@ import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.TableName; +import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HConnection; import org.apache.hadoop.hbase.client.HConnectionManager; @@ -92,7 +93,7 @@ public class RegionPlacementMaintainer { private Configuration conf; private final boolean enforceLocality; private final boolean enforceMinAssignmentMove; - private HBaseAdmin admin; + private Admin admin; private RackManager rackManager; private Set targetTableSet; @@ -127,7 +128,7 @@ public class RegionPlacementMaintainer { * @return the cached HBaseAdmin * @throws IOException */ - private HBaseAdmin getHBaseAdmin() throws IOException { + private Admin getHBaseAdmin() throws IOException { if (this.admin == null) { this.admin = new HBaseAdmin(this.conf); } http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableNamespaceManager.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableNamespaceManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableNamespaceManager.java index 6a83dc3..a78e225 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableNamespaceManager.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableNamespaceManager.java @@ -45,6 +45,7 @@ import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.constraint.ConstraintException; import org.apache.hadoop.hbase.master.handler.CreateTableHandler; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; @@ -107,7 +108,7 @@ public class TableNamespaceManager { isTableAvailableAndInitialized(); } - private synchronized HTable getNamespaceTable() throws IOException { + private synchronized Table getNamespaceTable() throws IOException { if (!isTableAvailableAndInitialized()) { throw new IOException(this.getClass().getName() + " isn't ready to serve"); } @@ -125,14 +126,14 @@ public class TableNamespaceManager { } public synchronized void update(NamespaceDescriptor ns) throws IOException { - HTable table = getNamespaceTable(); + Table table = getNamespaceTable(); if (get(table, ns.getName()) == null) { throw new NamespaceNotFoundException(ns.getName()); } upsert(table, ns); } - private NamespaceDescriptor get(HTable table, String name) throws IOException { + private NamespaceDescriptor get(Table table, String name) throws IOException { Result res = table.get(new Get(Bytes.toBytes(name))); if (res.isEmpty()) { return null; @@ -144,7 +145,7 @@ public class TableNamespaceManager { HBaseProtos.NamespaceDescriptor.parseFrom(val)); } - private void create(HTable table, NamespaceDescriptor ns) throws IOException { + private void create(Table table, NamespaceDescriptor ns) throws IOException { if (get(table, ns.getName()) != null) { throw new NamespaceExistException(ns.getName()); } @@ -154,7 +155,7 @@ public class TableNamespaceManager { upsert(table, ns); } - private void upsert(HTable table, NamespaceDescriptor ns) throws IOException { + private void upsert(Table table, NamespaceDescriptor ns) throws IOException { Put p = new Put(Bytes.toBytes(ns.getName())); p.addImmutable(HTableDescriptor.NAMESPACE_FAMILY_INFO_BYTES, HTableDescriptor.NAMESPACE_COL_DESC_BYTES, http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredNodeAssignmentHelper.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredNodeAssignmentHelper.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredNodeAssignmentHelper.java index 79fd21e..eb6878b 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredNodeAssignmentHelper.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/balancer/FavoredNodeAssignmentHelper.java @@ -41,6 +41,7 @@ import org.apache.hadoop.hbase.MetaTableAccessor; import org.apache.hadoop.hbase.client.HConnection; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.master.RackManager; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; @@ -120,7 +121,7 @@ public class FavoredNodeAssignmentHelper { } } // Write the region assignments to the meta table. - HTable metaTable = null; + Table metaTable = null; try { metaTable = new HTable(conf, TableName.META_TABLE_NAME); metaTable.put(puts); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/ModifyTableHandler.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/ModifyTableHandler.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/ModifyTableHandler.java index 48fb26f..591a1d8 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/ModifyTableHandler.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/handler/ModifyTableHandler.java @@ -36,6 +36,7 @@ import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.executor.EventType; import org.apache.hadoop.hbase.master.HMaster; import org.apache.hadoop.hbase.master.MasterCoprocessorHost; @@ -94,7 +95,7 @@ public class ModifyTableHandler extends TableEventHandler { Set tableRows = new HashSet(); Scan scan = MetaTableAccessor.getScanForTableName(table); scan.addColumn(HConstants.CATALOG_FAMILY, HConstants.REGIONINFO_QUALIFIER); - HTable htable = null; + Table htable = null; try { htable = new HTable(masterServices.getConfiguration(), TableName.META_TABLE_NAME); ResultScanner resScanner = htable.getScanner(scan); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSink.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSink.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSink.java index 4a4de86..d10dc48 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSink.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSink.java @@ -46,10 +46,10 @@ import org.apache.hadoop.hbase.Stoppable; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.HConnection; import org.apache.hadoop.hbase.client.HConnectionManager; -import org.apache.hadoop.hbase.client.HTableInterface; import org.apache.hadoop.hbase.client.Mutation; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Row; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.WALEntry; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos; @@ -230,7 +230,7 @@ public class ReplicationSink { if (allRows.isEmpty()) { return; } - HTableInterface table = null; + Table table = null; try { table = this.sharedHtableCon.getTable(tableName); for (List rows : allRows) { http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java index 9295e9d..5e7038d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java @@ -23,7 +23,7 @@ import java.io.IOException; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.client.HBaseAdmin; -import org.apache.hadoop.hbase.client.HTableInterface; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.filter.ParseFilter; import org.apache.hadoop.hbase.security.UserProvider; import org.apache.hadoop.hbase.util.ConnectionCache; @@ -100,7 +100,7 @@ public class RESTServlet implements Constants { /** * Caller closes the table afterwards. */ - HTableInterface getTable(String tableName) throws IOException { + Table getTable(String tableName) throws IOException { return connectionCache.getTable(tableName); } http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RowResource.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RowResource.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RowResource.java index 6f0a9eb..84564ed 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RowResource.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RowResource.java @@ -45,8 +45,8 @@ import org.apache.hadoop.hbase.CellUtil; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Delete; -import org.apache.hadoop.hbase.client.HTableInterface; import org.apache.hadoop.hbase.client.Put; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.rest.model.CellModel; import org.apache.hadoop.hbase.rest.model.CellSetModel; import org.apache.hadoop.hbase.rest.model.RowModel; @@ -182,7 +182,7 @@ public class RowResource extends ResourceBase { .build(); } - HTableInterface table = null; + Table table = null; try { List rows = model.getRows(); List puts = new ArrayList(); @@ -253,7 +253,7 @@ public class RowResource extends ResourceBase { .type(MIMETYPE_TEXT).entity("Forbidden" + CRLF) .build(); } - HTableInterface table = null; + Table table = null; try { byte[] row = rowspec.getRow(); byte[][] columns = rowspec.getColumns(); @@ -393,7 +393,7 @@ public class RowResource extends ResourceBase { } } } - HTableInterface table = null; + Table table = null; try { table = servlet.getTable(tableResource.getName()); table.delete(delete); @@ -422,7 +422,7 @@ public class RowResource extends ResourceBase { * @return Response 200 OK, 304 Not modified, 400 Bad request */ Response checkAndPut(final CellSetModel model) { - HTableInterface table = null; + Table table = null; try { table = servlet.getTable(tableResource.getName()); if (model.getRows().size() != 1) { @@ -513,7 +513,7 @@ public class RowResource extends ResourceBase { * @return Response 200 OK, 304 Not modified, 400 Bad request */ Response checkAndDelete(final CellSetModel model) { - HTableInterface table = null; + Table table = null; Delete delete = null; try { table = servlet.getTable(tableResource.getName()); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RowResultGenerator.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RowResultGenerator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RowResultGenerator.java index cdfc041..0b37d55 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RowResultGenerator.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/RowResultGenerator.java @@ -30,8 +30,8 @@ import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.DoNotRetryIOException; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Get; -import org.apache.hadoop.hbase.client.HTableInterface; import org.apache.hadoop.hbase.client.Result; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.filter.Filter; import org.apache.hadoop.util.StringUtils; @@ -45,7 +45,7 @@ public class RowResultGenerator extends ResultGenerator { public RowResultGenerator(final String tableName, final RowSpec rowspec, final Filter filter, final boolean cacheBlocks) throws IllegalArgumentException, IOException { - HTableInterface table = RESTServlet.getInstance().getTable(tableName); + Table table = RESTServlet.getInstance().getTable(tableName); try { Get get = new Get(rowspec.getRow()); if (rowspec.hasColumns()) { http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ScannerResultGenerator.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ScannerResultGenerator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ScannerResultGenerator.java index 32f145b..261feae 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ScannerResultGenerator.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/ScannerResultGenerator.java @@ -28,10 +28,10 @@ import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.UnknownScannerException; -import org.apache.hadoop.hbase.client.HTableInterface; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.filter.Filter; import org.apache.hadoop.hbase.rest.model.ScannerModel; import org.apache.hadoop.hbase.security.visibility.Authorizations; @@ -67,7 +67,7 @@ public class ScannerResultGenerator extends ResultGenerator { public ScannerResultGenerator(final String tableName, final RowSpec rowspec, final Filter filter, final int caching, final boolean cacheBlocks) throws IllegalArgumentException, IOException { - HTableInterface table = RESTServlet.getInstance().getTable(tableName); + Table table = RESTServlet.getInstance().getTable(tableName); try { Scan scan; if (rowspec.hasEndRow()) { http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/SchemaResource.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/SchemaResource.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/SchemaResource.java index ae03651..f85ddac 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/SchemaResource.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/SchemaResource.java @@ -44,7 +44,7 @@ import org.apache.hadoop.hbase.TableExistsException; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.TableNotFoundException; import org.apache.hadoop.hbase.client.HBaseAdmin; -import org.apache.hadoop.hbase.client.HTableInterface; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.rest.model.ColumnSchemaModel; import org.apache.hadoop.hbase.rest.model.TableSchemaModel; import org.apache.hadoop.hbase.util.Bytes; @@ -74,7 +74,7 @@ public class SchemaResource extends ResourceBase { private HTableDescriptor getTableSchema() throws IOException, TableNotFoundException { - HTableInterface table = servlet.getTable(tableResource.getName()); + Table table = servlet.getTable(tableResource.getName()); try { return table.getTableDescriptor(); } finally { http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java index 0627ed2..5eef017 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java @@ -35,8 +35,8 @@ import org.apache.commons.lang.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; -import org.apache.hadoop.hbase.client.HTableInterface; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.filter.Filter; import org.apache.hadoop.hbase.filter.FilterList; import org.apache.hadoop.hbase.filter.ParseFilter; @@ -144,7 +144,7 @@ public class TableResource extends ResourceBase { + " End Row => " + endRow + " Columns => " + column + " Start Time => " + startTime + " End Time => " + endTime + " Cache Blocks => " + cacheBlocks + " Max Versions => " + maxVersions + " Batch Size => " + batchSize); - HTableInterface hTable = RESTServlet.getInstance().getTable(this.table); + Table hTable = RESTServlet.getInstance().getTable(this.table); Scan tableScan = new Scan(); tableScan.setBatch(batchSize); tableScan.setMaxVersions(maxVersions); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlLists.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlLists.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlLists.java index 19e5136..16c2c3a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlLists.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessControlLists.java @@ -51,6 +51,7 @@ import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.exceptions.DeserializationException; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.filter.QualifierFilter; @@ -172,7 +173,7 @@ public class AccessControlLists { Bytes.toString(key)+": "+Bytes.toStringBinary(value) ); } - HTable acls = null; + Table acls = null; try { acls = new HTable(conf, ACL_TABLE_NAME); acls.put(p); @@ -203,7 +204,7 @@ public class AccessControlLists { LOG.debug("Removing permission "+ userPerm.toString()); } d.deleteColumns(ACL_LIST_FAMILY, key); - HTable acls = null; + Table acls = null; try { acls = new HTable(conf, ACL_TABLE_NAME); acls.delete(d); @@ -223,7 +224,7 @@ public class AccessControlLists { LOG.debug("Removing permissions of removed table "+ tableName); } - HTable acls = null; + Table acls = null; try { acls = new HTable(conf, ACL_TABLE_NAME); acls.delete(d); @@ -243,7 +244,7 @@ public class AccessControlLists { LOG.debug("Removing permissions of removed namespace "+ namespace); } - HTable acls = null; + Table acls = null; try { acls = new HTable(conf, ACL_TABLE_NAME); acls.delete(d); @@ -263,7 +264,7 @@ public class AccessControlLists { " from table "+ tableName); } - HTable acls = null; + Table acls = null; try { acls = new HTable(conf, ACL_TABLE_NAME); @@ -424,7 +425,7 @@ public class AccessControlLists { Scan scan = new Scan(); scan.addFamily(ACL_LIST_FAMILY); - HTable acls = null; + Table acls = null; ResultScanner scanner = null; try { acls = new HTable(conf, ACL_TABLE_NAME); @@ -467,7 +468,7 @@ public class AccessControlLists { // for normal user tables, we just read the table row from _acl_ ListMultimap perms = ArrayListMultimap.create(); - HTable acls = null; + Table acls = null; try { acls = new HTable(conf, ACL_TABLE_NAME); Get get = new Get(entryName); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java index cdc824d..78d2c14 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java @@ -30,6 +30,7 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.HTable; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos; @@ -53,7 +54,7 @@ public class TokenUtil { */ public static Token obtainToken( Configuration conf) throws IOException { - HTable meta = null; + Table meta = null; try { meta = new HTable(conf, TableName.META_TABLE_NAME); CoprocessorRpcChannel rpcChannel = meta.coprocessorService(HConstants.EMPTY_START_ROW); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java index 489aaf1..9b11291 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/Canary.java @@ -49,6 +49,7 @@ import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.util.Tool; import org.apache.hadoop.util.ToolRunner; @@ -326,7 +327,7 @@ public final class Canary implements Tool { public static abstract class Monitor implements Runnable { protected Configuration config; - protected HBaseAdmin admin; + protected Admin admin; protected String[] targets; protected boolean useRegExp; protected boolean initialized = false; @@ -476,7 +477,7 @@ public final class Canary implements Tool { */ private static void sniff(final Admin admin, final Sink sink, HTableDescriptor tableDesc) throws Exception { - HTable table = null; + Table table = null; try { table = new HTable(admin.getConfiguration(), tableDesc.getName()); @@ -506,7 +507,7 @@ public final class Canary implements Tool { final Admin admin, final Sink sink, HRegionInfo region, - HTable table) throws Exception { + Table table) throws Exception { HTableDescriptor tableDesc = table.getTableDescriptor(); byte[] startKey = null; Get get = null; @@ -607,7 +608,7 @@ public final class Canary implements Tool { String serverName = null; String tableName = null; HRegionInfo region = null; - HTable table = null; + Table table = null; Get get = null; byte[] startKey = null; Scan scan = null; http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java index efcd7cd..017153a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java @@ -73,6 +73,7 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.ZooKeeperConnectionException; import org.apache.hadoop.hbase.MetaTableAccessor; +import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HBaseAdmin; @@ -87,6 +88,7 @@ import org.apache.hadoop.hbase.client.MetaScanner.MetaScannerVisitorBase; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.RowMutations; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.io.hfile.CacheConfig; import org.apache.hadoop.hbase.io.hfile.HFile; import org.apache.hadoop.hbase.master.MasterFileSystem; @@ -184,8 +186,8 @@ public class HBaseFsck extends Configured { private static final Log LOG = LogFactory.getLog(HBaseFsck.class.getName()); private ClusterStatus status; private HConnection connection; - private HBaseAdmin admin; - private HTable meta; + private Admin admin; + private Table meta; // threads to do ||izable tasks: retrieve data from regionservers, handle overlapping regions protected ExecutorService executor; private long startMillis = System.currentTimeMillis(); @@ -2805,7 +2807,7 @@ public class HBaseFsck extends Configured { HTableDescriptor[] getHTableDescriptors(List tableNames) { HTableDescriptor[] htd = new HTableDescriptor[0]; - HBaseAdmin admin = null; + Admin admin = null; try { LOG.info("getHTableDescriptors == tableNames => " + tableNames); admin = new HBaseAdmin(getConf()); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsckRepair.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsckRepair.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsckRepair.java index 960d69c..5566e25 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsckRepair.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HBaseFsckRepair.java @@ -36,10 +36,10 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.ZooKeeperConnectionException; import org.apache.hadoop.hbase.MetaTableAccessor; import org.apache.hadoop.hbase.client.Admin; -import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HConnection; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.master.RegionState; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService; @@ -63,7 +63,7 @@ public class HBaseFsckRepair { * @param region Region to undeploy * @param servers list of Servers to undeploy from */ - public static void fixMultiAssignment(HBaseAdmin admin, HRegionInfo region, + public static void fixMultiAssignment(Admin admin, HRegionInfo region, List servers) throws IOException, KeeperException, InterruptedException { HRegionInfo actualRegion = new HRegionInfo(region); @@ -89,7 +89,7 @@ public class HBaseFsckRepair { * @throws IOException * @throws KeeperException */ - public static void fixUnassigned(HBaseAdmin admin, HRegionInfo region) + public static void fixUnassigned(Admin admin, HRegionInfo region) throws IOException, KeeperException { HRegionInfo actualRegion = new HRegionInfo(region); @@ -109,7 +109,7 @@ public class HBaseFsckRepair { * side-effect of requiring a HRegionInfo that considers regionId (timestamp) * in comparators that is addressed by HBASE-5563. */ - private static void forceOfflineInZK(HBaseAdmin admin, final HRegionInfo region) + private static void forceOfflineInZK(Admin admin, final HRegionInfo region) throws ZooKeeperConnectionException, KeeperException, IOException { admin.assign(region.getRegionName()); } @@ -117,7 +117,7 @@ public class HBaseFsckRepair { /* * Should we check all assignments or just not in RIT? */ - public static void waitUntilAssigned(HBaseAdmin admin, + public static void waitUntilAssigned(Admin admin, HRegionInfo region) throws IOException, InterruptedException { long timeout = admin.getConfiguration().getLong("hbase.hbck.assign.timeout", 120000); long expiration = timeout + System.currentTimeMillis(); @@ -179,7 +179,7 @@ public class HBaseFsckRepair { */ public static void fixMetaHoleOnlineAndAddReplicas(Configuration conf, HRegionInfo hri, Collection servers, int numReplicas) throws IOException { - HTable meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = new HTable(conf, TableName.META_TABLE_NAME); Put put = MetaTableAccessor.makePutFromRegionInfo(hri); if (numReplicas > 1) { Random r = new Random(); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HMerge.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HMerge.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HMerge.java index f3ac5cc..98eb7e2 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HMerge.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/HMerge.java @@ -36,6 +36,7 @@ import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.MetaTableAccessor; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.TableNotDisabledException; +import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HConnectable; @@ -44,6 +45,7 @@ import org.apache.hadoop.hbase.client.HConnectionManager; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.regionserver.wal.HLog; import org.apache.hadoop.hbase.regionserver.wal.HLogFactory; @@ -124,7 +126,7 @@ class HMerge { throw new IllegalStateException( "HBase instance must be running to merge a normal table"); } - HBaseAdmin admin = new HBaseAdmin(conf); + Admin admin = new HBaseAdmin(conf); try { if (!admin.isTableDisabled(tableName)) { throw new TableNotDisabledException(tableName); @@ -231,7 +233,7 @@ class HMerge { /** Instantiated to compact a normal user table */ private static class OnlineMerger extends Merger { private final TableName tableName; - private final HTable table; + private final Table table; private final ResultScanner metaScanner; private HRegionInfo latestRegion; http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSizeCalculator.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSizeCalculator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSizeCalculator.java index fa5500b..9284f03 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSizeCalculator.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/RegionSizeCalculator.java @@ -27,11 +27,11 @@ import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.RegionLoad; import org.apache.hadoop.hbase.ServerLoad; import org.apache.hadoop.hbase.ServerName; +import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HTable; import java.io.IOException; -import java.text.MessageFormat; import java.util.Arrays; import java.util.Collection; import java.util.Collections; @@ -65,7 +65,7 @@ public class RegionSizeCalculator { } /** ctor for unit testing */ - RegionSizeCalculator (HTable table, HBaseAdmin admin) throws IOException { + RegionSizeCalculator (HTable table, Admin admin) throws IOException { try { if (!enabled(table.getConfiguration())) { http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java index e08530d..db89a45 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestCase.java @@ -39,6 +39,7 @@ import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.regionserver.InternalScanner; import org.apache.hadoop.hbase.util.Bytes; @@ -467,12 +468,12 @@ public abstract class HBaseTestCase extends TestCase { * A class that makes a {@link Incommon} out of a {@link HTable} */ public static class HTableIncommon implements Incommon { - final HTable table; + final Table table; /** * @param table */ - public HTableIncommon(final HTable table) { + public HTableIncommon(final Table table) { super(); this.table = table; } http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java index 294f182..45bc524 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java @@ -63,11 +63,12 @@ import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HConnection; import org.apache.hadoop.hbase.client.HConnectionManager; import org.apache.hadoop.hbase.client.HTable; -import org.apache.hadoop.hbase.client.HTableInterface; import org.apache.hadoop.hbase.client.Put; +import org.apache.hadoop.hbase.client.RegionLocator; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.fs.HFileSystem; import org.apache.hadoop.hbase.io.compress.Compression; import org.apache.hadoop.hbase.io.compress.Compression.Algorithm; @@ -898,7 +899,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { this.hbaseCluster = new MiniHBaseCluster(c, numMasters, numSlaves, masterClass, regionserverClass); // Don't leave here till we've done a successful scan of the hbase:meta - HTable t = new HTable(c, TableName.META_TABLE_NAME); + Table t = new HTable(c, TableName.META_TABLE_NAME); ResultScanner s = t.getScanner(new Scan()); while (s.next() != null) { continue; @@ -920,7 +921,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { public void restartHBaseCluster(int servers) throws IOException, InterruptedException { this.hbaseCluster = new MiniHBaseCluster(this.conf, servers); // Don't leave here till we've done a successful scan of the hbase:meta - HTable t = new HTable(new Configuration(this.conf), TableName.META_TABLE_NAME); + Table t = new HTable(new Configuration(this.conf), TableName.META_TABLE_NAME); ResultScanner s = t.getScanner(new Scan()); while (s.next() != null) { // do nothing @@ -1912,7 +1913,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { return rowCount; } - public void loadNumericRows(final HTableInterface t, final byte[] f, int startRow, int endRow) + public void loadNumericRows(final Table t, final byte[] f, int startRow, int endRow) throws IOException { for (int i = startRow; i < endRow; i++) { byte[] data = Bytes.toBytes(String.valueOf(i)); @@ -1937,7 +1938,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { } } - public void deleteNumericRows(final HTable t, final byte[] f, int startRow, int endRow) + public void deleteNumericRows(final Table t, final byte[] f, int startRow, int endRow) throws IOException { for (int i = startRow; i < endRow; i++) { byte[] data = Bytes.toBytes(String.valueOf(i)); @@ -1950,7 +1951,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { /** * Return the number of rows in the given table. */ - public int countRows(final HTable table) throws IOException { + public int countRows(final Table table) throws IOException { Scan scan = new Scan(); ResultScanner results = table.getScanner(scan); int count = 0; @@ -1961,7 +1962,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { return count; } - public int countRows(final HTable table, final byte[]... families) throws IOException { + public int countRows(final Table table, final byte[]... families) throws IOException { Scan scan = new Scan(); for (byte[] family: families) { scan.addFamily(family); @@ -1978,7 +1979,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { /** * Return an md5 digest of the entire contents of a table. */ - public String checksumRows(final HTable table) throws Exception { + public String checksumRows(final Table table) throws Exception { Scan scan = new Scan(); ResultScanner results = table.getScanner(scan); MessageDigest digest = MessageDigest.getInstance("MD5"); @@ -2084,7 +2085,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { final byte[] columnFamily, byte [][] startKeys) throws IOException { Arrays.sort(startKeys, Bytes.BYTES_COMPARATOR); - HTable meta = new HTable(c, TableName.META_TABLE_NAME); + Table meta = new HTable(c, TableName.META_TABLE_NAME); HTableDescriptor htd = table.getTableDescriptor(); if(!htd.hasFamily(columnFamily)) { HColumnDescriptor hcd = new HColumnDescriptor(columnFamily); @@ -2149,7 +2150,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { public List createMultiRegionsInMeta(final Configuration conf, final HTableDescriptor htd, byte [][] startKeys) throws IOException { - HTable meta = new HTable(conf, TableName.META_TABLE_NAME); + Table meta = new HTable(conf, TableName.META_TABLE_NAME); Arrays.sort(startKeys, Bytes.BYTES_COMPARATOR); List newRegions = new ArrayList(startKeys.length); // add custom ones @@ -2172,7 +2173,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { */ public List getMetaTableRows() throws IOException { // TODO: Redo using MetaTableAccessor class - HTable t = new HTable(new Configuration(this.conf), TableName.META_TABLE_NAME); + Table t = new HTable(new Configuration(this.conf), TableName.META_TABLE_NAME); List rows = new ArrayList(); ResultScanner s = t.getScanner(new Scan()); for (Result result : s) { @@ -2192,7 +2193,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { */ public List getMetaTableRows(TableName tableName) throws IOException { // TODO: Redo using MetaTableAccessor. - HTable t = new HTable(new Configuration(this.conf), TableName.META_TABLE_NAME); + Table t = new HTable(new Configuration(this.conf), TableName.META_TABLE_NAME); List rows = new ArrayList(); ResultScanner s = t.getScanner(new Scan()); for (Result result : s) { @@ -2638,7 +2639,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { * @param table The table to find the region. * @throws IOException */ - public void closeRegionByRow(String row, HTable table) throws IOException { + public void closeRegionByRow(String row, RegionLocator table) throws IOException { closeRegionByRow(Bytes.toBytes(row), table); } @@ -2649,7 +2650,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { * @param table The table to find the region. * @throws IOException */ - public void closeRegionByRow(byte[] row, HTable table) throws IOException { + public void closeRegionByRow(byte[] row, RegionLocator table) throws IOException { HRegionLocation hrl = table.getRegionLocation(row); closeRegion(hrl.getRegionInfo().getRegionName()); } @@ -2975,7 +2976,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { */ public void waitUntilAllRegionsAssigned(final TableName tableName, final long timeout) throws IOException { - final HTable meta = new HTable(getConfiguration(), TableName.META_TABLE_NAME); + final Table meta = new HTable(getConfiguration(), TableName.META_TABLE_NAME); try { waitFor(timeout, 200, true, new Predicate() { @Override @@ -3371,7 +3372,7 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility { } public static int getMetaRSPort(Configuration conf) throws IOException { - HTable table = new HTable(conf, TableName.META_TABLE_NAME); + RegionLocator table = new HTable(conf, TableName.META_TABLE_NAME); HRegionLocation hloc = table.getRegionLocation(Bytes.toBytes("")); table.close(); return hloc.getPort(); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java index 2ac4135..936b181 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/PerformanceEvaluation.java @@ -55,11 +55,11 @@ import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HConnection; import org.apache.hadoop.hbase.client.HConnectionManager; import org.apache.hadoop.hbase.client.HTable; -import org.apache.hadoop.hbase.client.HTableInterface; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.filter.BinaryComparator; import org.apache.hadoop.hbase.filter.CompareFilter; import org.apache.hadoop.hbase.filter.Filter; @@ -923,7 +923,7 @@ public class PerformanceEvaluation extends Configured implements Tool { private final Sampler traceSampler; private final SpanReceiverHost receiverHost; protected HConnection connection; - protected HTableInterface table; + protected Table table; private String testName; private Histogram latency; @@ -1567,7 +1567,7 @@ public class PerformanceEvaluation extends Configured implements Tool { // Log the configuration we're going to run with. Uses JSON mapper because lazy. It'll do // the TestOptions introspection for us and dump the output in a readable format. LOG.info(cmd.getSimpleName() + " test run options=" + MAPPER.writeValueAsString(opts)); - HBaseAdmin admin = null; + Admin admin = null; try { admin = new HBaseAdmin(getConf()); checkTable(admin, opts); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/ScanPerformanceEvaluation.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/ScanPerformanceEvaluation.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/ScanPerformanceEvaluation.java index 484a00c..dfafa83 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/ScanPerformanceEvaluation.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/ScanPerformanceEvaluation.java @@ -30,6 +30,7 @@ import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.client.TableSnapshotScanner; import org.apache.hadoop.hbase.client.metrics.ScanMetrics; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; @@ -146,7 +147,7 @@ public class ScanPerformanceEvaluation extends AbstractHBaseTool { Stopwatch scanTimer = new Stopwatch(); tableOpenTimer.start(); - HTable table = new HTable(getConf(), TableName.valueOf(tablename)); + Table table = new HTable(getConf(), TableName.valueOf(tablename)); tableOpenTimer.stop(); Scan scan = getScan(); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestAcidGuarantees.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestAcidGuarantees.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestAcidGuarantees.java index d73dd38..172395a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestAcidGuarantees.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestAcidGuarantees.java @@ -35,6 +35,7 @@ import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.util.StringUtils; @@ -97,7 +98,7 @@ public class TestAcidGuarantees implements Tool { byte data[] = new byte[10]; byte targetRows[][]; byte targetFamilies[][]; - HTable table; + Table table; AtomicLong numWritten = new AtomicLong(); public AtomicityWriter(TestContext ctx, byte targetRows[][], @@ -131,7 +132,7 @@ public class TestAcidGuarantees implements Tool { public static class AtomicGetReader extends RepeatingTestThread { byte targetRow[]; byte targetFamilies[][]; - HTable table; + Table table; int numVerified = 0; AtomicLong numRead = new AtomicLong(); @@ -189,7 +190,7 @@ public class TestAcidGuarantees implements Tool { */ public static class AtomicScanReader extends RepeatingTestThread { byte targetFamilies[][]; - HTable table; + Table table; AtomicLong numScans = new AtomicLong(); AtomicLong numRowsScanned = new AtomicLong(); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestHBaseTestingUtility.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestHBaseTestingUtility.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestHBaseTestingUtility.java index 8e448e8..cb01558 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestHBaseTestingUtility.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestHBaseTestingUtility.java @@ -28,9 +28,9 @@ import org.apache.commons.logging.LogFactory; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.client.Get; -import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster; import org.apache.hadoop.hdfs.MiniDFSCluster; @@ -87,8 +87,8 @@ public class TestHBaseTestingUtility { final byte[] QUAL_NAME = Bytes.toBytes("qual"); final byte[] VALUE = Bytes.toBytes("value"); - HTable table1 = htu1.createTable(TABLE_NAME, FAM_NAME); - HTable table2 = htu2.createTable(TABLE_NAME, FAM_NAME); + Table table1 = htu1.createTable(TABLE_NAME, FAM_NAME); + Table table2 = htu2.createTable(TABLE_NAME, FAM_NAME); Put put = new Put(ROW); put.add(FAM_NAME, QUAL_NAME, VALUE); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableAccessor.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableAccessor.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableAccessor.java index 359bb0a..c812034 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableAccessor.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMetaTableAccessor.java @@ -36,6 +36,7 @@ import org.apache.hadoop.hbase.client.HConnection; import org.apache.hadoop.hbase.client.HConnectionManager; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Pair; import org.junit.AfterClass; @@ -333,7 +334,7 @@ public class TestMetaTableAccessor { long seqNum100 = random.nextLong(); - HTable meta = MetaTableAccessor.getMetaHTable(hConnection); + Table meta = MetaTableAccessor.getMetaHTable(hConnection); try { MetaTableAccessor.updateRegionLocation(hConnection, primary, serverName0, seqNum0); @@ -360,7 +361,7 @@ public class TestMetaTableAccessor { } } - public static void assertMetaLocation(HTable meta, byte[] row, ServerName serverName, + public static void assertMetaLocation(Table meta, byte[] row, ServerName serverName, long seqNum, int replicaId, boolean checkSeqNum) throws IOException { Get get = new Get(row); Result result = meta.get(get); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMultiVersions.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMultiVersions.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMultiVersions.java index c04edc1..e619d5e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMultiVersions.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestMultiVersions.java @@ -33,6 +33,7 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseTestCase.FlushCache; import org.apache.hadoop.hbase.HBaseTestCase.HTableIncommon; import org.apache.hadoop.hbase.HBaseTestCase.Incommon; +import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HTable; @@ -40,6 +41,7 @@ import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.util.Bytes; import org.junit.After; import org.junit.AfterClass; @@ -56,7 +58,7 @@ import org.junit.experimental.categories.Category; public class TestMultiVersions { private static final Log LOG = LogFactory.getLog(TestMultiVersions.class); private static final HBaseTestingUtility UTIL = new HBaseTestingUtility(); - private HBaseAdmin admin; + private Admin admin; private static final int NUM_SLAVES = 3; @@ -97,7 +99,7 @@ public class TestMultiVersions { hcd.setMaxVersions(3); desc.addFamily(hcd); this.admin.createTable(desc); - HTable table = new HTable(UTIL.getConfiguration(), desc.getTableName()); + Table table = new HTable(UTIL.getConfiguration(), desc.getTableName()); // TODO: Remove these deprecated classes or pull them in here if this is // only test using them. Incommon incommon = new HTableIncommon(table); @@ -140,7 +142,7 @@ public class TestMultiVersions { this.admin.createTable(desc); Put put = new Put(row, timestamp1); put.add(contents, contents, value1); - HTable table = new HTable(UTIL.getConfiguration(), tableName); + Table table = new HTable(UTIL.getConfiguration(), tableName); table.put(put); // Shut down and restart the HBase cluster table.close(); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestNamespace.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestNamespace.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestNamespace.java index a752f7a..79166a4 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestNamespace.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestNamespace.java @@ -37,6 +37,7 @@ import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.master.HMaster; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.FSUtils; @@ -249,7 +250,7 @@ public class TestNamespace { } //sanity check try to write and read from table - HTable table = new HTable(TEST_UTIL.getConfiguration(), desc.getTableName()); + Table table = new HTable(TEST_UTIL.getConfiguration(), desc.getTableName()); Put p = new Put(Bytes.toBytes("row1")); p.add(Bytes.toBytes("my_cf"),Bytes.toBytes("my_col"),Bytes.toBytes("value1")); table.put(p); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestRegionRebalancing.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestRegionRebalancing.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestRegionRebalancing.java index fdcb00c..fd7cd6f 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestRegionRebalancing.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestRegionRebalancing.java @@ -29,8 +29,10 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HTable; +import org.apache.hadoop.hbase.client.RegionLocator; import org.apache.hadoop.hbase.master.RegionStates; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.regionserver.HRegionServer; @@ -63,7 +65,7 @@ public class TestRegionRebalancing { private static final byte[] FAMILY_NAME = Bytes.toBytes("col"); public static final Log LOG = LogFactory.getLog(TestRegionRebalancing.class); private final HBaseTestingUtility UTIL = new HBaseTestingUtility(); - private HTable table; + private RegionLocator table; private HTableDescriptor desc; private String balancerName; @@ -94,7 +96,7 @@ public class TestRegionRebalancing { @Test (timeout=300000) public void testRebalanceOnRegionServerNumberChange() throws IOException, InterruptedException { - HBaseAdmin admin = new HBaseAdmin(UTIL.getConfiguration()); + Admin admin = new HBaseAdmin(UTIL.getConfiguration()); admin.createTable(this.desc, Arrays.copyOfRange(HBaseTestingUtility.KEYS, 1, HBaseTestingUtility.KEYS.length)); this.table = new HTable(UTIL.getConfiguration(), this.desc.getTableName()); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestZooKeeper.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestZooKeeper.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestZooKeeper.java index 9d502d0..80faad2 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestZooKeeper.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestZooKeeper.java @@ -34,6 +34,7 @@ import java.util.Map; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HConnection; @@ -43,6 +44,7 @@ import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.ResultScanner; import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.master.HMaster; import org.apache.hadoop.hbase.master.LoadBalancer; import org.apache.hadoop.hbase.master.balancer.SimpleLoadBalancer; @@ -249,14 +251,14 @@ public class TestZooKeeper { HColumnDescriptor family = new HColumnDescriptor("fam"); desc.addFamily(family); LOG.info("Creating table " + tableName); - HBaseAdmin admin = new HBaseAdmin(TEST_UTIL.getConfiguration()); + Admin admin = new HBaseAdmin(TEST_UTIL.getConfiguration()); try { admin.createTable(desc); } finally { admin.close(); } - HTable table = + Table table = new HTable(new Configuration(TEST_UTIL.getConfiguration()), tableName); Put put = new Put(Bytes.toBytes("testrow")); put.add(Bytes.toBytes("fam"), @@ -269,11 +271,11 @@ public class TestZooKeeper { @Test public void testMultipleZK() throws IOException, NoSuchMethodException, InvocationTargetException, IllegalAccessException { - HTable localMeta = + Table localMeta = new HTable(new Configuration(TEST_UTIL.getConfiguration()), TableName.META_TABLE_NAME); Configuration otherConf = new Configuration(TEST_UTIL.getConfiguration()); otherConf.set(HConstants.ZOOKEEPER_QUORUM, "127.0.0.1"); - HTable ipMeta = new HTable(otherConf, TableName.META_TABLE_NAME); + Table ipMeta = new HTable(otherConf, TableName.META_TABLE_NAME); // dummy, just to open the connection final byte [] row = new byte [] {'r'}; @@ -490,7 +492,7 @@ public class TestZooKeeper { ZooKeeperWatcher zkw = m.getZooKeeper(); int expectedNumOfListeners = zkw.getNumberOfListeners(); // now the cluster is up. So assign some regions. - HBaseAdmin admin = new HBaseAdmin(TEST_UTIL.getConfiguration()); + Admin admin = new HBaseAdmin(TEST_UTIL.getConfiguration()); try { byte[][] SPLIT_KEYS = new byte[][] { Bytes.toBytes("a"), Bytes.toBytes("b"), Bytes.toBytes("c"), Bytes.toBytes("d"), Bytes.toBytes("e"), Bytes.toBytes("f"), @@ -527,8 +529,8 @@ public class TestZooKeeper { cluster.startRegionServer(); HMaster m = cluster.getMaster(); // now the cluster is up. So assign some regions. - HBaseAdmin admin = new HBaseAdmin(TEST_UTIL.getConfiguration()); - HTable table = null; + Admin admin = new HBaseAdmin(TEST_UTIL.getConfiguration()); + Table table = null; try { byte[][] SPLIT_KEYS = new byte[][] { Bytes.toBytes("1"), Bytes.toBytes("2"), Bytes.toBytes("3"), Bytes.toBytes("4"), Bytes.toBytes("5") }; http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin.java index 664ae04..b91c962 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin.java @@ -246,7 +246,7 @@ public class TestAdmin { final byte [] qualifier = Bytes.toBytes("qualifier"); final byte [] value = Bytes.toBytes("value"); final TableName table = TableName.valueOf("testDisableAndEnableTable"); - HTable ht = TEST_UTIL.createTable(table, HConstants.CATALOG_FAMILY); + Table ht = TEST_UTIL.createTable(table, HConstants.CATALOG_FAMILY); Put put = new Put(row); put.add(HConstants.CATALOG_FAMILY, qualifier, value); ht.put(put); @@ -291,8 +291,8 @@ public class TestAdmin { final byte [] value = Bytes.toBytes("value"); final byte [] table1 = Bytes.toBytes("testDisableAndEnableTable1"); final byte [] table2 = Bytes.toBytes("testDisableAndEnableTable2"); - HTable ht1 = TEST_UTIL.createTable(table1, HConstants.CATALOG_FAMILY); - HTable ht2 = TEST_UTIL.createTable(table2, HConstants.CATALOG_FAMILY); + Table ht1 = TEST_UTIL.createTable(table1, HConstants.CATALOG_FAMILY); + Table ht2 = TEST_UTIL.createTable(table2, HConstants.CATALOG_FAMILY); Put put = new Put(row); put.add(HConstants.CATALOG_FAMILY, qualifier, value); ht1.put(put); @@ -400,7 +400,7 @@ public class TestAdmin { htd.addFamily(fam2); htd.addFamily(fam3); this.admin.createTable(htd); - HTable table = new HTable(TEST_UTIL.getConfiguration(), "myTestTable"); + Table table = new HTable(TEST_UTIL.getConfiguration(), "myTestTable"); HTableDescriptor confirmedHtd = table.getTableDescriptor(); assertEquals(htd.compareTo(confirmedHtd), 0); table.close(); @@ -816,7 +816,7 @@ public class TestAdmin { TableName TABLE_4 = TableName.valueOf(tableName.getNameAsString() + "_4"); desc = new HTableDescriptor(TABLE_4); desc.addFamily(new HColumnDescriptor(HConstants.CATALOG_FAMILY)); - HBaseAdmin ladmin = new HBaseAdmin(TEST_UTIL.getConfiguration()); + Admin ladmin = new HBaseAdmin(TEST_UTIL.getConfiguration()); try { ladmin.createTable(desc, splitKeys); assertTrue("Should not be able to create this table because of " + @@ -1238,7 +1238,7 @@ public class TestAdmin { // Use 80 bit numbers to make sure we aren't limited byte [] startKey = { 1, 1, 1, 1, 1, 1, 1, 1, 1, 1 }; byte [] endKey = { 9, 9, 9, 9, 9, 9, 9, 9, 9, 9 }; - HBaseAdmin hbaseadmin = new HBaseAdmin(TEST_UTIL.getConfiguration()); + Admin hbaseadmin = new HBaseAdmin(TEST_UTIL.getConfiguration()); HTableDescriptor htd = new HTableDescriptor(TableName.valueOf(name)); htd.addFamily(new HColumnDescriptor(HConstants.CATALOG_FAMILY)); hbaseadmin.createTable(htd, startKey, endKey, expectedRegions); @@ -1255,7 +1255,7 @@ public class TestAdmin { @Test (timeout=300000) public void testReadOnlyTable() throws Exception { byte [] name = Bytes.toBytes("testReadOnlyTable"); - HTable table = TEST_UTIL.createTable(name, HConstants.CATALOG_FAMILY); + Table table = TEST_UTIL.createTable(name, HConstants.CATALOG_FAMILY); byte[] value = Bytes.toBytes("somedata"); // This used to use an empty row... That must have been a bug Put put = new Put(value); @@ -1324,7 +1324,7 @@ public class TestAdmin { @Test (expected=TableNotDisabledException.class, timeout=300000) public void testTableNotDisabledExceptionWithATable() throws IOException { final TableName name = TableName.valueOf("testTableNotDisabledExceptionWithATable"); - HTable t = TEST_UTIL.createTable(name, HConstants.CATALOG_FAMILY); + Table t = TEST_UTIL.createTable(name, HConstants.CATALOG_FAMILY); try { this.admin.enableTable(name); }finally { @@ -1338,7 +1338,7 @@ public class TestAdmin { */ @Test (expected=TableNotFoundException.class, timeout=300000) public void testTableNotFoundExceptionWithoutAnyTables() throws IOException { - HTable ht = + Table ht = new HTable(TEST_UTIL.getConfiguration(),"testTableNotFoundExceptionWithoutAnyTables"); ht.get(new Get("e".getBytes())); } @@ -1659,7 +1659,7 @@ public class TestAdmin { HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(tableName)); desc.addFamily(new HColumnDescriptor(HConstants.CATALOG_FAMILY)); admin.createTable(desc); - HTable table = new HTable(TEST_UTIL.getConfiguration(), tableName); + Table table = new HTable(TEST_UTIL.getConfiguration(), tableName); HRegionServer regionServer = TEST_UTIL.getRSForFirstRegionInTable(Bytes.toBytes(tableName)); for (int i = 1; i <= 256; i++) { // 256 writes should cause 8 log rolls http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientOperationInterrupt.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientOperationInterrupt.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientOperationInterrupt.java index bcae508..b29456f 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientOperationInterrupt.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientOperationInterrupt.java @@ -83,7 +83,7 @@ public class TestClientOperationInterrupt { } util.createTable(tableName, new byte[][]{dummy, test}); - HTable ht = new HTable(conf, tableName); + Table ht = new HTable(conf, tableName); Put p = new Put(row1); p.add(dummy, dummy, dummy); ht.put(p); @@ -105,7 +105,7 @@ public class TestClientOperationInterrupt { @Override public void run() { try { - HTable ht = new HTable(conf, tableName); + Table ht = new HTable(conf, tableName); Result r = ht.get(new Get(row1)); noEx.incrementAndGet(); } catch (IOException e) { @@ -154,7 +154,7 @@ public class TestClientOperationInterrupt { Thread.sleep(1); } - HTable ht = new HTable(conf, tableName); + Table ht = new HTable(conf, tableName); Result r = ht.get(new Get(row1)); Assert.assertFalse(r.isEmpty()); } http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientScannerRPCTimeout.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientScannerRPCTimeout.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientScannerRPCTimeout.java index 95b9b21..546b9ea 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientScannerRPCTimeout.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientScannerRPCTimeout.java @@ -84,7 +84,7 @@ public class TestClientScannerRPCTimeout { @Test public void testScannerNextRPCTimesout() throws Exception { final byte[] TABLE_NAME = Bytes.toBytes("testScannerNextRPCTimesout"); - HTable ht = TEST_UTIL.createTable(TABLE_NAME, FAMILY); + Table ht = TEST_UTIL.createTable(TABLE_NAME, FAMILY); byte[] r1 = Bytes.toBytes("row-1"); byte[] r2 = Bytes.toBytes("row-2"); byte[] r3 = Bytes.toBytes("row-3"); @@ -123,7 +123,7 @@ public class TestClientScannerRPCTimeout { RSRpcServicesWithScanTimeout.tryNumber <= CLIENT_RETRIES_NUMBER); } - private void putToTable(HTable ht, byte[] rowkey) throws IOException { + private void putToTable(Table ht, byte[] rowkey) throws IOException { Put put = new Put(rowkey); put.add(FAMILY, QUALIFIER, VALUE); ht.put(put); http://git-wip-us.apache.org/repos/asf/hbase/blob/4995ed8a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientTimeouts.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientTimeouts.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientTimeouts.java index 1d6ef77..0bb73ca 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientTimeouts.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestClientTimeouts.java @@ -87,7 +87,7 @@ public class TestClientTimeouts { // Ensure the HBaseAdmin uses a new connection by changing Configuration. Configuration conf = HBaseConfiguration.create(TEST_UTIL.getConfiguration()); conf.set(HConstants.HBASE_CLIENT_INSTANCE_ID, String.valueOf(-1)); - HBaseAdmin admin = null; + Admin admin = null; try { admin = new HBaseAdmin(conf); HConnection connection = admin.getConnection();