Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 53CA2200C21 for ; Sun, 5 Feb 2017 16:49:36 +0100 (CET) Received: by cust-asf.ponee.io (Postfix) id 52A54160B32; Sun, 5 Feb 2017 15:49:36 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 0A03D160B70 for ; Sun, 5 Feb 2017 16:49:33 +0100 (CET) Received: (qmail 18558 invoked by uid 500); 5 Feb 2017 15:49:28 -0000 Mailing-List: contact commits-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list commits@hbase.apache.org Received: (qmail 15674 invoked by uid 99); 5 Feb 2017 15:49:24 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 05 Feb 2017 15:49:24 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id D70D5DFC31; Sun, 5 Feb 2017 15:49:24 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: tedyu@apache.org To: commits@hbase.apache.org Date: Sun, 05 Feb 2017 15:50:02 -0000 Message-Id: <27352ac4d55148cb95c83587d0a7a92f@git.apache.org> In-Reply-To: References: X-Mailer: ASF-Git Admin Mailer Subject: [40/52] [partial] hbase-site git commit: Published site at 26a94844f533b95db1f0a58d6a7cc3dc4a7a7098. archived-at: Sun, 05 Feb 2017 15:49:36 -0000 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/62e361eb/apidocs/org/apache/hadoop/hbase/client/Admin.html ---------------------------------------------------------------------- diff --git a/apidocs/org/apache/hadoop/hbase/client/Admin.html b/apidocs/org/apache/hadoop/hbase/client/Admin.html index 1868dd7..6d71c8d 100644 --- a/apidocs/org/apache/hadoop/hbase/client/Admin.html +++ b/apidocs/org/apache/hadoop/hbase/client/Admin.html @@ -18,8 +18,8 @@ catch(err) { } //--> -var methods = {"i0":6,"i1":6,"i2":6,"i3":38,"i4":6,"i5":6,"i6":6,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":38,"i33":6,"i34":6,"i35":6,"i36":6,"i37":6,"i38":6,"i39":6,"i40":6,"i41":6,"i42":6,"i43":6,"i44":6,"i45":6,"i46":6,"i47":6,"i48":6,"i49":6,"i50":6,"i51":6,"i52":6,"i53":6,"i54":6,"i55":6,"i56":6,"i57":6,"i58":6,"i59":6,"i60":38,"i61":6,"i62":6,"i63":6,"i64":6,"i65":6,"i66":6,"i67":6,"i68":6,"i69":6,"i70":6,"i71":6,"i72":6,"i73":6,"i74":6,"i75":6,"i76":6,"i77":6,"i78":6,"i79":6,"i80":6,"i81":6,"i82":6,"i83":6,"i84":6,"i85":6,"i86":6,"i87":6,"i88":6,"i89":6,"i90":6,"i91":6,"i92":6,"i93":6,"i94":6,"i95":6,"i96":6,"i97":6,"i98":6,"i99":6,"i100":6,"i101":6,"i102":6,"i103":6,"i104":6,"i105":6,"i106":6,"i107":6,"i108":6,"i109":6,"i110":6,"i111":6,"i112":6,"i113":6,"i114":6,"i115":6,"i116":6,"i117":6,"i118":6,"i119":6,"i120":38 ,"i121":6,"i122":6,"i123":38,"i124":6,"i125":6,"i126":6,"i127":6,"i128":6,"i129":6,"i130":6,"i131":6,"i132":6,"i133":6,"i134":6,"i135":6,"i136":6,"i137":6,"i138":6,"i139":6,"i140":6,"i141":6,"i142":6,"i143":6,"i144":6,"i145":6,"i146":6,"i147":6,"i148":6,"i149":6,"i150":6,"i151":6,"i152":6,"i153":6,"i154":6,"i155":6,"i156":6,"i157":6,"i158":6,"i159":6,"i160":6,"i161":6}; -var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],32:["t6","Deprecated Methods"]}; +var methods = {"i0":6,"i1":6,"i2":6,"i3":38,"i4":6,"i5":6,"i6":18,"i7":6,"i8":6,"i9":6,"i10":6,"i11":6,"i12":6,"i13":6,"i14":6,"i15":6,"i16":6,"i17":6,"i18":6,"i19":6,"i20":6,"i21":6,"i22":6,"i23":6,"i24":6,"i25":6,"i26":6,"i27":6,"i28":6,"i29":6,"i30":6,"i31":6,"i32":6,"i33":38,"i34":6,"i35":6,"i36":6,"i37":6,"i38":6,"i39":6,"i40":6,"i41":6,"i42":6,"i43":6,"i44":6,"i45":6,"i46":6,"i47":6,"i48":18,"i49":6,"i50":6,"i51":6,"i52":6,"i53":6,"i54":6,"i55":6,"i56":18,"i57":6,"i58":6,"i59":6,"i60":6,"i61":6,"i62":6,"i63":6,"i64":6,"i65":6,"i66":38,"i67":6,"i68":6,"i69":6,"i70":6,"i71":6,"i72":6,"i73":6,"i74":6,"i75":6,"i76":6,"i77":6,"i78":6,"i79":6,"i80":6,"i81":6,"i82":6,"i83":6,"i84":18,"i85":6,"i86":6,"i87":6,"i88":6,"i89":6,"i90":6,"i91":6,"i92":6,"i93":6,"i94":6,"i95":6,"i96":6,"i97":6,"i98":6,"i99":6,"i100":6,"i101":6,"i102":6,"i103":6,"i104":6,"i105":6,"i106":6,"i107":18,"i108":18,"i109":18,"i110":6,"i111":6,"i112":6,"i113":6,"i114":6,"i115":6,"i116":6,"i117":6,"i118":6,"i119":6,"i 120":6,"i121":6,"i122":6,"i123":6,"i124":6,"i125":6,"i126":6,"i127":6,"i128":6,"i129":6,"i130":6,"i131":6,"i132":6,"i133":38,"i134":6,"i135":6,"i136":38,"i137":6,"i138":6,"i139":6,"i140":6,"i141":6,"i142":6,"i143":6,"i144":6,"i145":6,"i146":6,"i147":18,"i148":6,"i149":6,"i150":6,"i151":6,"i152":6,"i153":6,"i154":6,"i155":6,"i156":6,"i157":6,"i158":6,"i159":6,"i160":6,"i161":6,"i162":6,"i163":6,"i164":6,"i165":6,"i166":6,"i167":6,"i168":6,"i169":6,"i170":6,"i171":6,"i172":6,"i173":6,"i174":6,"i175":6,"i176":6,"i177":6,"i178":6,"i179":18}; +var tabs = {65535:["t0","All Methods"],2:["t2","Instance Methods"],4:["t3","Abstract Methods"],16:["t5","Default Methods"],32:["t6","Deprecated Methods"]}; var altColor = "altColor"; var rowColor = "rowColor"; var tableTab = "tableTab"; @@ -107,7 +107,7 @@ var activeTableTab = "activeTableTab";
@InterfaceAudience.Public
  @InterfaceStability.Evolving
-public interface Admin
+public interface Admin
 extends org.apache.hadoop.hbase.Abortable, Closeable
The administrative API for HBase. Obtain an instance from an Connection.getAdmin() and call close() afterwards. @@ -134,7 +134,7 @@ extends org.apache.hadoop.hbase.Abortable, -All Methods Instance Methods Abstract Methods Deprecated Methods  +All Methods Instance Methods Abstract Methods Default Methods Deprecated Methods  Modifier and Type Method and Description @@ -187,36 +187,43 @@ extends org.apache.hadoop.hbase.Abortable, +default void +addReplicationPeer(String peerId, + ReplicationPeerConfig peerConfig) +
Add a new replication peer for replicating data to slave cluster
+ + + void assign(byte[] regionName)  - + boolean balancer()
Invoke the balancer.
- + boolean balancer(boolean force)
Invoke the balancer.
- + void cloneSnapshot(byte[] snapshotName, TableName tableName)
Create a new table by cloning the snapshot content.
- + void cloneSnapshot(String snapshotName, TableName tableName)
Create a new table by cloning the snapshot content.
- + Future<Void> cloneSnapshotAsync(String snapshotName, TableName tableName) @@ -224,52 +231,52 @@ extends org.apache.hadoop.hbase.Abortable, + void close()  - + void closeRegion(byte[] regionname, String serverName)
Close a region.
- + void closeRegion(ServerName sn, HRegionInfo hri)
Close a region.
- + void closeRegion(String regionname, String serverName)
Close a region.
- + boolean closeRegionWithEncodedRegionName(String encodedRegionName, String serverName)
For expert-admins.
- + void compact(TableName tableName)
Compact a table.
- + void compact(TableName tableName, byte[] columnFamily)
Compact a column family within a table.
- + void compact(TableName tableName, byte[] columnFamily, @@ -277,73 +284,73 @@ extends org.apache.hadoop.hbase.Abortable, Compact a column family within a table.
- + void compact(TableName tableName, CompactType compactType)
Compact a table.
- + void compactRegion(byte[] regionName)
Compact an individual region.
- + void compactRegion(byte[] regionName, byte[] columnFamily)
Compact a column family within a region.
- + void compactRegionServer(ServerName sn, boolean major)
Compact all regions on the region server
- + CoprocessorRpcChannel coprocessorService()
Creates and returns a RpcChannel instance connected to the active master.
- + CoprocessorRpcChannel coprocessorService(ServerName sn)
Creates and returns a RpcChannel instance connected to the passed region server.
- + void createNamespace(NamespaceDescriptor descriptor)
Create a new namespace.
- + Future<Void> createNamespaceAsync(NamespaceDescriptor descriptor)
Create a new namespace
- + void createTable(HTableDescriptor desc)
Creates a new table.
- + void createTable(HTableDescriptor desc, byte[][] splitKeys)
Creates a new table with an initial set of empty regions defined by the specified split keys.
- + void createTable(HTableDescriptor desc, byte[] startKey, @@ -352,14 +359,14 @@ extends org.apache.hadoop.hbase.Abortable, Creates a new table with the specified number of regions. - + Future<Void> createTableAsync(HTableDescriptor desc, byte[][] splitKeys)
Creates a new table but does not block and wait for it to come online.
- + void deleteColumn(TableName tableName, byte[] columnFamily) @@ -371,81 +378,81 @@ extends org.apache.hadoop.hbase.Abortable, + void deleteColumnFamily(TableName tableName, byte[] columnFamily)
Delete a column family from a table.
- + Future<Void> deleteColumnFamilyAsync(TableName tableName, byte[] columnFamily)
Delete a column family from a table.
- + void deleteNamespace(String name)
Delete an existing namespace.
- + Future<Void> deleteNamespaceAsync(String name)
Delete an existing namespace.
- + void deleteSnapshot(byte[] snapshotName)
Delete an existing snapshot.
- + void deleteSnapshot(String snapshotName)
Delete an existing snapshot.
- + void deleteSnapshots(Pattern pattern)
Delete existing snapshots whose names match the pattern passed.
- + void deleteSnapshots(String regex)
Delete existing snapshots whose names match the pattern passed.
- + void deleteTable(TableName tableName)
Deletes a table.
- + Future<Void> deleteTableAsync(TableName tableName)
Deletes the table but does not block and wait for it be completely removed.
- + HTableDescriptor[] deleteTables(Pattern pattern)
Delete tables matching the passed in pattern and wait on completion.
- + HTableDescriptor[] deleteTables(String regex)
Deletes tables matching the passed in pattern and wait on completion.
- + void deleteTableSnapshots(Pattern tableNamePattern, Pattern snapshotNamePattern) @@ -453,7 +460,7 @@ extends org.apache.hadoop.hbase.Abortable, + void deleteTableSnapshots(String tableNameRegex, String snapshotNameRegex) @@ -461,61 +468,91 @@ extends org.apache.hadoop.hbase.Abortable, + +default void +disableReplicationPeer(String peerId) +
Stop the replication stream to the specified peer
+ + + void disableTable(TableName tableName)
Disable table and wait on completion.
- + Future<Void> disableTableAsync(TableName tableName)
Disable the table but does not block and wait for it be completely disabled.
- + +void +disableTableReplication(TableName tableName) +
Disable a table's replication switch.
+ + + HTableDescriptor[] disableTables(Pattern pattern)
Disable tables matching the passed in pattern and wait on completion.
- + HTableDescriptor[] disableTables(String regex)
Disable tables matching the passed in pattern and wait on completion.
- + +void +drainRegionServers(List<ServerName> servers) +
Mark a region server as draining to prevent additional regions from getting assigned to it.
+ + + boolean enableCatalogJanitor(boolean enable)
Enable/Disable the catalog janitor
- + +default void +enableReplicationPeer(String peerId) +
Restart the replication stream to the specified peer
+ + + void enableTable(TableName tableName)
Enable a table.
- + Future<Void> enableTableAsync(TableName tableName)
Enable the table but does not block and wait for it be completely enabled.
- + +void +enableTableReplication(TableName tableName) +
Enable a table's replication switch.
+ + + HTableDescriptor[] enableTables(Pattern pattern)
Enable tables matching the passed in pattern and wait on completion.
- + HTableDescriptor[] enableTables(String regex)
Enable tables matching the passed in pattern and wait on completion.
- + void execProcedure(String signature, String instance, @@ -523,7 +560,7 @@ extends org.apache.hadoop.hbase.Abortable, Execute a distributed procedure on a cluster. - + byte[] execProcedureWithRet(String signature, String instance, @@ -531,19 +568,19 @@ extends org.apache.hadoop.hbase.Abortable, Execute a distributed procedure on a cluster. - + void flush(TableName tableName)
Flush a table.
- + void flushRegion(byte[] regionName)
Flush an individual region.
- + Pair<Integer,Integer> getAlterStatus(byte[] tableName)
Deprecated.  @@ -552,45 +589,45 @@ extends org.apache.hadoop.hbase.Abortable, + Pair<Integer,Integer> getAlterStatus(TableName tableName)
Get the status of alter command - indicates how many regions have received the updated schema Asynchronous operation.
- + ClusterStatus getClusterStatus()  - + CompactionState getCompactionState(TableName tableName)
Get the current compaction state of a table.
- + CompactionState getCompactionState(TableName tableName, CompactType compactType)
Get the current compaction state of a table.
- + CompactionState getCompactionStateForRegion(byte[] regionName)
Get the current compaction state of region.
- + org.apache.hadoop.conf.Configuration getConfiguration()  - + Connection getConnection()  - + long getLastMajorCompactionTimestamp(TableName tableName)
Get the timestamp of the last major compaction for the passed table @@ -599,120 +636,132 @@ extends org.apache.hadoop.hbase.Abortable, + long getLastMajorCompactionTimestampForRegion(byte[] regionName)
Get the timestamp of the last major compaction for the passed region.
- + String[] getMasterCoprocessors()
Helper delegage to getClusterStatus().getMasterCoprocessors().
- + int getMasterInfoPort()
Get the info port of the current master if one is available.
- + NamespaceDescriptor getNamespaceDescriptor(String name)
Get a namespace descriptor by name
- + List<HRegionInfo> getOnlineRegions(ServerName sn)
Get all the online regions on a region server.
- + int getOperationTimeout()  - + QuotaRetriever getQuotaRetriever(QuotaFilter filter)
Return a QuotaRetriever to list the quotas based on the filter.
- + Map<byte[],RegionLoad> getRegionLoad(ServerName sn)
Get RegionLoad of all regions hosted on a regionserver.
- + Map<byte[],RegionLoad> getRegionLoad(ServerName sn, TableName tableName)
Get RegionLoad of all regions hosted on a regionserver for a table.
- + +default ReplicationPeerConfig +getReplicationPeerConfig(String peerId) +
Returns the configured ReplicationPeerConfig for the specified peer
+ + + List<SecurityCapability> getSecurityCapabilities()
Return the set of supported security capabilities.
- + HTableDescriptor getTableDescriptor(TableName tableName)
Method for getting the tableDescriptor
- + HTableDescriptor[] getTableDescriptors(List<String> names)
Get tableDescriptors
- + HTableDescriptor[] getTableDescriptorsByTableName(List<TableName> tableNames)
Get tableDescriptors
- + List<HRegionInfo> getTableRegions(TableName tableName)
Get the regions of a given table.
- + boolean isAborted()
Check if the server or client was aborted.
- + boolean isBalancerEnabled()
Query the current state of the balancer
- + boolean isCatalogJanitorEnabled()
Query on the catalog janitor state (Enabled/Disabled?)
- + +boolean +isCleanerChoreEnabled() +
Query on the cleaner chore state (Enabled/Disabled?)
+ + + boolean isMasterInMaintenanceMode()
Check whether Master is in maintenance mode
- + boolean isNormalizerEnabled()
Query the current state of the region normalizer
- + boolean isProcedureFinished(String signature, String instance, @@ -720,23 +769,23 @@ extends org.apache.hadoop.hbase.Abortable, Check the current state of the specified procedure.
- + boolean isSnapshotFinished(SnapshotDescription snapshot)
Check the current state of the passed snapshot.
- + boolean isSplitOrMergeEnabled(MasterSwitchType switchType)
Query the current state of the switch
- + boolean isTableAvailable(TableName tableName)  - + boolean isTableAvailable(TableName tableName, byte[][] splitKeys) @@ -744,121 +793,151 @@ extends org.apache.hadoop.hbase.Abortable, + boolean isTableDisabled(TableName tableName)  - + boolean isTableEnabled(TableName tableName)  - + +List<ServerName> +listDrainingRegionServers() +
List region servers marked as draining to not get additional regions assigned to them.
+ + + NamespaceDescriptor[] listNamespaceDescriptors()
List available namespace descriptors
- + ProcedureInfo[] listProcedures()
List procedures
- + +List<TableCFs> +listReplicatedTableCFs() +
Find all table and column families that are replicated from this cluster
+ + + +default List<ReplicationPeerDescription> +listReplicationPeers() +
Return a list of replication peers.
+ + + +default List<ReplicationPeerDescription> +listReplicationPeers(Pattern pattern) +
Return a list of replication peers.
+ + + +default List<ReplicationPeerDescription> +listReplicationPeers(String regex) +
Return a list of replication peers.
+ + + List<SnapshotDescription> listSnapshots()
List completed snapshots.
- + List<SnapshotDescription> listSnapshots(Pattern pattern)
List all the completed snapshots matching the given pattern.
- + List<SnapshotDescription> listSnapshots(String regex)
List all the completed snapshots matching the given regular expression.
- + HTableDescriptor[] listTableDescriptorsByNamespace(String name)
Get list of table descriptors by namespace
- + TableName[] listTableNames()
List all of the names of userspace tables.
- + TableName[] listTableNames(Pattern pattern)
List all of the names of userspace tables.
- + TableName[] listTableNames(Pattern pattern, boolean includeSysTables)
List all of the names of userspace tables.
- + TableName[] listTableNames(String regex)
List all of the names of userspace tables.
- + TableName[] listTableNames(String regex, boolean includeSysTables)
List all of the names of userspace tables.
- + TableName[] listTableNamesByNamespace(String name)
Get list of table names by namespace
- + HTableDescriptor[] listTables()
List all the userspace tables.
- + HTableDescriptor[] listTables(Pattern pattern)
List all the userspace tables matching the given pattern.
- + HTableDescriptor[] listTables(Pattern pattern, boolean includeSysTables)
List all the tables matching the given pattern.
- + HTableDescriptor[] listTables(String regex)
List all the userspace tables matching the given regular expression.
- + HTableDescriptor[] listTables(String regex, boolean includeSysTables)
List all the tables matching the given pattern.
- + List<SnapshotDescription> listTableSnapshots(Pattern tableNamePattern, Pattern snapshotNamePattern) @@ -866,7 +945,7 @@ extends org.apache.hadoop.hbase.Abortable, + List<SnapshotDescription> listTableSnapshots(String tableNameRegex, String snapshotNameRegex) @@ -874,20 +953,20 @@ extends org.apache.hadoop.hbase.Abortable, + void majorCompact(TableName tableName)
Major compact a table.
- + void majorCompact(TableName tableName, byte[] columnFamily)
Major compact a column family within a table.
- + void majorCompact(TableName tableName, byte[] columnFamily, @@ -895,27 +974,27 @@ extends org.apache.hadoop.hbase.Abortable, Major compact a column family within a table.
- + void majorCompact(TableName tableName, CompactType compactType)
Major compact a table.
- + void majorCompactRegion(byte[] regionName)
Major compact a table or an individual region.
- + void majorCompactRegion(byte[] regionName, byte[] columnFamily)
Major compact a column family within region.
- + void mergeRegions(byte[] nameOfRegionA, byte[] nameOfRegionB, @@ -926,14 +1005,14 @@ extends org.apache.hadoop.hbase.Abortable, + Future<Void> mergeRegionsAsync(byte[][] nameofRegionsToMerge, boolean forcible)
Merge regions.
- + Future<Void> mergeRegionsAsync