-org.apache.hadoop.hbase.client.HBaseAdmin.addColumn(TableName, HColumnDescriptor)
-
- |
-
-
org.apache.hadoop.hbase.client.Admin.addColumn(TableName, HColumnDescriptor)
|
+
+org.apache.hadoop.hbase.client.HBaseAdmin.addColumn(TableName, HColumnDescriptor)
+
+ |
+
org.apache.hadoop.hbase.security.visibility.VisibilityClient.addLabel(Configuration, String)
@@ -397,10 +397,10 @@
| org.apache.hadoop.hbase.mapreduce.CellCreator.create(byte[], int, int, byte[], int, int, byte[], int, int, long, byte[], int, int, String) |
-org.apache.hadoop.hbase.regionserver.SimpleRpcSchedulerFactory.create(Configuration, PriorityFunction) |
+org.apache.hadoop.hbase.regionserver.RpcSchedulerFactory.create(Configuration, PriorityFunction) |
-org.apache.hadoop.hbase.regionserver.RpcSchedulerFactory.create(Configuration, PriorityFunction) |
+org.apache.hadoop.hbase.regionserver.SimpleRpcSchedulerFactory.create(Configuration, PriorityFunction) |
org.apache.hadoop.hbase.client.HConnection.deleteCachedRegionLocation(HRegionLocation)
@@ -408,12 +408,6 @@
|
-org.apache.hadoop.hbase.client.HBaseAdmin.deleteColumn(TableName, byte[])
-
- |
-
-
org.apache.hadoop.hbase.client.Admin.deleteColumn(TableName, byte[])
|
+
+org.apache.hadoop.hbase.client.HBaseAdmin.deleteColumn(TableName, byte[])
+
+ |
+
org.apache.hadoop.hbase.KeyValueUtil.ensureKeyValue(Cell)
without any replacement.
@@ -448,15 +448,15 @@
| org.apache.hadoop.hbase.rest.client.RemoteHTable.exists(List) |
-org.apache.hadoop.hbase.filter.Filter.filterRowKey(byte[], int, int)
+ | org.apache.hadoop.hbase.filter.FilterBase.filterRowKey(byte[], int, int)
+ Instead use FilterBase.filterRowKey(Cell)
|
-org.apache.hadoop.hbase.filter.FilterBase.filterRowKey(byte[], int, int)
+ | org.apache.hadoop.hbase.filter.Filter.filterRowKey(byte[], int, int)
+ Instead use Filter.filterRowKey(Cell)
|
@@ -543,56 +543,56 @@
+org.apache.hadoop.hbase.client.HConnection.getHTableDescriptor(byte[])
+ internal method, do not use through HConnection
+ |
+
+
org.apache.hadoop.hbase.client.ConnectionImplementation.getHTableDescriptor(byte[])
|
-
-org.apache.hadoop.hbase.client.HConnection.getHTableDescriptor(byte[])
+ |
+org.apache.hadoop.hbase.client.HConnection.getHTableDescriptor(TableName)
internal method, do not use through HConnection
|
-
+
org.apache.hadoop.hbase.client.ConnectionImplementation.getHTableDescriptor(TableName)
|
-
-org.apache.hadoop.hbase.client.HConnection.getHTableDescriptor(TableName)
- internal method, do not use through HConnection
+ |
+org.apache.hadoop.hbase.client.HConnection.getHTableDescriptors(List)
+ since 0.96.0
|
-
+
org.apache.hadoop.hbase.client.ConnectionImplementation.getHTableDescriptors(List)
|
-
-org.apache.hadoop.hbase.client.HConnection.getHTableDescriptors(List)
- since 0.96.0
- |
-
-org.apache.hadoop.hbase.client.ConnectionImplementation.getHTableDescriptorsByTableName(List)
-
+ | org.apache.hadoop.hbase.client.HConnection.getHTableDescriptorsByTableName(List)
+
|
-org.apache.hadoop.hbase.client.HConnection.getHTableDescriptorsByTableName(List)
-
+ | org.apache.hadoop.hbase.client.ConnectionImplementation.getHTableDescriptorsByTableName(List)
+
|
-org.apache.hadoop.hbase.client.ClusterConnection.getKeepAliveMasterService()
+ | org.apache.hadoop.hbase.client.HConnection.getKeepAliveMasterService()
Since 0.96.0
|
-org.apache.hadoop.hbase.client.HConnection.getKeepAliveMasterService()
+ | org.apache.hadoop.hbase.client.ClusterConnection.getKeepAliveMasterService()
Since 0.96.0
|
@@ -658,22 +658,22 @@
-org.apache.hadoop.hbase.client.ConnectionImplementation.getRegionCachePrefetch(byte[])
+ | org.apache.hadoop.hbase.client.HConnection.getRegionCachePrefetch(byte[])
always return false since 0.99
|
-org.apache.hadoop.hbase.client.HConnection.getRegionCachePrefetch(byte[])
+ | org.apache.hadoop.hbase.client.ConnectionImplementation.getRegionCachePrefetch(byte[])
always return false since 0.99
|
-org.apache.hadoop.hbase.client.ConnectionImplementation.getRegionCachePrefetch(TableName)
+ | org.apache.hadoop.hbase.client.HConnection.getRegionCachePrefetch(TableName)
always return false since 0.99
|
-org.apache.hadoop.hbase.client.HConnection.getRegionCachePrefetch(TableName)
+ | org.apache.hadoop.hbase.client.ConnectionImplementation.getRegionCachePrefetch(TableName)
always return false since 0.99
|
@@ -726,13 +726,13 @@
-org.apache.hadoop.hbase.util.Bytes.getSize()
-
+ | org.apache.hadoop.hbase.io.ImmutableBytesWritable.getSize()
+
|
-org.apache.hadoop.hbase.io.ImmutableBytesWritable.getSize()
-
+ | org.apache.hadoop.hbase.util.Bytes.getSize()
+
|
@@ -750,13 +750,13 @@
-org.apache.hadoop.hbase.client.ConnectionImplementation.getTableNames()
-
+ | org.apache.hadoop.hbase.client.HConnection.getTableNames()
+
|
-org.apache.hadoop.hbase.client.HConnection.getTableNames()
-
+ | org.apache.hadoop.hbase.client.ConnectionImplementation.getTableNames()
+
|
@@ -804,12 +804,12 @@
org.apache.hadoop.hbase.KeyValue.iscreate(InputStream) |
-org.apache.hadoop.hbase.client.ClusterConnection.isDeadServer(ServerName)
+ | org.apache.hadoop.hbase.client.HConnection.isDeadServer(ServerName)
internal method, do not use thru HConnection
|
-org.apache.hadoop.hbase.client.HConnection.isDeadServer(ServerName)
+ | org.apache.hadoop.hbase.client.ClusterConnection.isDeadServer(ServerName)
internal method, do not use thru HConnection
|
@@ -820,8 +820,8 @@
org.apache.hadoop.hbase.master.cleaner.BaseLogCleanerDelegate.isLogDeletable(FileStatus) |
-org.apache.hadoop.hbase.client.ClusterConnection.isMasterRunning()
- this has been deprecated without a replacement
+ | org.apache.hadoop.hbase.client.HConnection.isMasterRunning()
+ internal method, do not use thru HConnection
|
@@ -830,8 +830,8 @@
-org.apache.hadoop.hbase.client.HConnection.isMasterRunning()
- internal method, do not use thru HConnection
+ | org.apache.hadoop.hbase.client.ClusterConnection.isMasterRunning()
+ this has been deprecated without a replacement
|
@@ -865,23 +865,23 @@
-org.apache.hadoop.hbase.client.ConnectionImplementation.listTableNames()
-
+ | org.apache.hadoop.hbase.client.HConnection.listTableNames()
+
|
-org.apache.hadoop.hbase.client.HConnection.listTableNames()
-
+ | org.apache.hadoop.hbase.client.ConnectionImplementation.listTableNames()
+
|
-org.apache.hadoop.hbase.client.ConnectionImplementation.listTables()
-
+ | org.apache.hadoop.hbase.client.HConnection.listTables()
+
|
-org.apache.hadoop.hbase.client.HConnection.listTables()
-
+ | org.apache.hadoop.hbase.client.ConnectionImplementation.listTables()
+
|
@@ -926,12 +926,6 @@
-org.apache.hadoop.hbase.client.HBaseAdmin.modifyColumn(TableName, HColumnDescriptor)
-
- |
-
-
org.apache.hadoop.hbase.client.Admin.modifyColumn(TableName, HColumnDescriptor)
|
+
+org.apache.hadoop.hbase.client.HBaseAdmin.modifyColumn(TableName, HColumnDescriptor)
+
+ |
+
org.apache.hadoop.hbase.KeyValue.oswrite(KeyValue, OutputStream, boolean)
As of release 2.0.0, this will be removed in HBase 3.0.0.
@@ -1526,44 +1526,44 @@
|
-org.apache.hadoop.hbase.client.ConnectionImplementation.processBatch(List extends Row>, byte[], ExecutorService, Object[])
- Unsupported API
+ | org.apache.hadoop.hbase.client.HConnection.processBatch(List extends Row>, byte[], ExecutorService, Object[])
+ internal method, do not use through HConnection
|
-org.apache.hadoop.hbase.client.HConnection.processBatch(List extends Row>, byte[], ExecutorService, Object[])
- internal method, do not use through HConnection
+ | org.apache.hadoop.hbase.client.ConnectionImplementation.processBatch(List extends Row>, byte[], ExecutorService, Object[])
+ Unsupported API
|
-org.apache.hadoop.hbase.client.ConnectionImplementation.processBatch(List extends Row>, TableName, ExecutorService, Object[])
-
+ | org.apache.hadoop.hbase.client.HConnection.processBatch(List extends Row>, TableName, ExecutorService, Object[])
+
|
-org.apache.hadoop.hbase.client.HConnection.processBatch(List extends Row>, TableName, ExecutorService, Object[])
-
+ | org.apache.hadoop.hbase.client.ConnectionImplementation.processBatch(List extends Row>, TableName, ExecutorService, Object[])
+
|
-org.apache.hadoop.hbase.client.ConnectionImplementation.processBatchCallback(List extends Row>, byte[], ExecutorService, Object[], Batch.Callback)
+ | org.apache.hadoop.hbase.client.HConnection.processBatchCallback(List extends Row>, byte[], ExecutorService, Object[], Batch.Callback)
Unsupported API
|
-org.apache.hadoop.hbase.client.HConnection.processBatchCallback(List extends Row>, byte[], ExecutorService, Object[], Batch.Callback)
+ | org.apache.hadoop.hbase.client.ConnectionImplementation.processBatchCallback(List extends Row>, byte[], ExecutorService, Object[], Batch.Callback)
Unsupported API
|
-org.apache.hadoop.hbase.client.ConnectionImplementation.processBatchCallback(List extends Row>, TableName, ExecutorService, Object[], Batch.Callback)
-
+ | org.apache.hadoop.hbase.client.HConnection.processBatchCallback(List extends Row>, TableName, ExecutorService, Object[], Batch.Callback)
+
|
-org.apache.hadoop.hbase.client.HConnection.processBatchCallback(List extends Row>, TableName, ExecutorService, Object[], Batch.Callback)
-
+ | org.apache.hadoop.hbase.client.ConnectionImplementation.processBatchCallback(List extends Row>, TableName, ExecutorService, Object[], Batch.Callback)
+
|
@@ -1703,25 +1703,25 @@
org.apache.hadoop.hbase.http.HttpServer.Builder.setPort(int) |
-org.apache.hadoop.hbase.client.ConnectionImplementation.setRegionCachePrefetch(byte[], boolean)
- always return false since 0.99
- |
-
-
org.apache.hadoop.hbase.client.HConnection.setRegionCachePrefetch(byte[], boolean)
does nothing since 0.99
|
-
-org.apache.hadoop.hbase.client.ConnectionImplementation.setRegionCachePrefetch(TableName, boolean)
+ |
+org.apache.hadoop.hbase.client.ConnectionImplementation.setRegionCachePrefetch(byte[], boolean)
always return false since 0.99
|
-
+
org.apache.hadoop.hbase.client.HConnection.setRegionCachePrefetch(TableName, boolean)
does nothing since since 0.99
|
+
+org.apache.hadoop.hbase.client.ConnectionImplementation.setRegionCachePrefetch(TableName, boolean)
+ always return false since 0.99
+ |
+
org.apache.hadoop.hbase.client.HTableInterface.setWriteBufferSize(long)
|
-org.apache.hadoop.hbase.client.Table.setWriteBufferSize(long)
-
+ | org.apache.hadoop.hbase.client.BufferedMutatorImpl.setWriteBufferSize(long)
+
|
-org.apache.hadoop.hbase.client.BufferedMutatorImpl.setWriteBufferSize(long)
-
+ | org.apache.hadoop.hbase.client.Table.setWriteBufferSize(long)
+
|
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/96543536/devapidocs/index-all.html
----------------------------------------------------------------------
diff --git a/devapidocs/index-all.html b/devapidocs/index-all.html
index c0d61fa..acce8ec 100644
--- a/devapidocs/index-all.html
+++ b/devapidocs/index-all.html
@@ -16383,7 +16383,7 @@
createWAL() - Method in class org.apache.hadoop.hbase.wal.AsyncFSWALProvider
-createWAL() - Method in class org.apache.hadoop.hbase.wal.DefaultWALProvider
+createWAL() - Method in class org.apache.hadoop.hbase.wal.FSHLogProvider
createWALWriter(FileSystem, Path) - Method in class org.apache.hadoop.hbase.wal.WALFactory
@@ -16429,7 +16429,7 @@
createWriter() - Method in class org.apache.hadoop.hbase.regionserver.StripeStoreFlusher.StripeFlushRequest
-createWriter(Configuration, FileSystem, Path, boolean) - Static method in class org.apache.hadoop.hbase.wal.DefaultWALProvider
+createWriter(Configuration, FileSystem, Path, boolean) - Static method in class org.apache.hadoop.hbase.wal.FSHLogProvider
public because of FSHLog.
@@ -18685,14 +18685,6 @@
DefaultVisibilityLabelServiceImpl() - Constructor for class org.apache.hadoop.hbase.security.visibility.DefaultVisibilityLabelServiceImpl
-DefaultWALProvider - Class in org.apache.hadoop.hbase.wal
-
-A WAL provider that use
FSHLog
.
-
-DefaultWALProvider() - Constructor for class org.apache.hadoop.hbase.wal.DefaultWALProvider
-
-DefaultWALProvider.Writer - Interface in org.apache.hadoop.hbase.wal
-
DEFERRED_LOG_FLUSH - Static variable in class org.apache.hadoop.hbase.HTableDescriptor
Deprecated.
@@ -20253,7 +20245,7 @@
doInit(Configuration) - Method in class org.apache.hadoop.hbase.wal.AsyncFSWALProvider
-
doInit(Configuration) - Method in class org.apache.hadoop.hbase.wal.DefaultWALProvider
+
doInit(Configuration) - Method in class org.apache.hadoop.hbase.wal.FSHLogProvider
doKerberosAuth(HttpServletRequest) - Method in class org.apache.hadoop.hbase.thrift.ThriftHttpServlet
@@ -25900,6 +25892,14 @@
UPDATE!
+FSHLogProvider - Class in org.apache.hadoop.hbase.wal
+
+A WAL provider that use
FSHLog
.
+
+
FSHLogProvider() - Constructor for class org.apache.hadoop.hbase.wal.FSHLogProvider
+
+
FSHLogProvider.Writer - Interface in org.apache.hadoop.hbase.wal
+
FSMapRUtils - Class in org.apache.hadoop.hbase.util
@@ -26011,7 +26011,7 @@
FSWALEntry - Class in org.apache.hadoop.hbase.regionserver.wal
-A WAL Entry for
FSHLog
implementation.
+
FSWALEntry(long, WALKey, WALEdit, HRegionInfo, boolean) - Constructor for class org.apache.hadoop.hbase.regionserver.wal.FSWALEntry
@@ -34617,7 +34617,7 @@
Currently, we need to expose the writer's OutputStream to tests so that they can manipulate the
default behavior (such as setting the maxRecoveryErrorCount value for example (see
- TestWALReplay.testReplayEditsWrittenIntoWAL()
).
+ AbstractTestWALReplay.testReplayEditsWrittenIntoWAL()
).
getOutputStream() - Method in class org.apache.hadoop.hbase.rest.filter.GZIPResponseWrapper
@@ -46450,10 +46450,10 @@
init(Configuration, String) - Method in class org.apache.hadoop.hbase.wal.BoundedGroupingStrategy
-init(FileSystem, Path, Configuration, boolean) - Method in interface org.apache.hadoop.hbase.wal.DefaultWALProvider.Writer
-
init(WALFactory, Configuration, List<WALActionsListener>, String) - Method in class org.apache.hadoop.hbase.wal.DisabledWALProvider
+init(FileSystem, Path, Configuration, boolean) - Method in interface org.apache.hadoop.hbase.wal.FSHLogProvider.Writer
+
init(Configuration, String) - Method in class org.apache.hadoop.hbase.wal.NamespaceGroupingStrategy
init(Configuration, String) - Method in class org.apache.hadoop.hbase.wal.RegionGroupingProvider.IdentityGroupingStrategy
@@ -53949,10 +53949,10 @@
LOG - Static variable in class org.apache.hadoop.hbase.wal.AsyncFSWALProvider
-LOG - Static variable in class org.apache.hadoop.hbase.wal.DefaultWALProvider
-
LOG - Static variable in class org.apache.hadoop.hbase.wal.DisabledWALProvider
+LOG - Static variable in class org.apache.hadoop.hbase.wal.FSHLogProvider
+
LOG - Static variable in class org.apache.hadoop.hbase.wal.RegionGroupingProvider
LOG - Static variable in class org.apache.hadoop.hbase.wal.WALFactory
@@ -91216,6 +91216,8 @@ service.
tinfo - Variable in class org.apache.hadoop.hbase.ipc.RpcServer.Call
+TMP_DIR - Static variable in class org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles
+
TMP_DIR - Static variable in class org.apache.hadoop.hbase.util.FSTableDescriptors
TMP_JARS_DIR - Static variable in class org.apache.hadoop.hbase.util.CoprocessorClassLoader
@@ -92344,9 +92346,9 @@ service.
toString() - Method in class org.apache.hadoop.hbase.regionserver.TimeRangeTracker
-toString() - Method in class org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.Payload
+toString() - Method in class org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL
-toString() - Method in class org.apache.hadoop.hbase.regionserver.wal.FSHLog
+toString() - Method in class org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.Payload
toString() - Method in class org.apache.hadoop.hbase.regionserver.wal.FSWALEntry
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/96543536/devapidocs/org/apache/hadoop/hbase/HealthChecker.HealthCheckerExitStatus.html
----------------------------------------------------------------------
diff --git a/devapidocs/org/apache/hadoop/hbase/HealthChecker.HealthCheckerExitStatus.html b/devapidocs/org/apache/hadoop/hbase/HealthChecker.HealthCheckerExitStatus.html
index 97174a2..e386993 100644
--- a/devapidocs/org/apache/hadoop/hbase/HealthChecker.HealthCheckerExitStatus.html
+++ b/devapidocs/org/apache/hadoop/hbase/HealthChecker.HealthCheckerExitStatus.html
@@ -258,7 +258,7 @@ the order they are declared.