Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 31BAF200C48 for ; Thu, 6 Apr 2017 19:30:40 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id 30428160B81; Thu, 6 Apr 2017 17:30:40 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 2CBDA160B91 for ; Thu, 6 Apr 2017 19:30:38 +0200 (CEST) Received: (qmail 96162 invoked by uid 500); 6 Apr 2017 17:30:27 -0000 Mailing-List: contact commits-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list commits@hbase.apache.org Received: (qmail 91683 invoked by uid 99); 6 Apr 2017 17:30:23 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 06 Apr 2017 17:30:23 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id 90AE6F49E7; Thu, 6 Apr 2017 17:30:22 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: git-site-role@apache.org To: commits@hbase.apache.org Date: Thu, 06 Apr 2017 17:30:51 -0000 Message-Id: In-Reply-To: <91bd3fd8b37f4458b1c09883f7dcaf48@git.apache.org> References: <91bd3fd8b37f4458b1c09883f7dcaf48@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [31/51] [partial] hbase-site git commit: Published site at ec5188df3090d42088b6f4cb8f0c2fd49425f8c1. archived-at: Thu, 06 Apr 2017 17:30:40 -0000 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/b64f8dea/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html ---------------------------------------------------------------------- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html index 83dcf5c..1bee57f 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html @@ -262,16 +262,16 @@ service. -private static HRegionInfo -MetaTableAccessor.getHRegionInfo(Result r, +private static Optional<HRegionInfo> +AsyncMetaTableAccessor.getHRegionInfo(Result r, byte[] qualifier)
Returns the HRegionInfo object from the column HConstants.CATALOG_FAMILY and qualifier of the catalog table result.
-private static Optional<HRegionInfo> -AsyncMetaTableAccessor.getHRegionInfo(Result r, +private static HRegionInfo +MetaTableAccessor.getHRegionInfo(Result r, byte[] qualifier)
Returns the HRegionInfo object from the column HConstants.CATALOG_FAMILY and qualifier of the catalog table result.
@@ -286,7 +286,7 @@ service. private static HRegionLocation -MetaTableAccessor.getRegionLocation(Result r, +AsyncMetaTableAccessor.getRegionLocation(Result r, HRegionInfo regionInfo, int replicaId)
Returns the HRegionLocation parsed from the given meta row Result @@ -295,7 +295,7 @@ service.
private static HRegionLocation -AsyncMetaTableAccessor.getRegionLocation(Result r, +MetaTableAccessor.getRegionLocation(Result r, HRegionInfo regionInfo, int replicaId)
Returns the HRegionLocation parsed from the given meta row Result @@ -303,55 +303,55 @@ service.
-static RegionLocations -MetaTableAccessor.getRegionLocations(Result r) +private static Optional<RegionLocations> +AsyncMetaTableAccessor.getRegionLocations(Result r)
Returns an HRegionLocationList extracted from the result.
-private static Optional<RegionLocations> -AsyncMetaTableAccessor.getRegionLocations(Result r) +static RegionLocations +MetaTableAccessor.getRegionLocations(Result r)
Returns an HRegionLocationList extracted from the result.
private static long -MetaTableAccessor.getSeqNumDuringOpen(Result r, +AsyncMetaTableAccessor.getSeqNumDuringOpen(Result r, int replicaId)
The latest seqnum that the server writing to meta observed when opening the region.
private static long -AsyncMetaTableAccessor.getSeqNumDuringOpen(Result r, +MetaTableAccessor.getSeqNumDuringOpen(Result r, int replicaId)
The latest seqnum that the server writing to meta observed when opening the region.
-static ServerName -MetaTableAccessor.getServerName(Result r, +private static Optional<ServerName> +AsyncMetaTableAccessor.getServerName(Result r, int replicaId)
Returns a ServerName from catalog table Result.
-private static Optional<ServerName> -AsyncMetaTableAccessor.getServerName(Result r, +static ServerName +MetaTableAccessor.getServerName(Result r, int replicaId)
Returns a ServerName from catalog table Result.
+private static Optional<TableState> +AsyncMetaTableAccessor.getTableState(Result r)  + + static TableState MetaTableAccessor.getTableState(Result r)
Decode table state from META Result.
- -private static Optional<TableState> -AsyncMetaTableAccessor.getTableState(Result r)  - void AsyncMetaTableAccessor.MetaTableRawScanResultConsumer.onNext(Result[] results, @@ -447,13 +447,13 @@ service. ClientScanner.cache  -private Deque<Result> -BatchScanResultCache.partialResults  - - private List<Result> CompleteScanResultCache.partialResults  + +private Deque<Result> +BatchScanResultCache.partialResults  + private Queue<Result> AsyncTableResultScanner.queue  @@ -476,7 +476,7 @@ service. Result[] -BatchScanResultCache.addAndGet(Result[] results, +AllowPartialScanResultCache.addAndGet(Result[] results, boolean isHeartbeatMessage)  @@ -486,18 +486,18 @@ service. Result[] -AllowPartialScanResultCache.addAndGet(Result[] results, +BatchScanResultCache.addAndGet(Result[] results, boolean isHeartbeatMessage)  Result -HTable.append(Append append) +Table.append(Append append)
Appends values to one or more columns within a single row.
Result -Table.append(Append append) +HTable.append(Append append)
Appends values to one or more columns within a single row.
@@ -506,6 +506,10 @@ service. HTableWrapper.append(Append append)  +Result[] +ScannerCallableWithReplicas.call(int timeout)  + + Result RpcRetryingCallerWithReadReplicas.call(int operationTimeout)
@@ -513,10 +517,6 @@ service.
- we put the query into the execution pool. - -Result[] -ScannerCallableWithReplicas.call(int timeout)  - private Result[] ClientScanner.call(ScannerCallableWithReplicas callable, @@ -588,13 +588,13 @@ service. Result -HTable.get(Get get) +Table.get(Get get)
Extracts certain cells from a given row.
Result -Table.get(Get get) +HTable.get(Get get)
Extracts certain cells from a given row.
@@ -609,13 +609,13 @@ service. Result[] -HTable.get(List<Get> gets) +Table.get(List<Get> gets)
Extracts certain cells from the given rows, in batch.
Result[] -Table.get(List<Get> gets) +HTable.get(List<Get> gets)
Extracts certain cells from the given rows, in batch.
@@ -629,13 +629,13 @@ service. Result -HTable.increment(Increment increment) +Table.increment(Increment increment)
Increments one or more columns within a single row.
Result -Table.increment(Increment increment) +HTable.increment(Increment increment)
Increments one or more columns within a single row.
@@ -649,11 +649,11 @@ service. Result -ClientAsyncPrefetchScanner.next()  +ClientScanner.next()  Result -ClientScanner.next()  +ClientAsyncPrefetchScanner.next()  Result @@ -663,11 +663,11 @@ service. Result -ClientSideRegionScanner.next()  +TableSnapshotScanner.next()  Result -TableSnapshotScanner.next()  +ClientSideRegionScanner.next()  default Result[] @@ -693,13 +693,13 @@ service. BatchScanResultCache.regroupResults(Result result)  -protected Result -RpcRetryingCallerWithReadReplicas.ReplicaRegionServerCallable.rpcCall()  - - protected Result[] ScannerCallable.rpcCall()  + +protected Result +RpcRetryingCallerWithReadReplicas.ReplicaRegionServerCallable.rpcCall()  + private static Result RawAsyncTableImpl.toResult(HBaseRpcController controller, @@ -720,11 +720,11 @@ service. CompletableFuture<Result> -RawAsyncTableImpl.append(Append append)  +AsyncTableImpl.append(Append append)  CompletableFuture<Result> -AsyncTableImpl.append(Append append)  +RawAsyncTableImpl.append(Append append)  CompletableFuture<Result> @@ -734,11 +734,11 @@ service. CompletableFuture<Result> -RawAsyncTableImpl.get(Get get)  +AsyncTableImpl.get(Get get)  CompletableFuture<Result> -AsyncTableImpl.get(Get get)  +RawAsyncTableImpl.get(Get get)  CompletableFuture<Result> @@ -748,11 +748,11 @@ service. List<CompletableFuture<Result>> -RawAsyncTableImpl.get(List<Get> gets)  +AsyncTableImpl.get(List<Get> gets)  List<CompletableFuture<Result>> -AsyncTableImpl.get(List<Get> gets)  +RawAsyncTableImpl.get(List<Get> gets)  List<CompletableFuture<Result>> @@ -768,11 +768,11 @@ service. CompletableFuture<Result> -RawAsyncTableImpl.increment(Increment increment)  +AsyncTableImpl.increment(Increment increment)  CompletableFuture<Result> -AsyncTableImpl.increment(Increment increment)  +RawAsyncTableImpl.increment(Increment increment)  CompletableFuture<Result> @@ -786,11 +786,11 @@ service. CompletableFuture<List<Result>> -RawAsyncTableImpl.scanAll(Scan scan)  +AsyncTableImpl.scanAll(Scan scan)  CompletableFuture<List<Result>> -AsyncTableImpl.scanAll(Scan scan)  +RawAsyncTableImpl.scanAll(Scan scan)  CompletableFuture<List<Result>> @@ -816,7 +816,7 @@ service. Result[] -BatchScanResultCache.addAndGet(Result[] results, +AllowPartialScanResultCache.addAndGet(Result[] results, boolean isHeartbeatMessage)  @@ -826,7 +826,7 @@ service. Result[] -AllowPartialScanResultCache.addAndGet(Result[] results, +BatchScanResultCache.addAndGet(Result[] results, boolean isHeartbeatMessage)  @@ -884,11 +884,11 @@ service. private void -BatchScanResultCache.recordLastResult(Result result)  +AllowPartialScanResultCache.recordLastResult(Result result)  private void -AllowPartialScanResultCache.recordLastResult(Result result)  +BatchScanResultCache.recordLastResult(Result result)  private boolean @@ -1095,11 +1095,11 @@ service. Result -TableSnapshotInputFormat.TableSnapshotRecordReader.createValue()  +TableRecordReader.createValue()  Result -TableRecordReader.createValue()  +TableSnapshotInputFormat.TableSnapshotRecordReader.createValue()  Result @@ -1116,9 +1116,11 @@ service. org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result> -TableSnapshotInputFormat.getRecordReader(org.apache.hadoop.mapred.InputSplit split, +TableInputFormatBase.getRecordReader(org.apache.hadoop.mapred.InputSplit split, org.apache.hadoop.mapred.JobConf job, - org.apache.hadoop.mapred.Reporter reporter)  + org.apache.hadoop.mapred.Reporter reporter) +
Builds a TableRecordReader.
+ org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result> @@ -1128,11 +1130,9 @@ service. org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result> -TableInputFormatBase.getRecordReader(org.apache.hadoop.mapred.InputSplit split, +TableSnapshotInputFormat.getRecordReader(org.apache.hadoop.mapred.InputSplit split, org.apache.hadoop.mapred.JobConf job, - org.apache.hadoop.mapred.Reporter reporter) -
Builds a TableRecordReader.
- + org.apache.hadoop.mapred.Reporter reporter)
  @@ -1151,10 +1151,12 @@ service. void -RowCounter.RowCounterMapper.map(ImmutableBytesWritable row, - Result values, +IdentityTableMap.map(ImmutableBytesWritable key, + Result value, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output, - org.apache.hadoop.mapred.Reporter reporter)  + org.apache.hadoop.mapred.Reporter reporter) +
Pass the key, value to reduce
+ void @@ -1167,21 +1169,19 @@ service. void -IdentityTableMap.map(ImmutableBytesWritable key, - Result value, +RowCounter.RowCounterMapper.map(ImmutableBytesWritable row, + Result values, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output, - org.apache.hadoop.mapred.Reporter reporter) -
Pass the key, value to reduce
- + org.apache.hadoop.mapred.Reporter reporter)
  boolean -TableSnapshotInputFormat.TableSnapshotRecordReader.next(ImmutableBytesWritable key, +TableRecordReader.next(ImmutableBytesWritable key, Result value)  boolean -TableRecordReader.next(ImmutableBytesWritable key, +TableSnapshotInputFormat.TableSnapshotRecordReader.next(ImmutableBytesWritable key, Result value)  @@ -1200,10 +1200,12 @@ service. void -RowCounter.RowCounterMapper.map(ImmutableBytesWritable row, - Result values, +IdentityTableMap.map(ImmutableBytesWritable key, + Result value, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output, - org.apache.hadoop.mapred.Reporter reporter)  + org.apache.hadoop.mapred.Reporter reporter) +
Pass the key, value to reduce
+ void @@ -1216,12 +1218,10 @@ service. void -IdentityTableMap.map(ImmutableBytesWritable key, - Result value, +RowCounter.RowCounterMapper.map(ImmutableBytesWritable row, + Result values, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output, - org.apache.hadoop.mapred.Reporter reporter) -
Pass the key, value to reduce
- + org.apache.hadoop.mapred.Reporter reporter)
  @@ -1251,11 +1251,11 @@ service. private Result -TableRecordReaderImpl.value  +MultithreadedTableMapper.SubMapRecordReader.value  private Result -MultithreadedTableMapper.SubMapRecordReader.value  +TableRecordReaderImpl.value  @@ -1297,27 +1297,27 @@ service. Result -TableSnapshotInputFormat.TableSnapshotRegionRecordReader.getCurrentValue()  +MultithreadedTableMapper.SubMapRecordReader.getCurrentValue()  Result -TableSnapshotInputFormatImpl.RecordReader.getCurrentValue()  - - -Result TableRecordReader.getCurrentValue()
Returns the current value.
+ +Result +TableSnapshotInputFormat.TableSnapshotRegionRecordReader.getCurrentValue()  + Result -TableRecordReaderImpl.getCurrentValue() -
Returns the current value.
- +TableSnapshotInputFormatImpl.RecordReader.getCurrentValue()  Result -MultithreadedTableMapper.SubMapRecordReader.getCurrentValue()  +TableRecordReaderImpl.getCurrentValue() +
Returns the current value.
+ @@ -1330,23 +1330,23 @@ service. org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> -TableSnapshotInputFormat.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, - org.apache.hadoop.mapreduce.TaskAttemptContext context)  - - -org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> MultiTableInputFormatBase.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context)
Builds a TableRecordReader.
- + org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> TableInputFormatBase.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context) + +org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> +TableSnapshotInputFormat.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, + org.apache.hadoop.mapreduce.TaskAttemptContext context)  + org.apache.hadoop.io.serializer.Deserializer<Result> ResultSerialization.getDeserializer(Class<Result> c)  @@ -1401,14 +1401,14 @@ service. org.apache.hadoop.mapreduce.Mapper.Context context)
  -protected void -SyncTable.SyncMapper.map(ImmutableBytesWritable key, +void +Import.Importer.map(ImmutableBytesWritable row, Result value, org.apache.hadoop.mapreduce.Mapper.Context context)  -void -Import.Importer.map(