Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id EE117200B75 for ; Sun, 4 Sep 2016 17:54:11 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id ECA49160AD6; Sun, 4 Sep 2016 15:54:11 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id 91773160AD0 for ; Sun, 4 Sep 2016 17:54:09 +0200 (CEST) Received: (qmail 11548 invoked by uid 500); 4 Sep 2016 15:54:05 -0000 Mailing-List: contact commits-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list commits@hbase.apache.org Received: (qmail 10131 invoked by uid 99); 4 Sep 2016 15:54:04 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 04 Sep 2016 15:54:04 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id B879EDFDD3; Sun, 4 Sep 2016 15:54:04 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: dimaspivak@apache.org To: commits@hbase.apache.org Date: Sun, 04 Sep 2016 15:54:27 -0000 Message-Id: <24fc641d423e404d8e74e14972d3a31a@git.apache.org> In-Reply-To: <08f89282ffe740e6990312eb90b0aef7@git.apache.org> References: <08f89282ffe740e6990312eb90b0aef7@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [24/52] [partial] hbase-site git commit: Published site at ab07f0087b8c5efcb1b084e42f4941ddb0c8885e. archived-at: Sun, 04 Sep 2016 15:54:12 -0000 http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8ba76a42/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html ---------------------------------------------------------------------- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html index 3f5cc3e..1b44092 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Result.html @@ -521,7 +521,7 @@ service. Result -ClientSimpleScanner.next()  +ClientAsyncPrefetchScanner.next()  Result @@ -533,7 +533,7 @@ service. Result -ClientAsyncPrefetchScanner.next()  +ClientSimpleScanner.next()  Result @@ -568,17 +568,17 @@ service. ClientAsyncPrefetchScanner.pollCache()  -protected Result[] -ScannerCallable.rpcCall()  - - protected Result RpcRetryingCallerWithReadReplicas.ReplicaRegionServerCallable.rpcCall()  - + protected Result[] ClientSmallScanner.SmallScannerCallable.rpcCall()  + +protected Result[] +ScannerCallable.rpcCall()  + @@ -893,11 +893,11 @@ service. - + - + @@ -914,9 +914,11 @@ service. - + org.apache.hadoop.mapred.Reporter reporter) +
Builds a TableRecordReader.
+ @@ -926,11 +928,9 @@ service. - + org.apache.hadoop.mapred.Reporter reporter) 
ResultTableSnapshotInputFormat.TableSnapshotRecordReader.createValue() TableRecordReader.createValue() 
ResultTableRecordReader.createValue() TableSnapshotInputFormat.TableSnapshotRecordReader.createValue() 
Result
org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result>MultiTableSnapshotInputFormat.getRecordReader(org.apache.hadoop.mapred.InputSplit split, +TableInputFormatBase.getRecordReader(org.apache.hadoop.mapred.InputSplit split, org.apache.hadoop.mapred.JobConf job, - org.apache.hadoop.mapred.Reporter reporter) 
org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result>
org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result>TableInputFormatBase.getRecordReader(org.apache.hadoop.mapred.InputSplit split, +MultiTableSnapshotInputFormat.getRecordReader(org.apache.hadoop.mapred.InputSplit split, org.apache.hadoop.mapred.JobConf job, - org.apache.hadoop.mapred.Reporter reporter) -
Builds a TableRecordReader.
-
@@ -949,15 +949,6 @@ service. void -IdentityTableMap.map(ImmutableBytesWritable key, - Result value, - org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output, - org.apache.hadoop.mapred.Reporter reporter) -
Pass the key, value to reduce
- - - -void GroupingTableMap.map(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output, @@ -965,21 +956,30 @@ service.
Extract the grouping columns from value to construct a new key.
- + void RowCounter.RowCounterMapper.map(ImmutableBytesWritable row, Result values, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output, org.apache.hadoop.mapred.Reporter reporter)  + +void +IdentityTableMap.map(ImmutableBytesWritable key, + Result value, + org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output, + org.apache.hadoop.mapred.Reporter reporter) +
Pass the key, value to reduce
+ + boolean -TableSnapshotInputFormat.TableSnapshotRecordReader.next(ImmutableBytesWritable key, +TableRecordReader.next(ImmutableBytesWritable key, Result value)  boolean -TableRecordReader.next(ImmutableBytesWritable key, +TableSnapshotInputFormat.TableSnapshotRecordReader.next(ImmutableBytesWritable key, Result value)  @@ -998,15 +998,6 @@ service. void -IdentityTableMap.map(ImmutableBytesWritable key, - Result value, - org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output, - org.apache.hadoop.mapred.Reporter reporter) -
Pass the key, value to reduce
- - - -void GroupingTableMap.map(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output, @@ -1014,13 +1005,22 @@ service.
Extract the grouping columns from value to construct a new key.
- + void RowCounter.RowCounterMapper.map(ImmutableBytesWritable row, Result values, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output, org.apache.hadoop.mapred.Reporter reporter)  + +void +IdentityTableMap.map(ImmutableBytesWritable key, + Result value, + org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output, + org.apache.hadoop.mapred.Reporter reporter) +
Pass the key, value to reduce
+ + @@ -1095,18 +1095,18 @@ service. Result -TableSnapshotInputFormat.TableSnapshotRegionRecordReader.getCurrentValue()  - - -Result TableSnapshotInputFormatImpl.RecordReader.getCurrentValue()  - + Result TableRecordReader.getCurrentValue()
Returns the current value.
+ +Result +TableSnapshotInputFormat.TableSnapshotRegionRecordReader.getCurrentValue()  + Result TableRecordReaderImpl.getCurrentValue() @@ -1128,16 +1128,16 @@ service. org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> -TableSnapshotInputFormat.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, - org.apache.hadoop.mapreduce.TaskAttemptContext context)  - - -org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> TableInputFormatBase.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context) + +org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> +TableSnapshotInputFormat.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, + org.apache.hadoop.mapreduce.TaskAttemptContext context)  + org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> MultiTableInputFormatBase.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, @@ -1477,21 +1477,21 @@ service. void -OperationQuota.addGetResult(Result result) -
Add a get result.
- +DefaultOperationQuota.addGetResult(Result result)  void -OperationQuota.AvgOperationSize.addGetResult(Result result)  +NoopOperationQuota.addGetResult(Result result)  void -NoopOperationQuota.addGetResult(Result result)  +OperationQuota.addGetResult(Result result) +
Add a get result.
+ void -DefaultOperationQuota.addGetResult(Result result)  +OperationQuota.AvgOperationSize.addGetResult(Result result)  static long @@ -1546,21 +1546,21 @@ service. void -OperationQuota.addScanResult(List<Result> results) -
Add a scan result.
- +DefaultOperationQuota.addScanResult(List<Result> results)  void -OperationQuota.AvgOperationSize.addScanResult(List<Result> results)  +NoopOperationQuota.addScanResult(List<Result> results)  void -NoopOperationQuota.addScanResult(List<Result> results)  +OperationQuota.addScanResult(List<Result> results) +
Add a scan result.
+ void -DefaultOperationQuota.addScanResult(List<Result> results)  +OperationQuota.AvgOperationSize.addScanResult(List<Result> results)  static long @@ -1702,9 +1702,9 @@ service. private void -RSRpcServices.addResult(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutateResponse.Builder builder, +RSRpcServices.addResult(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutateResponse.Builder builder, Result result, - PayloadCarryingRpcController rpcc)  + HBaseRpcController rpcc)  (package private) Object http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8ba76a42/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallableBase.html ---------------------------------------------------------------------- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallableBase.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallableBase.html index d48cafe..70e9cac 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallableBase.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallableBase.html @@ -222,14 +222,6 @@ -RetryingCallerInterceptorContext -NoOpRetryingInterceptorContext.prepare(RetryingCallableBase callable)  - - -FastFailInterceptorContext -FastFailInterceptorContext.prepare(RetryingCallableBase callable)  - - abstract RetryingCallerInterceptorContext RetryingCallerInterceptorContext.prepare(RetryingCallableBase callable)
This prepares the context object by populating it with information specific @@ -238,14 +230,12 @@ -RetryingCallerInterceptorContext -NoOpRetryingInterceptorContext.prepare(RetryingCallableBase callable, - int tries)  +FastFailInterceptorContext +FastFailInterceptorContext.prepare(RetryingCallableBase callable)  -FastFailInterceptorContext -FastFailInterceptorContext.prepare(RetryingCallableBase callable, - int tries)  +RetryingCallerInterceptorContext +NoOpRetryingInterceptorContext.prepare(RetryingCallableBase callable)  abstract RetryingCallerInterceptorContext @@ -255,6 +245,16 @@ in.
+ +FastFailInterceptorContext +FastFailInterceptorContext.prepare(RetryingCallableBase callable, + int tries)  + + +RetryingCallerInterceptorContext +NoOpRetryingInterceptorContext.prepare(RetryingCallableBase callable, + int tries)  + http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8ba76a42/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallerInterceptor.html ---------------------------------------------------------------------- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallerInterceptor.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallerInterceptor.html index 190669f..d4a8e71 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallerInterceptor.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallerInterceptor.html @@ -120,11 +120,11 @@ private RetryingCallerInterceptor -RpcRetryingCallerFactory.interceptor  +ConnectionImplementation.interceptor  private RetryingCallerInterceptor -ConnectionImplementation.interceptor  +RpcRetryingCallerFactory.interceptor  private RetryingCallerInterceptor http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8ba76a42/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallerInterceptorContext.html ---------------------------------------------------------------------- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallerInterceptorContext.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallerInterceptorContext.html index d3f7486..24c6e4f 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallerInterceptorContext.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RetryingCallerInterceptorContext.html @@ -135,20 +135,16 @@ PreemptiveFastFailInterceptor.createEmptyContext()  -RetryingCallerInterceptorContext -NoOpRetryableCallerInterceptor.createEmptyContext()  - - abstract RetryingCallerInterceptorContext RetryingCallerInterceptor.createEmptyContext()
This returns the context object for the current call.
- + RetryingCallerInterceptorContext -NoOpRetryingInterceptorContext.prepare(RetryingCallableBase callable)  +NoOpRetryableCallerInterceptor.createEmptyContext()  - + abstract RetryingCallerInterceptorContext RetryingCallerInterceptorContext.prepare(RetryingCallableBase callable)
This prepares the context object by populating it with information specific @@ -156,12 +152,11 @@ which this will be used.
- + RetryingCallerInterceptorContext -NoOpRetryingInterceptorContext.prepare(RetryingCallableBase callable, - int tries)  +NoOpRetryingInterceptorContext.prepare(RetryingCallableBase callable)  - + abstract RetryingCallerInterceptorContext RetryingCallerInterceptorContext.prepare(RetryingCallableBase callable, int tries) @@ -169,6 +164,11 @@ in. + +RetryingCallerInterceptorContext +NoOpRetryingInterceptorContext.prepare(RetryingCallableBase callable, + int tries)  + @@ -184,45 +184,45 @@ Throwable t)  - - - - + + + + - - - - + + + + - - - - + + + +
voidNoOpRetryableCallerInterceptor.handleFailure(RetryingCallerInterceptorContext context, - Throwable t) 
abstract void RetryingCallerInterceptor.handleFailure(RetryingCallerInterceptorContext context, Throwable t)
Call this function in case we caught a failure during retries.
voidNoOpRetryableCallerInterceptor.handleFailure(RetryingCallerInterceptorContext context, + Throwable t) 
void PreemptiveFastFailInterceptor.intercept(RetryingCallerInterceptorContext context) 
voidNoOpRetryableCallerInterceptor.intercept(RetryingCallerInterceptorContext abstractRetryingCallerInterceptorContext) 
abstract void RetryingCallerInterceptor.intercept(RetryingCallerInterceptorContext abstractRetryingCallerInterceptorContext)
Call this function alongside the actual call done on the callable.
voidNoOpRetryableCallerInterceptor.intercept(RetryingCallerInterceptorContext abstractRetryingCallerInterceptorContext) 
void PreemptiveFastFailInterceptor.updateFailureInfo(RetryingCallerInterceptorContext context) 
voidNoOpRetryableCallerInterceptor.updateFailureInfo(RetryingCallerInterceptorContext context) 
abstract void RetryingCallerInterceptor.updateFailureInfo(RetryingCallerInterceptorContext context)
Call this function to update at the end of the retry.
voidNoOpRetryableCallerInterceptor.updateFailureInfo(RetryingCallerInterceptorContext context) 
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8ba76a42/devapidocs/org/apache/hadoop/hbase/client/class-use/Row.html ---------------------------------------------------------------------- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/Row.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/Row.html index 5a926d8..7278645 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/Row.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/Row.html @@ -263,11 +263,11 @@ int -Increment.compareTo(Row i)  +Mutation.compareTo(Row d)  int -Mutation.compareTo(Row d)  +Increment.compareTo(Row i)  int @@ -395,6 +395,12 @@ Object[] results)
  +void +HTable.batch(List<? extends Row> actions, + Object[] results, + int timeout)  + + <R> void Table.batchCallback(List<? extends Row> actions, Object[] results, @@ -402,7 +408,7 @@
Same as Table.batch(List, Object[]), but with a callback.
- + <R> void HTable.batchCallback(List<? extends Row> actions, Object[] results, @@ -410,13 +416,13 @@
Same as Table.batch(List, Object[]), but with a callback.
- + <R> void HTableWrapper.batchCallback(List<? extends Row> actions, Object[] results, Batch.Callback<R> callback)  - + protected <CResult> AsyncProcess.AsyncRequestFutureImpl<CResult> AsyncProcess.createAsyncRequestFuture(TableName tableName, List<Action<Row>> actions, @@ -426,7 +432,7 @@ Object[] results, boolean needResults)  - + protected <CResult> AsyncProcess.AsyncRequestFutureImpl<CResult> AsyncProcess.createAsyncRequestFuture(TableName tableName, List<Action<Row>> actions, @@ -438,7 +444,7 @@ CancellableRegionServerCallable callable, int curTimeout)  - + protected MultiServerCallable<Row> AsyncProcess.createCallable(ServerName server, TableName tableName, @@ -446,7 +452,7 @@
Create a callable.
- + static <R> void HTable.doBatchWithCallback(List<? extends Row> actions, Object[] results, @@ -455,47 +461,39 @@ ExecutorService pool, TableName tableName)  - + private RegionLocations AsyncProcess.AsyncRequestFutureImpl.findAllLocationsOrFail(Action<Row> action, boolean useCache)  - + static String RetriesExhaustedWithDetailsException.getDesc(List<Throwable> exceptions, List<? extends Row> actions, List<String> hostnamePort)  - + private Collection<? extends Runnable> AsyncProcess.AsyncRequestFutureImpl.getNewMultiActionRunnable(ServerName server, MultiAction<Row> multiAction, int numAttempt)  - + private HRegionLocation AsyncProcess.AsyncRequestFutureImpl.getReplicaLocationOrFail(Action<Row> action)  - + private void AsyncProcess.AsyncRequestFutureImpl.groupAndSendMultiAction(List<Action<Row>> currentActions, int numAttempt)
Group a list of actions per region servers, and send them.
- + private void AsyncProcess.AsyncRequestFutureImpl.manageLocationError(Action<Row> action, Exception ex)  - -void -HTable.processBatch(List<? extends Row> list, - Object[] results) -
Parameterized batch processing, allowing varying return types for different - Row implementations.
- - <R> void HTable.processBatchCallback(List<? extends Row> list, http://git-wip-us.apache.org/repos/asf/hbase-site/blob/8ba76a42/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html ---------------------------------------------------------------------- diff --git a/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html b/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html index 74c7a98..c1b0113 100644 --- a/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html +++ b/devapidocs/org/apache/hadoop/hbase/client/class-use/RpcRetryingCallerFactory.html @@ -120,19 +120,19 @@ private RpcRetryingCallerFactory -HTable.rpcCallerFactory  +ConnectionImplementation.rpcCallerFactory  -private RpcRetryingCallerFactory -HBaseAdmin.rpcCallerFactory  - - protected RpcRetryingCallerFactory AsyncProcess.rpcCallerFactory  + +private RpcRetryingCallerFactory +HTable.rpcCallerFactory  + private RpcRetryingCallerFactory -ConnectionImplementation.rpcCallerFactory  +HBaseAdmin.rpcCallerFactory  private RpcRetryingCallerFactory