hbase-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From git-site-r...@apache.org
Subject [23/51] [partial] hbase-site git commit: Published site at .
Date Sat, 09 Sep 2017 15:14:15 GMT
http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1490b3ab/book.html
----------------------------------------------------------------------
diff --git a/book.html b/book.html
index 0a365e5..5a80fe0 100644
--- a/book.html
+++ b/book.html
@@ -2584,7 +2584,7 @@ Some configurations would only appear in source code; the only way to
identify t
 <dd>
 <div class="paragraph">
 <div class="title">Description</div>
-<p>A split policy determines when a region should be split. The various other split
policies that are available currently are BusyRegionSplitPolicy, ConstantSizeRegionSplitPolicy,
DisabledRegionSplitPolicy, DelimitedKeyPrefixRegionSplitPolicy, and KeyPrefixRegionSplitPolicy.
DisabledRegionSplitPolicy blocks manual region splitting.</p>
+<p>A split policy determines when a region should be split. The various other split
policies that are available currently are BusyRegionSplitPolicy, ConstantSizeRegionSplitPolicy,
DisabledRegionSplitPolicy, DelimitedKeyPrefixRegionSplitPolicy, KeyPrefixRegionSplitPolicy,
and SteppingSplitPolicy. DisabledRegionSplitPolicy blocks manual region splitting.</p>
 </div>
 <div class="paragraph">
 <div class="title">Default</div>
@@ -2794,7 +2794,7 @@ Some configurations would only appear in source code; the only way to
identify t
 <dd>
 <div class="paragraph">
 <div class="title">Description</div>
-<p>Default size of the HTable client write buffer in bytes. A bigger buffer takes more
memory&#8201;&#8212;&#8201;on both the client and server side since server instantiates
the passed write buffer to process it&#8201;&#8212;&#8201;but a larger buffer
size reduces the number of RPCs made. For an estimate of server-side memory-used, evaluate
hbase.client.write.buffer * hbase.regionserver.handler.count</p>
+<p>Default size of the BufferedMutator write buffer in bytes. A bigger buffer takes
more memory&#8201;&#8212;&#8201;on both the client and server side since server
instantiates the passed write buffer to process it&#8201;&#8212;&#8201;but a larger
buffer size reduces the number of RPCs made. For an estimate of server-side memory-used, evaluate
hbase.client.write.buffer * hbase.regionserver.handler.count</p>
 </div>
 <div class="paragraph">
 <div class="title">Default</div>
@@ -5395,7 +5395,7 @@ It also makes it so region boundaries are known and invariant (if you
disable re
 </div>
 <div class="paragraph">
 <div class="title">Disable Automatic Splitting</div>
-<p>To disable automatic splitting, set <code>hbase.hregion.max.filesize</code>
to a very large value, such as <code>100 GB</code> It is not recommended to set
it to its absolute maximum value of <code>Long.MAX_VALUE</code>.</p>
+<p>To disable automatic splitting, you can set region split policy in either cluster
configuration or table configuration to be <code>org.apache.hadoop.hbase.regionserver.DisabledRegionSplitPolicy</code></p>
 </div>
 <div class="admonitionblock note">
 <table>
@@ -7485,7 +7485,7 @@ Gets are executed via <a href="http://hbase.apache.org/apidocs/org/apache/hadoop
 <div class="sect2">
 <h3 id="_put"><a class="anchor" href="#_put"></a>26.2. Put</h3>
 <div class="paragraph">
-<p><a href="http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/client/Put.html">Put</a>
either adds new rows to a table (if the key is new) or can update existing rows (if the key
already exists). Puts are executed via <a href="http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/client/Table.html#put(org.apache.hadoop.hbase.client.Put)">Table.put</a>
(writeBuffer) or <a href="http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/client/Table.html#batch(java.util.List,%20java.lang.Object%5B%5D)">Table.batch</a>
(non-writeBuffer).</p>
+<p><a href="http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/client/Put.html">Put</a>
either adds new rows to a table (if the key is new) or can update existing rows (if the key
already exists). Puts are executed via <a href="http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/client/Table.html#put(org.apache.hadoop.hbase.client.Put)">Table.put</a>
(non-writeBuffer) or <a href="http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/client/Table.html#batch(java.util.List,%20java.lang.Object%5B%5D)">Table.batch</a>
(non-writeBuffer).</p>
 </div>
 </div>
 <div class="sect2">
@@ -13206,7 +13206,7 @@ Please use <a href="http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/clie
 <p>In HBase 1.0 and later, <a href="http://hbase.apache.org/devapidocs/org/apache/hadoop/hbase/client/HTable.html">HTable</a>
is deprecated in favor of <a href="http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/client/Table.html">Table</a>.
<code>Table</code> does not use autoflush. To do buffered writes, use the BufferedMutator
class.</p>
 </div>
 <div class="paragraph">
-<p>Before a <code>Table</code> or <code>HTable</code> instance
is discarded, invoke either <code>close()</code> or <code>flushCommits()</code>,
so `Put`s will not be lost.</p>
+<p>In HBase 2.0 and later, <a href="http://hbase.apache.org/devapidocs/org/apache/hadoop/hbase/client/HTable.html">HTable</a>
does not use BufferedMutator to execute the <code>Put</code> operation. Refer
to <a href="https://issues.apache.org/jira/browse/HBASE-18500">HBASE-18500</a>
for more information.</p>
 </div>
 <div class="paragraph">
 <p>For additional information on write durability, review the <a href="/acid-semantics.html">ACID
semantics</a> page.</p>
@@ -13304,14 +13304,14 @@ The following example shows an 'or' between two Filters (checking
for either 'my
 SingleColumnValueFilter filter1 = <span class="keyword">new</span> SingleColumnValueFilter(
   cf,
   column,
-  CompareOp.EQUAL,
+  CompareOperator.EQUAL,
   Bytes.toBytes(<span class="string"><span class="delimiter">&quot;</span><span
class="content">my value</span><span class="delimiter">&quot;</span></span>)
   );
 list.add(filter1);
 SingleColumnValueFilter filter2 = <span class="keyword">new</span> SingleColumnValueFilter(
   cf,
   column,
-  CompareOp.EQUAL,
+  CompareOperator.EQUAL,
   Bytes.toBytes(<span class="string"><span class="delimiter">&quot;</span><span
class="content">my other value</span><span class="delimiter">&quot;</span></span>)
   );
 list.add(filter2);
@@ -13327,8 +13327,8 @@ scan.setFilter(list);</code></pre>
 <div class="paragraph">
 <p>A SingleColumnValueFilter (see:
 <a href="http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.html"
class="bare">http://hbase.apache.org/apidocs/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.html</a>)
-can be used to test column values for equivalence (<code>CompareOp.EQUAL</code>),
-inequality (<code>CompareOp.NOT_EQUAL</code>), or ranges (e.g., <code>CompareOp.GREATER</code>).
The following is an
+can be used to test column values for equivalence (<code>CompareOperaor.EQUAL</code>),
+inequality (<code>CompareOperaor.NOT_EQUAL</code>), or ranges (e.g., <code>CompareOperaor.GREATER</code>).
The following is an
 example of testing equivalence of a column to a String value "my value"&#8230;&#8203;</p>
 </div>
 <div class="listingblock">
@@ -13336,7 +13336,7 @@ example of testing equivalence of a column to a String value "my value"&#8230;&#
 <pre class="CodeRay highlight"><code data-lang="java">SingleColumnValueFilter
filter = <span class="keyword">new</span> SingleColumnValueFilter(
   cf,
   column,
-  CompareOp.EQUAL,
+  CompareOperaor.EQUAL,
   Bytes.toBytes(<span class="string"><span class="delimiter">&quot;</span><span
class="content">my value</span><span class="delimiter">&quot;</span></span>)
   );
 scan.setFilter(filter);</code></pre>
@@ -13361,7 +13361,7 @@ These Comparators are used in concert with other Filters, such as
<a href="#clie
 SingleColumnValueFilter filter = <span class="keyword">new</span> SingleColumnValueFilter(
   cf,
   column,
-  CompareOp.EQUAL,
+  CompareOperaor.EQUAL,
   comp
   );
 scan.setFilter(filter);</code></pre>
@@ -13383,7 +13383,7 @@ The comparison is case-insensitive.</p>
 SingleColumnValueFilter filter = <span class="keyword">new</span> SingleColumnValueFilter(
   cf,
   column,
-  CompareOp.EQUAL,
+  CompareOperaor.EQUAL,
   comp
   );
 scan.setFilter(filter);</code></pre>
@@ -24374,7 +24374,7 @@ If inconsistencies, run <code>hbck</code> a few times
because the inconsistency
 <div class="sect2">
 <h3 id="_wal_tools"><a class="anchor" href="#_wal_tools"></a>130.6. WAL
Tools</h3>
 <div class="sect3">
-<h4 id="hlog_tool"><a class="anchor" href="#hlog_tool"></a>130.6.1. <code>FSHLog</code>
tool</h4>
+<h4 id="hlog_tool"><a class="anchor" href="#hlog_tool"></a>130.6.1. FSHLog
tool</h4>
 <div class="paragraph">
 <p>The main method on <code>FSHLog</code> offers manual split and dump
facilities.
 Pass it WALs or the product of a split, the content of the <em>recovered.edits</em>.
@@ -24400,9 +24400,9 @@ directory.</p>
 </div>
 </div>
 <div class="sect4">
-<h5 id="hlog_tool.prettyprint"><a class="anchor" href="#hlog_tool.prettyprint"></a>WAL
Pretty Printer</h5>
+<h5 id="hlog_tool.prettyprint"><a class="anchor" href="#hlog_tool.prettyprint"></a>WALPrettyPrinter</h5>
 <div class="paragraph">
-<p>The WAL Pretty Printer is a tool with configurable options to print the contents
of a WAL.
+<p>The <code>WALPrettyPrinter</code> is a tool with configurable options
to print the contents of a WAL.
 You can invoke it via the HBase cli with the 'wal' command.</p>
 </div>
 <div class="listingblock">
@@ -24419,7 +24419,7 @@ You can invoke it via the HBase cli with the 'wal' command.</p>
 <td class="content">
 <div class="title">WAL Printing in older versions of HBase</div>
 <div class="paragraph">
-<p>Prior to version 2.0, the WAL Pretty Printer was called the <code>HLogPrettyPrinter</code>,
after an internal name for HBase&#8217;s write ahead log.
+<p>Prior to version 2.0, the <code>WALPrettyPrinter</code> was called the
<code>HLogPrettyPrinter</code>, after an internal name for HBase&#8217;s write
ahead log.
 In those versions, you can print the contents of a WAL using the same configuration as above,
but with the 'hlog' command.</p>
 </div>
 <div class="listingblock">
@@ -24526,13 +24526,83 @@ For performance consider the following general options:
 <h3 id="export"><a class="anchor" href="#export"></a>130.9. Export</h3>
 <div class="paragraph">
 <p>Export is a utility that will dump the contents of table to HDFS in a sequence file.
-Invoke via:</p>
+The Export can be run via a Coprocessor Endpoint or MapReduce. Invoke via:</p>
+</div>
+<div class="paragraph">
+<p><strong>mapreduce-based Export</strong></p>
 </div>
 <div class="listingblock">
 <div class="content">
 <pre>$ bin/hbase org.apache.hadoop.hbase.mapreduce.Export &lt;tablename&gt;
&lt;outputdir&gt; [&lt;versions&gt; [&lt;starttime&gt; [&lt;endtime&gt;]]]</pre>
 </div>
 </div>
+<div class="paragraph">
+<p><strong>endpoint-based Export</strong></p>
+</div>
+<div class="listingblock">
+<div class="content">
+<pre>$ bin/hbase org.apache.hadoop.hbase.coprocessor.Export &lt;tablename&gt;
&lt;outputdir&gt; [&lt;versions&gt; [&lt;starttime&gt; [&lt;endtime&gt;]]]</pre>
+</div>
+</div>
+<div class="paragraph">
+<p><strong>The Comparison of Endpoint-based Export And Mapreduce-based Export</strong></p>
+</div>
+<table class="tableblock frame-all grid-all spread">
+<colgroup>
+<col style="width: 33.3333%;">
+<col style="width: 33.3333%;">
+<col style="width: 33.3334%;">
+</colgroup>
+<thead>
+<tr>
+<th class="tableblock halign-left valign-top"></th>
+<th class="tableblock halign-left valign-top">Endpoint-based Export</th>
+<th class="tableblock halign-left valign-top">Mapreduce-based Export</th>
+</tr>
+</thead>
+<tbody>
+<tr>
+<td class="tableblock halign-left valign-top"><p class="tableblock">HBase version
requirement</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">2.0+</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">0.2.1+</p></td>
+</tr>
+<tr>
+<td class="tableblock halign-left valign-top"><p class="tableblock">Maven dependency</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">hbase-endpoint</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">hbase-mapreduce
(2.0+), hbase-server(prior to 2.0)</p></td>
+</tr>
+<tr>
+<td class="tableblock halign-left valign-top"><p class="tableblock">Requirement
before dump</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">mount the
endpoint.Export on the target table</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">deploy the
MapReduce framework</p></td>
+</tr>
+<tr>
+<td class="tableblock halign-left valign-top"><p class="tableblock">Read latency</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">low, directly
read the data from region</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">normal, traditional
RPC scan</p></td>
+</tr>
+<tr>
+<td class="tableblock halign-left valign-top"><p class="tableblock">Read Scalability</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">depend on
number of regions</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">depend on
number of mappers (see TableInputFormatBase#getSplits)</p></td>
+</tr>
+<tr>
+<td class="tableblock halign-left valign-top"><p class="tableblock">Timeout</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">operation
timeout. configured by hbase.client.operation.timeout</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">scan timeout.
configured by hbase.client.scanner.timeout.period</p></td>
+</tr>
+<tr>
+<td class="tableblock halign-left valign-top"><p class="tableblock">Permission
requirement</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">READ, EXECUTE</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">READ</p></td>
+</tr>
+<tr>
+<td class="tableblock halign-left valign-top"><p class="tableblock">Fault tolerance</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">no</p></td>
+<td class="tableblock halign-left valign-top"><p class="tableblock">depend on
MapReduce</p></td>
+</tr>
+</tbody>
+</table>
 <div class="admonitionblock note">
 <table>
 <tr>
@@ -35075,7 +35145,7 @@ The server will return cellblocks compressed using this same compressor
as long
 <div id="footer">
 <div id="footer-text">
 Version 3.0.0-SNAPSHOT<br>
-Last updated 2017-09-08 14:29:37 UTC
+Last updated 2017-09-09 14:29:38 UTC
 </div>
 </div>
 </body>

http://git-wip-us.apache.org/repos/asf/hbase-site/blob/1490b3ab/bulk-loads.html
----------------------------------------------------------------------
diff --git a/bulk-loads.html b/bulk-loads.html
index 32fe7a1..2bd58ba 100644
--- a/bulk-loads.html
+++ b/bulk-loads.html
@@ -7,7 +7,7 @@
   <head>
     <meta charset="UTF-8" />
     <meta name="viewport" content="width=device-width, initial-scale=1.0" />
-    <meta name="Date-Revision-yyyymmdd" content="20170908" />
+    <meta name="Date-Revision-yyyymmdd" content="20170909" />
     <meta http-equiv="Content-Language" content="en" />
     <title>Apache HBase &#x2013;  
       Bulk Loads in Apache HBase (TM)
@@ -311,7 +311,7 @@ under the License. -->
                         <a href="https://www.apache.org/">The Apache Software Foundation</a>.
             All rights reserved.      
                     
-                  <li id="publishDate" class="pull-right">Last Published: 2017-09-08</li>
+                  <li id="publishDate" class="pull-right">Last Published: 2017-09-09</li>
             </p>
                 </div>
 


Mime
View raw message