hbase-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From bus...@apache.org
Subject [3/3] hbase git commit: HBASE-13569 Correct Javadoc (for Java8)
Date Sun, 14 Jun 2015 03:27:12 GMT
HBASE-13569 Correct Javadoc (for Java8)

Signed-off-by: Sean Busbey <busbey@apache.org>

Conflicts:
	hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
	hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
	hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
	hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
	hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
	hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java
	hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
	hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
	hbase-server/src/main/java/org/apache/hadoop/hbase/constraint/ConstraintException.java
	hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java
	hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java
	hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java
	hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/MiniBatchOperationInProgress.java
	hbase-server/src/main/java/org/apache/hadoop/hbase/util/BloomFilterWriter.java


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/0a0ff335
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/0a0ff335
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/0a0ff335

Branch: refs/heads/branch-1
Commit: 0a0ff3354ae61cc17fe6db23852ac7b424f8bd6d
Parents: 4468f3c
Author: Gábor Lipták <gliptak@gmail.com>
Authored: Sun Apr 26 21:07:45 2015 -0400
Committer: Sean Busbey <busbey@apache.org>
Committed: Sat Jun 13 22:23:53 2015 -0500

----------------------------------------------------------------------
 .../apache/hadoop/hbase/HColumnDescriptor.java  |  4 +-
 .../org/apache/hadoop/hbase/HRegionInfo.java    |  7 ++-
 .../apache/hadoop/hbase/HTableDescriptor.java   |  2 +-
 .../apache/hadoop/hbase/MetaTableAccessor.java  |  2 +-
 .../org/apache/hadoop/hbase/ServerName.java     | 11 ++---
 .../hbase/client/AbstractClientScanner.java     |  4 +-
 .../org/apache/hadoop/hbase/client/Admin.java   | 15 ++++---
 .../client/ClientSmallReversedScanner.java      |  3 +-
 .../apache/hadoop/hbase/client/Durability.java  |  4 +-
 .../apache/hadoop/hbase/client/HBaseAdmin.java  |  3 +-
 .../hadoop/hbase/client/HTableInterface.java    |  2 +-
 .../hadoop/hbase/client/HTableMultiplexer.java  |  3 --
 .../org/apache/hadoop/hbase/client/Put.java     |  2 +-
 .../org/apache/hadoop/hbase/client/Result.java  |  8 ++--
 .../hadoop/hbase/client/ResultScanner.java      |  2 +-
 .../hadoop/hbase/client/RetryingCallable.java   |  4 +-
 .../RpcRetryingCallerWithReadReplicas.java      |  3 +-
 .../org/apache/hadoop/hbase/client/Scan.java    |  8 ++--
 .../client/coprocessor/AggregationClient.java   | 41 ++++++++---------
 .../hadoop/hbase/client/package-info.java       |  5 ---
 .../hbase/coprocessor/ColumnInterpreter.java    | 31 ++++++-------
 .../hadoop/hbase/filter/FamilyFilter.java       |  7 +--
 .../org/apache/hadoop/hbase/filter/Filter.java  |  1 -
 .../apache/hadoop/hbase/filter/FilterBase.java  | 18 ++++----
 .../apache/hadoop/hbase/filter/FilterList.java  |  8 ++--
 .../hbase/filter/MultiRowRangeFilter.java       |  2 +-
 .../apache/hadoop/hbase/filter/PageFilter.java  |  4 +-
 .../apache/hadoop/hbase/filter/SkipFilter.java  |  4 +-
 .../apache/hadoop/hbase/ipc/AsyncRpcClient.java |  1 -
 .../apache/hadoop/hbase/ipc/ConnectionId.java   |  2 +-
 .../hadoop/hbase/ipc/ServerRpcController.java   |  3 +-
 .../hadoop/hbase/protobuf/ProtobufUtil.java     |  6 +--
 .../hadoop/hbase/quotas/QuotaTableUtil.java     | 10 ++---
 .../hadoop/hbase/regionserver/BloomType.java    |  2 +-
 .../hbase/replication/ReplicationPeers.java     |  2 +-
 .../security/visibility/CellVisibility.java     |  7 +--
 .../apache/hadoop/hbase/zookeeper/ZKUtil.java   |  3 --
 .../hbase/zookeeper/ZooKeeperWatcher.java       |  1 -
 .../main/java/org/apache/hadoop/hbase/Cell.java | 47 +++++++++++---------
 .../org/apache/hadoop/hbase/CellComparator.java |  2 +-
 .../org/apache/hadoop/hbase/CellScannable.java  |  2 +-
 .../org/apache/hadoop/hbase/CellScanner.java    |  2 +-
 .../org/apache/hadoop/hbase/HConstants.java     |  3 +-
 .../java/org/apache/hadoop/hbase/KeyValue.java  | 30 +++++++------
 .../org/apache/hadoop/hbase/KeyValueUtil.java   |  2 +-
 .../java/org/apache/hadoop/hbase/TableName.java |  4 +-
 .../hadoop/hbase/io/CellOutputStream.java       |  2 +-
 .../org/apache/hadoop/hbase/io/TimeRange.java   |  2 +-
 .../hbase/io/crypto/KeyStoreKeyProvider.java    |  2 +-
 .../io/hadoopbackport/ThrottledInputStream.java |  8 ++--
 .../apache/hadoop/hbase/io/util/Dictionary.java |  4 +-
 .../org/apache/hadoop/hbase/security/User.java  |  1 -
 .../org/apache/hadoop/hbase/types/Struct.java   |  6 +--
 .../hadoop/hbase/util/AbstractByteRange.java    |  4 +-
 .../apache/hadoop/hbase/util/Addressing.java    |  8 ++--
 .../org/apache/hadoop/hbase/util/Base64.java    |  2 +-
 .../org/apache/hadoop/hbase/util/ByteRange.java |  4 +-
 .../org/apache/hadoop/hbase/util/Bytes.java     |  8 ++--
 .../org/apache/hadoop/hbase/util/ClassSize.java |  4 +-
 .../hbase/util/DefaultEnvironmentEdge.java      |  3 +-
 .../hbase/util/IncrementingEnvironmentEdge.java |  3 +-
 .../apache/hadoop/hbase/util/JenkinsHash.java   |  4 +-
 .../org/apache/hadoop/hbase/util/KeyLocker.java |  5 ++-
 .../apache/hadoop/hbase/util/OrderedBytes.java  |  9 ++--
 .../org/apache/hadoop/hbase/util/Sleeper.java   |  4 +-
 .../apache/hadoop/hbase/metrics/BaseSource.java |  2 +-
 .../hbase/codec/prefixtree/PrefixTreeCodec.java |  3 +-
 .../codec/prefixtree/PrefixTreeSeeker.java      |  8 ++--
 .../prefixtree/decode/ArraySearcherPool.java    |  3 +-
 .../decode/PrefixTreeArraySearcher.java         | 11 ++---
 .../prefixtree/encode/PrefixTreeEncoder.java    | 15 ++++---
 .../encode/column/ColumnNodeWriter.java         |  7 ++-
 .../encode/column/ColumnSectionWriter.java      |  6 ++-
 .../prefixtree/encode/row/RowNodeWriter.java    |  3 +-
 .../prefixtree/encode/tokenize/Tokenizer.java   |  2 +
 .../encode/tokenize/TokenizerNode.java          | 32 ++++++-------
 .../codec/prefixtree/scanner/CellSearcher.java  | 34 ++++++++------
 .../scanner/ReversibleCellScanner.java          |  4 +-
 .../hadoop/hbase/util/vint/UFIntTool.java       |  8 ++--
 .../hadoop/hbase/util/vint/UVIntTool.java       |  4 +-
 .../hadoop/hbase/util/vint/UVLongTool.java      |  4 +-
 .../hadoop/hbase/procedure2/Procedure.java      |  6 +--
 .../hbase/procedure2/SequentialProcedure.java   |  4 +-
 .../protobuf/HBaseZeroCopyByteString.java       |  6 +++
 .../apache/hadoop/hbase/InterProcessLock.java   |  2 +-
 .../example/HFileArchiveTableMonitor.java       |  2 +-
 .../hbase/constraint/ConstraintException.java   | 13 +++---
 .../hadoop/hbase/constraint/package-info.java   |  6 +--
 .../SplitLogManagerCoordination.java            |  2 +-
 .../ZKSplitLogManagerCoordination.java          |  5 +--
 .../coprocessor/AggregateImplementation.java    | 11 +++--
 .../coprocessor/MultiRowMutationEndpoint.java   | 10 ++---
 .../hbase/coprocessor/RegionObserver.java       |  7 +--
 .../hadoop/hbase/coprocessor/package-info.java  | 36 +++++++--------
 .../hbase/errorhandling/TimeoutException.java   |  3 +-
 .../apache/hadoop/hbase/http/HttpServer.java    | 13 +++---
 .../apache/hadoop/hbase/http/package-info.java  |  3 +-
 .../org/apache/hadoop/hbase/io/FileLink.java    |  2 +-
 .../org/apache/hadoop/hbase/io/HFileLink.java   |  1 -
 .../org/apache/hadoop/hbase/io/WALLink.java     |  4 +-
 .../org/apache/hadoop/hbase/io/hfile/HFile.java |  5 ++-
 .../hadoop/hbase/io/hfile/HFileBlock.java       | 12 ++---
 .../hadoop/hbase/io/hfile/HFileReaderV2.java    |  2 +-
 .../hadoop/hbase/io/hfile/HFileScanner.java     | 18 ++++----
 .../hadoop/hbase/io/hfile/LruBlockCache.java    |  2 +-
 .../hbase/io/hfile/LruCachedBlockQueue.java     |  2 +-
 .../hadoop/hbase/io/hfile/package-info.java     |  2 +-
 .../apache/hadoop/hbase/ipc/RpcCallContext.java |  2 +-
 .../hbase/mapreduce/MultiTableInputFormat.java  |  2 +-
 .../hbase/mapreduce/TableMapReduceUtil.java     |  9 ++--
 .../hbase/mapreduce/TableRecordReaderImpl.java  |  3 +-
 .../org/apache/hadoop/hbase/master/HMaster.java |  2 +-
 .../hadoop/hbase/master/MasterRpcServices.java  |  2 +-
 .../hadoop/hbase/master/RegionStateStore.java   |  2 +-
 .../master/balancer/SimpleLoadBalancer.java     |  4 +-
 .../master/balancer/StochasticLoadBalancer.java |  4 +-
 .../hbase/regionserver/ColumnTracker.java       | 14 +++---
 .../hbase/regionserver/CompactionRequestor.java |  4 +-
 .../hbase/regionserver/DefaultMemStore.java     |  2 +-
 .../hbase/regionserver/DeleteTracker.java       |  7 +--
 .../regionserver/ExplicitColumnTracker.java     |  7 +--
 .../hbase/regionserver/HeapMemoryManager.java   |  2 +-
 ...IncreasingToUpperBoundRegionSplitPolicy.java |  4 +-
 .../hadoop/hbase/regionserver/LruHashMap.java   |  6 +--
 .../hbase/regionserver/MemStoreChunkPool.java   |  2 +-
 .../hbase/regionserver/MetricsRegionServer.java |  3 +-
 .../MiniBatchOperationInProgress.java           | 10 ++---
 .../NonReversedNonLazyKeyValueScanner.java      |  2 +-
 .../hbase/regionserver/RSRpcServices.java       |  8 ++--
 .../hadoop/hbase/regionserver/Region.java       |  3 +-
 .../hbase/regionserver/RegionScanner.java       |  4 +-
 .../hbase/regionserver/ScanDeleteTracker.java   |  7 +--
 .../apache/hadoop/hbase/regionserver/Store.java |  7 +--
 .../hadoop/hbase/regionserver/StoreScanner.java |  2 +-
 .../hbase/regionserver/wal/ReplayHLogKey.java   |  1 -
 .../hadoop/hbase/regionserver/wal/WALEdit.java  | 14 +++---
 .../regionserver/wal/WALEditsReplaySink.java    |  4 +-
 .../replication/HBaseReplicationEndpoint.java   |  1 -
 .../HBaseInterClusterReplicationEndpoint.java   |  3 +-
 .../regionserver/ReplicationSink.java           |  7 +--
 .../regionserver/ReplicationSource.java         |  4 +-
 .../regionserver/ReplicationSourceManager.java  |  2 +
 .../regionserver/ReplicationThrottler.java      |  2 +-
 .../security/access/AccessControlLists.java     |  1 +
 .../hbase/security/access/AccessController.java |  2 +
 .../security/visibility/VisibilityUtils.java    |  2 +-
 .../snapshot/SnapshotDescriptionUtils.java      | 14 +++---
 .../hadoop/hbase/util/BloomFilterWriter.java    |  2 +-
 .../org/apache/hadoop/hbase/util/FSUtils.java   |  9 ++--
 .../org/apache/hadoop/hbase/util/HBaseFsck.java |  2 +-
 .../hadoop/hbase/util/HFileV1Detector.java      |  6 ++-
 .../hadoop/hbase/util/MultiHConnection.java     |  1 -
 .../hadoop/hbase/util/RegionSplitter.java       | 14 +++---
 .../hbase/util/ServerRegionReplicaUtil.java     |  2 +-
 .../org/apache/hadoop/hbase/wal/WALKey.java     |  4 +-
 .../hadoop/hbase/zookeeper/ZKSplitLog.java      |  8 ++--
 .../hadoop/hbase/thrift/ThriftServerRunner.java |  3 +-
 .../hadoop/hbase/thrift/generated/Hbase.java    |  4 --
 158 files changed, 518 insertions(+), 463 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
index ef6a4c9..0f6abf0 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java
@@ -91,8 +91,8 @@ public class HColumnDescriptor implements WritableComparable<HColumnDescriptor>
   /**
    * Key for cache data into L1 if cache is set up with more than one tier.
    * To set in the shell, do something like this:
-   * <code>hbase(main):003:0> create 't',
-   *    {NAME => 't', CONFIGURATION => {CACHE_DATA_IN_L1 => 'true'}}</code>
+   * <code>hbase(main):003:0&gt; create 't',
+   *    {NAME =&gt; 't', CONFIGURATION =&gt; {CACHE_DATA_IN_L1 =&gt; 'true'}}</code>
    */
   public static final String CACHE_DATA_IN_L1 = "CACHE_DATA_IN_L1";
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
index 175515c..31bb016 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java
@@ -57,17 +57,21 @@ import com.google.protobuf.InvalidProtocolBufferException;
  * about the region.
  *
  * The region has a unique name which consists of the following fields:
+ * <ul>
  * <li> tableName   : The name of the table </li>
  * <li> startKey    : The startKey for the region. </li>
  * <li> regionId    : A timestamp when the region is created. </li>
  * <li> replicaId   : An id starting from 0 to differentiate replicas of the same region range
  * but hosted in separated servers. The same region range can be hosted in multiple locations.</li>
  * <li> encodedName : An MD5 encoded string for the region name.</li>
+ * </ul>
  *
  * <br> Other than the fields in the region name, region info contains:
+ * <ul>
  * <li> endKey      : the endKey for the region (exclusive) </li>
  * <li> split       : Whether the region is split </li>
  * <li> offline     : Whether the region is offline </li>
+ * </ul>
  *
  * In 0.98 or before, a list of table's regions would fully cover the total keyspace, and at any
  * point in time, a row key always belongs to a single region, which is hosted in a single server.
@@ -711,7 +715,7 @@ public class HRegionInfo implements Comparable<HRegionInfo> {
    * by this region. For example, if the region is foo,a,g and this is
    * passed ["b","c"] or ["a","c"] it will return true, but if this is passed
    * ["b","z"] it will return false.
-   * @throws IllegalArgumentException if the range passed is invalid (ie end < start)
+   * @throws IllegalArgumentException if the range passed is invalid (ie. end &lt; start)
    */
   public boolean containsRange(byte[] rangeStartKey, byte[] rangeEndKey) {
     if (Bytes.compareTo(rangeStartKey, rangeEndKey) > 0) {
@@ -1148,7 +1152,6 @@ public class HRegionInfo implements Comparable<HRegionInfo> {
    * @param r Result to pull from
    * @return A pair of the {@link HRegionInfo} and the {@link ServerName}
    * (or null for server address if no address set in hbase:meta).
-   * @throws IOException
    * @deprecated use MetaTableAccessor methods for interacting with meta layouts
    */
   @Deprecated

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
index 947ac91..7f61bf2 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java
@@ -195,7 +195,7 @@ public class HTableDescriptor implements WritableComparable<HTableDescriptor> {
 
   /**
    * <em>INTERNAL</em> flag to indicate whether or not the memstore should be replicated
-   * for read-replicas (CONSISTENCY => TIMELINE).
+   * for read-replicas (CONSISTENCY =&gt; TIMELINE).
    */
   public static final String REGION_MEMSTORE_REPLICATION = "REGION_MEMSTORE_REPLICATION";
   private static final ImmutableBytesWritable REGION_MEMSTORE_REPLICATION_KEY =

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
index cf34de5..807f956 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/MetaTableAccessor.java
@@ -121,7 +121,7 @@ public class MetaTableAccessor {
       META_REGION_PREFIX, 0, len);
   }
 
-  /** The delimiter for meta columns for replicaIds > 0 */
+  /** The delimiter for meta columns for replicaIds &gt; 0 */
   protected static final char META_REPLICA_ID_DELIMITER = '_';
 
   /** A regex for parsing server columns from meta. See above javadoc for meta layout */

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerName.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerName.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerName.java
index dc5ba78..ef7ff0d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerName.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ServerName.java
@@ -41,7 +41,8 @@ import java.util.regex.Pattern;
  * servers on same hostname and port (startcode is usually timestamp of server startup). The
  * {@link #toString()} format of ServerName is safe to use in the  filesystem and as znode name
  * up in ZooKeeper.  Its format is:
- * <code>&lt;hostname> '{@link #SERVERNAME_SEPARATOR}' &lt;port> '{@link #SERVERNAME_SEPARATOR}' &lt;startcode></code>.
+ * <code>&lt;hostname&gt; '{@link #SERVERNAME_SEPARATOR}' &lt;port&gt;
+ * '{@link #SERVERNAME_SEPARATOR}' &lt;startcode&gt;</code>.
  * For example, if hostname is <code>www.example.org</code>, port is <code>1234</code>,
  * and the startcode for the regionserver is <code>1212121212</code>, then
  * the {@link #toString()} would be <code>www.example.org,1234,1212121212</code>.
@@ -224,7 +225,7 @@ public class ServerName implements Comparable<ServerName>, Serializable {
    * @param port
    * @param startcode
    * @return Server name made of the concatenation of hostname, port and
-   * startcode formatted as <code>&lt;hostname> ',' &lt;port> ',' &lt;startcode></code>
+   * startcode formatted as <code>&lt;hostname&gt; ',' &lt;port&gt; ',' &lt;startcode&gt;</code>
    */
   static String getServerName(String hostName, int port, long startcode) {
     final StringBuilder name = new StringBuilder(hostName.length() + 1 + 5 + 1 + 13);
@@ -237,10 +238,10 @@ public class ServerName implements Comparable<ServerName>, Serializable {
   }
 
   /**
-   * @param hostAndPort String in form of &lt;hostname> ':' &lt;port>
+   * @param hostAndPort String in form of &lt;hostname&gt; ':' &lt;port&gt;
    * @param startcode
    * @return Server name made of the concatenation of hostname, port and
-   * startcode formatted as <code>&lt;hostname> ',' &lt;port> ',' &lt;startcode></code>
+   * startcode formatted as <code>&lt;hostname&gt; ',' &lt;port&gt; ',' &lt;startcode&gt;</code>
    */
   public static String getServerName(final String hostAndPort,
       final long startcode) {
@@ -339,7 +340,7 @@ public class ServerName implements Comparable<ServerName>, Serializable {
 
   /**
    * @param str Either an instance of {@link ServerName#toString()} or a
-   * "'<hostname>' ':' '<port>'".
+   * "'&lt;hostname&gt;' ':' '&lt;port&gt;'".
    * @return A ServerName instance.
    */
   public static ServerName parseServerName(final String str) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AbstractClientScanner.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AbstractClientScanner.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AbstractClientScanner.java
index 440400a..f926fa9 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AbstractClientScanner.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/AbstractClientScanner.java
@@ -54,11 +54,11 @@ public abstract class AbstractClientScanner implements ResultScanner {
   }
 
   /**
-   * Get <param>nbRows</param> rows.
+   * Get nbRows rows.
    * How many RPCs are made is determined by the {@link Scan#setCaching(int)}
    * setting (or hbase.client.scanner.caching in hbase-site.xml).
    * @param nbRows number of rows to return
-   * @return Between zero and <param>nbRows</param> RowResults.  Scan is done
+   * @return Between zero and nbRows rowResults.  Scan is done
    * if returned array is of zero-length (We never return null).
    * @throws IOException
    */

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
index 4f89467..353938d 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Admin.java
@@ -197,7 +197,7 @@ public interface Admin extends Abortable, Closeable {
    *
    * @param desc table descriptor for table
    * @throws IllegalArgumentException if the table name is reserved
-   * @throws MasterNotRunningException if master is not running
+   * @throws org.apache.hadoop.hbase.MasterNotRunningException if master is not running
    * @throws org.apache.hadoop.hbase.TableExistsException if table already exists (If concurrent
    * threads, the table may have been created between test-for-existence and attempt-at-creation).
    * @throws IOException if a remote or network exception occurs
@@ -216,7 +216,7 @@ public interface Admin extends Abortable, Closeable {
    * @param endKey end of key range
    * @param numRegions the total number of regions to create
    * @throws IllegalArgumentException if the table name is reserved
-   * @throws MasterNotRunningException if master is not running
+   * @throws org.apache.hadoop.hbase.MasterNotRunningException if master is not running
    * @throws org.apache.hadoop.hbase.TableExistsException if table already exists (If concurrent
    * threads, the table may have been created between test-for-existence and attempt-at-creation).
    * @throws IOException
@@ -233,7 +233,7 @@ public interface Admin extends Abortable, Closeable {
    * @param splitKeys array of split keys for the initial regions of the table
    * @throws IllegalArgumentException if the table name is reserved, if the split keys are repeated
    * and if the split key has empty byte array.
-   * @throws MasterNotRunningException if master is not running
+   * @throws org.apache.hadoop.hbase.MasterNotRunningException if master is not running
    * @throws org.apache.hadoop.hbase.TableExistsException if table already exists (If concurrent
    * threads, the table may have been created between test-for-existence and attempt-at-creation).
    * @throws IOException
@@ -245,13 +245,14 @@ public interface Admin extends Abortable, Closeable {
    * To check if the table exists, use {@link #isTableAvailable} -- it is not safe to create an
    * HTable instance to this table before it is available. Note : Avoid passing empty split key.
    *
+   * Throws IllegalArgumentException Bad table name, if the split keys
+   *    are repeated and if the split key has empty byte array.
+   *
    * @param desc table descriptor for table
-   * @throws IllegalArgumentException Bad table name, if the split keys are repeated and if the
-   * split key has empty byte array.
    * @throws MasterNotRunningException if master is not running
    * @throws org.apache.hadoop.hbase.TableExistsException if table already exists (If concurrent
    * threads, the table may have been created between test-for-existence and attempt-at-creation).
-   * @throws IOException
+   * @throws IOException if a remote or network exception occurs
    */
   void createTableAsync(final HTableDescriptor desc, final byte[][] splitKeys) throws IOException;
 
@@ -648,7 +649,7 @@ public interface Admin extends Abortable, Closeable {
    * @param destServerName The servername of the destination regionserver.  If passed the empty byte
    * array we'll assign to a random server.  A server name is made of host, port and startcode.
    * Here is an example: <code> host187.example.com,60020,1289493121758</code>
-   * @throws UnknownRegionException Thrown if we can't find a region named
+   * @throws IOException if we can't find a region named
    * <code>encodedRegionName</code>
    */
   void move(final byte[] encodedRegionName, final byte[] destServerName)

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientSmallReversedScanner.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientSmallReversedScanner.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientSmallReversedScanner.java
index 434812f..d4de6a0 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientSmallReversedScanner.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ClientSmallReversedScanner.java
@@ -39,10 +39,11 @@ import java.io.IOException;
 import java.util.concurrent.ExecutorService;
 
 /**
+ * <p>
  * Client scanner for small reversed scan. Generally, only one RPC is called to fetch the
  * scan results, unless the results cross multiple regions or the row count of
  * results exceed the caching.
- * <p/>
+ * </p>
  * For small scan, it will get better performance than {@link ReversedClientScanner}
  */
 @InterfaceAudience.Private

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Durability.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Durability.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Durability.java
index bc9a4ed..9b35e04 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Durability.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Durability.java
@@ -47,13 +47,13 @@ public enum Durability {
    * Write the Mutation to the WAL synchronously.
    * The data is flushed to the filesystem implementation, but not necessarily to disk.
    * For HDFS this will flush the data to the designated number of DataNodes.
-   * See <a href="https://issues.apache.org/jira/browse/HADOOP-6313">HADOOP-6313<a/>
+   * See <a href="https://issues.apache.org/jira/browse/HADOOP-6313">HADOOP-6313</a>
    */
   SYNC_WAL,
   /**
    * Write the Mutation to the WAL synchronously and force the entries to disk.
    * (Note: this is currently not supported and will behave identical to {@link #SYNC_WAL})
-   * See <a href="https://issues.apache.org/jira/browse/HADOOP-6313">HADOOP-6313<a/>
+   * See <a href="https://issues.apache.org/jira/browse/HADOOP-6313">HADOOP-6313</a>
    */
   FSYNC_WAL
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
index 5a0def3..63d0b7b 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java
@@ -234,7 +234,8 @@ public class HBaseAdmin implements Admin {
    * The connection to master will be created when required by admin functions.
    *
    * @param connection The Connection instance to use
-   * @throws MasterNotRunningException, ZooKeeperConnectionException are not
+   * @throws MasterNotRunningException
+   * @throws ZooKeeperConnectionException are not
    *  thrown anymore but kept into the interface for backward api compatibility
    * @deprecated Constructing HBaseAdmin objects manually has been deprecated.
    * Use {@link Connection#getAdmin()} to obtain an instance of {@link Admin} instead.

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java
index 5823f69..99a02e9 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java
@@ -137,7 +137,7 @@ public interface HTableInterface extends Table {
    * Executes all the buffered {@link Put} operations.
    * <p>
    * This method gets called once automatically for every {@link Put} or batch
-   * of {@link Put}s (when <code>put(List<Put>)</code> is used) when
+   * of {@link Put}s (when <code>put(List&lt;Put&gt;)</code> is used) when
    * {@link #isAutoFlush} is {@code true}.
    * @throws IOException if a remote or network exception occurs.
    * @deprecated as of 1.0.0. Replaced by {@link BufferedMutator#flush()}

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java
index 7d61a0b..f6becc6 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/HTableMultiplexer.java
@@ -121,7 +121,6 @@ public class HTableMultiplexer {
    * @param tableName
    * @param put
    * @return true if the request can be accepted by its corresponding buffer queue.
-   * @throws IOException
    */
   public boolean put(TableName tableName, final Put put) {
     return put(tableName, put, this.retryNum);
@@ -133,7 +132,6 @@ public class HTableMultiplexer {
    * @param tableName
    * @param puts
    * @return the list of puts which could not be queued
-   * @throws IOException
    */
   public List<Put> put(TableName tableName, final List<Put> puts) {
     if (puts == null)
@@ -169,7 +167,6 @@ public class HTableMultiplexer {
    * retried before dropping the request.
    * Return false if the queue is already full.
    * @return true if the request can be accepted by its corresponding buffer queue.
-   * @throws IOException
    */
   public boolean put(final TableName tableName, final Put put, int retry) {
     if (retry <= 0) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Put.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Put.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Put.java
index 1fdc4e2..b1d6775 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Put.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Put.java
@@ -328,7 +328,7 @@ public class Put extends Mutation implements HeapSize, Comparable<Row> {
 
   /**
    * A convenience method to determine if this object's familyMap contains
-   * a value assigned to the given family & qualifier.
+   * a value assigned to the given family &amp; qualifier.
    * Both given arguments must match the KeyValue object to return true.
    *
    * @param family column family

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
index 68eae5a..9d67549 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Result.java
@@ -216,7 +216,7 @@ public class Result implements CellScannable, CellScanner {
   /**
    * Return the array of Cells backing this Result instance.
    *
-   * The array is sorted from smallest -> largest using the
+   * The array is sorted from smallest -&gt; largest using the
    * {@link KeyValue#COMPARATOR}.
    *
    * The array only contains what your Get or Scan specifies and no more.
@@ -673,7 +673,7 @@ public class Result implements CellScannable, CellScanner {
    * Map of families to all versions of its qualifiers and values.
    * <p>
    * Returns a three level Map of the form:
-   * <code>Map&amp;family,Map&lt;qualifier,Map&lt;timestamp,value>>></code>
+   * <code>Map&amp;family,Map&lt;qualifier,Map&lt;timestamp,value&gt;&gt;&gt;</code>
    * <p>
    * Note: All other map returning methods make use of this map internally.
    * @return map from families to qualifiers to versions
@@ -717,7 +717,7 @@ public class Result implements CellScannable, CellScanner {
   /**
    * Map of families to their most recent qualifiers and values.
    * <p>
-   * Returns a two level Map of the form: <code>Map&amp;family,Map&lt;qualifier,value>></code>
+   * Returns a two level Map of the form: <code>Map&amp;family,Map&lt;qualifier,value&gt;&gt;</code>
    * <p>
    * The most recent version of each qualifier will be used.
    * @return map from families to qualifiers and value
@@ -749,7 +749,7 @@ public class Result implements CellScannable, CellScanner {
   /**
    * Map of qualifiers to values.
    * <p>
-   * Returns a Map of the form: <code>Map&lt;qualifier,value></code>
+   * Returns a Map of the form: <code>Map&lt;qualifier,value&gt;</code>
    * @param family column family to get
    * @return map of qualifiers to values
    */

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java
index 8fd28b3..f3f437a 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java
@@ -42,7 +42,7 @@ public interface ResultScanner extends Closeable, Iterable<Result> {
 
   /**
    * @param nbRows number of rows to return
-   * @return Between zero and <param>nbRows</param> Results
+   * @return Between zero and nbRows results
    * @throws IOException e
    */
   Result [] next(int nbRows) throws IOException;

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RetryingCallable.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RetryingCallable.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RetryingCallable.java
index e468d3c..ea65fcf 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RetryingCallable.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RetryingCallable.java
@@ -24,7 +24,7 @@ import java.io.IOException;
 import org.apache.hadoop.hbase.classification.InterfaceAudience;
 
 /**
- * A Callable<T> that will be retried.  If {@link #call(int)} invocation throws exceptions,
+ * A Callable&lt;T&gt; that will be retried.  If {@link #call(int)} invocation throws exceptions,
  * we will call {@link #throwable(Throwable, boolean)} with whatever the exception was.
  * @param <T>
  */
@@ -42,7 +42,7 @@ public interface RetryingCallable<T> {
    * make it so we succeed on next call (clear caches, do relookup of locations, etc.).
    * @param t
    * @param retrying True if we are in retrying mode (we are not in retrying mode when max
-   * retries == 1; we ARE in retrying mode if retries > 1 even when we are the last attempt)
+   * retries == 1; we ARE in retrying mode if retries &gt; 1 even when we are the last attempt)
    */
   void throwable(final Throwable t, boolean retrying);
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java
index bb40dea..a2bedb4 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java
@@ -177,6 +177,7 @@ public class RpcRetryingCallerWithReadReplicas {
   }
 
   /**
+   * <p>
    * Algo:
    * - we put the query into the execution pool.
    * - after x ms, if we don't have a result, we add the queries for the secondary replicas
@@ -189,7 +190,7 @@ public class RpcRetryingCallerWithReadReplicas {
    * - a call is a thread. Let's not multiply the number of thread by the number of replicas.
    * Server side, if we can cancel when it's still in the handler pool, it's much better, as a call
    * can take some i/o.
-   * <p/>
+   * </p>
    * Globally, the number of retries, timeout and so on still applies, but it's per replica,
    * not global. We continue until all retries are done, or all timeouts are exceeded.
    */

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
index 28af4d2..04931c8 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Scan.java
@@ -714,10 +714,10 @@ public class Scan extends Query {
    * this can deliver huge perf gains when there's a cf with lots of data; however, it can
    * also lead to some inconsistent results, as follows:
    * - if someone does a concurrent update to both column families in question you may get a row
-   *   that never existed, e.g. for { rowKey = 5, { cat_videos => 1 }, { video => "my cat" } }
-   *   someone puts rowKey 5 with { cat_videos => 0 }, { video => "my dog" }, concurrent scan
-   *   filtering on "cat_videos == 1" can get { rowKey = 5, { cat_videos => 1 },
-   *   { video => "my dog" } }.
+   *   that never existed, e.g. for { rowKey = 5, { cat_videos =&gt; 1 }, { video =&gt; "my cat" } }
+   *   someone puts rowKey 5 with { cat_videos =&gt; 0 }, { video =&gt; "my dog" }, concurrent scan
+   *   filtering on "cat_videos == 1" can get { rowKey = 5, { cat_videos =&gt; 1 },
+   *   { video =&gt; "my dog" } }.
    * - if there's a concurrent split and you have more than 2 column families, some rows may be
    *   missing some column families.
    */

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
index 1e378e7..eb7cb61 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java
@@ -65,16 +65,17 @@ import com.google.protobuf.Message;
  * <p>
  * This will serve as the client side handler for invoking the aggregate
  * functions.
- * <ul>
  * For all aggregate functions,
- * <li>start row < end row is an essential condition (if they are not
+ * <ul>
+ * <li>start row &lt; end row is an essential condition (if they are not
  * {@link HConstants#EMPTY_BYTE_ARRAY})
  * <li>Column family can't be null. In case where multiple families are
  * provided, an IOException will be thrown. An optional column qualifier can
- * also be defined.
+ * also be defined.</li>
  * <li>For methods to find maximum, minimum, sum, rowcount, it returns the
  * parameter type. For average and std, it returns a double value. For row
- * count, it returns a long value.
+ * count, it returns a long value.</li>
+ * </ul>
  * <p>Call {@link #close()} when done.
  */
 @InterfaceAudience.Private
@@ -110,10 +111,10 @@ public class AggregationClient implements Closeable {
    * @param tableName
    * @param ci
    * @param scan
-   * @return max val <R>
+   * @return max val &lt;R&gt;
    * @throws Throwable
    *           The caller is supposed to handle the exception as they are thrown
-   *           & propagated to it.
+   *           &amp; propagated to it.
    */
   public <R, S, P extends Message, Q extends Message, T extends Message> R max(
       final TableName tableName, final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan)
@@ -130,10 +131,10 @@ public class AggregationClient implements Closeable {
    * @param table
    * @param ci
    * @param scan
-   * @return max val <R>
+   * @return max val &lt;&gt;
    * @throws Throwable
    *           The caller is supposed to handle the exception as they are thrown
-   *           & propagated to it.
+   *           &amp; propagated to it.
    */
   public <R, S, P extends Message, Q extends Message, T extends Message>
   R max(final Table table, final ColumnInterpreter<R, S, P, Q, T> ci,
@@ -201,7 +202,7 @@ public class AggregationClient implements Closeable {
    * @param tableName
    * @param ci
    * @param scan
-   * @return min val <R>
+   * @return min val &lt;R&gt;
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message> R min(
@@ -219,7 +220,7 @@ public class AggregationClient implements Closeable {
    * @param table
    * @param ci
    * @param scan
-   * @return min val <R>
+   * @return min val &lt;R&gt;
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message>
@@ -271,11 +272,11 @@ public class AggregationClient implements Closeable {
    * optimised the operation. In case qualifier is provided, I can't use the
    * filter as it may set the flag to skip to next row, but the value read is
    * not of the given filter: in this case, this particular row will not be
-   * counted ==> an error.
+   * counted ==&gt; an error.
    * @param tableName
    * @param ci
    * @param scan
-   * @return <R, S>
+   * @return &lt;R, S&gt;
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message> long rowCount(
@@ -292,11 +293,11 @@ public class AggregationClient implements Closeable {
    * optimised the operation. In case qualifier is provided, I can't use the
    * filter as it may set the flag to skip to next row, but the value read is
    * not of the given filter: in this case, this particular row will not be
-   * counted ==> an error.
+   * counted ==&gt; an error.
    * @param table
    * @param ci
    * @param scan
-   * @return <R, S>
+   * @return &lt;R, S&gt;
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message>
@@ -343,7 +344,7 @@ public class AggregationClient implements Closeable {
    * @param tableName
    * @param ci
    * @param scan
-   * @return sum <S>
+   * @return sum &lt;S&gt;
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message> S sum(
@@ -360,7 +361,7 @@ public class AggregationClient implements Closeable {
    * @param table
    * @param ci
    * @param scan
-   * @return sum <S>
+   * @return sum &lt;S&gt;
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message>
@@ -487,7 +488,7 @@ public class AggregationClient implements Closeable {
    * @param tableName
    * @param ci
    * @param scan
-   * @return <R, S>
+   * @return &lt;R, S&gt;
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message>
@@ -506,7 +507,7 @@ public class AggregationClient implements Closeable {
    * @param table
    * @param ci
    * @param scan
-   * @return <R, S>
+   * @return &lt;R, S&gt;
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message> double avg(
@@ -595,7 +596,7 @@ public class AggregationClient implements Closeable {
    * @param tableName
    * @param ci
    * @param scan
-   * @return <R, S>
+   * @return &lt;R, S&gt;
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message>
@@ -615,7 +616,7 @@ public class AggregationClient implements Closeable {
    * @param table
    * @param ci
    * @param scan
-   * @return <R, S>
+   * @return &lt;R, S&gt;
    * @throws Throwable
    */
   public <R, S, P extends Message, Q extends Message, T extends Message> double std(

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/client/package-info.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/package-info.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/package-info.java
index 10261cd..0e595a3 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/package-info.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/package-info.java
@@ -193,14 +193,9 @@ public class MyLittleHBaseClient {
   <li><a href="http://hbase.org">HBase Home Page</a>
   <li><a href="http://hadoop.apache.org/">Hadoop Home Page</a>
 </ul>
-</pre></code>
-</div>
-
   <p>See also the section in the HBase Reference Guide where it discusses
   <a href="http://hbase.apache.org/book.html#client">HBase Client</a>.  It
   has section on how to access HBase from inside your multithreaded environment
   how to control resources consumed client-side, etc.</p>
-</body>
-</html>
 */
 package org.apache.hadoop.hbase.client;

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java
index 8a0cb9f..6aae684 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java
@@ -35,23 +35,23 @@ import com.google.protobuf.Message;
  * example.
  * <p>
  * Takes two generic parameters and three Message parameters. 
- * The cell value type of the interpreter is <T>.
+ * The cell value type of the interpreter is &lt;T&gt;.
  * During some computations like sum, average, the return type can be different
  * than the cell value data type, for eg, sum of int cell values might overflow
  * in case of a int result, we should use Long for its result. Therefore, this
  * class mandates to use a different (promoted) data type for result of these
- * computations <S>. All computations are performed on the promoted data type
- * <S>. There is a conversion method
- * {@link ColumnInterpreter#castToReturnType(Object)} which takes a <T> type and
- * returns a <S> type.
- * The AggregateImplementation uses PB messages to initialize the
+ * computations &lt;S&gt;. All computations are performed on the promoted data type
+ * &lt;S&gt;. There is a conversion method
+ * {@link ColumnInterpreter#castToReturnType(Object)} which takes a &lt;T&gt; type and
+ * returns a &lt;S&gt; type.
+ * The AggregateIm&gt;lementation uses PB messages to initialize the
  * user's ColumnInterpreter implementation, and for sending the responses
  * back to AggregationClient.
- * @param <T> Cell value data type
- * @param <S> Promoted data type
- * @param <P> PB message that is used to transport initializer specific bytes
- * @param <Q> PB message that is used to transport Cell (<T>) instance
- * @param <R> PB message that is used to transport Promoted (<S>) instance
+ * @param T Cell value data type
+ * @param S Promoted data type
+ * @param P PB message that is used to transport initializer specific bytes
+ * @param Q PB message that is used to transport Cell (&lt;T&gt;) instance
+ * @param R PB message that is used to transport Promoted (&lt;S&gt;) instance
  */
 @InterfaceAudience.Private
 public abstract class ColumnInterpreter<T, S, P extends Message, 
@@ -109,14 +109,15 @@ Q extends Message, R extends Message> {
    * This takes care if either of arguments are null. returns 0 if they are
    * equal or both are null;
    * <ul>
-   * <li>>0 if l1 > l2 or l1 is not null and l2 is null.
-   * <li>< 0 if l1 < l2 or l1 is null and l2 is not null.
+   * <li>&gt; 0 if l1 &gt; l2 or l1 is not null and l2 is null.</li>
+   * <li>&lt; 0 if l1 &lt; l2 or l1 is null and l2 is not null.</li>
+   * </ul>
    */
   public abstract int compare(final T l1, final T l2);
 
   /**
-   * used for computing average of <S> data values. Not providing the divide
-   * method that takes two <S> values as it is not needed as of now.
+   * used for computing average of &lt;S&gt; data values. Not providing the divide
+   * method that takes two &lt;S&gt; values as it is not needed as of now.
    * @param o
    * @param l
    * @return Average

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java
index 2f40302..f9722d3 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java
@@ -32,15 +32,16 @@ import org.apache.hadoop.hbase.protobuf.generated.FilterProtos;
 import com.google.protobuf.InvalidProtocolBufferException;
 
 /**
+ * <p>
  * This filter is used to filter based on the column family. It takes an
  * operator (equal, greater, not equal, etc) and a byte [] comparator for the
  * column family portion of a key.
- * <p/>
+ * </p><p>
  * This filter can be wrapped with {@link org.apache.hadoop.hbase.filter.WhileMatchFilter} and {@link org.apache.hadoop.hbase.filter.SkipFilter}
  * to add more control.
- * <p/>
+ * </p><p>
  * Multiple filters can be combined using {@link org.apache.hadoop.hbase.filter.FilterList}.
- * <p/>
+ * </p>
  * If an already known column family is looked for, use {@link org.apache.hadoop.hbase.client.Get#addFamily(byte[])}
  * directly rather than a filter.
  */

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/Filter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/Filter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/Filter.java
index 784d145..774f268 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/Filter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/Filter.java
@@ -264,7 +264,6 @@ public abstract class Filter {
    * @param pbBytes A pb serialized {@link Filter} instance
    * @return An instance of {@link Filter} made from <code>bytes</code>
    * @throws DeserializationException
-   * @throws IOException in case an I/O or an filter specific failure needs to be signaled.
    * @see #toByteArray
    */
   public static Filter parseFrom(final byte [] pbBytes) throws DeserializationException {

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java
index e9fad92..c1ffc02 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java
@@ -43,7 +43,7 @@ public abstract class FilterBase extends Filter {
    * Filters that are purely stateless and do nothing in their reset() methods can inherit
    * this null/empty implementation.
    *
-   * @inheritDoc
+   * {@inheritDoc}
    */
   @Override
   public void reset() throws IOException {
@@ -53,7 +53,7 @@ public abstract class FilterBase extends Filter {
    * Filters that do not filter by row key can inherit this implementation that
    * never filters anything. (ie: returns false).
    *
-   * @inheritDoc
+   * {@inheritDoc}
    */
   @Override
   public boolean filterRowKey(byte[] buffer, int offset, int length) throws IOException {
@@ -64,7 +64,7 @@ public abstract class FilterBase extends Filter {
    * Filters that never filter all remaining can inherit this implementation that
    * never stops the filter early.
    *
-   * @inheritDoc
+   * {@inheritDoc}
    */
   @Override
   public boolean filterAllRemaining() throws IOException {
@@ -74,7 +74,7 @@ public abstract class FilterBase extends Filter {
   /**
    * By default no transformation takes place
    *
-   * @inheritDoc
+   * {@inheritDoc}
    */
   @Override
   public Cell transformCell(Cell v) throws IOException {
@@ -98,7 +98,7 @@ public abstract class FilterBase extends Filter {
    * Filters that never filter by modifying the returned List of Cells can
    * inherit this implementation that does nothing.
    *
-   * @inheritDoc
+   * {@inheritDoc}
    */
   @Override
   public void filterRowCells(List<Cell> ignored) throws IOException {
@@ -108,7 +108,7 @@ public abstract class FilterBase extends Filter {
    * Fitlers that never filter by modifying the returned List of Cells can
    * inherit this implementation that does nothing.
    *
-   * @inheritDoc
+   * {@inheritDoc}
    */
   @Override
   public boolean hasFilterRow() {
@@ -120,7 +120,7 @@ public abstract class FilterBase extends Filter {
    * {@link #filterKeyValue(Cell)} can inherit this implementation that
    * never filters a row.
    *
-   * @inheritDoc
+   * {@inheritDoc}
    */
   @Override
   public boolean filterRow() throws IOException {
@@ -140,7 +140,7 @@ public abstract class FilterBase extends Filter {
    * Filters that are not sure which key must be next seeked to, can inherit
    * this implementation that, by default, returns a null Cell.
    *
-   * @inheritDoc
+   * {@inheritDoc}
    */
   public Cell getNextCellHint(Cell currentKV) throws IOException {
     // Old filters based off of this class will override KeyValue getNextKeyHint(KeyValue).
@@ -152,7 +152,7 @@ public abstract class FilterBase extends Filter {
    * By default, we require all scan's column families to be present. Our
    * subclasses may be more precise.
    *
-   * @inheritDoc
+   * {@inheritDoc}
    */
   public boolean isFamilyEssential(byte[] name) throws IOException {
     return true;

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
index 579fe2c..7984d08 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java
@@ -41,14 +41,14 @@ import com.google.protobuf.InvalidProtocolBufferException;
  * Since you can use Filter Lists as children of Filter Lists, you can create a
  * hierarchy of filters to be evaluated.
  *
- * <br/>
+ * <br>
  * {@link Operator#MUST_PASS_ALL} evaluates lazily: evaluation stops as soon as one filter does
  * not include the KeyValue.
  *
- * <br/>
+ * <br>
  * {@link Operator#MUST_PASS_ONE} evaluates non-lazily: all filters are always evaluated.
  *
- * <br/>
+ * <br>
  * Defaults to {@link Operator#MUST_PASS_ALL}.
  */
 @InterfaceAudience.Public
@@ -316,7 +316,7 @@ final public class FilterList extends Filter {
    * Filters that never filter by modifying the returned List of Cells can
    * inherit this implementation that does nothing.
    *
-   * @inheritDoc
+   * {@inheritDoc}
    */
   @Override
   public void filterRowCells(List<Cell> cells) throws IOException {

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
index 24c8135..ca4f9b0 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/MultiRowRangeFilter.java
@@ -44,7 +44,7 @@ import com.google.protobuf.InvalidProtocolBufferException;
  * phoenix etc. However, both solutions are inefficient. Both of them can't utilize the range info
  * to perform fast forwarding during scan which is quite time consuming. If the number of ranges
  * are quite big (e.g. millions), join is a proper solution though it is slow. However, there are
- * cases that user wants to specify a small number of ranges to scan (e.g. <1000 ranges). Both
+ * cases that user wants to specify a small number of ranges to scan (e.g. &lt;1000 ranges). Both
  * solutions can't provide satisfactory performance in such case. MultiRowRangeFilter is to support
  * such usec ase (scan multiple row key ranges), which can construct the row key ranges from user
  * specified list and perform fast-forwarding during scan. Thus, the scan will be quite efficient.

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java
index d3bb89d..ee94d4c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java
@@ -31,11 +31,11 @@ import java.io.IOException;
 import java.util.ArrayList;
 /**
  * Implementation of Filter interface that limits results to a specific page
- * size. It terminates scanning once the number of filter-passed rows is >
+ * size. It terminates scanning once the number of filter-passed rows is &gt;
  * the given page size.
  * <p>
  * Note that this filter cannot guarantee that the number of results returned
- * to a client are <= page size. This is because the filter is applied
+ * to a client are &lt;= page size. This is because the filter is applied
  * separately on different region servers. It does however optimize the scan of
  * individual HRegions by making sure that the page size is never exceeded
  * locally.

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java
index f6d120b..dd06995 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java
@@ -39,8 +39,9 @@ import com.google.protobuf.InvalidProtocolBufferException;
  * entire row if any of its weights are zero.  In this case, we want to prevent
  * rows from being emitted if a single key is filtered.  Combine this filter
  * with a {@link ValueFilter}:
+ * </p>
  * <p>
- * <pre>
+ * <code>
  * scan.setFilter(new SkipFilter(new ValueFilter(CompareOp.NOT_EQUAL,
  *     new BinaryComparator(Bytes.toBytes(0))));
  * </code>
@@ -48,6 +49,7 @@ import com.google.protobuf.InvalidProtocolBufferException;
  * (since ValueFilter will not pass that Cell).
  * Without this filter, the other non-zero valued columns in the row would still
  * be emitted.
+ * </p>
  */
 @InterfaceAudience.Public
 @InterfaceStability.Stable

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcClient.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcClient.java
index 2e4d0a6..005f03c 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcClient.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/AsyncRpcClient.java
@@ -422,7 +422,6 @@ public class AsyncRpcClient extends AbstractRpcClient {
    * @param rpcTimeout default rpc operation timeout
    *
    * @return A rpc channel that goes via this rpc client instance.
-   * @throws IOException when channel could not be created
    */
   public RpcChannel createRpcChannel(final ServerName sn, final User user, int rpcTimeout) {
     return new RpcChannelImplementation(this, sn, user, rpcTimeout);

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ConnectionId.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ConnectionId.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ConnectionId.java
index bbd2fc7..33fc880 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ConnectionId.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ConnectionId.java
@@ -24,7 +24,7 @@ import java.net.InetSocketAddress;
 
 /**
  * This class holds the address and the user ticket, etc. The client connections
- * to servers are uniquely identified by <remoteAddress, ticket, serviceName>
+ * to servers are uniquely identified by &lt;remoteAddress, ticket, serviceName&gt;
  */
 @InterfaceAudience.Private
 public class ConnectionId {

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcController.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcController.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcController.java
index 76a88ea..afa1449 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcController.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ServerRpcController.java
@@ -39,7 +39,8 @@ import java.io.IOException;
  * coprocessor endpoints can use the following
  * pattern to pass exceptions back to the RPC client:
  * <code>
- * public void myMethod(RpcController controller, MyRequest request, RpcCallback<MyResponse> done) {
+ * public void myMethod(RpcController controller, MyRequest request,
+ *     RpcCallback&lt;MyResponse&gt; done) {
  *   MyResponse response = null;
  *   try {
  *     // do processing

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
index f6d08bb..ea1601b 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ProtobufUtil.java
@@ -260,7 +260,7 @@ public final class ProtobufUtil {
    * Prepend the passed bytes with four bytes of magic, {@link #PB_MAGIC}, to flag what
    * follows as a protobuf in hbase.  Prepend these bytes to all content written to znodes, etc.
    * @param bytes Bytes to decorate
-   * @return The passed <code>bytes</codes> with magic prepended (Creates a new
+   * @return The passed <code>bytes</code> with magic prepended (Creates a new
    * byte array that is <code>bytes.length</code> plus {@link #PB_MAGIC}.length.
    */
   public static byte [] prependPBMagic(final byte [] bytes) {
@@ -2129,7 +2129,7 @@ public final class ProtobufUtil {
   }
 
   /**
-   * Convert a ListMultimap<String, TablePermission> where key is username
+   * Convert a ListMultimap&lt;String, TablePermission&gt; where key is username
    * to a protobuf UserPermission
    *
    * @param perm the list of user and table permissions
@@ -2383,7 +2383,7 @@ public final class ProtobufUtil {
 
   /**
    * Convert a protobuf UserTablePermissions to a
-   * ListMultimap<String, TablePermission> where key is username.
+   * ListMultimap&lt;String, TablePermission&gt; where key is username.
    *
    * @param proto the protobuf UserPermission
    * @return the converted UserPermission

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaTableUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaTableUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaTableUtil.java
index b89f329..c750c8e 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaTableUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/quotas/QuotaTableUtil.java
@@ -45,11 +45,11 @@ import org.apache.hadoop.hbase.util.Strings;
  * 
  * <pre>
  *     ROW-KEY      FAM/QUAL        DATA
- *   n.<namespace> q:s         <global-quotas>
- *   t.<table>     q:s         <global-quotas>
- *   u.<user>      q:s         <global-quotas>
- *   u.<user>      q:s.<table> <table-quotas>
- *   u.<user>      q:s.<ns>:   <namespace-quotas>
+ *   n.&lt;namespace&gt; q:s         &lt;global-quotas&gt;
+ *   t.&lt;table&gt;     q:s         &lt;global-quotas&gt;
+ *   u.&lt;user&gt;      q:s         &lt;global-quotas&gt;
+ *   u.&lt;user&gt;      q:s.&lt;table&gt; &lt;table-quotas&gt;
+ *   u.&lt;user&gt;      q:s.&lt;ns&gt;:   &lt;namespace-quotas&gt;
  * </pre>
  */
 @InterfaceAudience.Private

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/regionserver/BloomType.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/regionserver/BloomType.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/regionserver/BloomType.java
index 75c967d..50b8b15 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/regionserver/BloomType.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/regionserver/BloomType.java
@@ -34,7 +34,7 @@ public enum BloomType {
    */
   ROW,
   /**
-   * Bloom enabled with Table row & column (family+qualifier) as Key
+   * Bloom enabled with Table row &amp; column (family+qualifier) as Key
    */
   ROWCOL
 }

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeers.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeers.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeers.java
index a601fb5..62585f1 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeers.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeers.java
@@ -125,7 +125,7 @@ public interface ReplicationPeers {
    * have to be connected. The state is read directly from the backing store.
    * @param peerId a short that identifies the cluster
    * @return true if replication is enabled, false otherwise.
-   * @throws IOException Throws if there's an error contacting the store
+   * @throws ReplicationException thrown if there's an error contacting the store
    */
   boolean getStatusOfPeerFromBackingStore(String peerId) throws ReplicationException;
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java
index 4968548..765559f 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/security/visibility/CellVisibility.java
@@ -24,7 +24,8 @@ import org.apache.hadoop.hbase.util.Bytes;
 /**
  * This contains a visibility expression which can be associated with a cell. When it is set with a
  * Mutation, all the cells in that mutation will get associated with this expression. A visibility
- * expression can contain visibility labels combined with logical operators AND(&), OR(|) and NOT(!)
+ * expression can contain visibility labels combined with logical
+ * operators AND(&amp;), OR(|) and NOT(!)
  */
 @InterfaceAudience.Public
 @InterfaceStability.Evolving
@@ -51,7 +52,7 @@ public class CellVisibility {
   /**
    * Helps in quoting authentication Strings. Use this if unicode characters to
    * be used in expression or special characters like '(', ')',
-   * '"','\','&','|','!'
+   * '"','\','&amp;','|','!'
    */
   public static String quote(String auth) {
     return quote(Bytes.toBytes(auth));
@@ -60,7 +61,7 @@ public class CellVisibility {
   /**
    * Helps in quoting authentication Strings. Use this if unicode characters to
    * be used in expression or special characters like '(', ')',
-   * '"','\','&','|','!'
+   * '"','\','&amp;','|','!'
    */
   public static String quote(byte[] auth) {
     int escapeChars = 0;

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
index 0f57206..7cd0446 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java
@@ -1208,8 +1208,6 @@ public class ZKUtil {
    * @param data data of node to create
    * @param cb
    * @param ctx
-   * @throws KeeperException if unexpected zookeeper exception
-   * @throws KeeperException.NodeExistsException if node already exists
    */
   public static void asyncCreate(ZooKeeperWatcher zkw,
       String znode, byte [] data, final AsyncCallback.StringCallback cb,
@@ -2067,7 +2065,6 @@ public class ZKUtil {
   /**
    * Recursively print the current state of ZK (non-transactional)
    * @param root name of the root directory in zk to print
-   * @throws KeeperException
    */
   public static void logZKTree(ZooKeeperWatcher zkw, String root) {
     if (!LOG.isDebugEnabled()) return;

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java
index 4dffcd3..9239b35 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java
@@ -670,7 +670,6 @@ public class ZooKeeperWatcher implements Watcher, Abortable, Closeable {
   /**
    * Close the connection to ZooKeeper.
    *
-   * @throws InterruptedException
    */
   @Override
   public void close() {

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java
index 8f299cc..46ebeeb 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/Cell.java
@@ -23,7 +23,8 @@ import org.apache.hadoop.hbase.classification.InterfaceStability;
 
 
 /**
- * The unit of storage in HBase consisting of the following fields:<br/>
+ * The unit of storage in HBase consisting of the following fields:
+ * <br>
  * <pre>
  * 1) row
  * 2) column family
@@ -33,30 +34,36 @@ import org.apache.hadoop.hbase.classification.InterfaceStability;
  * 6) MVCC version
  * 7) value
  * </pre>
- * <p/>
+ * <p>
  * Uniqueness is determined by the combination of row, column family, column qualifier,
  * timestamp, and type.
- * <p/>
+ * </p>
+ * <p>
  * The natural comparator will perform a bitwise comparison on row, column family, and column
  * qualifier. Less intuitively, it will then treat the greater timestamp as the lesser value with
  * the goal of sorting newer cells first.
- * <p/>
+ * </p>
+ * <p>
  * This interface should not include methods that allocate new byte[]'s such as those used in client
  * or debugging code. These users should use the methods found in the {@link CellUtil} class.
  * Currently for to minimize the impact of existing applications moving between 0.94 and 0.96, we
  * include the costly helper methods marked as deprecated.   
- * <p/>
- * Cell implements Comparable<Cell> which is only meaningful when comparing to other keys in the
+ * </p>
+ * <p>
+ * Cell implements Comparable&lt;Cell&gt; which is only meaningful when
+ * comparing to other keys in the
  * same table. It uses CellComparator which does not work on the -ROOT- and hbase:meta tables.
- * <p/>
+ * </p>
+ * <p>
  * In the future, we may consider adding a boolean isOnHeap() method and a getValueBuffer() method
  * that can be used to pass a value directly from an off-heap ByteBuffer to the network without
  * copying into an on-heap byte[].
- * <p/>
+ * </p>
+ * <p>
  * Historic note: the original Cell implementation (KeyValue) requires that all fields be encoded as
  * consecutive bytes in the same byte[], whereas this interface allows fields to reside in separate
  * byte[]'s.
- * <p/>
+ * </p>
  */
 @InterfaceAudience.Public
 @InterfaceStability.Evolving
@@ -77,7 +84,7 @@ public interface Cell {
   int getRowOffset();
 
   /**
-   * @return Number of row bytes. Must be < rowArray.length - offset.
+   * @return Number of row bytes. Must be &lt; rowArray.length - offset.
    */
   short getRowLength();
 
@@ -97,7 +104,7 @@ public interface Cell {
   int getFamilyOffset();
 
   /**
-   * @return Number of family bytes.  Must be < familyArray.length - offset.
+   * @return Number of family bytes.  Must be &lt; familyArray.length - offset.
    */
   byte getFamilyLength();
 
@@ -117,7 +124,7 @@ public interface Cell {
   int getQualifierOffset();
 
   /**
-   * @return Number of qualifier bytes.  Must be < qualifierArray.length - offset.
+   * @return Number of qualifier bytes.  Must be &lt; qualifierArray.length - offset.
    */
   int getQualifierLength();
 
@@ -148,7 +155,7 @@ public interface Cell {
    * cells in the memstore but is not retained forever. It may survive several flushes, but
    * generally becomes irrelevant after the cell's row is no longer involved in any operations that
    * require strict consistency.
-   * @return mvccVersion (always >= 0 if exists), or 0 if it no longer exists
+   * @return mvccVersion (always &gt;= 0 if exists), or 0 if it no longer exists
    */
   @Deprecated
   long getMvccVersion();
@@ -158,7 +165,7 @@ public interface Cell {
    * exists for cells in the memstore but is not retained forever. It will be kept for
    * {@link HConstants#KEEP_SEQID_PERIOD} days, but generally becomes irrelevant after the cell's
    * row is no longer involved in any operations that require strict consistency.
-   * @return seqId (always > 0 if exists), or 0 if it no longer exists
+   * @return seqId (always &gt; 0 if exists), or 0 if it no longer exists
    */
   long getSequenceId();
 
@@ -177,7 +184,7 @@ public interface Cell {
   int getValueOffset();
 
   /**
-   * @return Number of value bytes.  Must be < valueArray.length - offset.
+   * @return Number of value bytes.  Must be &lt; valueArray.length - offset.
    */
   int getValueLength();
   
@@ -199,7 +206,7 @@ public interface Cell {
   /**
    * WARNING do not use, expensive.  This gets an arraycopy of the cell's value.
    *
-   * Added to ease transition from  0.94 -> 0.96.
+   * Added to ease transition from  0.94 -&gt; 0.96.
    * 
    * @deprecated as of 0.96, use {@link CellUtil#cloneValue(Cell)}
    */
@@ -209,7 +216,7 @@ public interface Cell {
   /**
    * WARNING do not use, expensive.  This gets an arraycopy of the cell's family. 
    *
-   * Added to ease transition from  0.94 -> 0.96.
+   * Added to ease transition from  0.94 -&gt; 0.96.
    * 
    * @deprecated as of 0.96, use {@link CellUtil#cloneFamily(Cell)}
    */
@@ -219,7 +226,7 @@ public interface Cell {
   /**
    * WARNING do not use, expensive.  This gets an arraycopy of the cell's qualifier.
    *
-   * Added to ease transition from  0.94 -> 0.96.
+   * Added to ease transition from  0.94 -&gt; 0.96.
    * 
    * @deprecated as of 0.96, use {@link CellUtil#cloneQualifier(Cell)}
    */
@@ -229,10 +236,10 @@ public interface Cell {
   /**
    * WARNING do not use, expensive.  this gets an arraycopy of the cell's row.
    *
-   * Added to ease transition from  0.94 -> 0.96.
+   * Added to ease transition from  0.94 -&gt; 0.96.
    * 
    * @deprecated as of 0.96, use {@link CellUtil#getRowByte(Cell, int)}
    */
   @Deprecated
   byte[] getRow();
-}
\ No newline at end of file
+}

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparator.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparator.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparator.java
index 6536192..e9b934f 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparator.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellComparator.java
@@ -56,7 +56,7 @@ public class CellComparator implements Comparator<Cell>, Serializable {
    * @param b
    * @param ignoreSequenceid True if we are to compare the key portion only and ignore
    * the sequenceid. Set to false to compare key and consider sequenceid.
-   * @return 0 if equal, -1 if a < b, and +1 if a > b.
+   * @return 0 if equal, -1 if a &lt; b, and +1 if a &gt; b.
    */
   public static int compare(final Cell a, final Cell b, boolean ignoreSequenceid) {
     // row

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java
index 79e677e..0adb7b5 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScannable.java
@@ -22,7 +22,7 @@ import org.apache.hadoop.hbase.classification.InterfaceAudience;
 /**
  * Implementer can return a CellScanner over its Cell content.
  * Class name is ugly but mimicing java.util.Iterable only we are about the dumber
- * CellScanner rather than say Iterator<Cell>.  See CellScanner class comment for why we go
+ * CellScanner rather than say Iterator&lt;Cell&gt;.  See CellScanner class comment for why we go
  * dumber than java.util.Iterator.
  */
 @InterfaceAudience.Private

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java
index 3b5cdb9..75722a7 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/CellScanner.java
@@ -36,7 +36,7 @@ import org.apache.hadoop.hbase.Cell;
  * may or may not point to a reusable cell implementation, so users of the CellScanner should not,
  * for example, accumulate a List of Cells. All of the references may point to the same object,
  * which would be the latest state of the underlying Cell. In short, the Cell is mutable.
- * <p/>
+ * </p>
  * Typical usage:
  *
  * <pre>

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-common/src/main/java/org/apache/hadoop/hbase/HConstants.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/HConstants.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/HConstants.java
index 70562e2..99cfcdc 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/HConstants.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/HConstants.java
@@ -1122,7 +1122,8 @@ public final class HConstants {
 
   /**
    * When using bucket cache, this is a float that EITHER represents a percentage of total heap
-   * memory size to give to the cache (if < 1.0) OR, it is the capacity in megabytes of the cache.
+   * memory size to give to the cache (if &lt; 1.0) OR, it is the capacity in
+   * megabytes of the cache.
    */
   public static final String BUCKET_CACHE_SIZE_KEY = "hbase.bucketcache.size";
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/0a0ff335/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
index 26d215c..7e6a22b 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/KeyValue.java
@@ -63,18 +63,21 @@ import com.google.common.annotations.VisibleForTesting;
  * <p>
  * KeyValue wraps a byte array and takes offsets and lengths into passed array at where to start
  * interpreting the content as KeyValue. The KeyValue format inside a byte array is:
- * <code>&lt;keylength> &lt;valuelength> &lt;key> &lt;value></code> Key is further decomposed as:
- * <code>&lt;rowlength> &lt;row> &lt;columnfamilylength> &lt;columnfamily> &lt;columnqualifier>
- * &lt;timestamp> &lt;keytype></code>
+ * <code>&lt;keylength&gt; &lt;valuelength&gt; &lt;key&gt; &lt;value&gt;</code>
+ * Key is further decomposed as:
+ * <code>&lt;rowlength&gt; &lt;row&gt; &lt;columnfamilylength&gt;
+ * &lt;columnfamily&gt; &lt;columnqualifier&gt;
+ * &lt;timestamp&gt; &lt;keytype&gt;</code>
  * The <code>rowlength</code> maximum is <code>Short.MAX_SIZE</code>, column family length maximum
- * is <code>Byte.MAX_SIZE</code>, and column qualifier + key length must be <
+ * is <code>Byte.MAX_SIZE</code>, and column qualifier + key length must be &lt;
  * <code>Integer.MAX_SIZE</code>. The column does not contain the family/qualifier delimiter,
  * {@link #COLUMN_FAMILY_DELIMITER}<br>
  * KeyValue can optionally contain Tags. When it contains tags, it is added in the byte array after
- * the value part. The format for this part is: <code>&lt;tagslength>&lt;tagsbytes></code>.
+ * the value part. The format for this part is: <code>&lt;tagslength&gt;&lt;tagsbytes&gt;</code>.
  * <code>tagslength</code> maximum is <code>Short.MAX_SIZE</code>. The <code>tagsbytes</code>
  * contain one or more tags where as each tag is of the form
- * <code>&lt;taglength>&lt;tagtype>&lt;tagbytes></code>.  <code>tagtype</code> is one byte and
+ * <code>&lt;taglength&gt;&lt;tagtype&gt;&lt;tagbytes&gt;</code>.
+ * <code>tagtype</code> is one byte and
  * <code>taglength</code> maximum is <code>Short.MAX_SIZE</code> and it includes 1 byte type length
  * and actual tag bytes length.
  */
@@ -1198,7 +1201,7 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
    * as JSON. Values are left out due to their tendency to be large. If needed,
    * they can be added manually.
    *
-   * @return the Map<String,?> containing data from this key
+   * @return the Map&lt;String,?&gt; containing data from this key
    */
   public Map<String, Object> toStringMap() {
     Map<String, Object> stringMap = new HashMap<String, Object>();
@@ -1913,7 +1916,7 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
      * Compares the only the user specified portion of a Key.  This is overridden by MetaComparator.
      * @param left
      * @param right
-     * @return 0 if equal, <0 if left smaller, >0 if right smaller
+     * @return 0 if equal, &lt;0 if left smaller, &gt;0 if right smaller
      */
     protected int compareRowKey(final Cell left, final Cell right) {
       return CellComparator.compareRows(left, right);
@@ -1928,7 +1931,7 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
      * @param right
      * @param roffset
      * @param rlength
-     * @return  0 if equal, <0 if left smaller, >0 if right smaller
+     * @return  0 if equal, &lt;0 if left smaller, &gt;0 if right smaller
      */
     public int compareFlatKey(byte[] left, int loffset, int llength,
         byte[] right, int roffset, int rlength) {
@@ -2040,7 +2043,7 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
      * @param right
      * @param roffset
      * @param rlength
-     * @return 0 if equal, <0 if left smaller, >0 if right smaller
+     * @return 0 if equal, &lt;0 if left smaller, &gt;0 if right smaller
      */
     public int compareRows(byte [] left, int loffset, int llength,
         byte [] right, int roffset, int rlength) {
@@ -2089,7 +2092,7 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
      * @param right
      * @param roffset
      * @param rlength
-     * @return 0 if equal, <0 if left smaller, >0 if right smaller
+     * @return 0 if equal, &lt;0 if left smaller, &gt;0 if right smaller
      */
     @Override // SamePrefixComparator
     public int compareIgnoringPrefix(int commonPrefix, byte[] left,
@@ -2327,9 +2330,8 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
      * This is a HFile block index key optimization.
      * @param leftKey
      * @param rightKey
-     * @return 0 if equal, <0 if left smaller, >0 if right smaller
-     * @deprecated Since 0.99.2; Use
-     *             {@link CellComparator#getMidpoint(KeyValue.KVComparator, Cell, Cell) instead}
+     * @return 0 if equal, &lt;0 if left smaller, &gt;0 if right smaller
+     * @deprecated Since 0.99.2;
      */
     @Deprecated
     public byte[] getShortMidpointKey(final byte[] leftKey, final byte[] rightKey) {


Mime
View raw message